mirror of
https://github.com/Memo-2023/mana-monorepo.git
synced 2026-05-17 06:59:40 +02:00
Merged shared-feedback-types + shared-feedback-service + shared-feedback-ui into a single @manacore/feedback package. Updated imports in all 21 apps. Before: 3 packages (types, service, ui) with cross-dependencies After: 1 package with direct imports, no circular refs Note: ESLint warnings from pre-existing unused vars in chat/mukke servers are unrelated to this change. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
136 lines
4.4 KiB
TypeScript
136 lines
4.4 KiB
TypeScript
/**
|
|
* Chat Hono Server — LLM completions (sync + streaming)
|
|
*
|
|
* CRUD for conversations/messages handled by mana-sync.
|
|
* This server handles AI completions via mana-llm or OpenRouter.
|
|
*/
|
|
|
|
import { Hono } from 'hono';
|
|
import { cors } from 'hono/cors';
|
|
import { streamSSE } from 'hono/streaming';
|
|
import { authMiddleware, healthRoute, errorHandler, notFoundHandler } from '@manacore/shared-hono';
|
|
import { consumeCredits, validateCredits } from '@manacore/shared-hono/credits';
|
|
|
|
const PORT = parseInt(process.env.PORT || '3002', 10);
|
|
const LLM_URL = process.env.MANA_LLM_URL || 'http://localhost:3025';
|
|
const OPENROUTER_KEY = process.env.OPENROUTER_API_KEY || '';
|
|
const CORS_ORIGINS = (process.env.CORS_ORIGINS || 'http://localhost:5173').split(',');
|
|
|
|
const app = new Hono();
|
|
|
|
app.onError(errorHandler);
|
|
app.notFound(notFoundHandler);
|
|
app.use('*', cors({ origin: CORS_ORIGINS, credentials: true }));
|
|
app.route('/health', healthRoute('chat-server'));
|
|
app.use('/api/*', authMiddleware());
|
|
|
|
// ─── Chat Completion (sync) ──────────────────────────────────
|
|
|
|
app.post('/api/v1/chat/completions', async (c) => {
|
|
const userId = c.get('userId');
|
|
const { messages, model, temperature, maxTokens } = await c.req.json();
|
|
|
|
if (!messages?.length) return c.json({ error: 'messages required' }, 400);
|
|
|
|
const isLocal = !model || model.startsWith('ollama/') || model.startsWith('local/');
|
|
const cost = isLocal ? 0.1 : 5;
|
|
|
|
const validation = await validateCredits(userId, 'AI_CHAT', cost);
|
|
if (!validation.hasCredits) {
|
|
return c.json({ error: 'Insufficient credits', required: cost }, 402);
|
|
}
|
|
|
|
try {
|
|
const llmRes = await fetch(`${LLM_URL}/api/v1/chat/completions`, {
|
|
method: 'POST',
|
|
headers: { 'Content-Type': 'application/json' },
|
|
body: JSON.stringify({
|
|
messages,
|
|
model: model || 'gemma3:4b',
|
|
temperature: temperature || 0.7,
|
|
max_tokens: maxTokens || 2000,
|
|
}),
|
|
});
|
|
|
|
if (!llmRes.ok) return c.json({ error: 'LLM request failed' }, 502);
|
|
|
|
const data = await llmRes.json();
|
|
await consumeCredits(userId, 'AI_CHAT', cost, `Chat: ${model || 'gemma3:4b'}`);
|
|
|
|
return c.json(data);
|
|
} catch (_err) {
|
|
return c.json({ error: 'Chat completion failed' }, 500);
|
|
}
|
|
});
|
|
|
|
// ─── Chat Completion (streaming SSE) ─────────────────────────
|
|
|
|
app.post('/api/v1/chat/completions/stream', async (c) => {
|
|
const userId = c.get('userId');
|
|
const { messages, model, temperature, maxTokens } = await c.req.json();
|
|
|
|
if (!messages?.length) return c.json({ error: 'messages required' }, 400);
|
|
|
|
const isLocal = !model || model.startsWith('ollama/') || model.startsWith('local/');
|
|
const cost = isLocal ? 0.1 : 5;
|
|
|
|
const validation = await validateCredits(userId, 'AI_CHAT', cost);
|
|
if (!validation.hasCredits) {
|
|
return c.json({ error: 'Insufficient credits' }, 402);
|
|
}
|
|
|
|
return streamSSE(c, async (stream) => {
|
|
try {
|
|
const llmRes = await fetch(`${LLM_URL}/api/v1/chat/completions`, {
|
|
method: 'POST',
|
|
headers: { 'Content-Type': 'application/json' },
|
|
body: JSON.stringify({
|
|
messages,
|
|
model: model || 'gemma3:4b',
|
|
temperature: temperature || 0.7,
|
|
max_tokens: maxTokens || 2000,
|
|
stream: true,
|
|
}),
|
|
});
|
|
|
|
if (!llmRes.ok || !llmRes.body) {
|
|
await stream.writeSSE({ data: JSON.stringify({ error: 'LLM failed' }) });
|
|
return;
|
|
}
|
|
|
|
const reader = llmRes.body.getReader();
|
|
const decoder = new TextDecoder();
|
|
|
|
while (true) {
|
|
const { done, value } = await reader.read();
|
|
if (done) break;
|
|
const chunk = decoder.decode(value, { stream: true });
|
|
// Forward SSE chunks directly
|
|
for (const line of chunk.split('\n')) {
|
|
if (line.startsWith('data: ')) {
|
|
await stream.writeSSE({ data: line.slice(6) });
|
|
}
|
|
}
|
|
}
|
|
|
|
await stream.writeSSE({ data: '[DONE]' });
|
|
consumeCredits(userId, 'AI_CHAT', cost, `Chat stream: ${model || 'gemma3:4b'}`).catch(() => {});
|
|
} catch (_err) {
|
|
await stream.writeSSE({ data: JSON.stringify({ error: 'Stream failed' }) });
|
|
}
|
|
});
|
|
});
|
|
|
|
// ─── Models List ─────────────────────────────────────────────
|
|
|
|
app.get('/api/v1/chat/models', async (c) => {
|
|
try {
|
|
const res = await fetch(`${LLM_URL}/api/v1/models`);
|
|
if (res.ok) return c.json(await res.json());
|
|
} catch {
|
|
// Fallback
|
|
}
|
|
return c.json({ models: [] });
|
|
});
|
|
|
|
export default { port: PORT, fetch: app.fetch };
|