280 lines
7.3 KiB
TypeScript
280 lines
7.3 KiB
TypeScript
/* eslint-disable @typescript-eslint/no-explicit-any,no-console */
|
||
|
||
import { NextRequest, NextResponse } from 'next/server';
|
||
|
||
import { getAuthInfoFromCookie } from '@/lib/auth';
|
||
import {
|
||
orchestrateDataSources,
|
||
VideoContext,
|
||
} from '@/lib/ai-orchestrator';
|
||
import { getConfig } from '@/lib/config';
|
||
|
||
export const runtime = 'nodejs';
|
||
|
||
interface ChatMessage {
|
||
role: 'user' | 'assistant';
|
||
content: string;
|
||
}
|
||
|
||
interface ChatRequest {
|
||
message: string;
|
||
context?: VideoContext;
|
||
history?: ChatMessage[];
|
||
}
|
||
|
||
/**
|
||
* OpenAI兼容的流式聊天请求
|
||
*/
|
||
async function streamOpenAIChat(
|
||
messages: ChatMessage[],
|
||
config: {
|
||
apiKey: string;
|
||
baseURL: string;
|
||
model: string;
|
||
temperature: number;
|
||
maxTokens: number;
|
||
}
|
||
): Promise<ReadableStream> {
|
||
const response = await fetch(`${config.baseURL}/chat/completions`, {
|
||
method: 'POST',
|
||
headers: {
|
||
'Content-Type': 'application/json',
|
||
Authorization: `Bearer ${config.apiKey}`,
|
||
},
|
||
body: JSON.stringify({
|
||
model: config.model,
|
||
messages,
|
||
temperature: config.temperature,
|
||
max_tokens: config.maxTokens,
|
||
stream: true,
|
||
}),
|
||
});
|
||
|
||
if (!response.ok) {
|
||
throw new Error(
|
||
`OpenAI API error: ${response.status} ${response.statusText}`
|
||
);
|
||
}
|
||
|
||
return response.body!;
|
||
}
|
||
|
||
/**
|
||
* Claude API流式聊天请求
|
||
*/
|
||
async function streamClaudeChat(
|
||
messages: ChatMessage[],
|
||
systemPrompt: string,
|
||
config: {
|
||
apiKey: string;
|
||
model: string;
|
||
temperature: number;
|
||
maxTokens: number;
|
||
}
|
||
): Promise<ReadableStream> {
|
||
// Claude API格式: 移除system消息,单独传递
|
||
const userMessages = messages.filter((m) => m.role !== 'system');
|
||
|
||
const response = await fetch('https://api.anthropic.com/v1/messages', {
|
||
method: 'POST',
|
||
headers: {
|
||
'Content-Type': 'application/json',
|
||
'x-api-key': config.apiKey,
|
||
'anthropic-version': '2023-06-01',
|
||
},
|
||
body: JSON.stringify({
|
||
model: config.model,
|
||
max_tokens: config.maxTokens,
|
||
temperature: config.temperature,
|
||
system: systemPrompt,
|
||
messages: userMessages,
|
||
stream: true,
|
||
}),
|
||
});
|
||
|
||
if (!response.ok) {
|
||
throw new Error(
|
||
`Claude API error: ${response.status} ${response.statusText}`
|
||
);
|
||
}
|
||
|
||
return response.body!;
|
||
}
|
||
|
||
/**
|
||
* 转换流为SSE格式
|
||
*/
|
||
function transformToSSE(
|
||
stream: ReadableStream,
|
||
provider: 'openai' | 'claude' | 'custom'
|
||
): ReadableStream {
|
||
const reader = stream.getReader();
|
||
const decoder = new TextDecoder();
|
||
|
||
return new ReadableStream({
|
||
async start(controller) {
|
||
try {
|
||
while (true) {
|
||
const { done, value } = await reader.read();
|
||
if (done) break;
|
||
|
||
const chunk = decoder.decode(value, { stream: true });
|
||
const lines = chunk.split('\n').filter((line) => line.trim() !== '');
|
||
|
||
for (const line of lines) {
|
||
if (line.startsWith('data: ')) {
|
||
const data = line.slice(6);
|
||
|
||
if (data === '[DONE]') {
|
||
controller.enqueue(
|
||
new TextEncoder().encode('data: [DONE]\n\n')
|
||
);
|
||
continue;
|
||
}
|
||
|
||
try {
|
||
const json = JSON.parse(data);
|
||
|
||
// 提取文本内容
|
||
let text = '';
|
||
if (provider === 'claude') {
|
||
// Claude格式
|
||
if (json.type === 'content_block_delta') {
|
||
text = json.delta?.text || '';
|
||
}
|
||
} else {
|
||
// OpenAI格式
|
||
text = json.choices?.[0]?.delta?.content || '';
|
||
}
|
||
|
||
if (text) {
|
||
controller.enqueue(
|
||
new TextEncoder().encode(`data: ${JSON.stringify({ text })}\n\n`)
|
||
);
|
||
}
|
||
} catch (e) {
|
||
console.error('Parse stream chunk error:', e);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
} catch (error) {
|
||
console.error('Stream error:', error);
|
||
controller.error(error);
|
||
} finally {
|
||
controller.close();
|
||
}
|
||
},
|
||
});
|
||
}
|
||
|
||
export async function POST(request: NextRequest) {
|
||
try {
|
||
// 1. 验证用户登录
|
||
const authInfo = getAuthInfoFromCookie(request);
|
||
if (!authInfo || !authInfo.username) {
|
||
return NextResponse.json({ error: 'Unauthorized' }, { status: 401 });
|
||
}
|
||
|
||
// 2. 获取AI配置
|
||
const adminConfig = await getConfig();
|
||
const aiConfig = adminConfig.AIConfig;
|
||
|
||
if (!aiConfig || !aiConfig.Enabled) {
|
||
return NextResponse.json(
|
||
{ error: 'AI功能未启用' },
|
||
{ status: 400 }
|
||
);
|
||
}
|
||
|
||
// 3. 解析请求参数
|
||
const body = (await request.json()) as ChatRequest;
|
||
const { message, context, history = [] } = body;
|
||
|
||
if (!message || typeof message !== 'string') {
|
||
return NextResponse.json(
|
||
{ error: '消息内容不能为空' },
|
||
{ status: 400 }
|
||
);
|
||
}
|
||
|
||
console.log('📨 收到AI聊天请求:', {
|
||
message: message.slice(0, 50),
|
||
context,
|
||
historyLength: history.length,
|
||
});
|
||
|
||
// 4. 使用orchestrator协调数据源
|
||
const orchestrationResult = await orchestrateDataSources(
|
||
message,
|
||
context,
|
||
{
|
||
enableWebSearch: aiConfig.EnableWebSearch,
|
||
webSearchProvider: aiConfig.WebSearchProvider,
|
||
tavilyApiKey: aiConfig.TavilyApiKey,
|
||
serperApiKey: aiConfig.SerperApiKey,
|
||
serpApiKey: aiConfig.SerpApiKey,
|
||
// 决策模型配置(固定使用自定义provider)
|
||
enableDecisionModel: aiConfig.EnableDecisionModel,
|
||
decisionProvider: 'custom',
|
||
decisionApiKey: aiConfig.DecisionCustomApiKey,
|
||
decisionBaseURL: aiConfig.DecisionCustomBaseURL,
|
||
decisionModel: aiConfig.DecisionCustomModel,
|
||
}
|
||
);
|
||
|
||
console.log('🎯 数据协调完成, systemPrompt长度:', orchestrationResult.systemPrompt.length);
|
||
|
||
// 5. 构建消息列表
|
||
const systemPrompt = aiConfig.SystemPrompt
|
||
? `${aiConfig.SystemPrompt}\n\n${orchestrationResult.systemPrompt}`
|
||
: orchestrationResult.systemPrompt;
|
||
|
||
const messages: ChatMessage[] = [
|
||
{ role: 'user', content: systemPrompt },
|
||
{ role: 'assistant', content: '明白了,我会按照要求回答用户的问题。' },
|
||
...history,
|
||
{ role: 'user', content: message },
|
||
];
|
||
|
||
// 6. 调用自定义API
|
||
const temperature = aiConfig.Temperature ?? 0.7;
|
||
const maxTokens = aiConfig.MaxTokens ?? 1000;
|
||
|
||
if (!aiConfig.CustomApiKey || !aiConfig.CustomBaseURL) {
|
||
return NextResponse.json(
|
||
{ error: '自定义API配置不完整' },
|
||
{ status: 400 }
|
||
);
|
||
}
|
||
|
||
const stream = await streamOpenAIChat(messages, {
|
||
apiKey: aiConfig.CustomApiKey,
|
||
baseURL: aiConfig.CustomBaseURL,
|
||
model: aiConfig.CustomModel || 'gpt-3.5-turbo',
|
||
temperature,
|
||
maxTokens,
|
||
});
|
||
|
||
// 7. 转换为SSE格式并返回
|
||
const sseStream = transformToSSE(stream, 'openai');
|
||
|
||
return new NextResponse(sseStream, {
|
||
headers: {
|
||
'Content-Type': 'text/event-stream',
|
||
'Cache-Control': 'no-cache',
|
||
Connection: 'keep-alive',
|
||
},
|
||
});
|
||
} catch (error) {
|
||
console.error('❌ AI聊天API错误:', error);
|
||
return NextResponse.json(
|
||
{
|
||
error: 'AI聊天请求失败',
|
||
details: (error as Error).message,
|
||
},
|
||
{ status: 500 }
|
||
);
|
||
}
|
||
}
|