mirror of
https://github.com/hellodigua/ChatLab.git
synced 2026-05-04 12:11:18 +08:00
499 lines
13 KiB
TypeScript
499 lines
13 KiB
TypeScript
/**
|
||
* LLM 服务模块入口
|
||
* 提供统一的 LLM 服务管理(支持多配置)
|
||
*/
|
||
|
||
import * as fs from 'fs'
|
||
import * as path from 'path'
|
||
import { app } from 'electron'
|
||
import { randomUUID } from 'crypto'
|
||
import type {
|
||
LLMConfig,
|
||
LLMProvider,
|
||
ILLMService,
|
||
ProviderInfo,
|
||
ChatMessage,
|
||
ChatOptions,
|
||
ChatStreamChunk,
|
||
AIServiceConfig,
|
||
AIConfigStore,
|
||
} from './types'
|
||
import { MAX_CONFIG_COUNT } from './types'
|
||
import { DeepSeekService, DEEPSEEK_INFO } from './deepseek'
|
||
import { QwenService, QWEN_INFO } from './qwen'
|
||
import { GeminiService, GEMINI_INFO } from './gemini'
|
||
import { OpenAICompatibleService, OPENAI_COMPATIBLE_INFO } from './openai-compatible'
|
||
import { aiLogger } from '../logger'
|
||
|
||
// 导出类型
|
||
export * from './types'
|
||
|
||
// ==================== 新增提供商信息 ====================
|
||
|
||
/** MiniMax 提供商信息 */
|
||
const MINIMAX_INFO: ProviderInfo = {
|
||
id: 'minimax',
|
||
name: 'MiniMax',
|
||
description: 'MiniMax 大语言模型,支持多模态和长上下文',
|
||
defaultBaseUrl: 'https://api.minimaxi.com/v1',
|
||
models: [
|
||
{ id: 'MiniMax-M2', name: 'MiniMax-M2', description: '旗舰模型' },
|
||
{ id: 'MiniMax-M2-Stable', name: 'MiniMax-M2-Stable', description: '稳定版本' },
|
||
],
|
||
}
|
||
|
||
/** 智谱 GLM 提供商信息 */
|
||
const GLM_INFO: ProviderInfo = {
|
||
id: 'glm',
|
||
name: 'GLM',
|
||
description: '智谱 AI 大语言模型,ChatGLM 系列',
|
||
defaultBaseUrl: 'https://open.bigmodel.cn/api/paas/v4',
|
||
models: [
|
||
{ id: 'glm-4-plus', name: 'GLM-4-Plus', description: '旗舰模型,效果最佳' },
|
||
{ id: 'glm-4-flash', name: 'GLM-4-Flash', description: '高速模型,性价比高' },
|
||
{ id: 'glm-4', name: 'GLM-4', description: '标准模型' },
|
||
{ id: 'glm-4v-plus', name: 'GLM-4V-Plus', description: '多模态视觉模型' },
|
||
{ id: 'glm-4.6v-flash', name: '4.6V免费版', description: '4.6V免费版模型' },
|
||
{ id: 'glm-4.5-flash', name: '4.5免费版', description: '4.5免费版模型' },
|
||
],
|
||
}
|
||
|
||
/** Kimi (月之暗面 Moonshot) 提供商信息 */
|
||
const KIMI_INFO: ProviderInfo = {
|
||
id: 'kimi',
|
||
name: 'Kimi',
|
||
description: 'Moonshot AI 大语言模型,支持超长上下文',
|
||
defaultBaseUrl: 'https://api.moonshot.cn/v1',
|
||
models: [
|
||
{ id: 'moonshot-v1-8k', name: 'Moonshot-V1-8K', description: '8K 上下文' },
|
||
{ id: 'moonshot-v1-32k', name: 'Moonshot-V1-32K', description: '32K 上下文' },
|
||
{ id: 'moonshot-v1-128k', name: 'Moonshot-V1-128K', description: '128K 超长上下文' },
|
||
],
|
||
}
|
||
|
||
/** 豆包 (字节跳动 ByteDance) 提供商信息 */
|
||
const DOUBAO_INFO: ProviderInfo = {
|
||
id: 'doubao',
|
||
name: '豆包',
|
||
description: '字节跳动豆包 AI 大语言模型',
|
||
defaultBaseUrl: 'https://ark.cn-beijing.volces.com/api/v3',
|
||
models: [
|
||
{ id: 'doubao-seed-1-6-lite-251015', name: '性价比豆包1.6', description: '性价比豆包1.6模型' },
|
||
{ id: 'doubao-seed-1-6-251015', name: '更强豆包1.6', description: '更强豆包1.6模型' },
|
||
{ id: 'doubao-seed-1-6-flash-250828', name: '更快豆包1.6', description: '更快豆包1.6模型' },
|
||
{ id: 'doubao-1-5-lite-32k-250115', name: '豆包1.5Pro', description: '豆包1.5Pro模型' },
|
||
],
|
||
}
|
||
|
||
// 所有支持的提供商信息
|
||
export const PROVIDERS: ProviderInfo[] = [
|
||
DEEPSEEK_INFO,
|
||
QWEN_INFO,
|
||
GEMINI_INFO,
|
||
MINIMAX_INFO,
|
||
GLM_INFO,
|
||
KIMI_INFO,
|
||
DOUBAO_INFO,
|
||
OPENAI_COMPATIBLE_INFO,
|
||
]
|
||
|
||
// 配置文件路径
|
||
let CONFIG_PATH: string | null = null
|
||
|
||
function getConfigPath(): string {
|
||
if (CONFIG_PATH) return CONFIG_PATH
|
||
|
||
try {
|
||
const docPath = app.getPath('documents')
|
||
CONFIG_PATH = path.join(docPath, 'ChatLab', 'ai', 'llm-config.json')
|
||
} catch {
|
||
CONFIG_PATH = path.join(process.cwd(), 'ai', 'llm-config.json')
|
||
}
|
||
|
||
return CONFIG_PATH
|
||
}
|
||
|
||
// ==================== 旧配置格式(用于迁移)====================
|
||
|
||
interface LegacyStoredConfig {
|
||
provider: LLMProvider
|
||
apiKey: string
|
||
model?: string
|
||
maxTokens?: number
|
||
}
|
||
|
||
/**
|
||
* 检测是否为旧格式配置
|
||
*/
|
||
function isLegacyConfig(data: unknown): data is LegacyStoredConfig {
|
||
if (!data || typeof data !== 'object') return false
|
||
const obj = data as Record<string, unknown>
|
||
return 'provider' in obj && 'apiKey' in obj && !('configs' in obj)
|
||
}
|
||
|
||
/**
|
||
* 迁移旧配置到新格式
|
||
*/
|
||
function migrateLegacyConfig(legacy: LegacyStoredConfig): AIConfigStore {
|
||
const now = Date.now()
|
||
const newConfig: AIServiceConfig = {
|
||
id: randomUUID(),
|
||
name: getProviderInfo(legacy.provider)?.name || legacy.provider,
|
||
provider: legacy.provider,
|
||
apiKey: legacy.apiKey,
|
||
model: legacy.model,
|
||
maxTokens: legacy.maxTokens,
|
||
createdAt: now,
|
||
updatedAt: now,
|
||
}
|
||
|
||
return {
|
||
configs: [newConfig],
|
||
activeConfigId: newConfig.id,
|
||
}
|
||
}
|
||
|
||
// ==================== 多配置管理 ====================
|
||
|
||
/**
|
||
* 加载配置存储(自动处理迁移)
|
||
*/
|
||
export function loadConfigStore(): AIConfigStore {
|
||
const configPath = getConfigPath()
|
||
|
||
if (!fs.existsSync(configPath)) {
|
||
return { configs: [], activeConfigId: null }
|
||
}
|
||
|
||
try {
|
||
const content = fs.readFileSync(configPath, 'utf-8')
|
||
const data = JSON.parse(content)
|
||
|
||
// 检查是否需要迁移
|
||
if (isLegacyConfig(data)) {
|
||
aiLogger.info('LLM', '检测到旧配置格式,执行迁移')
|
||
const migrated = migrateLegacyConfig(data)
|
||
saveConfigStore(migrated)
|
||
return migrated
|
||
}
|
||
|
||
return data as AIConfigStore
|
||
} catch {
|
||
return { configs: [], activeConfigId: null }
|
||
}
|
||
}
|
||
|
||
/**
|
||
* 保存配置存储
|
||
*/
|
||
export function saveConfigStore(store: AIConfigStore): void {
|
||
const configPath = getConfigPath()
|
||
const dir = path.dirname(configPath)
|
||
|
||
if (!fs.existsSync(dir)) {
|
||
fs.mkdirSync(dir, { recursive: true })
|
||
}
|
||
|
||
fs.writeFileSync(configPath, JSON.stringify(store, null, 2), 'utf-8')
|
||
}
|
||
|
||
/**
|
||
* 获取所有配置列表
|
||
*/
|
||
export function getAllConfigs(): AIServiceConfig[] {
|
||
return loadConfigStore().configs
|
||
}
|
||
|
||
/**
|
||
* 获取当前激活的配置
|
||
*/
|
||
export function getActiveConfig(): AIServiceConfig | null {
|
||
const store = loadConfigStore()
|
||
if (!store.activeConfigId) return null
|
||
return store.configs.find((c) => c.id === store.activeConfigId) || null
|
||
}
|
||
|
||
/**
|
||
* 获取单个配置
|
||
*/
|
||
export function getConfigById(id: string): AIServiceConfig | null {
|
||
const store = loadConfigStore()
|
||
return store.configs.find((c) => c.id === id) || null
|
||
}
|
||
|
||
/**
|
||
* 添加新配置
|
||
*/
|
||
export function addConfig(config: Omit<AIServiceConfig, 'id' | 'createdAt' | 'updatedAt'>): {
|
||
success: boolean
|
||
config?: AIServiceConfig
|
||
error?: string
|
||
} {
|
||
const store = loadConfigStore()
|
||
|
||
if (store.configs.length >= MAX_CONFIG_COUNT) {
|
||
return { success: false, error: `最多只能添加 ${MAX_CONFIG_COUNT} 个配置` }
|
||
}
|
||
|
||
const now = Date.now()
|
||
const newConfig: AIServiceConfig = {
|
||
...config,
|
||
id: randomUUID(),
|
||
createdAt: now,
|
||
updatedAt: now,
|
||
}
|
||
|
||
store.configs.push(newConfig)
|
||
|
||
// 如果是第一个配置,自动设为激活
|
||
if (store.configs.length === 1) {
|
||
store.activeConfigId = newConfig.id
|
||
}
|
||
|
||
saveConfigStore(store)
|
||
return { success: true, config: newConfig }
|
||
}
|
||
|
||
/**
|
||
* 更新配置
|
||
*/
|
||
export function updateConfig(
|
||
id: string,
|
||
updates: Partial<Omit<AIServiceConfig, 'id' | 'createdAt' | 'updatedAt'>>
|
||
): { success: boolean; error?: string } {
|
||
const store = loadConfigStore()
|
||
const index = store.configs.findIndex((c) => c.id === id)
|
||
|
||
if (index === -1) {
|
||
return { success: false, error: '配置不存在' }
|
||
}
|
||
|
||
store.configs[index] = {
|
||
...store.configs[index],
|
||
...updates,
|
||
updatedAt: Date.now(),
|
||
}
|
||
|
||
saveConfigStore(store)
|
||
return { success: true }
|
||
}
|
||
|
||
/**
|
||
* 删除配置
|
||
*/
|
||
export function deleteConfig(id: string): { success: boolean; error?: string } {
|
||
const store = loadConfigStore()
|
||
const index = store.configs.findIndex((c) => c.id === id)
|
||
|
||
if (index === -1) {
|
||
return { success: false, error: '配置不存在' }
|
||
}
|
||
|
||
store.configs.splice(index, 1)
|
||
|
||
// 如果删除的是当前激活的配置,选择第一个作为新的激活配置
|
||
if (store.activeConfigId === id) {
|
||
store.activeConfigId = store.configs.length > 0 ? store.configs[0].id : null
|
||
}
|
||
|
||
saveConfigStore(store)
|
||
return { success: true }
|
||
}
|
||
|
||
/**
|
||
* 设置激活的配置
|
||
*/
|
||
export function setActiveConfig(id: string): { success: boolean; error?: string } {
|
||
const store = loadConfigStore()
|
||
const config = store.configs.find((c) => c.id === id)
|
||
|
||
if (!config) {
|
||
return { success: false, error: '配置不存在' }
|
||
}
|
||
|
||
store.activeConfigId = id
|
||
saveConfigStore(store)
|
||
return { success: true }
|
||
}
|
||
|
||
/**
|
||
* 检查是否有激活的配置
|
||
*/
|
||
export function hasActiveConfig(): boolean {
|
||
const config = getActiveConfig()
|
||
return config !== null
|
||
}
|
||
|
||
/**
|
||
* 扩展的 LLM 配置(包含本地服务特有选项)
|
||
*/
|
||
interface ExtendedLLMConfig extends LLMConfig {
|
||
disableThinking?: boolean
|
||
}
|
||
|
||
/**
|
||
* 创建 LLM 服务实例
|
||
*/
|
||
export function createLLMService(config: ExtendedLLMConfig): ILLMService {
|
||
// 获取提供商的默认 baseUrl
|
||
const providerInfo = getProviderInfo(config.provider)
|
||
const baseUrl = config.baseUrl || providerInfo?.defaultBaseUrl
|
||
|
||
switch (config.provider) {
|
||
case 'deepseek':
|
||
return new DeepSeekService(config.apiKey, config.model, config.baseUrl)
|
||
case 'qwen':
|
||
return new QwenService(config.apiKey, config.model, config.baseUrl)
|
||
case 'gemini':
|
||
return new GeminiService(config.apiKey, config.model, config.baseUrl)
|
||
// 新增的官方API都使用 OpenAI 兼容格式
|
||
case 'minimax':
|
||
case 'glm':
|
||
case 'kimi':
|
||
case 'doubao':
|
||
return new OpenAICompatibleService(config.apiKey, config.model, baseUrl)
|
||
case 'openai-compatible':
|
||
return new OpenAICompatibleService(config.apiKey, config.model, config.baseUrl, config.disableThinking)
|
||
default:
|
||
throw new Error(`Unknown LLM provider: ${config.provider}`)
|
||
}
|
||
}
|
||
|
||
/**
|
||
* 获取当前配置的 LLM 服务实例
|
||
*/
|
||
export function getCurrentLLMService(): ILLMService | null {
|
||
const activeConfig = getActiveConfig()
|
||
if (!activeConfig) {
|
||
return null
|
||
}
|
||
|
||
return createLLMService({
|
||
provider: activeConfig.provider,
|
||
apiKey: activeConfig.apiKey,
|
||
model: activeConfig.model,
|
||
baseUrl: activeConfig.baseUrl,
|
||
maxTokens: activeConfig.maxTokens,
|
||
disableThinking: activeConfig.disableThinking,
|
||
})
|
||
}
|
||
|
||
/**
|
||
* 获取提供商信息
|
||
*/
|
||
export function getProviderInfo(provider: LLMProvider): ProviderInfo | null {
|
||
return PROVIDERS.find((p) => p.id === provider) || null
|
||
}
|
||
|
||
/**
|
||
* 验证 API Key
|
||
*/
|
||
export async function validateApiKey(
|
||
provider: LLMProvider,
|
||
apiKey: string
|
||
): Promise<{ success: boolean; error?: string }> {
|
||
const service = createLLMService({ provider, apiKey })
|
||
return service.validateApiKey()
|
||
}
|
||
|
||
/**
|
||
* 发送聊天请求(使用当前配置)
|
||
* 返回完整的 ChatResponse 对象,包含 finishReason 和 tool_calls
|
||
*/
|
||
export async function chat(
|
||
messages: ChatMessage[],
|
||
options?: ChatOptions
|
||
): Promise<{ content: string; finishReason: string; tool_calls?: import('./types').ToolCall[] }> {
|
||
aiLogger.info('LLM', '开始非流式聊天请求', {
|
||
messagesCount: messages.length,
|
||
firstMessageRole: messages[0]?.role,
|
||
firstMessageLength: messages[0]?.content?.length,
|
||
options,
|
||
})
|
||
|
||
const service = getCurrentLLMService()
|
||
if (!service) {
|
||
aiLogger.error('LLM', '服务未配置')
|
||
throw new Error('LLM 服务未配置,请先在设置中配置 API Key')
|
||
}
|
||
|
||
aiLogger.info('LLM', `使用提供商: ${service.getProvider()}`)
|
||
|
||
try {
|
||
const response = await service.chat(messages, options)
|
||
aiLogger.info('LLM', '非流式请求成功', {
|
||
contentLength: response.content?.length,
|
||
finishReason: response.finishReason,
|
||
usage: response.usage,
|
||
})
|
||
return response
|
||
} catch (error) {
|
||
aiLogger.error('LLM', '非流式请求失败', { error: String(error) })
|
||
throw error
|
||
}
|
||
}
|
||
|
||
/**
|
||
* 发送聊天请求(流式,使用当前配置)
|
||
*/
|
||
export async function* chatStream(messages: ChatMessage[], options?: ChatOptions): AsyncGenerator<ChatStreamChunk> {
|
||
aiLogger.info('LLM', '开始流式聊天请求', {
|
||
messagesCount: messages.length,
|
||
firstMessageRole: messages[0]?.role,
|
||
firstMessageLength: messages[0]?.content?.length,
|
||
options,
|
||
})
|
||
|
||
const service = getCurrentLLMService()
|
||
if (!service) {
|
||
aiLogger.error('LLM', '服务未配置(流式)')
|
||
throw new Error('LLM 服务未配置,请先在设置中配置 API Key')
|
||
}
|
||
|
||
aiLogger.info('LLM', `使用提供商(流式): ${service.getProvider()}`)
|
||
|
||
let chunkCount = 0
|
||
let totalContent = ''
|
||
|
||
let receivedFinish = false
|
||
let contentChunkCount = 0
|
||
|
||
try {
|
||
for await (const chunk of service.chatStream(messages, options)) {
|
||
chunkCount++
|
||
totalContent += chunk.content
|
||
|
||
// 追踪内容 chunk
|
||
if (chunk.content) {
|
||
contentChunkCount++
|
||
}
|
||
|
||
yield chunk
|
||
|
||
if (chunk.isFinished) {
|
||
receivedFinish = true
|
||
aiLogger.info('LLM', '流式请求完成', {
|
||
chunkCount,
|
||
contentChunkCount,
|
||
totalContentLength: totalContent.length,
|
||
finishReason: chunk.finishReason,
|
||
})
|
||
}
|
||
}
|
||
|
||
// 如果循环正常结束但没有收到 isFinished 的 chunk,记录警告
|
||
if (chunkCount > 0 && !receivedFinish) {
|
||
aiLogger.warn('LLM', '流式请求循环结束但未收到完成信号', {
|
||
chunkCount,
|
||
totalContentLength: totalContent.length,
|
||
})
|
||
}
|
||
} catch (error) {
|
||
aiLogger.error('LLM', '流式请求失败', {
|
||
error: String(error),
|
||
chunkCountBeforeError: chunkCount,
|
||
})
|
||
throw error
|
||
}
|
||
}
|