From 9192f6f7f7008555acf2533bbf5fbcf5755a7fe5 Mon Sep 17 00:00:00 2001 From: zhayujie Date: Thu, 19 Mar 2026 10:46:13 +0800 Subject: [PATCH] feat: add MiniMax-M2.7 and glm-5-turbo --- README.md | 22 +++++++++++----------- common/const.py | 12 +++++++----- config-template.json | 2 +- docs/agent.md | 4 ++-- docs/en/README.md | 4 ++-- docs/en/models/glm.mdx | 6 +++--- docs/en/models/index.mdx | 6 +++--- docs/en/models/minimax.mdx | 6 +++--- docs/ja/README.md | 4 ++-- docs/ja/models/glm.mdx | 6 +++--- docs/ja/models/index.mdx | 6 +++--- docs/ja/models/minimax.mdx | 6 +++--- docs/models/glm.mdx | 6 +++--- docs/models/index.mdx | 6 +++--- docs/models/minimax.mdx | 6 +++--- run.sh | 16 ++++++++-------- 16 files changed, 60 insertions(+), 58 deletions(-) diff --git a/README.md b/README.md index 959d334..42d3387 100644 --- a/README.md +++ b/README.md @@ -102,7 +102,7 @@ bash <(curl -fsSL https://cdn.link-ai.tech/code/cow/run.sh) 项目支持国内外主流厂商的模型接口,可选模型及配置说明参考:[模型说明](#模型说明)。 -> 注:Agent模式下推荐使用以下模型,可根据效果及成本综合选择:MiniMax-M2.5、glm-5、kimi-k2.5、qwen3.5-plus、claude-sonnet-4-6、gemini-3.1-pro-preview、gpt-5.4、gpt-5.4-mini +> 注:Agent模式下推荐使用以下模型,可根据效果及成本综合选择:MiniMax-M2.7、glm-5-turbo、kimi-k2.5、qwen3.5-plus、claude-sonnet-4-6、gemini-3.1-pro-preview、gpt-5.4、gpt-5.4-mini 同时支持使用 **LinkAI平台** 接口,支持上述全部模型,并支持知识库、工作流、插件等Agent技能,参考 [接口文档](https://docs.link-ai.tech/platform/api)。 @@ -148,7 +148,7 @@ pip3 install -r requirements-optional.txt # config.json 文件内容示例 { "channel_type": "web", # 接入渠道类型,默认为web,支持修改为:feishu,dingtalk,wecom_bot,qq,wechatcom_app,wechatmp_service,wechatmp,terminal - "model": "MiniMax-M2.5", # 模型名称 + "model": "MiniMax-M2.7", # 模型名称 "minimax_api_key": "", # MiniMax API Key "zhipu_ai_api_key": "", # 智谱GLM API Key "moonshot_api_key": "", # Kimi/Moonshot API Key @@ -187,7 +187,7 @@ pip3 install -r requirements-optional.txt
2. 其他配置 -+ `model`: 模型名称,Agent模式下推荐使用 `MiniMax-M2.5`、`glm-5`、`kimi-k2.5`、`qwen3.5-plus`、`claude-sonnet-4-6`、`gemini-3.1-pro-preview`,全部模型名称参考[common/const.py](https://github.com/zhayujie/chatgpt-on-wechat/blob/master/common/const.py)文件 ++ `model`: 模型名称,Agent模式下推荐使用 `MiniMax-M2.7`、`glm-5-turbo`、`kimi-k2.5`、`qwen3.5-plus`、`claude-sonnet-4-6`、`gemini-3.1-pro-preview`,全部模型名称参考[common/const.py](https://github.com/zhayujie/chatgpt-on-wechat/blob/master/common/const.py)文件 + `character_desc`:普通对话模式下的机器人系统提示词。在Agent模式下该配置不生效,由工作空间中的文件内容构成。 + `subscribe_msg`:订阅消息,公众号和企业微信channel中请填写,当被订阅时会自动回复, 可使用特殊占位符。目前支持的占位符有{trigger_prefix},在程序中它会自动替换成bot的触发词。
@@ -312,24 +312,24 @@ sudo docker logs -f chatgpt-on-wechat ```json { - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "minimax_api_key": "" } ``` - - `model`: 可填写 `MiniMax-M2.5、MiniMax-M2.1、MiniMax-M2.1-lightning、MiniMax-M2、abab6.5-chat` 等 + - `model`: 可填写 `MiniMax-M2.7、MiniMax-M2.5、MiniMax-M2.1、MiniMax-M2.1-lightning、MiniMax-M2、abab6.5-chat` 等 - `minimax_api_key`:MiniMax平台的API-KEY,在 [控制台](https://platform.minimaxi.com/user-center/basic-information/interface-key) 创建 方式二:OpenAI兼容方式接入,配置如下: ```json { "bot_type": "openai", - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "open_ai_api_base": "https://api.minimaxi.com/v1", "open_ai_api_key": "" } ``` - `bot_type`: OpenAI兼容方式 -- `model`: 可填 `MiniMax-M2.5、MiniMax-M2.1、MiniMax-M2.1-lightning、MiniMax-M2`,参考[API文档](https://platform.minimaxi.com/document/%E5%AF%B9%E8%AF%9D?key=66701d281d57f38758d581d0#QklxsNSbaf6kM4j6wjO5eEek) +- `model`: 可填 `MiniMax-M2.7、MiniMax-M2.5、MiniMax-M2.1、MiniMax-M2.1-lightning、MiniMax-M2`,参考[API文档](https://platform.minimaxi.com/document/%E5%AF%B9%E8%AF%9D?key=66701d281d57f38758d581d0#QklxsNSbaf6kM4j6wjO5eEek) - `open_ai_api_base`: MiniMax平台API的 BASE URL - `open_ai_api_key`: MiniMax平台的API-KEY @@ -341,24 +341,24 @@ sudo docker logs -f chatgpt-on-wechat ```json { - "model": "glm-5", + "model": "glm-5-turbo", "zhipu_ai_api_key": "" } ``` - - `model`: 可填 `glm-5、glm-4.7、glm-4-plus、glm-4-flash、glm-4-air、glm-4-airx、glm-4-long` 等, 参考 [glm系列模型编码](https://bigmodel.cn/dev/api/normal-model/glm-4) + - `model`: 可填 `glm-5-turbo、glm-5、glm-4.7、glm-4-plus、glm-4-flash、glm-4-air、glm-4-airx、glm-4-long` 等, 参考 [glm系列模型编码](https://bigmodel.cn/dev/api/normal-model/glm-4) - `zhipu_ai_api_key`: 智谱AI平台的 API KEY,在 [控制台](https://www.bigmodel.cn/usercenter/proj-mgmt/apikeys) 创建 方式二:OpenAI兼容方式接入,配置如下: ```json { "bot_type": "openai", - "model": "glm-5", + "model": "glm-5-turbo", "open_ai_api_base": "https://open.bigmodel.cn/api/paas/v4", "open_ai_api_key": "" } ``` - `bot_type`: OpenAI兼容方式 -- `model`: 可填 `glm-5、glm-4.7、glm-4-plus、glm-4-flash、glm-4-air、glm-4-airx、glm-4-long` 等 +- `model`: 可填 `glm-5-turbo、glm-5、glm-4.7、glm-4-plus、glm-4-flash、glm-4-air、glm-4-airx、glm-4-long` 等 - `open_ai_api_base`: 智谱AI平台的 BASE URL - `open_ai_api_key`: 智谱AI平台的 API KEY diff --git a/common/const.py b/common/const.py index 290cf5d..0e53701 100644 --- a/common/const.py +++ b/common/const.py @@ -92,14 +92,16 @@ QWEN35_PLUS = "qwen3.5-plus" # Qwen3.5 Plus - Omni model (MultiModalConversatio QWQ_PLUS = "qwq-plus" # MiniMax -MINIMAX_M2_5 = "MiniMax-M2.5" # MiniMax M2.5 - Latest -MINIMAX_M2_1 = "MiniMax-M2.1" # MiniMax M2.1 - Agent推荐模型 +MINIMAX_M2_7 = "MiniMax-M2.7" # MiniMax M2.7 - Latest +MINIMAX_M2_5 = "MiniMax-M2.5" # MiniMax M2.5 +MINIMAX_M2_1 = "MiniMax-M2.1" # MiniMax M2.1 MINIMAX_M2_1_LIGHTNING = "MiniMax-M2.1-lightning" # MiniMax M2.1 极速版 MINIMAX_M2 = "MiniMax-M2" # MiniMax M2 MINIMAX_ABAB6_5 = "abab6.5-chat" # MiniMax abab6.5 # GLM (智谱AI) -GLM_5 = "glm-5" # 智谱 GLM-5 - Latest +GLM_5_TURBO = "glm-5-turbo" # 智谱 GLM-5-Turbo - Latest +GLM_5 = "glm-5" # 智谱 GLM-5 GLM_4 = "glm-4" GLM_4_PLUS = "glm-4-plus" GLM_4_flash = "glm-4-flash" @@ -166,10 +168,10 @@ MODEL_LIST = [ QWEN, QWEN_TURBO, QWEN_PLUS, QWEN_MAX, QWEN_LONG, QWEN3_MAX, QWEN35_PLUS, # MiniMax - MiniMax, MINIMAX_M2_5, MINIMAX_M2_1, MINIMAX_M2_1_LIGHTNING, MINIMAX_M2, MINIMAX_ABAB6_5, + MiniMax, MINIMAX_M2_7, MINIMAX_M2_5, MINIMAX_M2_1, MINIMAX_M2_1_LIGHTNING, MINIMAX_M2, MINIMAX_ABAB6_5, # GLM - ZHIPU_AI, GLM_5, GLM_4, GLM_4_PLUS, GLM_4_flash, GLM_4_LONG, GLM_4_ALLTOOLS, + ZHIPU_AI, GLM_5_TURBO, GLM_5, GLM_4, GLM_4_PLUS, GLM_4_flash, GLM_4_LONG, GLM_4_ALLTOOLS, GLM_4_0520, GLM_4_AIR, GLM_4_AIRX, GLM_4_7, # Kimi diff --git a/config-template.json b/config-template.json index b73f436..5e6d269 100644 --- a/config-template.json +++ b/config-template.json @@ -1,6 +1,6 @@ { "channel_type": "web", - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "minimax_api_key": "", "zhipu_ai_api_key": "", "ark_api_key": "", diff --git a/docs/agent.md b/docs/agent.md index 51a9468..da8b0f0 100644 --- a/docs/agent.md +++ b/docs/agent.md @@ -137,8 +137,8 @@ bash <(curl -fsSL https://cdn.link-ai.tech/code/cow/run.sh) Agent模式推荐使用以下模型,可根据效果及成本综合选择: -- **MiniMax**: `MiniMax-M2.5` -- **GLM**: `glm-5` +- **MiniMax**: `MiniMax-M2.7` +- **GLM**: `glm-5-turbo` - **Kimi**: `kimi-k2.5` - **Doubao**: `doubao-seed-2-0-code-preview-260215` - **Qwen**: `qwen3.5-plus` diff --git a/docs/en/README.md b/docs/en/README.md index a25de13..7cb52ee 100644 --- a/docs/en/README.md +++ b/docs/en/README.md @@ -121,8 +121,8 @@ Supports mainstream model providers. Recommended models for Agent mode: | Provider | Recommended Model | | --- | --- | -| MiniMax | `MiniMax-M2.5` | -| GLM | `glm-5` | +| MiniMax | `MiniMax-M2.7` | +| GLM | `glm-5-turbo` | | Kimi | `kimi-k2.5` | | Doubao | `doubao-seed-2-0-code-preview-260215` | | Qwen | `qwen3.5-plus` | diff --git a/docs/en/models/glm.mdx b/docs/en/models/glm.mdx index 56ea819..cfa2a27 100644 --- a/docs/en/models/glm.mdx +++ b/docs/en/models/glm.mdx @@ -5,14 +5,14 @@ description: Zhipu AI GLM model configuration ```json { - "model": "glm-5", + "model": "glm-5-turbo", "zhipu_ai_api_key": "YOUR_API_KEY" } ``` | Parameter | Description | | --- | --- | -| `model` | Options include `glm-5`, `glm-4.7`, `glm-4-plus`, `glm-4-flash`, `glm-4-air`, etc. See [model codes](https://bigmodel.cn/dev/api/normal-model/glm-4) | +| `model` | Options include `glm-5-turbo`, `glm-5`, `glm-4.7`, `glm-4-plus`, `glm-4-flash`, `glm-4-air`, etc. See [model codes](https://bigmodel.cn/dev/api/normal-model/glm-4) | | `zhipu_ai_api_key` | Create at [Zhipu AI Console](https://www.bigmodel.cn/usercenter/proj-mgmt/apikeys) | OpenAI-compatible configuration is also supported: @@ -20,7 +20,7 @@ OpenAI-compatible configuration is also supported: ```json { "bot_type": "openai", - "model": "glm-5", + "model": "glm-5-turbo", "open_ai_api_base": "https://open.bigmodel.cn/api/paas/v4", "open_ai_api_key": "YOUR_API_KEY" } diff --git a/docs/en/models/index.mdx b/docs/en/models/index.mdx index 1ac9506..8d76af8 100644 --- a/docs/en/models/index.mdx +++ b/docs/en/models/index.mdx @@ -6,7 +6,7 @@ description: Supported models and recommended choices for CowAgent CowAgent supports mainstream LLMs from domestic and international providers. Model interfaces are implemented in the project's `models/` directory. - For Agent mode, the following models are recommended based on quality and cost: MiniMax-M2.5, glm-5, kimi-k2.5, qwen3.5-plus, claude-sonnet-4-6, gemini-3.1-pro-preview + For Agent mode, the following models are recommended based on quality and cost: MiniMax-M2.7, glm-5-turbo, kimi-k2.5, qwen3.5-plus, claude-sonnet-4-6, gemini-3.1-pro-preview ## Configuration @@ -19,10 +19,10 @@ You can also use the [LinkAI](https://link-ai.tech) platform interface to flexib - MiniMax-M2.5 and other series models + MiniMax-M2.7 and other series models - glm-5, glm-4.7 and other series models + glm-5-turbo, glm-5 and other series models qwen3.5-plus, qwen3-max and more diff --git a/docs/en/models/minimax.mdx b/docs/en/models/minimax.mdx index 981153d..c3137ca 100644 --- a/docs/en/models/minimax.mdx +++ b/docs/en/models/minimax.mdx @@ -5,14 +5,14 @@ description: MiniMax model configuration ```json { - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "minimax_api_key": "YOUR_API_KEY" } ``` | Parameter | Description | | --- | --- | -| `model` | Options include `MiniMax-M2.5`, `MiniMax-M2.1`, `MiniMax-M2.1-lightning`, `MiniMax-M2`, etc. | +| `model` | Options include `MiniMax-M2.7`, `MiniMax-M2.5`, `MiniMax-M2.1`, `MiniMax-M2.1-lightning`, `MiniMax-M2`, etc. | | `minimax_api_key` | Create at [MiniMax Console](https://platform.minimaxi.com/user-center/basic-information/interface-key) | OpenAI-compatible configuration is also supported: @@ -20,7 +20,7 @@ OpenAI-compatible configuration is also supported: ```json { "bot_type": "openai", - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "open_ai_api_base": "https://api.minimaxi.com/v1", "open_ai_api_key": "YOUR_API_KEY" } diff --git a/docs/ja/README.md b/docs/ja/README.md index 591cb89..cb14691 100644 --- a/docs/ja/README.md +++ b/docs/ja/README.md @@ -121,8 +121,8 @@ sudo docker logs -f chatgpt-on-wechat | プロバイダー | 推奨モデル | | --- | --- | -| MiniMax | `MiniMax-M2.5` | -| GLM | `glm-5` | +| MiniMax | `MiniMax-M2.7` | +| GLM | `glm-5-turbo` | | Kimi | `kimi-k2.5` | | Doubao | `doubao-seed-2-0-code-preview-260215` | | Qwen | `qwen3.5-plus` | diff --git a/docs/ja/models/glm.mdx b/docs/ja/models/glm.mdx index 2ab907d..3309fcf 100644 --- a/docs/ja/models/glm.mdx +++ b/docs/ja/models/glm.mdx @@ -5,14 +5,14 @@ description: 智谱AI GLMモデルの設定 ```json { - "model": "glm-5", + "model": "glm-5-turbo", "zhipu_ai_api_key": "YOUR_API_KEY" } ``` | パラメータ | 説明 | | --- | --- | -| `model` | `glm-5`、`glm-4.7`、`glm-4-plus`、`glm-4-flash`、`glm-4-air`などから選択可能。[モデルコード](https://bigmodel.cn/dev/api/normal-model/glm-4)を参照 | +| `model` | `glm-5-turbo`、`glm-5`、`glm-4.7`、`glm-4-plus`、`glm-4-flash`、`glm-4-air`などから選択可能。[モデルコード](https://bigmodel.cn/dev/api/normal-model/glm-4)を参照 | | `zhipu_ai_api_key` | [智谱AI Console](https://www.bigmodel.cn/usercenter/proj-mgmt/apikeys)で作成 | OpenAI互換の設定もサポートしています: @@ -20,7 +20,7 @@ OpenAI互換の設定もサポートしています: ```json { "bot_type": "openai", - "model": "glm-5", + "model": "glm-5-turbo", "open_ai_api_base": "https://open.bigmodel.cn/api/paas/v4", "open_ai_api_key": "YOUR_API_KEY" } diff --git a/docs/ja/models/index.mdx b/docs/ja/models/index.mdx index 8861641..8d79416 100644 --- a/docs/ja/models/index.mdx +++ b/docs/ja/models/index.mdx @@ -6,7 +6,7 @@ description: CowAgentがサポートするモデルとおすすめの選択肢 CowAgentは国内外の主要なLLMをサポートしています。モデルインターフェースはプロジェクトの`models/`ディレクトリに実装されています。 - Agent モードでは、品質とコストのバランスから以下のモデルをおすすめします: MiniMax-M2.5、glm-5、kimi-k2.5、qwen3.5-plus、claude-sonnet-4-6、gemini-3.1-pro-preview + Agent モードでは、品質とコストのバランスから以下のモデルをおすすめします: MiniMax-M2.7、glm-5-turbo、kimi-k2.5、qwen3.5-plus、claude-sonnet-4-6、gemini-3.1-pro-preview ## 設定 @@ -19,10 +19,10 @@ CowAgentは国内外の主要なLLMをサポートしています。モデルイ - MiniMax-M2.5およびその他のシリーズモデル + MiniMax-M2.7およびその他のシリーズモデル - glm-5、glm-4.7およびその他のシリーズモデル + glm-5-turbo、glm-5およびその他のシリーズモデル qwen3.5-plus、qwen3-maxなど diff --git a/docs/ja/models/minimax.mdx b/docs/ja/models/minimax.mdx index 1b4338d..c1e7477 100644 --- a/docs/ja/models/minimax.mdx +++ b/docs/ja/models/minimax.mdx @@ -5,14 +5,14 @@ description: MiniMaxモデルの設定 ```json { - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "minimax_api_key": "YOUR_API_KEY" } ``` | パラメータ | 説明 | | --- | --- | -| `model` | `MiniMax-M2.5`、`MiniMax-M2.1`、`MiniMax-M2.1-lightning`、`MiniMax-M2`などから選択可能 | +| `model` | `MiniMax-M2.7`、`MiniMax-M2.5`、`MiniMax-M2.1`、`MiniMax-M2.1-lightning`、`MiniMax-M2`などから選択可能 | | `minimax_api_key` | [MiniMax Console](https://platform.minimaxi.com/user-center/basic-information/interface-key)で作成 | OpenAI互換の設定もサポートしています: @@ -20,7 +20,7 @@ OpenAI互換の設定もサポートしています: ```json { "bot_type": "openai", - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "open_ai_api_base": "https://api.minimaxi.com/v1", "open_ai_api_key": "YOUR_API_KEY" } diff --git a/docs/models/glm.mdx b/docs/models/glm.mdx index 1d37230..fe10f34 100644 --- a/docs/models/glm.mdx +++ b/docs/models/glm.mdx @@ -5,14 +5,14 @@ description: 智谱AI GLM 模型配置 ```json { - "model": "glm-5", + "model": "glm-5-turbo", "zhipu_ai_api_key": "YOUR_API_KEY" } ``` | 参数 | 说明 | | --- | --- | -| `model` | 可填 `glm-5`、`glm-4.7`、`glm-4-plus`、`glm-4-flash`、`glm-4-air` 等,参考 [模型编码](https://bigmodel.cn/dev/api/normal-model/glm-4) | +| `model` | 可填 `glm-5-turbo`、`glm-5`、`glm-4.7`、`glm-4-plus`、`glm-4-flash`、`glm-4-air` 等,参考 [模型编码](https://bigmodel.cn/dev/api/normal-model/glm-4) | | `zhipu_ai_api_key` | 在 [智谱AI 控制台](https://www.bigmodel.cn/usercenter/proj-mgmt/apikeys) 创建 | 也支持 OpenAI 兼容方式接入: @@ -20,7 +20,7 @@ description: 智谱AI GLM 模型配置 ```json { "bot_type": "openai", - "model": "glm-5", + "model": "glm-5-turbo", "open_ai_api_base": "https://open.bigmodel.cn/api/paas/v4", "open_ai_api_key": "YOUR_API_KEY" } diff --git a/docs/models/index.mdx b/docs/models/index.mdx index 76827f8..6f20d1b 100644 --- a/docs/models/index.mdx +++ b/docs/models/index.mdx @@ -6,7 +6,7 @@ description: CowAgent 支持的模型及推荐选择 CowAgent 支持国内外主流厂商的大语言模型,模型接口实现在项目的 `models/` 目录下。 - Agent 模式下推荐使用以下模型,可根据效果及成本综合选择:MiniMax-M2.5、glm-5、kimi-k2.5、qwen3.5-plus、claude-sonnet-4-6、gemini-3.1-pro-preview + Agent 模式下推荐使用以下模型,可根据效果及成本综合选择:MiniMax-M2.7、glm-5-turbo、kimi-k2.5、qwen3.5-plus、claude-sonnet-4-6、gemini-3.1-pro-preview ## 配置方式 @@ -23,10 +23,10 @@ CowAgent 支持国内外主流厂商的大语言模型,模型接口实现在 - MiniMax-M2.5 等系列模型 + MiniMax-M2.7 等系列模型 - glm-5、glm-4.7 等系列模型 + glm-5-turbo、glm-5 等系列模型 qwen3.5-plus、qwen3-max 等 diff --git a/docs/models/minimax.mdx b/docs/models/minimax.mdx index b842f88..299a706 100644 --- a/docs/models/minimax.mdx +++ b/docs/models/minimax.mdx @@ -5,14 +5,14 @@ description: MiniMax 模型配置 ```json { - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "minimax_api_key": "YOUR_API_KEY" } ``` | 参数 | 说明 | | --- | --- | -| `model` | 可填 `MiniMax-M2.5`、`MiniMax-M2.1`、`MiniMax-M2.1-lightning`、`MiniMax-M2` 等 | +| `model` | 可填 `MiniMax-M2.7`、`MiniMax-M2.5`、`MiniMax-M2.1`、`MiniMax-M2.1-lightning`、`MiniMax-M2` 等 | | `minimax_api_key` | 在 [MiniMax 控制台](https://platform.minimaxi.com/user-center/basic-information/interface-key) 创建 | 也支持 OpenAI 兼容方式接入: @@ -20,7 +20,7 @@ description: MiniMax 模型配置 ```json { "bot_type": "openai", - "model": "MiniMax-M2.5", + "model": "MiniMax-M2.7", "open_ai_api_base": "https://api.minimaxi.com/v1", "open_ai_api_key": "YOUR_API_KEY" } diff --git a/run.sh b/run.sh index 34ac2bf..bbb1217 100644 --- a/run.sh +++ b/run.sh @@ -270,8 +270,8 @@ select_model() { echo -e "${CYAN}${BOLD}=========================================${NC}" echo -e "${CYAN}${BOLD} Select AI Model${NC}" echo -e "${CYAN}${BOLD}=========================================${NC}" - echo -e "${YELLOW}1) MiniMax (MiniMax-M2.5, MiniMax-M2.1, etc.)${NC}" - echo -e "${YELLOW}2) Zhipu AI (glm-5, glm-4.7, etc.)${NC}" + echo -e "${YELLOW}1) MiniMax (MiniMax-M2.7, MiniMax-M2.5, etc.)${NC}" + echo -e "${YELLOW}2) Zhipu AI (glm-5-turbo, glm-5, etc.)${NC}" echo -e "${YELLOW}3) Kimi (kimi-k2.5, kimi-k2, etc.)${NC}" echo -e "${YELLOW}4) Doubao (doubao-seed-2-0-code-preview-260215, etc.)${NC}" echo -e "${YELLOW}5) Qwen (qwen3.5-plus, qwen3-max, qwq-plus, etc.)${NC}" @@ -302,8 +302,8 @@ configure_model() { # MiniMax echo -e "${GREEN}Configuring MiniMax...${NC}" read -p "Enter MiniMax API Key: " minimax_key - read -p "Enter model name [press Enter for default: MiniMax-M2.5]: " model_name - model_name=${model_name:-MiniMax-M2.5} + read -p "Enter model name [press Enter for default: MiniMax-M2.7]: " model_name + model_name=${model_name:-MiniMax-M2.7} MODEL_NAME="$model_name" MINIMAX_KEY="$minimax_key" @@ -312,8 +312,8 @@ configure_model() { # Zhipu AI echo -e "${GREEN}Configuring Zhipu AI...${NC}" read -p "Enter Zhipu AI API Key: " zhipu_key - read -p "Enter model name [press Enter for default: glm-5]: " model_name - model_name=${model_name:-glm-5} + read -p "Enter model name [press Enter for default: glm-5-turbo]: " model_name + model_name=${model_name:-glm-5-turbo} MODEL_NAME="$model_name" ZHIPU_KEY="$zhipu_key" @@ -391,8 +391,8 @@ configure_model() { # LinkAI echo -e "${GREEN}Configuring LinkAI...${NC}" read -p "Enter LinkAI API Key: " linkai_key - read -p "Enter model name [press Enter for default: MiniMax-M2.5]: " model_name - model_name=${model_name:-MiniMax-M2.5} + read -p "Enter model name [press Enter for default: MiniMax-M2.7]: " model_name + model_name=${model_name:-MiniMax-M2.7} MODEL_NAME="$model_name" USE_LINKAI="true"