10 KiB
CowAgent is an AI super assistant powered by LLMs, capable of autonomous task planning, operating computers and external resources, creating and executing Skills, and continuously growing with long-term memory. It supports flexible model switching, handles text, voice, images, and files, and can be integrated into Web, Feishu, DingTalk, WeCom Bot, WeCom App, and WeChat Official Account — running 7×24 hours on your personal computer or server.
🌐 Website · 📖 Docs · 🚀 Quick Start · ☁️ Try Online
Introduction
CowAgent is both an out-of-the-box AI super assistant and a highly extensible Agent framework. You can extend it with new model interfaces, channels, built-in tools, and the Skills system to flexibly implement various customization needs.
- ✅ Autonomous Task Planning: Understands complex tasks and autonomously plans execution, continuously thinking and invoking tools until goals are achieved. Supports accessing files, terminal, browser, schedulers, and other system resources via tools.
- ✅ Long-term Memory: Automatically persists conversation memory to local files and databases, including core memory and daily memory, with keyword and vector retrieval support.
- ✅ Skills System: Implements a Skills creation and execution engine with multiple built-in skills, and supports custom Skills development through natural language conversation.
- ✅ Multimodal Messages: Supports parsing, processing, generating, and sending text, images, voice, files, and other message types.
- ✅ Multiple Model Support: Supports OpenAI, Claude, Gemini, DeepSeek, MiniMax, GLM, Qwen, Kimi, Doubao, and other mainstream model providers.
- ✅ Multi-platform Deployment: Runs on local computers or servers, integrable into Web, Feishu, DingTalk, WeChat Official Account, and WeCom applications.
- ✅ Knowledge Base: Integrates enterprise knowledge base capabilities via the LinkAI platform.
Disclaimer
- This project follows the MIT License and is intended for technical research and learning. Users must comply with local laws, regulations, policies, and corporate bylaws. Any illegal or rights-infringing use is prohibited.
- Agent mode consumes more tokens than normal chat mode. Choose models based on effectiveness and cost. Agent has access to the host OS — please deploy in trusted environments.
- CowAgent focuses on open-source development and does not participate in, authorize, or issue any cryptocurrency.
Demo
Try online (no deployment needed): CowAgent
Changelog
2026.02.27: v2.0.2 — Web console overhaul (streaming chat, model/skill/memory/channel/scheduler/log management), multi-channel concurrent running, session persistence, new models including Gemini 3.1 Pro / Claude 4.6 Sonnet / Qwen3.5 Plus.
2026.02.13: v2.0.1 — Built-in Web Search tool, smart context trimming, runtime info dynamic update, Windows compatibility, fixes for scheduler memory loss, Feishu connection issues, and more.
2026.02.03: v2.0.0 — Full upgrade to AI super assistant with multi-step task planning, long-term memory, built-in tools, Skills framework, new models, and optimized channels.
2025.05.23: v1.7.6 — Web channel optimization, AgentMesh multi-agent plugin, Baidu TTS, claude-4-sonnet/opus support.
2025.04.11: v1.7.5 — wechatferry protocol, DeepSeek model, Tencent Cloud voice, ModelScope and Gitee-AI support.
2024.12.13: v1.7.4 — Gemini 2.0 model, Web channel, memory leak fix.
Full changelog: Release Notes
🚀 Quick Start
The project provides a one-click script for installation, configuration, startup, and management:
bash <(curl -fsSL https://cdn.link-ai.tech/code/cow/run.sh)
After running, the Web service starts by default. Access http://localhost:9899/chat to chat.
Script usage: One-click Install
Manual Installation
1. Clone the project
git clone https://github.com/zhayujie/chatgpt-on-wechat
cd chatgpt-on-wechat/
2. Install dependencies
pip3 install -r requirements.txt
pip3 install -r requirements-optional.txt # optional but recommended
3. Configure
cp config-template.json config.json
Fill in your model API key and channel type in config.json. See the configuration docs for details.
4. Run
python3 app.py
For server background run:
nohup python3 app.py & tail -f nohup.out
Docker Deployment
curl -O https://cdn.link-ai.tech/code/cow/docker-compose.yml
# Edit docker-compose.yml with your config
sudo docker compose up -d
sudo docker logs -f chatgpt-on-wechat
Models
Supports mainstream model providers. Recommended models for Agent mode:
| Provider | Recommended Model |
|---|---|
| MiniMax | MiniMax-M2.5 |
| GLM | glm-5 |
| Kimi | kimi-k2.5 |
| Doubao | doubao-seed-2-0-code-preview-260215 |
| Qwen | qwen3.5-plus |
| Claude | claude-sonnet-4-6 |
| Gemini | gemini-3.1-pro-preview |
| OpenAI | gpt-5.4 |
| DeepSeek | deepseek-chat |
For detailed configuration of each model, see the Models documentation.
Coding Plan
Coding Plan is a monthly subscription package offered by various providers, ideal for high-frequency Agent usage. All providers can be accessed via OpenAI-compatible mode:
{
"bot_type": "openai",
"model": "MODEL_NAME",
"open_ai_api_base": "PROVIDER_CODING_PLAN_API_BASE",
"open_ai_api_key": "YOUR_API_KEY"
}
bot_type: Must beopenaimodel: Model name supported by the provideropen_ai_api_base: Provider's Coding Plan API Base (different from standard pay-as-you-go)open_ai_api_key: Provider's Coding Plan API Key
Note: Coding Plan API Base and API Key are usually separate from standard pay-as-you-go ones. Please obtain them from each provider's platform.
Supported providers include Alibaba Cloud, MiniMax, Zhipu GLM, Kimi, Volcengine, and more. For detailed configuration of each provider, see the Coding Plan documentation.
Channels
Supports multiple platforms. Set channel_type in config.json to switch:
| Channel | channel_type |
Docs |
|---|---|---|
| Web (default) | web |
Web Channel |
| Feishu | feishu |
Feishu Setup |
| DingTalk | dingtalk |
DingTalk Setup |
| WeCom Bot | wecom_bot |
WeCom Bot Setup |
| WeCom App | wechatcom_app |
WeCom Setup |
| WeChat MP | wechatmp / wechatmp_service |
WeChat MP Setup |
| Terminal | terminal |
— |
Multiple channels can be enabled simultaneously, separated by commas: "channel_type": "feishu,dingtalk".
Enterprise Services
LinkAI is a one-stop AI agent platform for enterprises and developers, integrating multimodal LLMs, knowledge bases, Agent plugins, and workflows. Supports one-click integration with mainstream platforms, SaaS and private deployment.
🔗 Related Projects
- bot-on-anything: Lightweight and highly extensible LLM application framework supporting Slack, Telegram, Discord, Gmail, and more.
- AgentMesh: Open-source Multi-Agent framework for complex problem solving through agent team collaboration.
🔎 FAQ
FAQs: https://github.com/zhayujie/chatgpt-on-wechat/wiki/FAQs
🛠️ Contributing
Welcome to add new channels, referring to the Feishu channel as an example. Also welcome to contribute new Skills, referring to the Skill Creator docs.
✉ Contact
Welcome to submit PRs and Issues, and support the project with a 🌟 Star. For questions, check the FAQ list or search Issues.
