Dolphins are good friends of whales, and DeepChat is your good assistant
中文 / English
- 🌐 Supports multiple model cloud services: DeepSeek, OpenAI, Silicon Flow, etc.
- 🏠 Supports local model deployment: Ollama
- 🚀 Multi-channel chat concurrency support, switch to other conversations without waiting for the model to finish generating, efficiency Max
- 💻 Supports multiple platforms: Windows, macOS, Linux
- 📄 Complete Markdown rendering, excellent code module rendering
- 🌟 Easy to use, with a complete guide page, you can get started immediately without understanding complex concepts
Ollama |
Deepseek |
Silicon |
QwenLM |
Doubao |
MiniMax |
Fireworks |
PPIO |
OpenAI |
Gemini |
GitHub Models |
Moonshot |
OpenRouter |
Azure OpenAI |
Compatible with any model provider in openai/gemini API format |
- Support for Ollama local model management
- Support for llama.cpp local model
- Support for local file processing
- Mermaid chart visualization
- Search integration (local + cloud API)
- MCP support
- Multi-modal model support
- Local chat data synchronization and encryption
Please read the Contribution Guidelines
$ npm install
# for windows x64
$ npm install --cpu=x64 --os=win32 sharp
# for mac apple silicon
$ npm install --cpu=arm64 --os=darwin sharp
# for mac intel
$ npm install --cpu=x64 --os=darwin sharp
# for linux x64
$ npm install --cpu=x64 --os=linux sharp
$ npm run dev
# For windows
$ npm run build:win
# For macOS
$ npm run build:mac
# For Linux
$ npm run build:linux
# Specify architecture packaging
$ npm run build:win:x64
$ npm run build:win:arm64
$ npm run build:mac:x64
$ npm run build:mac:arm64
$ npm run build:linux:x64
$ npm run build:linux:arm64