DeepChat logo

DeepChat

by ThinkInAIXYZ

DeepChat is a versatile assistant that supports multiple model cloud services and local model deployment. It offers multi-channel chat concurrency and supports various platforms like Windows, macOS, and Linux.

View on GitHub

Last updated: N/A

<p align='center'> <img src='./build/icon.png' width="150" height="150" alt="logo" /> </p> <h1 align="center">DeepChat</h1> <p align="center">Dolphins are good friends of whales, and DeepChat is your good assistant</p> <div align="center"> <a href="./README.zh.md">δΈ­ζ–‡</a> / English </div>

Reasoning

<p align='center'> <img src='./build/screen.jpg'/> </p>

Search

<p align='center'> <img src='./build/screen.search.jpg'/> </p>

Latex

<p align='center'> <img src='./build/screen.latex.jpg'/> </p>

Artifacts support

<p align='center'> <img src='./build/screen.artifacts.jpg'/> </p>

Main Features

  • 🌐 Supports multiple model cloud services: DeepSeek, OpenAI, Silicon Flow, etc.
  • 🏠 Supports local model deployment: Ollama
  • πŸš€ Multi-channel chat concurrency support, switch to other conversations without waiting for the model to finish generating, efficiency Max
  • πŸ’» Supports multiple platforms: Windows, macOS, Linux
  • πŸ“„ Complete Markdown rendering, excellent code module rendering
  • 🌟 Easy to use, with a complete guide page, you can get started immediately without understanding complex concepts

Currently Supported Model Providers

<table> <tr align="center"> <td> <img src="./src/renderer/src/assets/llm-icons/ollama.svg" width="50" height="50"><br/> <a href="https://ollama.com">Ollama</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/deepseek-color.svg" width="50" height="50"><br/> <a href="https://deepseek.com/">Deepseek</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/siliconcloud.svg" width="50" height="50"><br/> <a href="https://www.siliconflow.cn/">Silicon</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/qwen-color.svg" width="50" height="50"><br/> <a href="https://chat.qwenlm.ai">QwenLM</a> </td> </tr> <tr align="center"> <td> <img src="./src/renderer/src/assets/llm-icons/doubao-color.svg" width="50" height="50"><br/> <a href="https://console.volcengine.com/ark/">Doubao</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/minimax-color.svg" width="50" height="50"><br/> <a href="https://platform.minimaxi.com/">MiniMax</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/fireworks-color.svg" width="50" height="50"><br/> <a href="https://fireworks.ai/">Fireworks</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/ppio-color.svg" width="50" height="50"><br/> <a href="https://ppinfra.com/">PPIO</a> </td> </tr> <tr align="center"> <td> <img src="./src/renderer/src/assets/llm-icons/openai.svg" width="50" height="50"><br/> <a href="https://openai.com/">OpenAI</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/gemini-color.svg" width="50" height="50"><br/> <a href="https://gemini.google.com/">Gemini</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/github.svg" width="50" height="50"><br/> <a href="https://github.com/marketplace/models">GitHub Models</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/moonshot.svg" width="50" height="50"><br/> <a href="https://moonshot.ai/">Moonshot</a> </td> </tr> <tr align="center"> <td> <img src="./src/renderer/src/assets/llm-icons/openrouter.svg" width="50" height="50"><br/> <a href="https://openrouter.ai/">OpenRouter</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/azure-color.svg" width="50" height="50"><br/> <a href="https://azure.microsoft.com/en-us/products/ai-services/openai-service">Azure OpenAI</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/qiniu.svg" width="50" height="50"><br/> <a href="https://www.qiniu.com/products/ai-token-api">Qiniu</a> </td> <td> <img src="./src/renderer/src/assets/llm-icons/grok.svg" width="50" height="50"><br/> <a href="https://x.ai/">Grok</a> </td> </tr> </table>

Compatible with any model provider in OpenAI/Gemini/Anthropic API format

Other Features

  • Support for local model management with Ollama
  • Support for local file processing
  • Artifacts support
  • Customizable search engines (parsed through models, no API adaptation required)
  • MCP support (built-in npx, no additional node environment installation needed)
  • Support for multimodality models
  • Local chat data backup and recovery
  • Compatibility with any model provider in OpenAI, Gemini, and Anthropic API formats

Development

Please read the Contribution Guidelines Windows and Linux are packaged by GitHub Action. For Mac-related signing and packaging, please refer to the Mac Release Guide.

Install dependencies

$ npm install
$ npm run installRuntime
# if got err: No module named 'distutils'
$ pip install setuptools
# for windows x64
$ npm install --cpu=x64 --os=win32 sharp
# for mac apple silicon
$ npm install --cpu=arm64 --os=darwin sharp
# for mac intel
$ npm install --cpu=x64 --os=darwin sharp
# for linux x64
$ npm install --cpu=x64 --os=linux sharp

Start development

$ npm run dev

Build

# For windows
$ npm run build:win

# For macOS
$ npm run build:mac

# For Linux
$ npm run build:linux

# Specify architecture packaging
$ npm run build:win:x64
$ npm run build:win:arm64
$ npm run build:mac:x64
$ npm run build:mac:arm64
$ npm run build:linux:x64
$ npm run build:linux:arm64

Star History

Contributors

Thank you for considering contributing to deepchat! The contribution guide can be found in the Contribution Guidelines.

<a href="https://github.com/ThinkInAIXYZ/deepchat/graphs/contributors"> <img src="https://contrib.rocks/image?repo=ThinkInAIXYZ/deepchat" /> </a>

πŸ“ƒ License

LICENSE