mcp-server-deepseek
by tizee
A Model Context Protocol (MCP) server that provides access to DeepSeek-R1's reasoning capabilities, allowing non-reasoning models to generate better responses with enhanced thinking. It acts as a bridge between LLM applications and DeepSeek's reasoning capabilities.
Last updated: N/A
mcp-server-deepseek
A Model Context Protocol (MCP) server that provides access to DeepSeek-R1's reasoning capabilities, allowing non-reasoning models to generate better responses with enhanced thinking.
Overview
This server acts as a bridge between LLM applications and DeepSeek's reasoning capabilities. It exposes DeepSeek-R1's reasoning content through an MCP tool, which can be used by any MCP-compatible client.
The server is particularly useful for:
- Enhancing responses from models without native reasoning capabilities
- Accessing DeepSeek-R1's thinking process for complex problem solving
- Adding structured reasoning to Claude or other LLMs that support MCP
Features
- Access to DeepSeek-R1: Connects to DeepSeek's API to leverage their reasoning model
- Structured Thinking: Returns reasoning in a structured
<thinking>
format - Integration with MCP: Fully compatible with the Model Context Protocol
- Error Handling: Robust error handling with detailed logging
Installation
Prerequisites
- Python 3.13 or higher
- An API key for DeepSeek
Setup
-
Clone the repository:
git clone https://github.com/yourusername/mcp-server-deepseek.git cd mcp-server-deepseek
-
Create a virtual environment:
python -m venv venv source venv/bin/activate # On Windows: venv\Scripts\activate
-
Install the package:
pip install -e .
-
Create a
.env
file with your DeepSeek API credentials:cp .env.example .env
-
Edit the
.env
file with your API key and model details:MCP_SERVER_DEEPSEEK_MODEL_NAME=deepseek-reasoner MCP_SERVER_DEEPSEEK_API_KEY=your_api_key_here MCP_SERVER_DEEPSEEK_API_BASE_URL=https://api.deepseek.com
Usage
Running the Server
You can run the server directly:
mcp-server-deepseek
Or use the development mode with the MCP Inspector:
make dev
MCP Tool
The server exposes a single tool:
think_with_deepseek_r1
This tool sends a prompt to DeepSeek-R1 and returns its reasoning content.
Arguments:
prompt
(string): The full user prompt to process
Returns:
- String containing DeepSeek-R1's reasoning wrapped in
<thinking>
tags
Example Usage
When used with Claude or another LLM that supports MCP, you can trigger the thinking process by calling the tool:
Please use the think_with_deepseek_r1 tool with the following prompt:
"How can I optimize a neural network for time series forecasting?"
Development
Testing
For development and testing, use the MCP Inspector:
npx @modelcontextprotocol/inspector uv run mcp-server-deepseek
Logging
Logs are stored in ~/.cache/mcp-server-deepseek/server.log
The log level can be configured using the LOG_LEVEL
environment variable (defaults to DEBUG
).
Troubleshooting
Common Issues
- API Key Issues: Ensure your DeepSeek API key is correctly set in the
.env
file - Timeout Errors: Complex prompts may cause timeouts. Try simplifying your prompt
- Missing Reasoning: Some queries might not generate reasoning content. Try rephrasing
Error Logs
Check the logs for detailed error messages:
cat ~/.cache/mcp-server-deepseek/server.log
License
MIT
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
Acknowledgements
- Thanks to the DeepSeek team for their powerful reasoning model
- Built with the Model Context Protocol framework