
DeepSeek Thinker
Connects to DeepSeek's reasoning model to expose its step-by-step thinking process and chain-of-thought reasoning capabilities. Works with both cloud API and local Ollama deployments.
Integrates with DeepSeek Thinker model to enable chain-of-thought reasoning and complex problem-solving for applications requiring advanced cognitive capabilities.
What it does
- Access DeepSeek's internal reasoning process
- Generate step-by-step problem solving explanations
- Connect to DeepSeek via OpenAI-compatible API
- Run DeepSeek models locally through Ollama
- Capture and display chain-of-thought reasoning
Best for
About DeepSeek Thinker
DeepSeek Thinker is a community-built MCP server published by ruixingshi that provides AI assistants with tools and capabilities via the Model Context Protocol. Integrate DeepSeek Thinker for AI problem solving and chain-of-thought reasoning in advanced artificial intelligence app It is categorized under ai ml.
How to install
You can install DeepSeek Thinker in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
License
DeepSeek Thinker is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
Deepseek Thinker MCP Server
A MCP (Model Context Protocol) provider Deepseek reasoning content to MCP-enabled AI Clients, like Claude Desktop. Supports access to Deepseek's thought processes from the Deepseek API service or from a local Ollama server.
Core Features
-
🤖 Dual Mode Support
- OpenAI API mode support
- Ollama local mode support
-
🎯 Focused Reasoning
- Captures Deepseek's thinking process
- Provides reasoning output
Available Tools
get-deepseek-thinker
- Description: Perform reasoning using the Deepseek model
- Input Parameters:
originPrompt(string): User's original prompt
- Returns: Structured text response containing the reasoning process
Environment Configuration
OpenAI API Mode
Set the following environment variables:
API_KEY=<Your OpenAI API Key>
BASE_URL=<API Base URL>
Ollama Mode
Set the following environment variable:
USE_OLLAMA=true
Usage
Integration with AI Client, like Claude Desktop
Add the following configuration to your claude_desktop_config.json:
{
"mcpServers": {
"deepseek-thinker": {
"command": "npx",
"args": [
"-y",
"deepseek-thinker-mcp"
],
"env": {
"API_KEY": "<Your API Key>",
"BASE_URL": "<Your Base URL>"
}
}
}
}
Using Ollama Mode
{
"mcpServers": {
"deepseek-thinker": {
"command": "npx",
"args": [
"-y",
"deepseek-thinker-mcp"
],
"env": {
"USE_OLLAMA": "true"
}
}
}
}
Local Server Configuration
{
"mcpServers": {
"deepseek-thinker": {
"command": "node",
"args": [
"/your-path/deepseek-thinker-mcp/build/index.js"
],
"env": {
"API_KEY": "<Your API Key>",
"BASE_URL": "<Your Base URL>"
}
}
}
}
Development Setup
# Install dependencies
npm install
# Build project
npm run build
# Run service
node build/index.js
FAQ
Response like this: "MCP error -32001: Request timed out"
This error occurs when the Deepseek API response is too slow or when the reasoning content output is too long, causing the MCP server to timeout.
Tech Stack
- TypeScript
- @modelcontextprotocol/sdk
- OpenAI API
- Ollama
- Zod (parameter validation)
License
This project is licensed under the MIT License. See the LICENSE file for details.
Alternatives
Related Skills
Browse all skillsAdvanced context management with auto-compaction and dynamic context optimization for DeepSeek's 64k context window. Features intelligent compaction (merging, summarizing, extracting), query-aware relevance scoring, and hierarchical memory system with context archive. Logs optimization events to chat.
Smart LLM router — save 78% on inference costs. Routes every request to the cheapest capable model across 30+ models from OpenAI, Anthropic, Google, DeepSeek, and xAI.
This skill should be used when configuring or using the OpenCode CLI for headless LLM automation. Use when the user asks to "configure opencode", "use opencode cli", "set up opencode", "opencode run command", "opencode model selection", "opencode providers", "opencode vertex ai", "opencode mcp servers", "opencode ollama", "opencode local models", "opencode deepseek", "opencode kimi", "opencode mistral", "fallback cli tool", or "headless llm cli". Covers command syntax, provider configuration, Vertex AI setup, MCP servers, local models, cloud providers, and subprocess integration patterns.
Provides PyTorch-native distributed LLM pretraining using torchtitan with 4D parallelism (FSDP2, TP, PP, CP). Use when pretraining Llama 3.1, DeepSeek V3, or custom models at scale from 8 to 512+ GPUs with Float8, torch.compile, and distributed checkpointing.
为 OpenClaw 配置硅基流动 (SiliconFlow) 作为模型源。SiliconFlow 是国内领先的 AI 模型推理平台,提供 98+ 个 chat 模型,包含多个免费模型(Qwen3-8B、DeepSeek-R1-8B 等)。使用标准 OpenAI 协议(openai-completions)。包含 provider 注册、模型定义、别名配置、fallback 链接入和验证的完整流程。当管理员说想"加硅基流动"、"配 SiliconFlow"、"接入 SF 模型"、"加 Kimi"、"加 Qwen3"、"加免费模型"、"接入 DeepSeek V3.2"时使用此 skill。
Use this skill when developing Node.js backend services or CloudBase cloud functions (Express/Koa/NestJS, serverless, backend APIs) that need AI capabilities. Features text generation (generateText), streaming (streamText), AND image generation (generateImage) via @cloudbase/node-sdk ≥3.16.0. Built-in models include Hunyuan (hunyuan-2.0-instruct-20251111 recommended), DeepSeek (deepseek-v3.2 recommended), and hunyuan-image for images. This is the ONLY SDK that supports image generation. NOT for browser/Web apps (use ai-model-web) or WeChat Mini Program (use ai-model-wechat).