
Unichat (TS)
Provides a unified interface to interact with multiple AI language models (OpenAI, Anthropic, MistralAI, xAI, Google AI, DeepSeek) through a single tool. Includes pre-built prompts for common code analysis tasks.
Integrates multiple language models via the unified Unichat tool, enabling seamless interaction across OpenAI, MistralAI, Anthropic, xAI, and Google AI platforms.
What it does
- Send requests to 6+ AI providers through unified interface
- Review code for best practices and issues
- Generate documentation and comments for code
- Explain how code works in detail
- Apply changes to existing code
- Switch between AI providers without changing workflow
Best for
About Unichat (TS)
Unichat (TS) is a community-built MCP server published by amidabuddha that provides AI assistants with tools and capabilities via the Model Context Protocol. Unichat: a powerful AI chat platform integrating leading models like OpenAI, Anthropic, xAI, and more into one unified c It is categorized under ai ml, developer tools.
How to install
You can install Unichat (TS) in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
License
Unichat (TS) is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
Unichat MCP Server in TypeScript
Also available in Python
Send requests to OpenAI, MistralAI, Anthropic, xAI, Google AI or DeepSeek using MCP protocol via tool or predefined prompts. Vendor API key required.
Both STDIO and SSE transport mechanisms supported via arguments.
Tools
The server implements one tool:
unichat: Send a request to unichat- Takes "messages" as required string arguments
- Returns a response
Prompts
code_review- Review code for best practices, potential issues, and improvements
- Arguments:
code(string, required): The code to review"
document_code- Generate documentation for code including docstrings and comments
- Arguments:
code(string, required): The code to comment"
explain_code- Explain how a piece of code works in detail
- Arguments:
code(string, required): The code to explain"
code_rework- Apply requested changes to the provided code
- Arguments:
changes(string, optional): The changes to apply"code(string, required): The code to rework"
Development
Install dependencies:
npm install
Build the server:
npm run build
For development with auto-rebuild:
npm run watch
Running evals
The evals package loads an mcp client that then runs the index.ts file, so there is no need to rebuild between tests. You can load environment variables by prefixing the npx command. Full documentation can be found here.
OPENAI_API_KEY=your-key npx mcp-eval src/evals/evals.ts src/server.ts
Installation
Installing via Smithery
To install Unichat MCP Server for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install unichat-ts-mcp-server --client claude
Installing manually
To use with Claude Desktop, add the server config:
On MacOS: ~/Library/Application Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
Run locally:
{
"mcpServers": {
"unichat-ts-mcp-server": {
"command": "node",
"args": [
"{{/path/to}}/unichat-ts-mcp-server/build/index.js"
],
"env": {
"UNICHAT_MODEL": "YOUR_PREFERRED_MODEL_NAME",
"UNICHAT_API_KEY": "YOUR_VENDOR_API_KEY"
}
}
}
Run published:
{
"mcpServers": {
"unichat-ts-mcp-server": {
"command": "npx",
"args": [
"-y",
"unichat-ts-mcp-server"
],
"env": {
"UNICHAT_MODEL": "YOUR_PREFERRED_MODEL_NAME",
"UNICHAT_API_KEY": "YOUR_VENDOR_API_KEY"
}
}
}
Runs in STDIO by default or with argument
--stdio. To run in SSE add argument--sse
npx -y unichat-ts-mcp-server --sse
Supported Models:
A list of currently supported models to be used as
"YOUR_PREFERRED_MODEL_NAME"may be found here. Please make sure to add the relevant vendor API key as"YOUR_VENDOR_API_KEY"
Example:
"env": {
"UNICHAT_MODEL": "gpt-4o-mini",
"UNICHAT_API_KEY": "YOUR_OPENAI_API_KEY"
}
Debugging
Since MCP servers communicate over stdio, debugging can be challenging. We recommend using the MCP Inspector, which is available as a package script:
npm run inspector
The Inspector will provide a URL to access debugging tools in your browser.
If you experience timeouts during testing in SSE mode change the request URL on the inspector interface to: http://localhost:3001/sse?timeout=600000
Alternatives
Related Skills
Browse all skillsUI design system toolkit for Senior UI Designer including design token generation, component documentation, responsive design calculations, and developer handoff tools. Use for creating design systems, maintaining visual consistency, and facilitating design-dev collaboration.
Answer questions about the AI SDK and help build AI-powered features. Use when developers: (1) Ask about AI SDK functions like generateText, streamText, ToolLoopAgent, embed, or tools, (2) Want to build AI agents, chatbots, RAG systems, or text generation features, (3) Have questions about AI providers (OpenAI, Anthropic, Google, etc.), streaming, tool calling, structured output, or embeddings, (4) Use React hooks like useChat or useCompletion. Triggers on: "AI SDK", "Vercel AI SDK", "generateText", "streamText", "add AI to my app", "build an agent", "tool calling", "structured output", "useChat".
Master API documentation with OpenAPI 3.1, AI-powered tools, and modern developer experience practices. Create interactive docs, generate SDKs, and build comprehensive developer portals. Use PROACTIVELY for API documentation or developer portal creation.
Use when working with the OpenAI API (Responses API) or OpenAI platform features (tools, streaming, Realtime API, auth, models, rate limits, MCP) and you need authoritative, up-to-date documentation (schemas, examples, limits, edge cases). Prefer the OpenAI Developer Documentation MCP server tools when available; otherwise guide the user to enable `openaiDeveloperDocs`.
Guide for building TypeScript CLIs with Bun. Use when creating command-line tools, adding subcommands to existing CLIs, or building developer tooling. Covers argument parsing, subcommand patterns, output formatting, and distribution.
Integrate Vercel AI SDK applications with You.com tools (web search, AI agent, content extraction). Use when developer mentions AI SDK, Vercel AI SDK, generateText, streamText, or You.com integration with AI SDK.
