
n8n
Connects AI assistants to n8n workflow automation platform, allowing natural language control of workflows and execution monitoring. Acts as a bridge between AI capabilities and n8n's automation tools.
Integrates with n8n workflow automation platform to provide conversational access to 525+ nodes including AI-capable nodes and triggers, enabling natural language workflow creation, validation, and management without requiring direct platform knowledge.
What it does
- Manage n8n workflows through natural language
- Monitor workflow executions and status
- Control n8n automation tasks programmatically
- Query workflow data and results
- Execute workflows on demand
Best for
About n8n
n8n is a community-built MCP server published by czlonkowski that provides AI assistants with tools and capabilities via the Model Context Protocol. n8n offers conversational workflow automation, enabling seamless software workflow creation and management without platf It is categorized under developer tools.
How to install
You can install n8n in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
License
n8n is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
n8n-MCP
A Model Context Protocol (MCP) server that provides AI assistants with comprehensive access to n8n node documentation, properties, and operations. Deploy in minutes to give Claude and other AI assistants deep knowledge about n8n's 1,236 workflow automation nodes (806 core + 430 community).
Overview
n8n-MCP serves as a bridge between n8n's workflow automation platform and AI models, enabling them to understand and work with n8n nodes effectively. It provides structured access to:
- π 1,084 n8n nodes - 537 core nodes + 547 community nodes (301 verified)
- π§ Node properties - 99% coverage with detailed schemas
- β‘ Node operations - 63.6% coverage of available actions
- π Documentation - 87% coverage from official n8n docs (including AI nodes)
- π€ AI tools - 265 AI-capable tool variants detected with full documentation
- π‘ Real-world examples - 2,646 pre-extracted configurations from popular templates
- π― Template library - 2,709 workflow templates with 100% metadata coverage
- π Community nodes - Search verified community integrations with
sourcefilter (NEW!)
β οΈ Important Safety Warning
NEVER edit your production workflows directly with AI! Always:
- π Make a copy of your workflow before using AI tools
- π§ͺ Test in development environment first
- πΎ Export backups of important workflows
- β‘ Validate changes before deploying to production
AI results can be unpredictable. Protect your work!
π Quick Start
Option 1: Hosted Service (Easiest - No Setup!) βοΈ
The fastest way to try n8n-MCP - no installation, no configuration:
- β Free tier: 100 tool calls/day
- β Instant access: Start building workflows immediately
- β Always up-to-date: Latest n8n nodes and templates
- β No infrastructure: We handle everything
Just sign up, get your API key, and connect your MCP client.
π Self-Hosting Options
Prefer to run n8n-MCP yourself? Choose your deployment method:
Option A: npx (Quick Local Setup) π
Get n8n-MCP running in minutes:
Prerequisites: Node.js installed on your system
# Run directly with npx (no installation needed!)
npx n8n-mcp
Add to Claude Desktop config:
β οΈ Important: The
MCP_MODE: "stdio"environment variable is required for Claude Desktop. Without it, you will see JSON parsing errors like"Unexpected token..."in the UI. This variable ensures that only JSON-RPC messages are sent to stdout, preventing debug logs from interfering with the protocol.
Basic configuration (documentation tools only):
{
"mcpServers": {
"n8n-mcp": {
"command": "npx",
"args": ["n8n-mcp"],
"env": {
"MCP_MODE": "stdio",
"LOG_LEVEL": "error",
"DISABLE_CONSOLE_OUTPUT": "true"
}
}
}
}
Full configuration (with n8n management tools):
{
"mcpServers": {
"n8n-mcp": {
"command": "npx",
"args": ["n8n-mcp"],
"env": {
"MCP_MODE": "stdio",
"LOG_LEVEL": "error",
"DISABLE_CONSOLE_OUTPUT": "true",
"N8N_API_URL": "https://your-n8n-instance.com",
"N8N_API_KEY": "your-api-key"
}
}
}
}
Note: npx will download and run the latest version automatically. The package includes a pre-built database with all n8n node information.
Configuration file locations:
- macOS:
~/Library/Application Support/Claude/claude_desktop_config.json - Windows:
%APPDATA%\Claude\claude_desktop_config.json - Linux:
~/.config/Claude/claude_desktop_config.json
Restart Claude Desktop after updating configuration - That's it! π
Option B: Docker (Isolated & Reproducible) π³
Prerequisites: Docker installed on your system
π¦ Install Docker (click to expand)
macOS:
# Using Homebrew
brew install --cask docker
# Or download from https://www.docker.com/products/docker-desktop/
Linux (Ubuntu/Debian):
# Update package index
sudo apt-get update
# Install Docker
sudo apt-get install docker.io
# Start Docker service
sudo systemctl start docker
sudo systemctl enable docker
# Add your user to docker group (optional, to run without sudo)
sudo usermod -aG docker $USER
# Log out and back in for this to take effect
Windows:
# Option 1: Using winget (Windows Package Manager)
winget install Docker.DockerDesktop
# Option 2: Using Chocolatey
choco install docker-desktop
# Option 3: Download installer from https://www.docker.com/products/docker-desktop/
Verify installation:
docker --version
# Pull the Docker image (~280MB, no n8n dependencies!)
docker pull ghcr.io/czlonkowski/n8n-mcp:latest
β‘ Ultra-optimized: Our Docker image is 82% smaller than typical n8n images because it contains NO n8n dependencies - just the runtime MCP server with a pre-built database!
Add to Claude Desktop config:
Basic configuration (documentation tools only):
{
"mcpServers": {
"n8n-mcp": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"--init",
"-e", "MCP_MODE=stdio",
"-e", "LOG_LEVEL=error",
"-e", "DISABLE_CONSOLE_OUTPUT=true",
"ghcr.io/czlonkowski/n8n-mcp:latest"
]
}
}
}
Full configuration (with n8n management tools):
{
"mcpServers": {
"n8n-mcp": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"--init",
"-e", "MCP_MODE=stdio",
"-e", "LOG_LEVEL=error",
"-e", "DISABLE_CONSOLE_OUTPUT=true",
"-e", "N8N_API_URL=https://your-n8n-instance.com",
"-e", "N8N_API_KEY=your-api-key",
"ghcr.io/czlonkowski/n8n-mcp:latest"
]
}
}
}
π‘ Tip: If you're running n8n locally on the same machine (e.g., via Docker), use http://host.docker.internal:5678 as the N8N_API_URL.
Note: The n8n API credentials are optional. Without them, you'll have access to all documentation and validation tools. With them, you'll additionally get workflow management capabilities (create, update, execute workflows).
π Local n8n Instance Configuration
If you're running n8n locally (e.g., http://localhost:5678 or Docker), you need to allow localhost webhooks:
{
"mcpServers": {
"n8n-mcp": {
"command": "docker",
"args": [
"run", "-i", "--rm", "--init",
"-e", "MCP_MODE=stdio",
"-e", "LOG_LEVEL=error",
"-e", "DISABLE_CONSOLE_OUTPUT=true",
"-e", "N8N_API_URL=http://host.docker.internal:5678",
"-e", "N8N_API_KEY=your-api-key",
"-e", "WEBHOOK_SECURITY_MODE=moderate",
"ghcr.io/czlonkowski/n8n-mcp:latest"
]
}
}
}
β οΈ Important: Set
WEBHOOK_SECURITY_MODE=moderateto allow webhooks to your local n8n instance. This is safe for local development while still blocking private networks and cloud metadata.
Important: The -i flag is required for MCP stdio communication.
π§ If you encounter any issues with Docker, check our Docker Troubleshooting Guide.
Configuration file locations:
- macOS:
~/Library/Application Support/Claude/claude_desktop_config.json - Windows:
%APPDATA%\Claude\claude_desktop_config.json - Linux:
~/.config/Claude/claude_desktop_config.json
Restart Claude Desktop after updating configuration - That's it! π
π Privacy & Telemetry
n8n-mcp collects anonymous usage statistics to improve the tool. View our privacy policy.
Opting Out
For npx users:
npx n8n-mcp telemetry disable
For Docker users: Add the following environment variable to your Docker configuration:
"-e", "N8N_MCP_TELEMETRY_DISABLED=true"
Example in Claude Desktop config:
{
"mcpServers": {
"n8n-mcp": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"--init",
"-e", "MCP_MODE=stdio",
"-e", "LOG_LEVEL=error",
"-e", "N8N_MCP_TELEMETRY_DISABLED=true",
"ghcr.io/czlonkowski/n8n-mcp:latest"
]
}
}
}
For docker-compose users: Set in your environment file or docker-compose.yml:
environment:
N8N_MCP_TELEMETRY_DISABLED: "true"
βοΈ Database & Memory Configuration
Database Adapters
n8n-mcp uses SQLite for storing node documentation. Two adapters are available:
-
better-sqlite3 (Default in Docker)
- Native C++ bindings for best performance
- Direct disk writes (no memory overhead)
- Now enabled by default in Docker images (v2.20.2+)
- Memory usage: ~100-120 MB stable
-
sql.js (Fallback)
- Pure JavaScript
README truncated. View full README on GitHub.
Alternatives
Related Skills
Browse all skillsUI design system toolkit for Senior UI Designer including design token generation, component documentation, responsive design calculations, and developer handoff tools. Use for creating design systems, maintaining visual consistency, and facilitating design-dev collaboration.
Expert MCP (Model Context Protocol) orchestration with n8n workflow automation. Master bidirectional MCP integration, expose n8n workflows as AI agent tools, consume MCP servers in workflows, build agentic systems, orchestrate multi-agent workflows, and create production-ready AI-powered automation pipelines with Claude Code integration.
Expert guide for using n8n-mcp MCP tools effectively. Use when searching for nodes, validating configurations, accessing templates, managing workflows, or using any n8n-mcp tool. Provides tool selection guidance, parameter formats, and common patterns.
Answer questions about the AI SDK and help build AI-powered features. Use when developers: (1) Ask about AI SDK functions like generateText, streamText, ToolLoopAgent, embed, or tools, (2) Want to build AI agents, chatbots, RAG systems, or text generation features, (3) Have questions about AI providers (OpenAI, Anthropic, Google, etc.), streaming, tool calling, structured output, or embeddings, (4) Use React hooks like useChat or useCompletion. Triggers on: "AI SDK", "Vercel AI SDK", "generateText", "streamText", "add AI to my app", "build an agent", "tool calling", "structured output", "useChat".
Master API documentation with OpenAPI 3.1, AI-powered tools, and modern developer experience practices. Create interactive docs, generate SDKs, and build comprehensive developer portals. Use PROACTIVELY for API documentation or developer portal creation.
Use when working with the OpenAI API (Responses API) or OpenAI platform features (tools, streaming, Realtime API, auth, models, rate limits, MCP) and you need authoritative, up-to-date documentation (schemas, examples, limits, edge cases). Prefer the OpenAI Developer Documentation MCP server tools when available; otherwise guide the user to enable `openaiDeveloperDocs`.