Executes SQL queries and analyzes data in DuckDB databases through a single unified query interface. Supports both read-only and read-write modes for local data analysis.

Execute SQL queries and analyze data in DuckDB databases.

173581 views22Local (stdio)

What it does

  • Execute SQL queries on DuckDB databases
  • Create and modify database tables
  • Inspect database schemas
  • Analyze local data files
  • Perform joins and aggregations
  • Run data transformations

Best for

Data analysts working with local datasetsDevelopers prototyping SQL queriesLocal data exploration and analysisETL pipeline development
Single unified query interfaceRead-only mode for data protectionLocal analysis optimized

About DuckDB

DuckDB is a community-built MCP server published by ktanaka101 that provides AI assistants with tools and capabilities via the Model Context Protocol. Execute SQL queries and analyze data efficiently in DuckDB databases. Unlock powerful analytics with DuckDB. It is categorized under databases, analytics data.

How to install

You can install DuckDB in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

License

DuckDB is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

mcp-server-duckdb

PyPI - Version PyPI - License smithery badge

A Model Context Protocol (MCP) server implementation for DuckDB, providing database interaction capabilities through MCP tools. It would be interesting to have LLM analyze it. DuckDB is suitable for local analysis.

mcp-server-duckdb MCP server

Overview

This server enables interaction with a DuckDB database through the Model Context Protocol, allowing for database operations like querying, table creation, and schema inspection.

Components

Resources

Currently, no custom resources are implemented.

Prompts

Currently, no custom prompts are implemented.

Tools

The server implements the following database interaction tool:

  • query: Execute any SQL query on the DuckDB database
    • Input: query (string) - Any valid DuckDB SQL statement
    • Output: Query results as text (or success message for operations like CREATE/INSERT)

[!NOTE] The server provides a single unified query function rather than separate specialized functions, as modern LLMs can generate appropriate SQL for any database operation (SELECT, CREATE TABLE, JOIN, etc.) without requiring separate endpoints.

[!NOTE] When the server is running in readonly mode, DuckDB's native readonly protection is enforced. This ensures that the Language Model (LLM) cannot perform any write operations (CREATE, INSERT, UPDATE, DELETE), maintaining data integrity and preventing unintended changes.

Configuration

Required Parameters

  • db-path (string): Path to the DuckDB database file
    • The server will automatically create the database file and parent directories if they don't exist
    • If --readonly is specified and the database file doesn't exist, the server will fail to start with an error

Optional Parameters

  • --readonly: Run server in read-only mode (default: false)
    • Description: When this flag is set, the server operates in read-only mode. This means:
      • The DuckDB database will be opened with read_only=True, preventing any write operations.
      • If the specified database file does not exist, it will not be created.
      • Security Benefit: Prevents the Language Model (LLM) from performing any write operations, ensuring that the database remains unaltered.
    • Reference: For more details on read-only connections in DuckDB, see the DuckDB Python API documentation.
  • --keep-connection: Re-uses a single DuckDB connection mode (default: false)
    • Description: When this flag is set, Re-uses a single DuckDB connection for the entire server lifetime. Enables TEMP objects & slightly faster queries, but can hold an exclusive lock on the file.

Installation

Installing via Smithery

To install DuckDB Server for Claude Desktop automatically via Smithery:

npx -y @smithery/cli install mcp-server-duckdb --client claude

Claude Desktop Integration

Configure the MCP server in Claude Desktop's configuration file:

MacOS

Location: ~/Library/Application Support/Claude/claude_desktop_config.json

Windows

Location: %APPDATA%/Claude/claude_desktop_config.json

{
  "mcpServers": {
    "duckdb": {
      "command": "uvx",
      "args": [
        "mcp-server-duckdb",
        "--db-path",
        "~/mcp-server-duckdb/data/data.db"
      ]
    }
  }
}
  • Note: ~/mcp-server-duckdb/data/data.db should be replaced with the actual path to the DuckDB database file.

Development

Prerequisites

  • Python with uv package manager
  • DuckDB Python package
  • MCP server dependencies

Debugging

Debugging MCP servers can be challenging due to their stdio-based communication. We recommend using the MCP Inspector for the best debugging experience.

Using MCP Inspector

  1. Install the inspector using npm:
npx @modelcontextprotocol/inspector uv --directory ~/codes/mcp-server-duckdb run mcp-server-duckdb --db-path ~/mcp-server-duckdb/data/data.db
  1. Open the provided URL in your browser to access the debugging interface

The inspector provides visibility into:

  • Request/response communication
  • Tool execution
  • Server state
  • Error messages

Alternatives

Related Skills

Browse all skills
sql-splitter

High-performance CLI for working with SQL dump files: split/merge by table, analyze contents, validate integrity, convert between MySQL/PostgreSQL/SQLite/MSSQL, create FK-safe samples, shard multi-tenant dumps, generate ERD diagrams, reorder for safe imports, and run SQL analytics with embedded DuckDB. Use when working with .sql dump files for migrations, dev seeding, CI validation, schema visualization, data extraction, or ad-hoc analytics.

0
literature-review

Conduct comprehensive, systematic literature reviews using multiple academic databases (PubMed, arXiv, bioRxiv, Semantic Scholar, etc.). This skill should be used when conducting systematic literature reviews, meta-analyses, research synthesis, or comprehensive literature searches across biomedical, scientific, and technical domains. Creates professionally formatted markdown documents and PDFs with verified citations in multiple citation styles (APA, Nature, Vancouver, etc.).

377
postgresql-psql

Comprehensive guide for PostgreSQL psql - the interactive terminal client for PostgreSQL. Use when connecting to PostgreSQL databases, executing queries, managing databases/tables, configuring connection options, formatting output, writing scripts, managing transactions, and using advanced psql features for database administration and development.

38
data-storytelling

Transform data into compelling narratives using visualization, context, and persuasive structure. Use when presenting analytics to stakeholders, creating data reports, or building executive presentations.

27
content-trend-researcher

Advanced content and topic research skill that analyzes trends across Google Analytics, Google Trends, Substack, Medium, Reddit, LinkedIn, X, blogs, podcasts, and YouTube to generate data-driven article outlines based on user intent analysis

23
data-scientist

Expert data scientist for advanced analytics, machine learning, and statistical modeling. Handles complex data analysis, predictive modeling, and business intelligence. Use PROACTIVELY for data analysis tasks, ML modeling, statistical analysis, and data-driven insights.

13