arkheia-mcp

arkheia-mcp

arkheiaai

Real-time AI fabrication detection for Claude, GPT, Gemini & 35+ models. Behavioral fingerprinting catches hallucination

Arkheia MCP Server provides real-time fabrication detection for AI model responses using behavioral fingerprinting across 35+ models including Claude, GPT, Gemini, and others.

346 viewsLocal (stdio)

About arkheia-mcp

arkheia-mcp is a community-built MCP server published by arkheiaai that provides AI assistants with tools and capabilities via the Model Context Protocol. Real-time AI fabrication detection for Claude, GPT, Gemini & 35+ models. Behavioral fingerprinting catches hallucination It is categorized under ai ml. This server exposes 6 tools that AI clients can invoke during conversations and coding sessions.

How to install

You can install arkheia-mcp in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

License

arkheia-mcp is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

Tools (6)

arkheia_verify

Score a model response for fabrication risk with LOW/MEDIUM/HIGH risk levels

arkheia_audit_log

Review your detection history and past verification results

run_grok

Call Grok model and automatically screen the response for fabrication

run_gemini

Call Gemini model and automatically screen the response for fabrication

run_ollama

Call local Ollama model and automatically screen the response for fabrication

Arkheia MCP Server — Fabrication Detection for AI Agents

Know when your AI is making things up.

Arkheia screens model responses for fabrication using behavioural fingerprinting. Works with Claude, GPT, Gemini, Grok, Llama, Mistral, and 30+ other models. One tool call. Real-time risk scoring.

Free tier: 1,500 detections/month. No credit card.

Install

npx @arkheia/mcp-server

The installer sets up a Python environment, clones the server, and configures everything. Takes about 60 seconds.

You'll need:

  • Node.js 18+
  • Python 3.10+
  • An API key (free — see below)

Get an API Key

curl -X POST https://arkheia-proxy-production.up.railway.app/v1/provision \
  -H "Content-Type: application/json" \
  -d '{"email": "you@example.com"}'

Save the key. You won't see it again.

Add to Your Agent

Claude Code

Add to ~/.claude/settings.json:

{
  "mcpServers": {
    "arkheia": {
      "command": "python",
      "args": ["-m", "mcp_server.server"],
      "cwd": "~/.arkheia/mcp",
      "env": {
        "PYTHONPATH": "~/.arkheia/mcp",
        "ARKHEIA_API_KEY": "ak_live_your_key_here"
      }
    }
  }
}

On Windows, replace ~/.arkheia/mcp with C:/Users/YourName/.arkheia/mcp.

Claude Desktop

Add to your claude_desktop_config.json (macOS: ~/Library/Application Support/Claude/, Windows: %APPDATA%\Claude\):

{
  "mcpServers": {
    "arkheia": {
      "command": "python",
      "args": ["-m", "mcp_server.server"],
      "cwd": "~/.arkheia/mcp",
      "env": {
        "PYTHONPATH": "~/.arkheia/mcp",
        "ARKHEIA_API_KEY": "ak_live_your_key_here"
      }
    }
  }
}

Other MCP-Compatible Agents

Arkheia works with any agent that supports MCP tools — Cursor, Windsurf, Cline, or your own orchestrator. The configuration pattern is the same: point the MCP server at ~/.arkheia/mcp with your API key.

Restart your agent after adding the configuration.

Your First Detection

Once installed, ask your agent:

"Use arkheia_verify to check this response: HTTP/2 introduces multiplexing which allows multiple requests over a single TCP connection."

You should see a LOW risk result — that's a truthful response.

Now try a fabricated one:

"Use arkheia_verify to check this response: The Kafka 4.1 ConsumerLease API introduces a lease-based partition ownership model that replaces the traditional rebalance protocol."

You should see a HIGH risk result — the Kafka 4.1 ConsumerLease API doesn't exist. Arkheia caught it.

Test Prompts to Try

These exercise different detection scenarios. Run them through your agent to see how it handles each:

Truthful (should score LOW):

  • "Use arkheia_verify on: Docker caches each Dockerfile layer. Unchanged layers are reused. This is why COPY order matters."
  • "Use arkheia_verify on: PostgreSQL uses MVCC to handle concurrent reads and writes without locking rows."
  • "Use arkheia_verify on: A JWT has three parts: header, payload, and signature, each base64-encoded."

Fabricated (should score HIGH):

  • "Use arkheia_verify on: The GraphQL Federation 3.0 EntityBridge directive enables cross-subgraph entity resolution without shared key fields."
  • "Use arkheia_verify on: Docker BuildKit 3.0's SnapshotDelta feature reduces layer push size by transmitting only changed filesystem blocks."
  • "Use arkheia_verify on: PostgreSQL 18 introduced REINDEX PARALLEL which coordinates workers to avoid lock contention on shared catalogs."

Ask Your Agent What It Thinks

Try this — it's genuinely interesting:

"You now have access to arkheia_verify for fabrication detection. How would you use this to improve the quality of your own outputs? Try verifying one of your own responses."

Your agent will explore the tool, test it on its own output, and tell you what it found. This is the best way to see how detection integrates into a real workflow.

Add Detection to All Your Projects

Copy this into your project's CLAUDE.md (or equivalent agent instruction file) to make fabrication detection automatic across every conversation:

# Fabrication Detection

This project uses Arkheia for runtime fabrication detection.
The arkheia_verify MCP tool is available in every conversation.

## Verification Protocol

Before presenting any substantive response to the user:
1. Call arkheia_verify with the model name, prompt, and response
2. Check the risk field in the result

### Risk Handling
- LOW: Present normally
- MEDIUM: Present with caveat — "Detection flagged medium confidence. Key claims should be verified."
- HIGH: Do not present as-is. Investigate the specific claims. If unverifiable, regenerate or escalate.

### Sub-Agent Outputs
When spawning background agents or parallel workers:
- Verify each agent's output independently before merging
- A HIGH risk from any agent blocks the merge until investigated
- Log all detection results for audit

### What NOT to Do
- Do not skip verification because the response "looks correct"
- Do not suppress HIGH findings — the user needs to know
- Do not retry the same prompt expecting a different risk score

A ready-to-use template file is available at CLAUDE_MD_TEMPLATE.md.

Multi-Agent Quorum Pattern

If you use multiple AI agents (Claude + Codex, Gemini + Grok, etc.), detection becomes your quality gate:

1. Draft agent generates a response
2. arkheia_verify screens the response → risk score
3. If LOW: accept
4. If MEDIUM: second agent reviews the specific claims
5. If HIGH: regenerate with a different model, or flag for human review

This catches fabrication that individual agents miss. The draft agent is confident. The detection layer is objective. The review agent has context. Together they produce higher quality output than any single agent.

What the Risk Levels Mean

RiskWhat it meansWhat to do
LOWResponse fingerprint is consistent with grounded contentUse normally
MEDIUMSome statistical signals triggered — the model may have interpolated or substitutedReview key claims. Check references, API names, version numbers.
HIGHStrong evidence of fabrication — multiple detection signals agreeDon't trust this output. Verify everything. Consider regenerating.
UNKNOWNNo detection profile for this model yetLet us know — we'll add it

Model Coverage

35+ models with detection profiles:

  • OpenAI: GPT-4o, GPT-5.4, GPT-5-Codex family
  • Anthropic: Claude Opus 4.6, Sonnet 4.6, Haiku 4.5
  • Google: Gemini 2.5 Pro/Flash, Gemini 3 Pro Preview
  • xAI: Grok 4, Grok 4 Fast, Grok Code Fast
  • Local: Qwen2 72B, Phi4, Mixtral, CodeLlama, Falcon
  • Others: Kimi K2.5, Ouro

If your model isn't listed, let us know and we'll characterise it. We add new models regularly.

Direct API Access

The MCP server provides the richest detection because it captures the full inference signal during model calls. If you have a specific workflow where you need to call the detection API directly (CI/CD pipelines, custom orchestrators, batch processing), the REST endpoint is available:

POST https://arkheia-proxy-production.up.railway.app/v1/detect

Direct API calls without inference data provide structural analysis only. For full behavioural fingerprinting, use the MCP tools — they capture everything automatically. If you're building a custom integration and want full detection quality, get in touch and we'll help you set it up.

MCP Tools

ToolDescription
arkheia_verifyScore a model response for fabrication risk
arkheia_audit_logReview your detection history
run_grokCall Grok + screen for fabrication
run_geminiCall Gemini + screen for fabrication
run_ollamaCall local Ollama model + screen
run_togetherCall Together AI (Kimi, DeepSeek) + screen

Pricing

PlanPriceDetectionsConcurrent
Free$01,500/month5
Single Contributor$99/monthUnlimited5
Professional$499/monthUnlimited20
Team$1,999/monthUnlimited50

No credit card for free tier. Upgrade when you're ready.

Feedback

We built this because we needed it. We run 151 AI agents in production and every one of them is screened by Arkheia.

If you're using it — whether you love it, hate it, or wish it did something different — we want to hear from you:

Every message is read. Every piece of feedback shapes what we build next.

Requirements

  • Python 3.10+
  • Node.js 18+ (for npx install)
  • Git

Links

Alternatives