Langfuse

Langfuse

z9905080

Connects AI models to Langfuse analytics workspaces for querying LLM performance metrics. Requires Langfuse project setup with public/private keys.

Connects AI models to Langfuse analytics workspaces, enabling access to LLM performance metrics by time range for monitoring and analysis.

3858 views3Local (stdio)

What it does

  • Query LLM metrics by time range
  • Connect to Langfuse workspaces
  • Access performance analytics data
  • Retrieve model execution metrics

Best for

AI developers monitoring model performanceTeams tracking LLM usage analyticsAnalyzing AI assistant effectiveness over time
Direct Langfuse workspace integrationTime-based metrics queries

About Langfuse

Langfuse is a community-built MCP server published by z9905080 that provides AI assistants with tools and capabilities via the Model Context Protocol. Monitor LLM performance with Langfuse: advanced ai data analytics and data analysis ai for actionable insights and impro It is categorized under ai ml, analytics data.

How to install

You can install Langfuse in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

License

Langfuse is released under the Apache-2.0 license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

MCP Server for langfuse

npm version

A Model Context Protocol (MCP) server implementation for integrating AI assistants with Langfuse workspaces.

Overview

This package provides an MCP server that enables AI assistants to interact with Langfuse workspaces. It allows AI models to:

  • Query LLM Metrics by Time Range

Installation

# Install from npm
npm install shouting-mcp-langfuse

# Or install globally
npm install -g shouting-mcp-langfuse

You can find the package on npm: shouting-mcp-langfuse

Prerequisites

Before using the server, you need to create a Langfuse project and obtain your project's public and private keys. You can find these keys in the Langfuse dashboard.

  1. set up a Langfuse project
  2. get the public and private keys
  3. set the environment variables

Configuration

The server requires the following environment variables:

  • LANGFUSE_DOMAIN: The Langfuse domain (default: https://api.langfuse.com)
  • LANGFUSE_PUBLIC_KEY: Your Langfuse Project Public Key
  • LANGFUSE_PRIVATE_KEY: Your Langfuse Project Private Key

Usage

Running as a CLI Tool

# Set environment variables
export LANGFUSE_DOMAIN="https://api.langfuse.com"
export LANGFUSE_PUBLIC_KEY="your-public-key"
export LANGFUSE_PRIVATE_KEY="your-private

# Run the server
mcp-server-langfuse

Using in Your Code

import { Server } from "@modelcontextprotocol/sdk/server/index.js";
import { langfuseClient } from "shouting-mcp-langfuse";

// Initialize the server and client
const server = new Server({...});
const langfuseClient = new LangfuseClient(process.env.LANGFUSE_DOMAIN, process.env.LANGFUSE_PUBLIC_KEY, process.env.LANGFUSE_PRIVATE_KEY);

// Register your custom handlers
// ...

Available Tools

The server provides the following langfuse integration tools:

  • getLLMMetricsByTimeRange: Get LLM Metrics by Time Range

License

ISC

Author

[email protected]

Repository

https://github.com/z9905080/mcp-langfuse

Alternatives

Related Skills

Browse all skills
backend-dev-guidelines

Comprehensive backend development guide for Langfuse's Next.js 14/tRPC/Express/TypeScript monorepo. Use when creating tRPC routers, public API endpoints, BullMQ queue processors, services, or working with tRPC procedures, Next.js API routes, Prisma database access, ClickHouse analytics queries, Redis queues, OpenTelemetry instrumentation, Zod v4 validation, env.mjs configuration, tenant isolation patterns, or async patterns. Covers layered architecture (tRPC procedures → services, queue processors → services), dual database system (PostgreSQL + ClickHouse), projectId filtering for multi-tenant isolation, traceException error handling, observability patterns, and testing strategies (Jest for web, vitest for worker).

7
data-storytelling

Transform data into compelling narratives using visualization, context, and persuasive structure. Use when presenting analytics to stakeholders, creating data reports, or building executive presentations.

27
content-trend-researcher

Advanced content and topic research skill that analyzes trends across Google Analytics, Google Trends, Substack, Medium, Reddit, LinkedIn, X, blogs, podcasts, and YouTube to generate data-driven article outlines based on user intent analysis

23
data-scientist

Expert data scientist for advanced analytics, machine learning, and statistical modeling. Handles complex data analysis, predictive modeling, and business intelligence. Use PROACTIVELY for data analysis tasks, ML modeling, statistical analysis, and data-driven insights.

13
google-analytics

Analyze Google Analytics data, review website performance metrics, identify traffic patterns, and suggest data-driven improvements. Use when the user asks about analytics, website metrics, traffic analysis, conversion rates, user behavior, or performance optimization.

13
senior-data-scientist

World-class data science skill for statistical modeling, experimentation, causal inference, and advanced analytics. Expertise in Python (NumPy, Pandas, Scikit-learn), R, SQL, statistical methods, A/B testing, time series, and business intelligence. Includes experiment design, feature engineering, model evaluation, and stakeholder communication. Use when designing experiments, building predictive models, performing causal analysis, or driving data-driven decisions.

8