trulens-evaluation-setup
Configure feedback functions and selectors for TruLens evaluations
Install
mkdir -p .claude/skills/trulens-evaluation-setup && curl -L -o skill.zip "https://mcp.directory/api/skills/download/3472" && unzip -o skill.zip -d .claude/skills/trulens-evaluation-setup && rm skill.zipInstalls to .claude/skills/trulens-evaluation-setup
About this skill
TruLens Evaluation Setup
Configure feedback functions to evaluate your LLM app's quality.
Interactive Evaluation Selection
Before proceeding, let's determine the right evaluations for your app.
Question 1: What type of app are you building?
Option A: RAG (Retrieval-Augmented Generation)
- Your app retrieves documents/chunks from a knowledge base
- It generates responses grounded in the retrieved context
- Examples: Q&A systems, document search, knowledge assistants
→ Recommended: RAG Triad metrics
- Context Relevance
- Groundedness
- Answer Relevance
Option B: Agent
- Your app uses tools to accomplish tasks
- It may involve multi-step reasoning or planning
- Examples: research agents, coding assistants, task automation
→ Recommended: Agent GPA metrics (continue to Question 2)
Question 2 (Agents only): Does your agent do explicit planning?
Yes, my agent creates plans before executing:
- Agent outputs a plan/strategy before taking actions
- Agent references its plan during execution
→ Use all Agent GPA metrics:
- Logical Consistency
- Plan Quality
- Plan Adherence
- Execution Efficiency
- Tool Selection
- Tool Calling
- Tool Quality
No, my agent acts without explicit planning:
- Agent takes actions directly without stating a plan
- Agent uses reactive decision-making
→ Use Agent GPA metrics (excluding plan metrics):
- Logical Consistency
- Execution Efficiency
- Tool Selection
- Tool Calling
- Tool Quality
Question 3: Do you want to add any additional evaluations?
Consider adding these based on your needs:
| Evaluation | Use Case |
|---|---|
| Coherence | Check if output is well-structured and readable |
| Conciseness | Ensure responses aren't unnecessarily verbose |
| Harmlessness | Detect potentially harmful content |
| Sentiment | Analyze emotional tone of responses |
| Custom metrics | Domain-specific evaluations (see below) |
Creating Custom Metrics
If you need domain-specific evaluations, describe what you want to measure:
What aspect of your app do you want to evaluate?
Examples:
- "Check if the response follows our brand voice guidelines"
- "Verify the output contains required legal disclaimers"
- "Measure technical accuracy for code generation"
- "Evaluate if customer support responses show empathy"
Template for custom metrics:
def my_custom_metric(input_text: str, output_text: str) -> float:
"""
Describe what this metric evaluates.
Returns:
float: Score between 0.0 (worst) and 1.0 (best)
"""
# Option 1: Rule-based logic
# score = 1.0 if "required phrase" in output_text else 0.0
# Option 2: Use LLM-as-judge
# provider = OpenAI()
# response = provider.client.chat.completions.create(
# model="gpt-4o",
# messages=[{
# "role": "user",
# "content": f"Rate this response on [YOUR CRITERIA]. Input: {input_text} Output: {output_text}. Return only a number 0-10."
# }]
# )
# score = float(response.choices[0].message.content) / 10.0
return score
f_custom = (
Feedback(my_custom_metric, name="My Custom Metric")
.on_input()
.on_output()
)
Custom metric with context:
def custom_with_context(query: str, context: str, response: str) -> float:
"""Evaluate using query, retrieved context, and response."""
# Your evaluation logic
return score
f_custom_context = (
Feedback(custom_with_context, name="Custom Context Metric")
.on_input()
.on_context(collect_list=True)
.on_output()
)
Tell me what you want to evaluate and I'll help you create the metric!
Overview
Feedback functions evaluate specific aspects of your app by:
- Selecting data from instrumented spans (inputs, outputs, retrieved contexts)
- Applying evaluation logic (LLM-as-judge, similarity metrics, etc.)
- Returning scores between 0.0 and 1.0
Prerequisites
pip install trulens trulens-providers-openai
Instructions
Step 1: Initialize a Feedback Provider
from trulens.providers.openai import OpenAI
provider = OpenAI(model_engine="gpt-4o")
Step 2: Create Feedback Functions with Selector Shortcuts
TruLens provides shortcuts for common selection patterns:
from trulens.core import Feedback
# Answer relevance: input → output
f_answer_relevance = (
Feedback(provider.relevance_with_cot_reasons, name="Answer Relevance")
.on_input()
.on_output()
)
# Context relevance: input → each context chunk
f_context_relevance = (
Feedback(provider.context_relevance_with_cot_reasons, name="Context Relevance")
.on_input()
.on_context(collect_list=False) # Evaluates each context individually
)
# Groundedness: all contexts → output
f_groundedness = (
Feedback(provider.groundedness_measure_with_cot_reasons, name="Groundedness")
.on_context(collect_list=True) # Concatenates all contexts
.on_output()
)
Shortcut Reference:
| Shortcut | Selects | Required Span Type |
|---|---|---|
on_input() | App input | RECORD_ROOT |
on_output() | App output | RECORD_ROOT |
on_context() | Retrieved contexts | RETRIEVAL |
⚠️ IMPORTANT: .on_input() and .on_output() require RECORD_ROOT spans!
These shortcuts look for spans with span_type=SpanAttributes.SpanType.RECORD_ROOT. If you use manual instrumentation with a different span type (like AGENT), the shortcuts will not find any data.
Solutions:
- Use framework wrappers (
TruGraph,TruChain,TruLlama) which createRECORD_ROOTautomatically - Use explicit
@instrument(span_type=SpanAttributes.SpanType.RECORD_ROOT, ...)on your entry point - Use explicit
Selectorobjects instead of shortcuts (see Step 3)
Step 3: Using Explicit Selectors
For more control, use Selector to target specific span attributes:
from trulens.core import Feedback
from trulens.core.feedback.selector import Selector
from trulens.otel.semconv.trace import SpanAttributes
f_answer_relevance = (
Feedback(provider.relevance_with_cot_reasons, name="Answer Relevance")
.on({
"prompt": Selector(
span_type=SpanAttributes.SpanType.RECORD_ROOT,
span_attribute=SpanAttributes.RECORD_ROOT.INPUT,
),
})
.on({
"response": Selector(
span_type=SpanAttributes.SpanType.RECORD_ROOT,
span_attribute=SpanAttributes.RECORD_ROOT.OUTPUT,
),
})
)
Step 4: Understanding collect_list
The collect_list parameter controls how multiple values are handled:
| Setting | Behavior | Use Case |
|---|---|---|
collect_list=False | Evaluate each value individually | Context relevance (score each chunk) |
collect_list=True | Concatenate all values | Groundedness (check against all context) |
# Evaluate each retrieved context individually (returns multiple scores)
f_context_relevance = (
Feedback(provider.context_relevance_with_cot_reasons, name="Context Relevance")
.on_input()
.on({
"context": Selector(
span_type=SpanAttributes.SpanType.RETRIEVAL,
span_attribute=SpanAttributes.RETRIEVAL.RETRIEVED_CONTEXTS,
collect_list=False # Each context evaluated separately
),
})
)
# Evaluate against all contexts combined (returns single score)
f_groundedness = (
Feedback(provider.groundedness_measure_with_cot_reasons, name="Groundedness")
.on({
"context": Selector(
span_type=SpanAttributes.SpanType.RETRIEVAL,
span_attribute=SpanAttributes.RETRIEVAL.RETRIEVED_CONTEXTS,
collect_list=True # All contexts concatenated
),
})
.on_output()
)
Step 5: Aggregating Multiple Scores
When collect_list=False produces multiple scores, aggregate them:
import numpy as np
f_context_relevance = (
Feedback(provider.context_relevance_with_cot_reasons, name="Context Relevance")
.on_input()
.on_context(collect_list=False)
.aggregate(np.mean) # Average all context relevance scores
)
Common aggregation functions:
np.mean- Average scorenp.min- Worst score (conservative)np.max- Best score (optimistic)
Common Patterns
RAG Triad Setup
import numpy as np
from trulens.core import Feedback
from trulens.providers.openai import OpenAI
provider = OpenAI()
# Context Relevance: Is each retrieved chunk relevant to the query?
f_context_relevance = (
Feedback(provider.context_relevance_with_cot_reasons, name="Context Relevance")
.on_input()
.on_context(collect_list=False)
.aggregate(np.mean)
)
# Groundedness: Is the response grounded in the retrieved context?
f_groundedness = (
Feedback(provider.groundedness_measure_with_cot_reasons, name="Groundedness")
.on_context(collect_list=True)
.on_output()
)
# Answer Relevance: Does the response answer the original question?
f_answer_relevance = (
Feedback(provider.relevance_with_cot_reasons, name="Answer Relevance")
.on_input()
.on_output()
)
rag_feedbacks = [f_context_relevance, f_groundedness, f_answer_relevance]
Agent GPA Setup
from trulens.core import Feedback
from trulens.core.feedback.selector import Selector
from trulens.providers.openai import OpenAI
provider = OpenAI()
# All Agent GPA metrics use trace-level selection
trace_selector = {"trace": Selector(trace_level=True)}
# Logical Consistency
f_logical_consistency = (
Feedback(provider.logical_consistency_with_cot_reasons, name="Logical Consistency")
.on(trace_selector)
)
# Plan Quality (exclude if agent doesn't do explicit planning)
f_plan_quality = (
Feedback(provider.plan_quality_with_cot_reasons, name="Plan Quality")
.on(trace_selector)
)
# Plan Adherence (exclude if agent doesn't do explicit planning)
f_plan_adherence = (
Feedback(provider.plan_adherence_with_cot_reasons, name="Plan Adherence")
.on(trace_selector)
)
# Execution Efficiency
f_execution_efficiency = (
Feedback(provider.execution_efficiency_with_cot_reasons, name="Execution Efficiency")
.on(trace_selector)
)
# Tool Selection
f_tool_selection = (
Feedback(provider.tool_selection_with_cot_reasons, name="Tool Selection")
.on(trace_selector)
)
# Tool Calling
f_tool_calling = (
Feedback(provider.tool_calling_with_cot_reasons, name="Tool Calling")
.on(trace_selector)
)
# Tool Quality
f_tool_quality = (
Feedback(provider.tool_quality_with_cot_reasons, name="Tool Quality")
.on(trace_selector)
)
# Use all for agents with planning
agent_feedbacks_with_planning = [
f_logical_consistency,
f_plan_quality,
f_plan_adherence,
f_execution_efficiency,
f_tool_selection,
f_tool_calling,
f_tool_quality,
]
# For agents without explicit planning, exclude plan metrics
agent_feedbacks_no_planning = [
f_logical_consistency,
f_execution_efficiency,
f_tool_selection,
f_tool_calling,
f_tool_quality,
]
Custom Feedback Function
def my_custom_metric(input_text: str, output_text: str) -> float:
"""Custom evaluation returning score between 0.0 and 1.0."""
# Your evaluation logic here
score = len(output_text) / (len(input_text) + len(output_text))
return min(max(score, 0.0), 1.0)
f_custom = (
Feedback(my_custom_metric, name="Custom Metric")
.on_input()
.on_output()
)
Troubleshooting
- Selector not finding data: Ensure the span attribute was set during instrumentation
- Empty context: Verify
RETRIEVAL.RETRIEVED_CONTEXTSis mapped in your@instrument()decorator - Aggregation errors: Check that
collect_list=Falseis set when using.aggregate() .on_input()/.on_output()returning no data: These shortcuts requireRECORD_ROOTspan type. Use framework wrappers or explicit@instrument(span_type=SpanAttributes.SpanType.RECORD_ROOT, ...). See the instrumentation skill for details.- Feedback columns show empty/null: Verify your instrumentation creates
RECORD_ROOTspans withINPUTandOUTPUTattributes
More by truera
View all →You might also like
flutter-development
aj-geddes
Build beautiful cross-platform mobile apps with Flutter and Dart. Covers widgets, state management with Provider/BLoC, navigation, API integration, and material design.
drawio-diagrams-enhanced
jgtolentino
Create professional draw.io (diagrams.net) diagrams in XML format (.drawio files) with integrated PMP/PMBOK methodologies, extensive visual asset libraries, and industry-standard professional templates. Use this skill when users ask to create flowcharts, swimlane diagrams, cross-functional flowcharts, org charts, network diagrams, UML diagrams, BPMN, project management diagrams (WBS, Gantt, PERT, RACI), risk matrices, stakeholder maps, or any other visual diagram in draw.io format. This skill includes access to custom shape libraries for icons, clipart, and professional symbols.
godot
bfollington
This skill should be used when working on Godot Engine projects. It provides specialized knowledge of Godot's file formats (.gd, .tscn, .tres), architecture patterns (component-based, signal-driven, resource-based), common pitfalls, validation tools, code templates, and CLI workflows. The `godot` command is available for running the game, validating scripts, importing resources, and exporting builds. Use this skill for tasks involving Godot game development, debugging scene/resource files, implementing game systems, or creating new Godot components.
nano-banana-pro
garg-aayush
Generate and edit images using Google's Nano Banana Pro (Gemini 3 Pro Image) API. Use when the user asks to generate, create, edit, modify, change, alter, or update images. Also use when user references an existing image file and asks to modify it in any way (e.g., "modify this image", "change the background", "replace X with Y"). Supports both text-to-image generation and image-to-image editing with configurable resolution (1K default, 2K, or 4K for high resolution). DO NOT read the image file first - use this skill directly with the --input-image parameter.
ui-ux-pro-max
nextlevelbuilder
"UI/UX design intelligence. 50 styles, 21 palettes, 50 font pairings, 20 charts, 8 stacks (React, Next.js, Vue, Svelte, SwiftUI, React Native, Flutter, Tailwind). Actions: plan, build, create, design, implement, review, fix, improve, optimize, enhance, refactor, check UI/UX code. Projects: website, landing page, dashboard, admin panel, e-commerce, SaaS, portfolio, blog, mobile app, .html, .tsx, .vue, .svelte. Elements: button, modal, navbar, sidebar, card, table, form, chart. Styles: glassmorphism, claymorphism, minimalism, brutalism, neumorphism, bento grid, dark mode, responsive, skeuomorphism, flat design. Topics: color palette, accessibility, animation, layout, typography, font pairing, spacing, hover, shadow, gradient."
rust-coding-skill
UtakataKyosui
Guides Claude in writing idiomatic, efficient, well-structured Rust code using proper data modeling, traits, impl organization, macros, and build-speed best practices.
Stay ahead of the MCP ecosystem
Get weekly updates on new skills and servers.