trulens-notebook-execution
Execute and display Jupyter notebooks for TruLens demos and quickstarts
Install
mkdir -p .claude/skills/trulens-notebook-execution && curl -L -o skill.zip "https://mcp.directory/api/skills/download/4605" && unzip -o skill.zip -d .claude/skills/trulens-notebook-execution && rm skill.zipInstalls to .claude/skills/trulens-notebook-execution
About this skill
TruLens Notebook Execution
Execute Jupyter notebooks, display progress to the user, and handle API key requirements.
When to Use This Skill
Use this skill when:
- Running TruLens quickstart or example notebooks
- Demonstrating TruLens functionality via notebooks
- Testing notebook examples end-to-end
- User asks to "run the notebook" or "execute the notebook"
Execution Method
Always use jupyter nbconvert --execute to run notebooks. This:
- Maintains state across cells (variables persist)
- Captures all output properly
- Handles async operations correctly
- Works with OTEL tracing
DO NOT try to run notebooks by:
- Extracting cells and running them individually in bash
- Using
python -cwith heredocs - Running as a standalone Python script (loses notebook context)
Basic Execution Command
jupyter nbconvert --to notebook --execute --inplace <notebook_path>
Execution with Timeout (for long-running notebooks)
jupyter nbconvert --to notebook --execute --inplace \
--ExecutePreprocessor.timeout=600 \
<notebook_path>
Execution with Output to stdout
jupyter nbconvert --to notebook --execute --stdout <notebook_path>
Displaying Progress to User
When running a notebook, display section headers as each cell executes - NOT generic "BASH_OUTPUT" messages.
Step 1: Parse the Notebook Structure First
Before executing, read the notebook JSON to build a map of:
- Markdown headers (## Section Name)
- Which code cells belong to which section
import json
with open('notebook.ipynb') as f:
nb = json.load(f)
sections = []
current_section = "Setup"
for i, cell in enumerate(nb['cells']):
if cell['cell_type'] == 'markdown':
source = ''.join(cell['source'])
# Extract header
for line in source.split('\n'):
if line.startswith('## '):
current_section = line.replace('## ', '').strip()
sections.append((i, current_section))
Step 2: Display Section Headers During Execution
When checking output or between cell groups, display the section name:
=== Step 1: Create the Search Tool ===
[cell output here]
=== Step 2: Create the Deep Agent ===
[cell output here]
=== Step 3: Set Up TruLens Session ===
[cell output here]
Key Rule: Never Show "BASH_OUTPUT" to User
When polling for bash output during notebook execution:
- DO: Print the current section header before showing output
- DON'T: Just say "BASH_OUTPUT" or "checking output..."
Example Display Implementation
# When starting a section:
print(f"\n=== {section_name} ===")
# When showing cell output:
print(output)
# When section completes:
print("✓ Complete")
Progress Display Pattern
Running notebook: deep_agents_quickstart.ipynb
=== Step 1: Create the Search Tool ===
✓ Complete
=== Step 2: Create the Deep Agent ===
✓ Complete
=== Step 3: Set Up TruLens Session ===
Starting dashboard...
Dashboard started at http://localhost:8501
✓ Complete
=== Step 4: Define Agent GPA Feedback Functions ===
✓ Complete
=== Step 5: Instrument the Agent with TruGraph ===
✓ Complete
=== Step 6: Run and Evaluate ===
Running agent with question: "What is the weather in San Francisco?"
Agent response: "The weather in San Francisco is..."
Waiting for evaluation results...
✓ Evaluations complete
=== Results ===
Answer Relevance: 1.0
Tool Selection: 1.0
...
Handling API Keys
Critical: Check environment first, then prompt for keys ONE AT A TIME
Step 1: Check Environment
env | grep -E "OPENAI|TAVILY|ANTHROPIC" || echo "No API keys found"
Step 2: If Keys Not Found, Prompt Individually
When prompting for keys:
- Ask for ONE key at a time
- Use the key prefix as a hint in the option label (e.g., "sk-proj-..." for OpenAI)
- Let users paste directly - don't rely on complex "Other" field workflows
Example prompt pattern:
Question: "Paste your OPENAI_API_KEY:"
Header: "OpenAI"
Options: [{"label": "sk-proj-...", "description": "Paste your sk-... key"}]
The user will paste their actual key by selecting "Other" or the option itself will be replaced with their input.
Step 3: Set Keys When Running
OPENAI_API_KEY="sk-..." TAVILY_API_KEY="tvly-..." \
jupyter nbconvert --execute ...
Common API Keys for TruLens Notebooks
| Key | Used For |
|---|---|
OPENAI_API_KEY | OpenAI LLM calls, embeddings, feedback provider |
TAVILY_API_KEY | Web search tool (Deep Agents, research agents) |
ANTHROPIC_API_KEY | Anthropic/Claude models |
HUGGINGFACE_API_KEY | HuggingFace models |
Keeping the Dashboard Alive
Critical: The notebook execution process ends, killing any dashboard started within it.
After notebook execution completes, launch the dashboard separately using TruLens's run_dashboard() function.
Important: Database Location
The notebook writes its database to ./default.sqlite relative to the notebook's directory. The run_dashboard() function reads from ./default.sqlite relative to the current working directory.
This means you MUST cd to the notebook's directory before launching the dashboard.
Correct Pattern for Dashboard Persistence
cd /path/to/notebook/directory && \
python3 << 'EOF'
from trulens.core import TruSession
from trulens.dashboard import run_dashboard
session = TruSession()
run_dashboard(session)
EOF
Use run_in_background=true with the bash tool so the dashboard stays alive.
Why NOT to Use Native Streamlit Commands
DO NOT try to launch the dashboard with native streamlit commands like:
# WRONG - will connect to wrong/empty database!
streamlit run /path/to/trulens/src/dashboard/trulens/dashboard/main.py
This fails because:
- Streamlit runs from the current working directory (likely repo root)
- It looks for
./default.sqliterelative to that directory - The actual database is in the notebook's directory
- Result: "No apps found" in the dashboard
Full Example Workflow
# Step 1: Execute notebook
OPENAI_API_KEY="sk-..." jupyter nbconvert --execute --inplace \
/path/to/examples/notebook.ipynb
# Step 2: Launch persistent dashboard FROM THE NOTEBOOK'S DIRECTORY
cd /path/to/examples && \
python3 << 'EOF'
from trulens.core import TruSession
from trulens.dashboard import run_dashboard
session = TruSession()
run_dashboard(session)
EOF
# Use run_in_background=true for this command
The dashboard will output its URL (e.g., http://localhost:55872) and remain running until explicitly stopped.
Post-Execution
After notebook execution:
- Show the leaderboard/results if available
- Provide the dashboard URL if one was launched
- Summarize what was evaluated (metrics used, scores achieved)
Troubleshooting
| Issue | Solution |
|---|---|
| Notebook times out | Increase timeout: --ExecutePreprocessor.timeout=1200 |
| Kernel not found | Ensure correct Python environment is active |
| Import errors | Run pip install cell first or install dependencies |
| API key errors | Verify keys are set correctly in environment |
| Dashboard doesn't start | Check if port is already in use |
'id' was unexpected error | Remove id fields from cells (see fix below) |
Fixing Invalid Notebook JSON
If you see Additional properties are not allowed ('id' was unexpected):
import json
with open('notebook.ipynb', 'r') as f:
nb = json.load(f)
# Remove 'id' fields from cells (not valid in nbformat 4)
for cell in nb['cells']:
if 'id' in cell:
del cell['id']
with open('notebook.ipynb', 'w') as f:
json.dump(nb, f, indent=1)
Example Workflow
User: "Run the deep agents quickstart notebook"
1. Read notebook to identify:
- Section headers (for progress display)
- Required API keys (OPENAI_API_KEY, TAVILY_API_KEY)
2. Check environment for existing keys:
env | grep -E "OPENAI|TAVILY"
3. Prompt for missing keys (ONE AT A TIME):
"Please provide your OPENAI_API_KEY:"
[User enters key]
"Please provide your TAVILY_API_KEY:"
[User enters key]
4. Execute notebook, displaying section headers:
=== Step 1: Create the Search Tool ===
✓ Complete
=== Step 2: Create the Deep Agent ===
✓ Complete
=== Step 3: Set Up TruLens Session ===
✓ Complete
=== Step 4: Define Agent GPA Feedback Functions ===
✓ Complete
=== Step 5: Instrument the Agent ===
✓ Complete
=== Step 6: Run and Evaluate ===
Running agent...
Waiting for evaluation results...
✓ Complete
5. Launch dashboard in background FROM THE NOTEBOOK'S DIRECTORY:
cd /path/to/notebook/directory && python3 -c "
from trulens.core import TruSession
from trulens.dashboard import run_dashboard
session = TruSession()
run_dashboard(session)
"
[run_in_background=true]
6. Display results summary:
"✓ Notebook execution complete!
Evaluation Results:
- Answer Relevance: 1.0
- Tool Selection: 1.0
- Tool Calling: 1.0
- Execution Efficiency: 0.33
- Plan Quality: 1.0
- Plan Adherence: 1.0
Dashboard running at: http://localhost:8501
(Dashboard will stay alive until you stop it)"
Integration with Other Skills
This skill works alongside:
instrumentation/- for understanding what's being tracedevaluation-setup/- for understanding feedback functionsrunning-evaluations/- for interpreting results
More by truera
View all →You might also like
flutter-development
aj-geddes
Build beautiful cross-platform mobile apps with Flutter and Dart. Covers widgets, state management with Provider/BLoC, navigation, API integration, and material design.
drawio-diagrams-enhanced
jgtolentino
Create professional draw.io (diagrams.net) diagrams in XML format (.drawio files) with integrated PMP/PMBOK methodologies, extensive visual asset libraries, and industry-standard professional templates. Use this skill when users ask to create flowcharts, swimlane diagrams, cross-functional flowcharts, org charts, network diagrams, UML diagrams, BPMN, project management diagrams (WBS, Gantt, PERT, RACI), risk matrices, stakeholder maps, or any other visual diagram in draw.io format. This skill includes access to custom shape libraries for icons, clipart, and professional symbols.
godot
bfollington
This skill should be used when working on Godot Engine projects. It provides specialized knowledge of Godot's file formats (.gd, .tscn, .tres), architecture patterns (component-based, signal-driven, resource-based), common pitfalls, validation tools, code templates, and CLI workflows. The `godot` command is available for running the game, validating scripts, importing resources, and exporting builds. Use this skill for tasks involving Godot game development, debugging scene/resource files, implementing game systems, or creating new Godot components.
nano-banana-pro
garg-aayush
Generate and edit images using Google's Nano Banana Pro (Gemini 3 Pro Image) API. Use when the user asks to generate, create, edit, modify, change, alter, or update images. Also use when user references an existing image file and asks to modify it in any way (e.g., "modify this image", "change the background", "replace X with Y"). Supports both text-to-image generation and image-to-image editing with configurable resolution (1K default, 2K, or 4K for high resolution). DO NOT read the image file first - use this skill directly with the --input-image parameter.
ui-ux-pro-max
nextlevelbuilder
"UI/UX design intelligence. 50 styles, 21 palettes, 50 font pairings, 20 charts, 8 stacks (React, Next.js, Vue, Svelte, SwiftUI, React Native, Flutter, Tailwind). Actions: plan, build, create, design, implement, review, fix, improve, optimize, enhance, refactor, check UI/UX code. Projects: website, landing page, dashboard, admin panel, e-commerce, SaaS, portfolio, blog, mobile app, .html, .tsx, .vue, .svelte. Elements: button, modal, navbar, sidebar, card, table, form, chart. Styles: glassmorphism, claymorphism, minimalism, brutalism, neumorphism, bento grid, dark mode, responsive, skeuomorphism, flat design. Topics: color palette, accessibility, animation, layout, typography, font pairing, spacing, hover, shadow, gradient."
rust-coding-skill
UtakataKyosui
Guides Claude in writing idiomatic, efficient, well-structured Rust code using proper data modeling, traits, impl organization, macros, and build-speed best practices.
Stay ahead of the MCP ecosystem
Get weekly updates on new skills and servers.