
LLM Code Context
Provides smart file selection and code context management for sharing relevant project files with LLMs, avoiding token limits while ensuring complete coverage.
Streamlines code context sharing with LLMs by implementing smart file selection, code outlining, and multi-language support for efficient code reviews and documentation generation.
What it does
- Find files modified since a specific timestamp
- Generate code outlines highlighting important sections
- Retrieve missing context files on demand
- Create custom file selection rules
- Extract excerpted content from codebases
Best for
About LLM Code Context
LLM Code Context is a community-built MCP server published by cyberchitta that provides AI assistants with tools and capabilities via the Model Context Protocol. LLM Code Context boosts code reviews and documentation with smart file selection, code outlining, and multi-language sup It is categorized under developer tools. This server exposes 4 tools that AI clients can invoke during conversations and coding sessions.
How to install
You can install LLM Code Context in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
License
LLM Code Context is released under the Apache-2.0 license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
Tools (4)
Returns list of files modified since given timestamp. Args: root_path: Root directory path (e.g. '/home/user/projects/myproject') timestamp: Unix timestamp to check modifications since
Returns excerpted content highlighting important sections in all supported files. Args: root_path: Root directory path rule_name: Rule to use for file selection rules timestamp: Context generation timestamp to check against existing selections
Provides step-by-step instructions for creating custom rules. Args: root_path: Root directory path
Unified tool for retrieving missing context (files, implementations, or excluded sections). Args: root_path: Root directory path (e.g. '/home/user/projects/myproject') param_type: Type of data - 'f' for files, 'i' for implementations, 'e' for excluded sections data: JSON string containing the data (file paths in /{project-name}/ format or implementation queries) timestamp: Context generation timestamp
LLM Context
Smart context management for LLM development workflows. Share relevant project files instantly through intelligent selection and rule-based filtering.
The Problem
Getting the right context into LLM conversations is friction-heavy:
- Manually finding and copying relevant files wastes time
- Too much context hits token limits, too little misses important details
- AI requests for additional files require manual fetching
- Hard to track what changed during development sessions
The Solution
llm-context provides focused, task-specific project context through composable rules.
For humans using chat interfaces:
lc-select # Smart file selection
lc-context # Copy formatted context to clipboard
# Paste and work - AI can access additional files via MCP
For AI agents with CLI access:
lc-preview tmp-prm-auth # Validate rule selects right files
lc-context tmp-prm-auth # Get focused context for sub-agent
For AI agents in chat (MCP tools):
lc_outlines- Generate excerpted context from current rulelc_preview- Validate rule effectiveness before uselc_missing- Fetch specific files/implementations on demand
Note: This project was developed in collaboration with several Claude Sonnets (3.5, 3.6, 3.7, 4.0) and Groks (3, 4), using LLM Context itself to share code during development. All code is heavily human-curated by @restlessronin.
Installation
uv tool install "llm-context>=0.6.0"
Quick Start
Human Workflow (Clipboard)
# One-time setup
cd your-project
lc-init
# Daily usage
lc-select
lc-context
# Paste into your LLM chat
MCP Integration (Recommended)
Add to Claude Desktop config (~/Library/Application Support/Claude/claude_desktop_config.json):
{
"mcpServers": {
"llm-context": {
"command": "uvx",
"args": ["--from", "llm-context", "lc-mcp"]
}
}
}
Restart Claude Desktop. Now AI can access additional files during conversations without manual copying.
Agent Workflow (CLI)
AI agents with shell access use llm-context to create focused contexts:
# Agent explores codebase
lc-outlines
# Agent creates focused rule for specific task
# (via Skill or lc-rule-instructions)
# Agent validates rule
lc-preview tmp-prm-oauth-task
# Agent uses context for sub-task
lc-context tmp-prm-oauth-task
Agent Workflow (MCP)
AI agents in chat environments use MCP tools:
# Explore codebase structure
lc_outlines(root_path, rule_name)
# Validate rule effectiveness
lc_preview(root_path, rule_name)
# Fetch specific files/implementations
lc_missing(root_path, param_type, data, timestamp)
Core Concepts
Rules: Task-Specific Context Descriptors
Rules are YAML+Markdown files that describe what context to provide for a task:
---
description: "Debug API authentication"
compose:
filters: [lc/flt-no-files]
excerpters: [lc/exc-base]
also-include:
full-files: ["/src/auth/**", "/tests/auth/**"]
---
Focus on authentication system and related tests.
Five Rule Categories
- Prompt Rules (
prm-): Generate project contexts (e.g.,lc/prm-developer) - Filter Rules (
flt-): Control file inclusion (e.g.,lc/flt-base,lc/flt-no-files) - Instruction Rules (
ins-): Provide guidelines (e.g.,lc/ins-developer) - Style Rules (
sty-): Enforce coding standards (e.g.,lc/sty-python) - Excerpt Rules (
exc-): Configure content extraction (e.g.,lc/exc-base)
Rule Composition
Build complex rules from simpler ones:
---
instructions: [lc/ins-developer, lc/sty-python]
compose:
filters: [lc/flt-base, project-filters]
excerpters: [lc/exc-base]
---
Essential Commands
| Command | Purpose |
|---|---|
lc-init | Initialize project configuration |
lc-select | Select files based on current rule |
lc-context | Generate and copy context |
lc-context -p | Include prompt instructions |
lc-context -m | Format as separate message |
lc-context -nt | No tools (manual workflow) |
lc-set-rule <name> | Switch active rule |
lc-preview <rule> | Validate rule selection and size |
lc-outlines | Get code structure excerpts |
lc-missing | Fetch files/implementations (manual MCP) |
AI-Assisted Rule Creation
Let AI help create focused, task-specific rules. Two approaches depending on your environment:
Claude Skill (Interactive, Claude Desktop/Code)
How it works: Global skill guides you through creating rules interactively. Examines your codebase as needed using MCP tools.
Setup:
lc-init # Installs skill to ~/.claude/skills/
# Restart Claude Desktop or Claude Code
Usage:
# 1. Share project context
lc-context # Any rule - overview included
# 2. Paste into Claude, then ask:
# "Create a rule for refactoring authentication to JWT"
# "I need a rule to debug the payment processing"
Claude will:
- Use project overview already in context
- Examine specific files via
lc-missingas needed - Ask clarifying questions about scope
- Generate optimized rule (
tmp-prm-<task>.md) - Provide validation instructions
Skill documentation (progressively disclosed):
Skill.md- Quick workflow, decision patternsPATTERNS.md- Common rule patternsSYNTAX.md- Detailed referenceEXAMPLES.md- Complete walkthroughsTROUBLESHOOTING.md- Problem solving
Instruction Rules (Works Anywhere)
How it works: Load comprehensive rule-creation documentation into context, work with any LLM.
Usage:
# 1. Load framework
lc-set-rule lc/prm-rule-create
lc-select
lc-context -nt
# 2. Paste into any LLM
# "I need a rule for adding OAuth integration"
# 3. LLM generates focused rule using framework
# 4. Use the new rule
lc-set-rule tmp-prm-oauth
lc-select
lc-context
Included documentation:
lc/ins-rule-intro- Introduction and overviewlc/ins-rule-framework- Complete decision framework
Comparison
| Aspect | Skill | Instruction Rules |
|---|---|---|
| Setup | Automatic with lc-init | Already available |
| Interaction | Interactive, uses lc-missing | Static documentation |
| File examination | Automatic via MCP | Manual or via AI |
| Best for | Claude Desktop/Code | Any LLM, any environment |
| Updates | Automatic with version upgrades | Built-in to rules |
Both require sharing project context first. Both produce equivalent results.
Project Customization
Create Base Filters
cat > .llm-context/rules/flt-repo-base.md << 'EOF'
---
description: "Repository-specific exclusions"
compose:
filters: [lc/flt-base]
gitignores:
full-files: ["*.md", "/tests", "/node_modules"]
excerpted-files: ["*.md", "/tests"]
---
EOF
Create Development Rule
cat > .llm-context/rules/prm-code.md << 'EOF'
---
description: "Main development rule"
instructions: [lc/ins-developer, lc/sty-python]
compose:
filters: [flt-repo-base]
excerpters: [lc/exc-base]
---
Additional project-specific guidelines and context.
EOF
lc-set-rule prm-code
Deployment Patterns
Choose format based on your LLM environment:
| Pattern | Command | Use Case |
|---|---|---|
| System Message | lc-context -p | AI Studio, etc. |
| Single User Message | lc-context -p -m | Grok, etc. |
| Separate Messages | lc-prompt + lc-context -m | Flexible placement |
| Project Files (included) | lc-context | Claude Projects, etc. |
| Project Files (searchable) | lc-context -m | Force into context |
See Deployment Patterns for details.
Key Features
- Intelligent Selection: Rules automatically include/exclude appropriate files
- Context Validation: Preview size and selection before generation
- Code Excerpting: Extract structure while reducing tokens (15+ languages)
- MCP Integration: AI accesses additional files without manual intervention
- Composable Rules: Build complex contexts from reusable patterns
- AI-Assisted Creation: Interactive skill or documentation-based approaches
- Agent-Friendly: CLI and MCP interfaces for autonomous operation
Common Workflows
Daily Development (Human)
lc-set-rule prm-code
lc-select
lc-context
# Paste into chat - AI accesses more files via MCP if needed
Focused Task (Human or Agent)
# Share project context first
lc-context
# Then create focused rule:
# Via Skill: "Create a rule for [task]"
# Via Instructions: lc-set-rule lc/prm-rule-create && lc-context -nt
# Validate and use
lc-preview tmp-prm-task
lc-context tmp-prm-task
Agent Context Provisioning (CLI)
# Agent validates rule effectiveness
lc-preview tmp-prm-refactor-auth
# Agent generates context for sub-agent
lc-context tmp-prm-refactor-auth > /tmp/context.md
# Sub-agent reads context and executes task
Agent Context Provisioning (MCP)
# Agent validates rule
preview
---
*README truncated. [View full README on GitHub](https://github.com/cyberchitta/llm-context.py).*
Alternatives
Related Skills
Browse all skillsGuide for creating high-quality MCP (Model Context Protocol) servers that enable LLMs to interact with external services through well-designed tools. Use when building MCP servers to integrate external APIs or services, whether in Python (FastMCP) or Node/TypeScript (MCP SDK).
Syncs Claude Skills with other AI coding tools like Cursor, Copilot, and Codeium by creating cross-references and shared knowledge bases. Invoke when user wants to leverage skills across multiple tools or create unified AI context.
Manage MCP (Model Context Protocol) servers in Claude Code projects. Use this skill when the user requests enabling, installing, disabling, or removing specific MCP servers like context7 or chrome-devtools. Always operates at project level (local scope only).
Expert MCP (Model Context Protocol) orchestration with n8n workflow automation. Master bidirectional MCP integration, expose n8n workflows as AI agent tools, consume MCP servers in workflows, build agentic systems, orchestrate multi-agent workflows, and create production-ready AI-powered automation pipelines with Claude Code integration.
Builds remote MCP (Model Context Protocol) servers on Cloudflare Workers with tools, OAuth authentication, and production deployment. Generates server code, configures auth providers, and deploys to Workers. Use when: user wants to "build MCP server", "create MCP tools", "remote MCP", "deploy MCP", add "OAuth to MCP", or mentions Model Context Protocol on Cloudflare. Also triggers on "MCP authentication" or "MCP deployment".
Creates educational Teams channel posts for internal knowledge sharing about Claude Code features, tools, and best practices. Applies when writing posts, announcements, or documentation to teach colleagues effective Claude Code usage, announce new features, share productivity tips, or document lessons learned. Provides templates, writing guidelines, and structured approaches emphasizing concrete examples, underlying principles, and connections to best practices like context engineering. Activates for content involving Teams posts, channel announcements, feature documentation, or tip sharing.