Background Job

Background Job

dylan-gluck

Execute shell commands asynchronously in the background and manage them with full process lifecycle control. Monitor output, send input, and kill jobs as needed.

Enables coding agents to execute long-running shell commands asynchronously with full process management capabilities.

10561 views6Local (stdio)

What it does

  • Execute shell commands as background jobs
  • Monitor job status and output in real-time
  • Send input to running processes via stdin
  • Kill or terminate background processes
  • List all active and completed jobs
  • Tail recent output from job logs

Best for

Running long build processes or test suitesManaging development servers and servicesAutomating deployment and CI/CD workflowsInteractive shell session management
Full process lifecycle managementReal-time stdout/stderr monitoringInteractive stdin support

About Background Job

Background Job is a community-built MCP server published by dylan-gluck that provides AI assistants with tools and capabilities via the Model Context Protocol. Enhance your business management process software with Background Job, enabling BPM in software via asynchronous long-ru It is categorized under developer tools. This server exposes 7 tools that AI clients can invoke during conversations and coding sessions.

How to install

You can install Background Job in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

License

Background Job is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

Tools (7)

list_jobs

List all background jobs with their status. Returns a list of all background jobs, including their job ID, status, command, and start time. Jobs are sorted by start time (newest first).

get_job_status

Get the current status of a background job. Args: job_id: The UUID of the job to check Returns: The current status of the job (running, completed, failed, or killed)

get_job_output

Get the complete stdout and stderr output of a job. Args: job_id: The UUID of the job to get output from Returns: ProcessOutput containing the complete stdout and stderr content

tail_job_output

Get the last N lines of stdout and stderr from a job. Args: job_id: The UUID of the job to tail lines: Number of lines to return (1-1000, default 50) Returns: ProcessOutput containing the last N lines of stdout and stderr

execute_command

Execute a command as a background job and return job ID. Args: command: Shell command to execute in the background Returns: ExecuteOutput containing the job ID (UUID) of the started job

MCP Background Job Server

Python 3.12+ FastMCP PyPI version License

An MCP (Model Context Protocol) server that enables coding agents to execute long-running shell commands asynchronously with full process management capabilities.

Overview

The MCP Background Job Server provides a robust solution for running shell commands in the background, allowing agents to start processes, monitor their status, interact with them, and manage their lifecycle. This is particularly useful for development workflows involving build processes, test suites, servers, or any long-running operations.

Features

  • Asynchronous Process Execution: Execute shell commands as background jobs with unique job IDs
  • Process Lifecycle Management: Start, monitor, interact with, and terminate background processes
  • Real-time Output Monitoring: Capture and retrieve stdout/stderr with buffering and tailing capabilities
  • Interactive Process Support: Send input to running processes via stdin
  • Resource Management: Configurable job limits and automatic cleanup of completed processes
  • MCP Protocol Integration: Full integration with Model Context Protocol for agent interactions

Installation

Quick Install (Recommended)

Install directly from PyPI using uvx:

# Install and run the MCP server
uvx mcp-background-job

Claude Code Integration

Add the server to your Claude Code configuration:

  1. Option A: Using Claude Code Desktop

    • Open Claude Code settings/preferences
    • Navigate to MCP Servers section
    • Add a new server:
      • Name: background-job
      • Command: uvx
      • Args: ["mcp-background-job"]
  2. Option B: Configuration File Add to your Claude Code configuration file:

    {
      "mcpServers": {
        "background-job": {
          "command": "uvx",
          "args": ["mcp-background-job"]
        }
      }
    }
    
  3. Restart Claude Code to load the new MCP server.

Development Setup

For local development or contributing:

Prerequisites

  • Python 3.12 or higher
  • uv package manager

Setup Steps

  1. Clone and navigate to the project directory:

    git clone https://github.com/dylan-gluck/mcp-background-job.git
    cd mcp-background-job
    
  2. Install dependencies:

    uv sync
    
  3. Install in development mode:

    uv add -e .
    

Quick Start

Using with Claude Code

Once configured, ask Claude to help you with background tasks:

You: "Start my development server in the background and monitor it"

Claude: I'll start your development server using the background job server.

[Uses the execute tool to run your dev server]
[Shows job ID and monitors startup progress]
[Provides status updates]

Claude: "Your development server is now running on http://localhost:3000. 
The job ID is abc123-def456 if you need to control it later."

Manual Server Usage

For development or direct usage:

# Run with stdio transport (most common)
uvx mcp-background-job

# Or for development:
uv run python -m mcp_background_job

Basic Usage Example

# 1. Execute a long-running command
execute_result = await execute_command("npm run dev")
job_id = execute_result.job_id

# 2. Check job status
status = await get_job_status(job_id)
print(f"Job status: {status.status}")

# 3. Get recent output
output = await tail_job_output(job_id, lines=20)
print("Recent output:", output.stdout)

# 4. Interact with the process
interaction = await interact_with_job(job_id, "some input\n")
print("Process response:", interaction.stdout)

# 5. Kill the job when done
result = await kill_job(job_id)
print(f"Kill result: {result.status}")

MCP Tools Reference

The server exposes 7 MCP tools for process management:

Read-only Tools

ToolDescriptionParametersReturns
listList all background jobsNone{jobs: [JobSummary]}
statusGet job statusjob_id: str{status: JobStatus}
outputGet complete job outputjob_id: str{stdout: str, stderr: str}
tailGet recent output linesjob_id: str, lines: int{stdout: str, stderr: str}

Interactive Tools

ToolDescriptionParametersReturns
executeStart new background jobcommand: str{job_id: str}
interactSend input to job stdinjob_id: str, input: str{stdout: str, stderr: str}
killTerminate running jobjob_id: str{status: str}

Job Status Values

  • running - Process is currently executing
  • completed - Process finished successfully
  • failed - Process terminated with error
  • killed - Process was terminated by user

Configuration

Environment Variables

Configure the server behavior using these environment variables:

# Maximum concurrent jobs (default: 10)
export MCP_BG_MAX_JOBS=20

# Maximum output buffer per job (default: 10MB)
export MCP_BG_MAX_OUTPUT_SIZE=20MB
# or in bytes:
export MCP_BG_MAX_OUTPUT_SIZE=20971520

# Default job timeout in seconds (default: no timeout)
export MCP_BG_JOB_TIMEOUT=3600

# Cleanup interval for completed jobs in seconds (default: 300)
export MCP_BG_CLEANUP_INTERVAL=600

# Working directory for jobs (default: current directory)
export MCP_BG_WORKING_DIR=/path/to/project

# Allowed command patterns (optional security restriction)
export MCP_BG_ALLOWED_COMMANDS="^npm ,^python ,^echo ,^ls"

Claude Code Configuration with Environment Variables

{
  "mcpServers": {
    "background-job": {
      "command": "uvx",
      "args": ["mcp-background-job"],
      "env": {
        "MCP_BG_MAX_JOBS": "20",
        "MCP_BG_MAX_OUTPUT_SIZE": "20MB"
      }
    }
  }
}

Programmatic Configuration

from mcp_background_job.config import BackgroundJobConfig

config = BackgroundJobConfig(
    max_concurrent_jobs=20,
    max_output_size_bytes=20 * 1024 * 1024,  # 20MB
    default_job_timeout=7200,  # 2 hours
    cleanup_interval_seconds=600  # 10 minutes
)

Architecture

The server is built with a modular architecture:

  • JobManager: Central service for job lifecycle management
  • ProcessWrapper: Abstraction layer for subprocess handling with I/O buffering
  • FastMCP Server: MCP protocol implementation with tool definitions
  • Pydantic Models: Type-safe data validation and serialization

Key Components

src/mcp_background_job/
├── server.py          # FastMCP server and tool definitions
├── service.py         # JobManager service implementation  
├── process.py         # ProcessWrapper for subprocess management
├── models.py          # Pydantic data models
├── config.py          # Configuration management
└── logging_config.py  # Logging setup

Development

Running Tests

# Run all tests
uv run pytest tests/

# Run unit tests only
uv run pytest tests/unit/ -v

# Run integration tests only  
uv run pytest tests/integration/ -v

Code Formatting

# Format code with ruff
uv run ruff format

# Run type checking
uv run mypy src/

Development Workflow

  1. Make your changes
  2. Run tests: uv run pytest tests/
  3. Format code: uv run ruff format
  4. Commit changes

Examples

Development Server Workflow

# Start a development server
job_id=$(echo '{"command": "npm run dev"}' | mcp-tool execute)

# Monitor the startup
mcp-tool tail --job_id "$job_id" --lines 10

# Check if server is ready
mcp-tool status --job_id "$job_id"

# Stop the server
mcp-tool kill --job_id "$job_id"

Long-running Build Process

# Start a build process
job_id=$(echo '{"command": "docker build -t myapp ."}' | mcp-tool execute)

# Monitor build progress
while true; do
  status=$(mcp-tool status --job_id "$job_id")
  if [[ "$status" != "running" ]]; then break; fi
  mcp-tool tail --job_id "$job_id" --lines 5
  sleep 10
done

# Get final build output
mcp-tool output --job_id "$job_id"

Interactive Process Example

# Start Python REPL
job_id=$(echo '{"command": "python -i"}' | mcp-tool execute)

# Send Python code
mcp-tool interact --job_id "$job_id" --input "print('Hello, World!')\n"

# Send more commands
mcp-tool interact --job_id "$job_id" --input "import sys; print(sys.version)\n"

# Exit REPL
mcp-tool interact --job_id "$job_id" --input "exit()\n"

Security Considerations

  • Process Isolation: Each job runs as a separate subprocess
  • Resource Limits: Configurable limits on concurrent jobs and memory usage
  • Input Validation: All parameters are validated using Pydantic models
  • Command Restrictions: Consider implementing command allowlists in production
  • Output Sanitization: Be aware that process output may contain sensitive information

Transport Support

The server supports multiple MCP transports:

  • stdio: Default transport for local development and agent integration
  • HTTP: For remote access (requires additional setup)

For stdio transport, ensure logging goes to stderr only to avoid protocol conflicts.

Troubleshooting

Common Issues

Import Errors: Ensure the package is installed in development mode:

uv add -e .

Tests Not Running: Install the package first, then run tests:

uv sync
uv add -e .
uv run pytest tests/

Permission Errors: Ensure proper permissions for the commands you're trying to execute.

Memory Issues: Adjust MCP_BG_MAX_OUTPUT_SIZE if dealing with processes that generat


README truncated. View full README on GitHub.

Alternatives

Related Skills

Browse all skills
trigger-dev

Trigger.dev expert for background jobs, AI workflows, and reliable async execution with excellent developer experience and TypeScript-first design. Use when: trigger.dev, trigger dev, background task, ai background job, long running task.

2
trigger-dev-tasks

Use this skill when writing, designing, or optimizing Trigger.dev background tasks and workflows. This includes creating reliable async tasks, implementing AI workflows, setting up scheduled jobs, structuring complex task hierarchies with subtasks, configuring build extensions for tools like ffmpeg or Puppeteer/Playwright, and handling task schemas with Zod validation.

0
dotnet-backend

.NET/C# backend developer for ASP.NET Core APIs with Entity Framework Core. Builds REST APIs, minimal APIs, gRPC services, authentication with Identity/JWT, authorization, database operations, background services, SignalR real-time features. Activates for: .NET, C#, ASP.NET Core, Entity Framework Core, EF Core, .NET Core, minimal API, Web API, gRPC, authentication .NET, Identity, JWT .NET, authorization, LINQ, async/await C#, background service, IHostedService, SignalR, SQL Server, PostgreSQL .NET, dependency injection, middleware .NET.

109
interactive-portfolio

Expert in building portfolios that actually land jobs and clients - not just showing work, but creating memorable experiences. Covers developer portfolios, designer portfolios, creative portfolios, and portfolios that convert visitors into opportunities. Use when: portfolio, personal website, showcase work, developer portfolio, designer portfolio.

22
ui-design-system

UI design system toolkit for Senior UI Designer including design token generation, component documentation, responsive design calculations, and developer handoff tools. Use for creating design systems, maintaining visual consistency, and facilitating design-dev collaboration.

18
bullmq-specialist

BullMQ expert for Redis-backed job queues, background processing, and reliable async execution in Node.js/TypeScript applications. Use when: bullmq, bull queue, redis queue, background job, job queue.

8