numerai-model-upload

0
0
Source

Create Numerai Tournament model upload pickles (.pkl) with a self-contained predict() function. Use when preparing upload artifacts, debugging numerai_predict import errors, or documenting model-upload requirements and testing steps.

Install

mkdir -p .claude/skills/numerai-model-upload && curl -L -o skill.zip "https://mcp.directory/api/skills/download/4864" && unzip -o skill.zip -d .claude/skills/numerai-model-upload && rm skill.zip

Installs to .claude/skills/numerai-model-upload

About this skill

Numerai Model Upload

Overview

Create a portable predict(live_features, live_benchmark_models) pickle that runs inside Numerai's numerai_predict container without repo dependencies.

CRITICAL: Python Version Compatibility

Before creating any pkl file, you must ensure your Python environment matches Numerai's compute environment. Mismatched versions cause segfaults and validation failures due to binary incompatibility (especially with numpy).

Step 1: Query the Default Docker Image (MCP Required)

If the numerai MCP server is available, always query the default Python version first:

query { computePickleDockerImages { id name image tag default } }

Look for the entry with default: true. The image name indicates the Python version:

  • numerai_predict_py_3_12:a78dedd → Python 3.12 (current default as of 2026)
  • numerai_predict_py_3_11:a78dedd → Python 3.11
  • numerai_predict_py_3_10:a78dedd → Python 3.10

If the numerai MCP is not installed, it can be installed through our install script via curl -sL https://numer.ai/install-mcp.sh | bash, this script guides the user through installing the MCP for Codex CLI and configuring an API key with the correct scopes that are required by MCP.

You can find more documentation about Numerai MCP here: https://docs.numer.ai/numerai-tournament/mcp

Step 2: Create Matching Virtual Environment with pyenv

Use pyenv to create a virtual environment with the exact Python version:

# 1. List available pyenv Python versions
ls ~/.pyenv/versions/

# 2. Find the matching minor version (e.g., for Python 3.12)
PYENV_PY=$(ls -d ~/.pyenv/versions/3.12.* 2>/dev/null | head -1)

# 3. Create the virtual environment
$PYENV_PY/bin/python -m venv ./venv

# 4. Activate and install pkl dependencies
source ./venv/bin/activate
pip install --upgrade pip
pip install numpy pandas cloudpickle scipy
# Add lightgbm, torch, etc. only if your model needs them

Step 3: Create pkl in the Correct Environment

Always create pkl files using the matching venv:

./venv/bin/python create_model_pkl.py

Requirements

  • Implement predict(live_features, live_benchmark_models) and return a DataFrame with a prediction column aligned to the input index.
  • Preserve training-time preprocessing (feature order, imputation values, scaling params) inside the pickle.
  • Avoid imports from local repo modules (no agents.*), because Numerai's container will not have them.
  • Prefer numpy/pandas/scipy-only inference; do not rely on torch/xgboost unless you verify the container has those packages.
  • Move any trained model to CPU before exporting and store plain numpy weights.
  • Validate required columns (era for per-era ranking, benchmark column if used).

Workflow

  1. Query the default Docker image from the MCP to determine the required Python version.
  2. Create/activate a matching venv using pyenv (see above).
  3. Train on the desired full dataset (train + validation) with the same preprocessing and early-stopping scheme as the best model only after your research has plateaued and you have selected the final configuration to deploy.
  4. Export an inference bundle from the trained model:
    • Feature list and ordering
    • Imputation values and scaling stats
    • Model weights/biases (numpy arrays)
    • Activation name and any constants
    • Benchmark column name if needed as a feature
  5. Build a predict function that:
    • Reads only from the bundle and standard libraries
    • Applies preprocessing and a numpy forward pass
    • Ranks predictions per era to [0, 1] when required
  6. cloudpickle.dump(predict, "model.pkl") using the matching venv's Python.
  7. Test the pickle with the Numerai container before uploading.

Testing

Run the Numerai debug container locally (use the same image tag as the default):

# Get the default image tag from MCP query, then test:
docker run -i --rm -v "$PWD:$PWD" ghcr.io/numerai/numerai_predict_py_3_12:a78dedd --debug --model $PWD/[PICKLE_FILE]

Common Pitfalls

  • Segmentation fault / numpy binary incompatibility: The pkl was created with a different Python version than Numerai's container. Always query the default docker image first and create pkl files using a matching pyenv-based venv.
  • ImportError: No module named 'agents': occurs when the pickle references repo classes. Fix by exporting a pure-numpy inference bundle and rebuilding predict without repo imports.
  • Missing era column: per-era ranking requires live_features["era"].
  • Benchmark misalignment: ensure live_benchmark_models is reindexed to live_features (by id) before use.
  • Feature drift: ensure feature order in inference matches training order exactly.

Debugging Validation Failures

If your pickle fails validation, query the trigger status and logs:

query {
  account {
    models {
      username
      computePickleUpload {
        filename
        validationStatus
        triggerStatus
        triggers {
          id
          status
          statuses {
            status
            description
            insertedAt
          }
        }
      }
    }
  }
}

Common error descriptions:

  • "Segmentation fault! Ensure python and library versions match our environment." → Python/numpy version mismatch
  • "No currently open rounds!" → Model validated successfully but no round is open for submission

Reference

  • Use numerai/example_model.ipynb for the expected predict signature and output format.

Deploying to Numerai via MCP Server

After creating and testing your pkl file, you can deploy it to Numerai using the Numerai MCP server. The MCP server provides tools for creating models and uploading pkl files programmatically.

Available MCP Tools

The numerai MCP server provides these key tools:

  1. check_api_credentials - Verify your API token and see granted scopes
  2. create_model - Create a new model in a tournament
  3. upload_model - Upload pkl files (multi-step workflow)
  4. graphql_query - List existing models and perform custom queries

Authentication

All authenticated operations require a Numerai API token with upload_submission scope:

Option 1: Upload to an Existing Model

If you already have a model slot you want to use:

  1. List your models using graphql_query:

    query {
      account { models { id name } }
    }
    
  2. Get upload authorization for your pkl file:

    • Call upload_model with operation: "get_upload_auth", modelId: "<model_uuid>", filename: "model.pkl"
    • This returns a presigned URL for uploading
  3. Upload the pkl file

    • Call a PUT file upload on the pre-signed URL with the path to the pkl file.
  4. Register the upload with Numerai:

    • Call upload_model with operation: "create", modelId: "<model_uuid>", filename: "model.pkl"
    • This triggers validation of your pickle
  5. Check validation status:

    • Call upload_model with operation: "list" to see all pickles and their status
    • Wait for validation to complete successfully
  6. Assign the pickle to the model slot:

    • Call upload_model with operation: "assign", modelId: "<model_uuid>", pickleId: "<pickle_uuid>"
    • This makes the pickle active for automated submissions

Option 2: Create a New Model and Upload

If you want to create a new model slot:

  1. Create the model:

    • Call create_model with name: "<unique_model_name>", tournament: 8 (for Classic)
    • Note: Model names must be unique within the tournament
  2. Get the model ID from the response

  3. Follow steps 2-6 from Option 1 to upload and assign the pkl file

Upload Workflow Summary

┌─────────────────────────────────────────────────────────────────┐
│                    PKL DEPLOYMENT WORKFLOW                       │
├─────────────────────────────────────────────────────────────────┤
│                                                                  │
│  1. Create pkl file (this skill's main workflow)                │
│  2. Test pkl locally with numerai_predict container             │
│  3. Choose: create new model OR use existing model              │
│                                                                  │
│  For new model:                                                  │
│    └─> create_model(name, tournament=8)                         │
│                                                                  │
│  For existing model:                                             │
│    └─> graphql_query to list models and get model ID            │
│                                                                  │
│  4. upload_model(operation="get_upload_auth", modelId, filename)│
│  5. upload_model(operation="put_file", presignedUrl, localPath) │
│  6. upload_model(operation="create", modelId, filename)         │
│  7. upload_model(operation="list") - wait for validation        │
│  8. upload_model(operation="assign", modelId, pickleId)         │
│                                                                  │
│  Optional:                                                       │
│  - upload_model(operation="trigger", pickleId) to test          │
│  - upload_model(operation="get_logs", pickleId, triggerId)      │
│                                                                  │
└─────────────────────────────────────────────────────────────────┘

Important Notes

  • Only the Classic tournament (tournament=8) supports pickle uploads
  • The model must have its submission webhook disabled before uploading
  • CRITICAL: Before creating pkl files, query the default docker image to ensure Python version compatibility
  • Use this GraphQL query to check available runtimes and the default:
    query { computePickleDockerImages { id name image tag default } }
    

Content truncated.

You might also like

flutter-development

aj-geddes

Build beautiful cross-platform mobile apps with Flutter and Dart. Covers widgets, state management with Provider/BLoC, navigation, API integration, and material design.

642969

drawio-diagrams-enhanced

jgtolentino

Create professional draw.io (diagrams.net) diagrams in XML format (.drawio files) with integrated PMP/PMBOK methodologies, extensive visual asset libraries, and industry-standard professional templates. Use this skill when users ask to create flowcharts, swimlane diagrams, cross-functional flowcharts, org charts, network diagrams, UML diagrams, BPMN, project management diagrams (WBS, Gantt, PERT, RACI), risk matrices, stakeholder maps, or any other visual diagram in draw.io format. This skill includes access to custom shape libraries for icons, clipart, and professional symbols.

590705

godot

bfollington

This skill should be used when working on Godot Engine projects. It provides specialized knowledge of Godot's file formats (.gd, .tscn, .tres), architecture patterns (component-based, signal-driven, resource-based), common pitfalls, validation tools, code templates, and CLI workflows. The `godot` command is available for running the game, validating scripts, importing resources, and exporting builds. Use this skill for tasks involving Godot game development, debugging scene/resource files, implementing game systems, or creating new Godot components.

339397

ui-ux-pro-max

nextlevelbuilder

"UI/UX design intelligence. 50 styles, 21 palettes, 50 font pairings, 20 charts, 8 stacks (React, Next.js, Vue, Svelte, SwiftUI, React Native, Flutter, Tailwind). Actions: plan, build, create, design, implement, review, fix, improve, optimize, enhance, refactor, check UI/UX code. Projects: website, landing page, dashboard, admin panel, e-commerce, SaaS, portfolio, blog, mobile app, .html, .tsx, .vue, .svelte. Elements: button, modal, navbar, sidebar, card, table, form, chart. Styles: glassmorphism, claymorphism, minimalism, brutalism, neumorphism, bento grid, dark mode, responsive, skeuomorphism, flat design. Topics: color palette, accessibility, animation, layout, typography, font pairing, spacing, hover, shadow, gradient."

318396

nano-banana-pro

garg-aayush

Generate and edit images using Google's Nano Banana Pro (Gemini 3 Pro Image) API. Use when the user asks to generate, create, edit, modify, change, alter, or update images. Also use when user references an existing image file and asks to modify it in any way (e.g., "modify this image", "change the background", "replace X with Y"). Supports both text-to-image generation and image-to-image editing with configurable resolution (1K default, 2K, or 4K for high resolution). DO NOT read the image file first - use this skill directly with the --input-image parameter.

451339

fastapi-templates

wshobson

Create production-ready FastAPI projects with async patterns, dependency injection, and comprehensive error handling. Use when building new FastAPI applications or setting up backend API projects.

304231

Stay ahead of the MCP ecosystem

Get weekly updates on new skills and servers.