Novita AI GPU Cloud

Novita AI GPU Cloud

Official
novitalabs

Connects to Novita AI's GPU cloud platform to deploy, monitor, and manage GPU-accelerated containers and workloads. Currently in beta with GPU instance management capabilities.

Provides direct access to Novita AI's GPU cloud infrastructure for deploying, monitoring, and managing GPU-accelerated workloads and containers without leaving your conversation context.

12382 views10Local (stdio)

What it does

  • Deploy GPU instances with custom configurations
  • Monitor and manage running GPU containers
  • Create and manage container templates
  • Configure network storage for GPU workloads
  • Control instance lifecycle (start, stop, restart, delete)
  • Manage container registry authentication

Best for

AI/ML developers needing GPU compute resourcesTeams deploying containerized GPU workloadsResearchers running compute-intensive experiments
Direct GPU cloud access from chat interfaceBeta release with active development

About Novita AI GPU Cloud

Novita AI GPU Cloud is an official MCP server published by novitalabs that provides AI assistants with tools and capabilities via the Model Context Protocol. Access Novita AI GPU cloud for cloud-based GPU computing. Deploy, monitor, and manage GPU workloads and containers with It is categorized under cloud infrastructure, ai ml.

How to install

You can install Novita AI GPU Cloud in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

License

Novita AI GPU Cloud is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

Novita MCP Server

smithery badge

novita-mcp-server is a Model Context Protocol (MCP) server that provides seamless interaction with Novita AI platform resources. We recommend accessing this server through Claude Desktop, Cursor, or any other compatible MCP client.

Novita Server MCP server

Features

⚠️ Beta Notice: novita-mcp-server is currently in beta and only supports GPU instance management. Additional resource types will be supported in future releases.

Currently, novita-mcp-server enables management the resources of GPU instances product.

Supported operations are as follows:

  • Cluster(/Region): List;
  • Product: List;
  • GPU Instance: List, Get, Create, Start, Stop, Delete, Restart;
  • Template: List, Get, Create, Delete;
  • Container Registry Auth: List, Create, Delete;
  • Network Storage: List, Create, Update, Delete;

Installation

You can install the package using npm, or Smithery:

Using npm

npm install -g @novitalabs/novita-mcp-server

Using Smithery

Visit the https://smithery.ai/server/@novitalabs/novita-mcp-server and follow the "Install" instructions to install the server.

Configuration to use novita-mcp-server

First, you need to get your Novita API key from the Novita AI Key Management.

And next, you can use the following configuration for both Claude Desktop and Cursor:

📌 Tips

For Claude Desktop, you can refer to the Claude Desktop MCP Quickstart guide to learn how to configure the MCP server.

For Cursor, you can refer to the Cursor MCP Quickstart guide to learn how to configure the MCP server.

{
  "mcpServers": {
    "@novitalabs/novita-mcp-server": {
      "command": "npx",
      "args": ["-y", "@novitalabs/novita-mcp-server"],
      "env": {
        "NOVITA_API_KEY": "your_api_key_here"
      }
    }
  }
}

Examples

Here are some examples of how to use the novita-mcp-server to manage your resources with Claude Desktop or Cursor:

List clusters

List all the Novita clusters

List products

List all available Novita GPU instance products

List GPU instances

List all my running Novita GPU instances

Create a new GPU instance

Create a new Novita GPU instance:

Name: test-novita-mcp-server-01
Product: any available product
GPU Number: 1
Image: A standard public PyTorch/CUDA image
Container Disk: 60GB

Testing

This project uses Jest for testing. The tests are located in the src/tests directory.

You can run the tests using one of the following commands:

npm test

Alternatives

Related Skills

Browse all skills
hugging-face-jobs

This skill should be used when users want to run any workload on Hugging Face Jobs infrastructure. Covers UV scripts, Docker-based jobs, hardware selection, cost estimation, authentication with tokens, secrets management, timeout configuration, and result persistence. Designed for general-purpose compute workloads including data processing, inference, experiments, batch jobs, and any Python-based tasks. Should be invoked for tasks involving cloud compute, GPU workloads, or when users mention running jobs on Hugging Face infrastructure without local setup.

1
modal-serverless-gpu

Serverless GPU cloud platform for running ML workloads. Use when you need on-demand GPU access without infrastructure management, deploying ML models as APIs, or running batch jobs with automatic scaling.

1
hugging-face-model-trainer

This skill should be used when users want to train or fine-tune language models using TRL (Transformer Reinforcement Learning) on Hugging Face Jobs infrastructure. Covers SFT, DPO, GRPO and reward modeling training methods, plus GGUF conversion for local deployment. Includes guidance on the TRL Jobs package, UV scripts with PEP 723 format, dataset preparation and validation, hardware selection, cost estimation, Trackio monitoring, Hub authentication, and model persistence. Should be invoked for tasks involving cloud GPU training, GGUF conversion, or when users mention training on Hugging Face Jobs without local GPU setup.

1
database-cloud-optimization-cost-optimize

You are a cloud cost optimization expert specializing in reducing infrastructure expenses while maintaining performance and reliability. Analyze cloud spending, identify savings opportunities, and implement cost-effective architectures across AWS, Azure, and GCP.

6
aws-skills

AWS development with infrastructure automation and cloud architecture patterns

6
senior-devops

Comprehensive DevOps skill for CI/CD, infrastructure automation, containerization, and cloud platforms (AWS, GCP, Azure). Includes pipeline setup, infrastructure as code, deployment automation, and monitoring. Use when setting up pipelines, deploying applications, managing infrastructure, implementing monitoring, or optimizing deployment processes.

5