
Supabase Management
Manage Supabase projects and organizations programmatically through the Management API. Create, delete, and retrieve details for projects and organizations.
Control Supabase projects and organizations.
What it does
- List all Supabase projects
- Create new Supabase projects
- Delete existing projects
- Retrieve project API keys
- Manage Supabase organizations
- Get project and organization details
Best for
About Supabase Management
Supabase Management is a community-built MCP server published by joshuarileydev that provides AI assistants with tools and capabilities via the Model Context Protocol. Manage Supabase projects and organizations effortlessly — centralized control, secure access, role management, and real- It is categorized under databases, cloud infrastructure.
How to install
You can install Supabase Management in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
License
Supabase Management is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
Supabase MCP Server
A Model Context Protocol (MCP) server that provides programmatic access to the Supabase Management API. This server allows AI models and other clients to manage Supabase projects and organizations through a standardized interface.
Features
Project Management
- List all projects
- Get project details
- Create new projects
- Delete projects
- Retrieve project API keys
Organization Management
- List all organizations
- Get organization details
- Create new organizations
Installation
Add the following to your Claude Config JSON file
{
"mcpServers": {
"supabase": {
"command": "npx",
"args": [
"y",
"@joshuarileydev/supabase-mcp-server"
],
"env": {
"SUPABASE_API_KEY": "API_KEY_HERE"
}
}
}
}
Alternatives
Related Skills
Browse all skillsBuild comprehensive ML pipelines, experiment tracking, and model registries with MLflow, Kubeflow, and modern MLOps tools. Implements automated training, deployment, and monitoring across cloud platforms. Use PROACTIVELY for ML infrastructure, experiment management, or pipeline automation.
This skill should be used when users want to run any workload on Hugging Face Jobs infrastructure. Covers UV scripts, Docker-based jobs, hardware selection, cost estimation, authentication with tokens, secrets management, timeout configuration, and result persistence. Designed for general-purpose compute workloads including data processing, inference, experiments, batch jobs, and any Python-based tasks. Should be invoked for tasks involving cloud compute, GPU workloads, or when users mention running jobs on Hugging Face infrastructure without local setup.
AWS CloudFormation infrastructure as code for stack management. Use when writing templates, deploying stacks, managing drift, troubleshooting deployments, or organizing infrastructure with nested stacks.
This skill enables Claude to audit access control implementations in various systems. It uses the access-control-auditor plugin to identify potential vulnerabilities and misconfigurations related to access control. Use this skill when the user asks to "audit access control", "check permissions", "assess access rights", or requests a "security review" focused on access management. It's particularly useful for analyzing IAM policies, ACLs, and other access control mechanisms in cloud environments, applications, or infrastructure. The skill helps ensure compliance with security best practices and identify potential privilege escalation paths.
Expert Terraform/OpenTofu specialist mastering advanced IaC automation, state management, and enterprise infrastructure patterns. Handles complex module design, multi-cloud deployments, GitOps workflows, policy as code, and CI/CD integration. Covers migration strategies, security best practices, and modern IaC ecosystems. Use PROACTIVELY for advanced IaC, state management, or infrastructure automation.
Serverless GPU cloud platform for running ML workloads. Use when you need on-demand GPU access without infrastructure management, deploying ML models as APIs, or running batch jobs with automatic scaling.