
DataWorks
OfficialConnects AI agents to Alibaba Cloud DataWorks APIs for managing data workflows and projects. Uses TypeScript with Zod validation for reliable cloud operations.
Exports Alibaba Cloud DataWorks Open API tools for dynamic project management and data workflow interactions using TypeScript and Zod for robust schema validation.
What it does
- Manage DataWorks projects and resources
- Execute data workflow operations
- Query DataWorks Open API endpoints
- Validate API requests with Zod schemas
- Configure cloud resource settings
Best for
About DataWorks
DataWorks is an official MCP server published by aliyun that provides AI assistants with tools and capabilities via the Model Context Protocol. Leverage Alibaba Cloud DataWorks Open API tools with TypeScript and Zod for dynamic project management and schema-valida It is categorized under cloud infrastructure, analytics data.
How to install
You can install DataWorks in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
License
DataWorks is released under the Apache-2.0 license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
DataWorks MCP Server
A Model Context Protocol (MCP) server that provides tools for AI, allowing it to interact with the DataWorks Open API through a standardized interface. This implementation is based on the Aliyun Open API and enables AI agents to perform cloud resources operations seamlessly.
Overview
This MCP server:
- Interact with DataWorks Open API
- Manage DataWorks resources
The server implements the Model Context Protocol specification to standardize cloud resource interactions for AI agents.
Prerequisites
- Node.js (v16 or higher)
- pnpm (recommended), npm, or yarn
- DataWorks Open API with access key and secret key
Installation
Option 1: Install from npm (recommend for clients like Cursor/Cline)
# Install globally
npm install -g alibabacloud-dataworks-mcp-server
# Or install locally in your project
npm install alibabacloud-dataworks-mcp-server
Option 2: Build from Source (for developers)
- Clone this repository:
git clone https://github.com/aliyun/alibabacloud-dataworks-mcp-server
cd alibabacloud-dataworks-mcp-server
- Install dependencies (pnpm is recommended, npm is supported):
pnpm install
- Build the project:
pnpm run build
- Development the project (by @modelcontextprotocol/inspector):
pnpm run dev
Configuration
MCP Server Configuration
If you installed via npm (Option 1):
{
"mcpServers": {
"alibabacloud-dataworks-mcp-server": {
"command": "npx",
"args": ["alibabacloud-dataworks-mcp-server"],
"env": {
"REGION": "your_dataworks_open_api_region_id_here",
"ALIBABA_CLOUD_ACCESS_KEY_ID": "your_alibaba_cloud_access_key_id",
"ALIBABA_CLOUD_ACCESS_KEY_SECRET": "your_alibaba_cloud_access_key_secret",
"TOOL_CATEGORIES": "optional_your_tool_categories_here_ex_UTILS",
"TOOL_NAMES": "optional_your_tool_names_here_ex_ListProjects"
},
"disabled": false,
"autoApprove": []
}
}
}
If you built from source (Option 2):
{
"mcpServers": {
"alibabacloud-dataworks-mcp-server": {
"command": "node",
"args": ["/path/to/alibabacloud-dataworks-mcp-server/build/index.js"],
"env": {
"REGION": "your_dataworks_open_api_region_id_here",
"ALIBABA_CLOUD_ACCESS_KEY_ID": "your_alibaba_cloud_access_key_id",
"ALIBABA_CLOUD_ACCESS_KEY_SECRET": "your_alibaba_cloud_access_key_secret",
"TOOL_CATEGORIES": "optional_your_tool_categories_here_ex_SERVER_IDE_DEFAULT",
"TOOL_NAMES": "optional_your_tool_names_here_ex_ListProjects"
},
"disabled": false,
"autoApprove": []
}
}
}
Environment Setup
init variables in your environment:
# DataWorks Configuration
REGION=your_dataworks_open_api_region_id_here
ALIBABA_CLOUD_ACCESS_KEY_ID=your_alibaba_cloud_access_key_id
ALIBABA_CLOUD_ACCESS_KEY_SECRET=your_alibaba_cloud_access_key_secret
TOOL_CATEGORIES=optional_your_tool_categories_here_ex_SERVER_IDE_DEFAULT
TOOL_NAMES=optional_your_tool_names_here_ex_ListProjects
Configuration Description
- Use Guide Description Link
Project Structure
alibabacloud-dataworks-mcp-server/
├── src/
│ ├── index.ts # Main entry point
├── package.json
└── tsconfig.json
Available Tools
The MCP server provides the following DataWorks tools:
See this link
Security Considerations
- Keep your private key secure and never share it
- Use environment variables for sensitive information
- Regularly monitor and audit AI agent activities
Troubleshooting
If you encounter issues:
- Verify your Aliyun Open API access key and secret key are correct
- Check your region id is correct
- Ensure you're on the intended network (mainnet, testnet, or devnet)
- Verify the build was successful
Dependencies
Key dependencies include:
Contributing
Contributions are welcome! Please feel free to submit a Pull Request.
- Fork the repository
- Create your feature branch (
git checkout -b feature/amazing-feature) - Commit your changes (
git commit -m 'Add some amazing feature') - Push to the branch (
git push origin feature/amazing-feature) - Open a Pull Request
License
This project is licensed under the Apache 2.0 License.
Alternatives
Related Skills
Browse all skillsBuild scalable data pipelines, modern data warehouses, and real-time streaming architectures. Implements Apache Spark, dbt, Airflow, and cloud-native data platforms. Use PROACTIVELY for data pipeline design, analytics infrastructure, or modern data stack implementation.
You are a cloud cost optimization expert specializing in reducing infrastructure expenses while maintaining performance and reliability. Analyze cloud spending, identify savings opportunities, and implement cost-effective architectures across AWS, Azure, and GCP.
This skill should be used when the user asks to "pentest AWS", "test AWS security", "enumerate IAM", "exploit cloud infrastructure", "AWS privilege escalation", "S3 bucket testing", "metadata SSRF", "Lambda exploitation", or needs guidance on Amazon Web Services security assessment.
Configure secure, high-performance connectivity between on-premises infrastructure and cloud platforms using VPN and dedicated connections. Use when building hybrid cloud architectures, connecting data centers to cloud, or implementing secure cross-premises networking.
Execute Hugging Face Hub operations using the `hf` CLI. Use when the user needs to download models/datasets/spaces, upload files to Hub repositories, create repos, manage local cache, or run compute jobs on HF infrastructure. Covers authentication, file transfers, repository creation, cache operations, and cloud compute.
YouTube Data API v3 analytics toolkit. Analyze YouTube channels, videos, and search results. Use when the user asks to: check YouTube channel stats, analyze video performance, compare channels, search for videos, get subscriber counts, view engagement metrics, find trending videos, get channel uploads, or analyze YouTube competition. Requires a YouTube Data API v3 key from Google Cloud Console.
