DataWorks

DataWorks

Official
aliyun

Connects AI agents to Alibaba Cloud DataWorks APIs for managing data workflows and projects. Uses TypeScript with Zod validation for reliable cloud operations.

Exports Alibaba Cloud DataWorks Open API tools for dynamic project management and data workflow interactions using TypeScript and Zod for robust schema validation.

29406 views11Local (stdio)

What it does

  • Manage DataWorks projects and resources
  • Execute data workflow operations
  • Query DataWorks Open API endpoints
  • Validate API requests with Zod schemas
  • Configure cloud resource settings

Best for

Data engineers working with Alibaba CloudAutomating DataWorks project managementAI agents managing data pipelines
TypeScript with Zod validationSecurity assessment badge included

About DataWorks

DataWorks is an official MCP server published by aliyun that provides AI assistants with tools and capabilities via the Model Context Protocol. Leverage Alibaba Cloud DataWorks Open API tools with TypeScript and Zod for dynamic project management and schema-valida It is categorized under cloud infrastructure, analytics data.

How to install

You can install DataWorks in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

License

DataWorks is released under the Apache-2.0 license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

MseeP.ai Security Assessment Badge

DataWorks MCP Server

A Model Context Protocol (MCP) server that provides tools for AI, allowing it to interact with the DataWorks Open API through a standardized interface. This implementation is based on the Aliyun Open API and enables AI agents to perform cloud resources operations seamlessly.

Overview

This MCP server:

  • Interact with DataWorks Open API
  • Manage DataWorks resources

The server implements the Model Context Protocol specification to standardize cloud resource interactions for AI agents.

Prerequisites

  • Node.js (v16 or higher)
  • pnpm (recommended), npm, or yarn
  • DataWorks Open API with access key and secret key

Installation

Option 1: Install from npm (recommend for clients like Cursor/Cline)

# Install globally
npm install -g alibabacloud-dataworks-mcp-server

# Or install locally in your project
npm install alibabacloud-dataworks-mcp-server

Option 2: Build from Source (for developers)

  1. Clone this repository:
git clone https://github.com/aliyun/alibabacloud-dataworks-mcp-server
cd alibabacloud-dataworks-mcp-server
  1. Install dependencies (pnpm is recommended, npm is supported):
pnpm install
  1. Build the project:
pnpm run build
  1. Development the project (by @modelcontextprotocol/inspector):
pnpm run dev

open http://localhost:5173

Configuration

MCP Server Configuration

If you installed via npm (Option 1):

{
  "mcpServers": {
    "alibabacloud-dataworks-mcp-server": {
      "command": "npx",
      "args": ["alibabacloud-dataworks-mcp-server"],
      "env": {
        "REGION": "your_dataworks_open_api_region_id_here",
        "ALIBABA_CLOUD_ACCESS_KEY_ID": "your_alibaba_cloud_access_key_id",
        "ALIBABA_CLOUD_ACCESS_KEY_SECRET": "your_alibaba_cloud_access_key_secret",
        "TOOL_CATEGORIES": "optional_your_tool_categories_here_ex_UTILS",
        "TOOL_NAMES": "optional_your_tool_names_here_ex_ListProjects"
      },
      "disabled": false,
      "autoApprove": []
    }
  }
}

If you built from source (Option 2):

{
  "mcpServers": {
    "alibabacloud-dataworks-mcp-server": {
      "command": "node",
      "args": ["/path/to/alibabacloud-dataworks-mcp-server/build/index.js"],
      "env": {
        "REGION": "your_dataworks_open_api_region_id_here",
        "ALIBABA_CLOUD_ACCESS_KEY_ID": "your_alibaba_cloud_access_key_id",
        "ALIBABA_CLOUD_ACCESS_KEY_SECRET": "your_alibaba_cloud_access_key_secret",
        "TOOL_CATEGORIES": "optional_your_tool_categories_here_ex_SERVER_IDE_DEFAULT",
        "TOOL_NAMES": "optional_your_tool_names_here_ex_ListProjects"
      },
      "disabled": false,
      "autoApprove": []
    }
  }
}

Environment Setup

init variables in your environment:

# DataWorks Configuration
REGION=your_dataworks_open_api_region_id_here
ALIBABA_CLOUD_ACCESS_KEY_ID=your_alibaba_cloud_access_key_id
ALIBABA_CLOUD_ACCESS_KEY_SECRET=your_alibaba_cloud_access_key_secret
TOOL_CATEGORIES=optional_your_tool_categories_here_ex_SERVER_IDE_DEFAULT
TOOL_NAMES=optional_your_tool_names_here_ex_ListProjects

Configuration Description

  • Use Guide Description Link

Project Structure

alibabacloud-dataworks-mcp-server/
├── src/
│   ├── index.ts          # Main entry point
├── package.json
└── tsconfig.json

Available Tools

The MCP server provides the following DataWorks tools:

See this link

Security Considerations

  • Keep your private key secure and never share it
  • Use environment variables for sensitive information
  • Regularly monitor and audit AI agent activities

Troubleshooting

If you encounter issues:

  1. Verify your Aliyun Open API access key and secret key are correct
  2. Check your region id is correct
  3. Ensure you're on the intended network (mainnet, testnet, or devnet)
  4. Verify the build was successful

Dependencies

Key dependencies include:

Contributing

Contributions are welcome! Please feel free to submit a Pull Request.

  1. Fork the repository
  2. Create your feature branch (git checkout -b feature/amazing-feature)
  3. Commit your changes (git commit -m 'Add some amazing feature')
  4. Push to the branch (git push origin feature/amazing-feature)
  5. Open a Pull Request

License

This project is licensed under the Apache 2.0 License.

Alternatives

Related Skills

Browse all skills
data-engineer

Build scalable data pipelines, modern data warehouses, and real-time streaming architectures. Implements Apache Spark, dbt, Airflow, and cloud-native data platforms. Use PROACTIVELY for data pipeline design, analytics infrastructure, or modern data stack implementation.

2
database-cloud-optimization-cost-optimize

You are a cloud cost optimization expert specializing in reducing infrastructure expenses while maintaining performance and reliability. Analyze cloud spending, identify savings opportunities, and implement cost-effective architectures across AWS, Azure, and GCP.

6
aws-penetration-testing

This skill should be used when the user asks to "pentest AWS", "test AWS security", "enumerate IAM", "exploit cloud infrastructure", "AWS privilege escalation", "S3 bucket testing", "metadata SSRF", "Lambda exploitation", or needs guidance on Amazon Web Services security assessment.

3
hybrid-cloud-networking

Configure secure, high-performance connectivity between on-premises infrastructure and cloud platforms using VPN and dedicated connections. Use when building hybrid cloud architectures, connecting data centers to cloud, or implementing secure cross-premises networking.

2
hugging-face-cli

Execute Hugging Face Hub operations using the `hf` CLI. Use when the user needs to download models/datasets/spaces, upload files to Hub repositories, create repos, manage local cache, or run compute jobs on HF infrastructure. Covers authentication, file transfers, repository creation, cache operations, and cloud compute.

2
youtube-analytics

YouTube Data API v3 analytics toolkit. Analyze YouTube channels, videos, and search results. Use when the user asks to: check YouTube channel stats, analyze video performance, compare channels, search for videos, get subscriber counts, view engagement metrics, find trending videos, get channel uploads, or analyze YouTube competition. Requires a YouTube Data API v3 key from Google Cloud Console.

2