Android MCP

Android MCP

CursorTouch

Connects AI agents to Android devices for automated app testing and UI interaction using ADB and Android Accessibility API. Works with any LLM without requiring computer vision models.

A lightweight bridge enabling AI agents to perform real-world tasks on Android devices such as app navigation, UI interaction, and automated QA testing without requiring computer-vision pipelines or preprogrammed scripts.

439517 views67Local (stdio)

What it does

  • Launch Android apps and navigate interfaces
  • Tap, swipe, and input text on Android screens
  • Read UI element hierarchies and device state
  • Capture screenshots and device information
  • Execute shell commands on Android devices
  • Automate gesture sequences and keystrokes

Best for

QA engineers automating mobile app testingDevelopers building Android automation workflowsAI agents performing mobile device tasks
No computer vision pipeline requiredWorks with any LLM2-4 second action latency

About Android MCP

Android MCP is a community-built MCP server published by CursorTouch that provides AI assistants with tools and capabilities via the Model Context Protocol. Android MCP — lightweight bridge enabling AI agents for Android to perform Android automation and Android UI testing: ap It is categorized under developer tools.

How to install

You can install Android MCP in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

License

Android MCP is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

🤖 Android MCP

License Python Platform Last Commit
Follow on Twitter Join us on Discord

Android-MCP is a lightweight, open-source tool that bridge between AI agents and Android devices. Running as an MCP server, it lets LLM agents perform real-world tasks such as app navigation, UI interaction and automated QA testing without relying on traditional computer-vision pipelines or preprogramed scripts.

https://github.com/user-attachments/assets/cf9a5e4e-b69f-46d4-8487-0f61a7a86d67

✨ Features

  • Native Android Integration
    Interact with UI elements via ADB and the Android Accessibility API: launch apps, tap, swipe, input text, and read view hierarchies.

  • Bring Your Own LLM/VLM
    Works with any language model, no fine-tuned CV model or OCR pipeline required.

  • Rich Toolset for Mobile Automation
    Pre-built tools for gestures, keystrokes, capture, device state, shell commands execution.

  • Real-Time Interaction
    Typical latency between actions (e.g., two taps) ranges 2-4s depending on device specs and load.

Supported Operating Systems

  • Android 10+

Installation

📦 Prerequisites

  • Python 3.10+
  • ADB (Android Debug Bridge)
  • Android 10+ (Emulator/ Android Device)

📲 Testing ADB Connection

Before running the server, ensure your Android device is connected and recognized by ADB:

  1. Connect your Android device via USB or ensure your emulator is running.
  2. Open a terminal and run:
    adb devices
    
  3. You should see your device listed:
    List of devices attached
    emulator-5554   device
    
    If the list is empty or shows "unauthorized", check your USB debugging settings on the device.

🏁 Getting Started

You can run the Android MCP server using UVX (recommended) or UV (for local development).

Option 1: UVX (Recommended)

No need to install dependencies manually. Just configure Claude Desktop:

  1. Locate your config file

    • Windows: %APPDATA%\Claude\claude_desktop_config.json
    • macOS: ~/Library/Application Support/Claude/claude_desktop_config.json
  2. Add the configuration

    {
      "mcpServers": {
        "android-mcp": {
          "command": "uvx",
          "args": [
            "android-mcp"
          ]
        }
      }
    }
    

    Note: By default, it connects to emulator-5554. To connect to a specific device, add "--device", "<YOUR_DEVICE_serial>" to the args list.

Option 2: UV Mode (Local Development)

  1. Clone and Install

    git clone https://github.com/CursorTouch/Android-MCP.git
    cd Android-MCP
    uv sync
    
  2. Configure Claude Desktop

    {
      "mcpServers": {
        "android-mcp": {
          "command": "uv",
          "args": [
            "--directory",
            "</PATH/TO/Android-MCP>",
            "run",
            "android-mcp"
          ]
        }
      }
    }
    

    Note: Replace </PATH/TO/Android-MCP> with the full path to your cloned directory. Add "--device", "<YOUR_DEVICE_serial>" to args to target a specific device.

  3. Restart the Claude Desktop

Restart your Claude Desktop. You should see "android-mcp" listed as an available integration. That's it, now you're ready to start controlling your Android device with natural language.

For troubleshooting tips (log locations, common ADB issues), see the MCP docs.


🛠️ Available Tools

Claude can access the following tools to interact with Windows:

  • State-Tool: To understand the state of the device.
  • Click-Tool: Click on the screen at the given coordinates.
  • Long-Click-Tool: Perform long click on the screen at the given coordinates.
  • Type-Tool: Type text on the specified coordinates (optionally clears existing text).
  • Swipe-Tool: Perform swipe from one location to other.
  • Drag-Tool: Drag from one point to another.
  • Press-Tool: To press the keys on the mobile device (Back, Volume Up, ...etc).
  • Wait-Tool: Pause for a defined duration.
  • State-Tool: Combined snapshot of active apps and interactive UI elements.
  • Notification-Tool: To access the notifications seen on the device.
  • Shell-Tool: To execute shell commands on the android device.

⚙️ Environment Variables

  • SCREENSHOT_QUANTIZED: Set to true to quantize the screenshot to reduce input tokens.

⚠️ Caution

Android-MCP can execute arbitrary UI actions on your mobile device. Use it in controlled environments (emulators, test devices) when running untrusted prompts or agents.

🪪 License

This project is licensed under the MIT License. See LICENSE for details.

🤝 Contributing

Contributions are welcome! Please read CONTRIBUTING for dev setup and PR guidelines.

Made with ❤️ by CursorTouch,

developers: Jeomon George, Muhammad Yaseen

Citation

@misc{
  author       = {cursortouch},
  title        = {Android-MCP},
  year         = {2025},
  publisher    = {GitHub},
  howpublished = {\url{https://github.com/CursorTouch/Android-MCP}},
  note         = {Lightweight open-source bridge between LLM agents and Android},
}

Alternatives

Related Skills

Browse all skills
penpot-uiux-design

Comprehensive guide for creating professional UI/UX designs in Penpot using MCP tools. Use this skill when: (1) Creating new UI/UX designs for web, mobile, or desktop applications, (2) Building design systems with components and tokens, (3) Designing dashboards, forms, navigation, or landing pages, (4) Applying accessibility standards and best practices, (5) Following platform guidelines (iOS, Android, Material Design), (6) Reviewing or improving existing Penpot designs for usability. Triggers: "design a UI", "create interface", "build layout", "design dashboard", "create form", "design landing page", "make it accessible", "design system", "component library".

19
ui-design-system

UI design system toolkit for Senior UI Designer including design token generation, component documentation, responsive design calculations, and developer handoff tools. Use for creating design systems, maintaining visual consistency, and facilitating design-dev collaboration.

18
ai-sdk

Answer questions about the AI SDK and help build AI-powered features. Use when developers: (1) Ask about AI SDK functions like generateText, streamText, ToolLoopAgent, embed, or tools, (2) Want to build AI agents, chatbots, RAG systems, or text generation features, (3) Have questions about AI providers (OpenAI, Anthropic, Google, etc.), streaming, tool calling, structured output, or embeddings, (4) Use React hooks like useChat or useCompletion. Triggers on: "AI SDK", "Vercel AI SDK", "generateText", "streamText", "add AI to my app", "build an agent", "tool calling", "structured output", "useChat".

6
api-documenter

Master API documentation with OpenAPI 3.1, AI-powered tools, and modern developer experience practices. Create interactive docs, generate SDKs, and build comprehensive developer portals. Use PROACTIVELY for API documentation or developer portal creation.

4
openai-knowledge

Use when working with the OpenAI API (Responses API) or OpenAI platform features (tools, streaming, Realtime API, auth, models, rate limits, MCP) and you need authoritative, up-to-date documentation (schemas, examples, limits, edge cases). Prefer the OpenAI Developer Documentation MCP server tools when available; otherwise guide the user to enable `openaiDeveloperDocs`.

4
cli-builder

Guide for building TypeScript CLIs with Bun. Use when creating command-line tools, adding subcommands to existing CLIs, or building developer tooling. Covers argument parsing, subcommand patterns, output formatting, and distribution.

3