Vision
OfficialGives AI assistants the ability to analyze and understand images and videos from local files or remote URLs.
Image Analysis - Supports intelligent analysis and content understanding of multiple image formats, giving your AI Agent visual capabilities. Video Understanding - Supports visual understanding of both local and remote videos. Easy Integration - One-click installation, quick integration with Claude Code and other MCP-compatible clients.
What it does
- Analyze images in multiple formats
- Extract content and context from images
- Process local and remote videos
- Understand visual content in videos
- Provide detailed descriptions of visual media
Best for
About Vision
Vision is an official MCP server published by z_ai that provides AI assistants with tools and capabilities via the Model Context Protocol. Vision: Add visual intelligence to your AI agents - image and video analysis with one-click integration for Claude Code It is categorized under ai ml, developer tools.
How to install
You can install Vision in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
License
Vision is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
Alternatives
Related Skills
Browse all skillsUI design system toolkit for Senior UI Designer including design token generation, component documentation, responsive design calculations, and developer handoff tools. Use for creating design systems, maintaining visual consistency, and facilitating design-dev collaboration.
Answer questions about the AI SDK and help build AI-powered features. Use when developers: (1) Ask about AI SDK functions like generateText, streamText, ToolLoopAgent, embed, or tools, (2) Want to build AI agents, chatbots, RAG systems, or text generation features, (3) Have questions about AI providers (OpenAI, Anthropic, Google, etc.), streaming, tool calling, structured output, or embeddings, (4) Use React hooks like useChat or useCompletion. Triggers on: "AI SDK", "Vercel AI SDK", "generateText", "streamText", "add AI to my app", "build an agent", "tool calling", "structured output", "useChat".
Master API documentation with OpenAPI 3.1, AI-powered tools, and modern developer experience practices. Create interactive docs, generate SDKs, and build comprehensive developer portals. Use PROACTIVELY for API documentation or developer portal creation.
23 production-ready engineering skills covering architecture, frontend, backend, fullstack, QA, DevOps, security, AI/ML, data engineering, computer vision, and specialized tools like Playwright Pro, Stripe integration, AWS, and MS365. 30+ Python automation tools (all stdlib-only). Works with Claude Code, Codex CLI, and OpenClaw.
Use when working with the OpenAI API (Responses API) or OpenAI platform features (tools, streaming, Realtime API, auth, models, rate limits, MCP) and you need authoritative, up-to-date documentation (schemas, examples, limits, edge cases). Prefer the OpenAI Developer Documentation MCP server tools when available; otherwise guide the user to enable `openaiDeveloperDocs`.
Guide for building TypeScript CLIs with Bun. Use when creating command-line tools, adding subcommands to existing CLIs, or building developer tooling. Covers argument parsing, subcommand patterns, output formatting, and distribution.