
Glif
OfficialConnects to Glif.app to run AI workflows like image generators, meme creators, and chained LLM processes directly from your MCP client.
Run AI workflows hosted on Glif.app via MCP, including ComfyUI-based image generators, meme generators, selfies, chained LLM calls, and more
What it does
- Run AI workflows with custom inputs
- Search and discover available workflows
- Generate images using ComfyUI-based tools
- Create memes and process selfies
- Chain multiple LLM calls together
- Browse workflow details and user information
Best for
About Glif
Glif is an official MCP server published by glifxyz that provides AI assistants with tools and capabilities via the Model Context Protocol. Run AI workflows on Glif.app via MCP — ComfyUI image generators, meme & selfie creators, chained LLM calls and more for It is categorized under ai ml, developer tools.
How to install
You can install Glif in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.
License
Glif is released under the MIT license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.
glif-mcp-server
MCP server for running AI workflows from glif.app.
For more info check out https://glif.app or join our Discord server: https://discord.gg/glif
Features
- Run workflows with inputs
- Get detailed information about workflows, runs, and users
- Search and discover workflows
- Browse and load agents
Setup
Running via npx (recommended)
If you have nodejs installed, you can run our @glifxyz/glif-mcp-server package via npx:
-
Get your API token from https://glif.app/settings/api-tokens
-
Add the server in your Claude Desktop config file. On macOS this is:
~/Library/Application Support/Claude/claude_desktop_config.json{ "mcpServers": { "glif": { "command": "npx", "args": ["-y", "@glifxyz/glif-mcp-server@latest"], "env": { "GLIF_API_TOKEN": "your-token-here" } } } }
Running from a local checkout
git clone https://github.com/glifxyz/glif-mcp-server
cd glif-mcp-server
npm install
npm run build
Then configure your MCP client:
{
"mcpServers": {
"glif": {
"command": "node",
"args": ["/path/to/glif-mcp/build/index.js"],
"env": {
"GLIF_API_TOKEN": "your-token-here"
}
}
}
}
You can also specify workflow IDs (comma-separated) which will be loaded automatically:
{
"mcpServers": {
"glif": {
"command": "node",
"args": ["/path/to/glif-mcp/build/index.js"],
"env": {
"GLIF_API_TOKEN": "your-token-here",
"GLIF_IDS": "cm2v9aiga00008vfqdiximl2m,cm2v98jk6000r11afslqvooil"
}
}
}
}
Configuration
Environment variables:
GLIF_API_TOKEN- Required. Your API token from https://glif.app/settings/api-tokensGLIF_IDS- Optional. Comma-separated workflow IDs to load as tools automaticallyIGNORE_DISCOVERY_TOOLS- Set totrueto disable discovery toolsAGENT_TOOLS- Set totrueto enable agent tools
Tools
Core Tools
run_workflow- Run a workflow with the specified ID and inputsworkflow_info- Get detailed information about a workflow
Discovery Tools (enabled by default)
list_featured_workflows- Get a curated list of featured workflowssearch_workflows- Search for workflows by name or descriptionmy_workflows- Get a list of your workflowsmy_user_info- Get detailed information about your account
Agent Tools (disabled by default, enable with AGENT_TOOLS=true)
list_agents- Get a list of agents with optional filteringload_agent- Load an agent and get its details including personality and workflows
Resources
glif://{id}- Get workflow metadataglifRun://{id}- Get run detailsglifUser://{id}- Get user profile
Development
npm install
npm run build
npm run dev # auto-rebuild
npm run test # run tests
Debugging
npm run inspector
Releasing
- Edit
package.jsonandsrc/index.tsand bump the version - Run
npm installto update lockfile - Commit and push to main
- Run
npm run release(requiresghCLI)
License
MIT - see LICENSE
Alternatives
Related Skills
Browse all skillsUI design system toolkit for Senior UI Designer including design token generation, component documentation, responsive design calculations, and developer handoff tools. Use for creating design systems, maintaining visual consistency, and facilitating design-dev collaboration.
Answer questions about the AI SDK and help build AI-powered features. Use when developers: (1) Ask about AI SDK functions like generateText, streamText, ToolLoopAgent, embed, or tools, (2) Want to build AI agents, chatbots, RAG systems, or text generation features, (3) Have questions about AI providers (OpenAI, Anthropic, Google, etc.), streaming, tool calling, structured output, or embeddings, (4) Use React hooks like useChat or useCompletion. Triggers on: "AI SDK", "Vercel AI SDK", "generateText", "streamText", "add AI to my app", "build an agent", "tool calling", "structured output", "useChat".
Master API documentation with OpenAPI 3.1, AI-powered tools, and modern developer experience practices. Create interactive docs, generate SDKs, and build comprehensive developer portals. Use PROACTIVELY for API documentation or developer portal creation.
Use when working with the OpenAI API (Responses API) or OpenAI platform features (tools, streaming, Realtime API, auth, models, rate limits, MCP) and you need authoritative, up-to-date documentation (schemas, examples, limits, edge cases). Prefer the OpenAI Developer Documentation MCP server tools when available; otherwise guide the user to enable `openaiDeveloperDocs`.
Guide for building TypeScript CLIs with Bun. Use when creating command-line tools, adding subcommands to existing CLIs, or building developer tooling. Covers argument parsing, subcommand patterns, output formatting, and distribution.
Integrate Vercel AI SDK applications with You.com tools (web search, AI agent, content extraction). Use when developer mentions AI SDK, Vercel AI SDK, generateText, streamText, or You.com integration with AI SDK.