Best AI Agent 2026: Cowork vs ChatGPT vs Operator vs Manus
The best AI agent in 2026 depends on the kind of “autonomous work” you want — and the four serious managed agents have very different answers. Claude Cowork is Anthropic’s pair-programmer for knowledge work, bundled into paid plans. OpenAI’s Operator (browser-only) and ChatGPT Agent (general) are slowly converging. Manus shipped one of the most viral demos of 2025 and turned it into a credit-metered general-purpose agent. We compare capability, pricing, supervision, and the failure modes you should plan for before committing real workflows.

On this page · 13 sections▾
TL;DR + decision tree
- If your team is already paying Anthropic, start with Claude Cowork. It’s bundled into Pro, Max, Team, and Enterprise, and the pair-programming framing translates well to knowledge work: give it a folder and a goal, walk away, come back to deliverables. The same MCP servers you’ve already wired to Claude Desktop are available inside Cowork.
- If your team is already paying OpenAI, start with ChatGPT Agents (agent mode in the ChatGPT app). It pairs with custom GPTs, has the broadest tool surface of the four (browser, code interpreter, file analysis, custom GPT actions), and is available on Pro, Plus, Team, and Enterprise with usage caps that scale with plan.
- If you specifically need a browser-driving agent for tasks with no API — booking, comparison shopping, form filling on legacy admin panels — pick Operator. It’s the most specialised of the four and OpenAI’s Computer-Using Agent (CUA) is purpose-built for visual UI interaction. Expect CAPTCHA-shaped frustration.
- If you want the most autonomous, deliverable-shaped output and you’re comfortable with a credit-metered SaaS, try Manus. Its “give it a goal and walk away” positioning is the strongest in the category, the workspace produces tangible artefacts, and it’s the only one of the four that isn’t attached to an existing chat product.
None of these is a strict superset of the others. Cowork wins on desktop integration and MCP compatibility. ChatGPT Agents wins on tool breadth and custom GPT pairing. Operator wins on browser autonomy. Manus wins on workspace-shaped deliverables. The rest of this piece covers what each one actually does, where each one fails, and how to pick.
What “managed agent” actually means
The term gets thrown around loosely. For this comparison we use a strict definition: a managed agent is a product where the vendor runs the agent loop on their infrastructure. You submit a goal through a UI or an API; the vendor’s sandbox executes tool calls, holds session state, manages credentials, and produces output; you supervise and intervene as needed. The agent loop itself — the “plan, act, observe, repeat” cycle that defines modern agents — runs on machines you don’t own.
That distinguishes managed agents from three other shapes of agentic system you might confuse them with:
- CLI agents like Claude Code, Aider, Goose, OpenCode, and Cline run the loop on your machine. Your terminal is the harness; your filesystem is the workspace. The vendor sells you a model API; everything else is yours. Our deep dive at Goose vs Cline vs Aider vs Claude Code vs OpenCode walks through the five most popular options.
- IDE agents like Cursor, Windsurf, Antigravity, and Kiro run the loop inside an editor. They’re managed at the chat level but the file edits, terminal commands, and Git operations happen in your local workspace. See Cursor vs Windsurf vs Antigravity vs Kiro for that side of the market.
- DIY agent frameworks like LangGraph, AutoGen, smolagents, and the open Anthropic Agents SDK give you the building blocks to host your own loop. The vendor sells you a library; you sell yourself the runtime. This is the right answer when vendor neutrality matters more than time-to-production.
Managed agents trade flexibility for convenience. You give up control over how the loop runs and where state lives in exchange for credential storage, long-running execution that survives your laptop going to sleep, sandboxing that someone else maintains, and a UI surface that’s already integrated with the rest of the vendor’s product. For most non-developer users — the marketing-ops person, the financial analyst, the legal associate — managed is the only shape that makes sense. For developers, it’s often a complement to a CLI agent rather than a replacement: Cowork for the marketing-deck work, Claude Code for the actual code.
The other thing worth flagging up front: these four products all run closed-source agent loops on closed infrastructure. None of them is open source. None of them lets you self-host the runtime. If that’s a hard requirement — for compliance, data residency, or independence reasons — the answer isn’t a managed agent at all. It’s LangGraph or the open Agents SDK running on your own VPC.
Capability axes that matter
When teams shortlist managed agents they tend to focus on the demos. The demos are misleading. The axes that actually separate the four products in production are the boring ones:
- Browser autonomy. Can the agent click, type, and navigate real websites that don’t have an API? Operator was purpose-built for this; ChatGPT Agent inherited the capability; Cowork drives a desktop via Computer Use; Manus has its own Browser Operator. They all do it, but the failure modes are different (see CAPTCHAs in the pitfalls section).
- Async and long-running. Can the agent work for an hour while you do something else? Cowork is the most explicit about this — Anthropic’s framing is “Claude works while you do other things.” ChatGPT Agent supports long-running tasks with the “Tasks” surface for scheduling. Manus leans heavily into this with its workspace. Operator is the most session-bound of the four.
- Multi-step planning. All four can plan multi-step. The differentiator is how legibly the plan surfaces to you — Cowork shows a task plan you can intervene in; ChatGPT Agent does the same; Operator shows you the browser; Manus shows you the workspace.
- Code execution. Can the agent run code? Cowork has a sandboxed Linux VM. ChatGPT Agent has the code interpreter and a virtual computer. Operator is browser-only. Manus has its own execution environment.
- Third-party integrations. What outside systems can the agent talk to? Cowork uses MCP servers plus native connectors (Chrome, Slack, Drive, Zoom). ChatGPT Agent talks to custom GPT actions plus a growing list of native connectors. Operator stays inside the browser sandbox. Manus has a connector model including Slack and Mail.
- Pricing model. Bundled into a flat monthly plan (Cowork, ChatGPT Agents) or metered by credits/usage (Manus, increasingly ChatGPT Agent at the top end). This matters for predictability.
- Supervision affordances. What does intervention look like? Are there approval gates for irreversible actions? Per-tool permission controls? This is the axis enterprise buyers care about most and marketing pages talk about least.
Pick your top three axes from that list before reading the per-tool sections — you’ll get a sharper answer than if you let the demos decide for you.
Side-by-side matrix
Every cell is sourced from the vendor’s public product page, help docs, or reputable third-party coverage. Where current details are volatile (pricing, region availability) we’ve used evergreen phrasing — check the vendor page at purchase time.
| Dimension | Cowork | ChatGPT Agents | Operator | Manus |
|---|---|---|---|---|
| Vendor | Anthropic | OpenAI | OpenAI | Butterfly Effect |
| Plan / access | Pro, Max, Team, Enterprise | Pro, Plus, Team, Enterprise | Pro (and broader plans over time) | Free + paid (credit-based) |
| Browser autonomy | Yes (Computer Use + VM) | Yes (inherited from Operator) | Yes — flagship capability (CUA) | Yes (Browser Operator) |
| Async / long-running | Strong — pair-programming framing | Strong — Tasks scheduling + Agent | Session-bound | Strong — workspace deliverables |
| Code execution | Sandboxed Linux VM | Virtual computer + code interpreter | Browser sandbox only | Workspace execution |
| Third-party integrations | MCP + connectors (Chrome, Slack, Drive, Zoom) | Custom GPT actions + native connectors | In-browser (limited) | Manus connectors + Slack/Mail |
| Region availability | Global on paid plans | Global on paid plans | US first; expanded since | Global |
| Killer use case | Knowledge-work pair programming | ChatGPT-stack power user | Tasks without an API | Goal-then-walk-away delivery |
| Current limitation | Anthropic ecosystem lock-in | Tool sprawl in agent mode | CAPTCHA / 2FA failures | Credit metering anxiety |
Three takeaways from the matrix. First, three of the four are bundled into a flat subscription (Cowork into Claude paid plans, ChatGPT Agents into ChatGPT paid plans, Operator into ChatGPT Pro). Only Manus meters by credits — and credit anxiety is the most-cited complaint in user threads, fairly or not. Second, browser autonomy is now table stakes; the differentiation is in what the agent does around the browser session. Third, the “killer use case” column is the most important row in the table — pick the agent whose strength matches your most common task type, not the one that scores highest across every axis.
Claude Cowork
Managed agent
Claude Cowork
Anthropic · Pro / Max / Team / Enterprise (bundled)
What it does best
Cowork’s strongest framing is pair-programming for knowledge work. The product page describes it as taking a desired outcome and cadence, then handling execution while keeping you informed — recurring tasks, file organisation, data extraction from receipts and invoices, branded report generation, cross-device messaging where you queue work from your phone and the desktop completes it. Under the hood it’s an agent loop with file access, a sandboxed Linux VM, browser control via Computer Use, and MCP server compatibility. The MCP angle matters: anything you’ve already wired into Claude Desktop — Notion, Drive, GitHub, Stripe, your custom servers — shows up in Cowork the same way.
Pick this if you...
- Already pay for a Claude plan (Pro, Max, Team, Enterprise) and want a managed agent at zero marginal cost
- Have invested in MCP servers and want to reuse them in a managed-agent surface
- Prefer Anthropic’s approach to safety and alignment, including the “approve before destructive action” ethos
- Want desktop-native integration — local files, Chrome, Slack, Zoom — rather than a pure-cloud workspace
- Are comfortable being inside the Anthropic ecosystem for the foreseeable future
Getting started: first task in five minutes
Open the Claude desktop app on Mac or Windows. If you’re on a paid plan, Cowork is in the sidebar — no separate signup, no install steps. Point it at a folder of source files (PDFs, screenshots, spreadsheets) and give it a goal. A good first task is something boring and repetitive that produces a clean artefact:
Folder: ~/Downloads/2026-q1-receipts
Goal: Read every receipt and invoice in that folder.
Extract: vendor, date, amount in USD, category
(travel / saas / meals / hardware / other), and notes.
Produce a single Google Sheet with one row per receipt,
sorted by date descending. Flag any that look duplicate
or that I should question for tax purposes.
When done, send me a Slack DM with the spreadsheet link
and a 3-bullet summary of total spend by category.Cowork will surface the plan, ask for permission on the destructive bits (creating the sheet, sending the DM), and execute. Expect to supervise the first few runs closely — the value comes once you trust it on a known shape of task and can queue work from your phone while you’re away from the desk.
Skip it if...
You don’t use Claude and don’t plan to. Cowork is bundled with paid Claude plans — buying it standalone isn’t an option, and the lock-in is real. If your stack is ChatGPT-centric, ChatGPT Agent gives you a similar shape with better compatibility with custom GPTs.
ChatGPT Agents
Managed agent
ChatGPT Agents
OpenAI · Pro / Plus / Team / Enterprise (agent mode)
What it does best
ChatGPT Agent — agent mode in the ChatGPT app — has the broadest tool surface of the four. It can browse, run code in a virtual computer, analyse files, call custom GPT actions (the single biggest ecosystem advantage of the four), and chain everything in multi-step plans. The Tasks surface inside ChatGPT lets you schedule recurring or one-shot agent work without staying in the chat. For organisations already on ChatGPT — which, at this point, is a lot — Agent mode is the lowest-friction entry to the managed-agent category.
Pick this if you...
- Already pay for ChatGPT Pro, Plus, Team, or Enterprise and want a managed agent that’s included
- Have built custom GPTs with OpenAPI actions and want the agent to call them
- Need the broadest tool catalogue out of the box — browser, code interpreter, file analysis, image generation, native connectors
- Value the ChatGPT-app surface for context — the conversation, the memory, the prior chats
- Are comfortable with OpenAI’s pace of product change (high; this surface has reshaped twice in 18 months)
Getting started: first task in five minutes
In the ChatGPT app, switch to agent mode (the tool-and-gear control in the message composer — exact UI shifts with model picker updates). Type a goal that involves multiple steps and at least two tools:
Research the top three competitors of [our product].
For each one, capture:
- Pricing tiers and what each tier includes
- Their most recent product launch in the last 90 days
- Their public security/compliance posture
(SOC 2, ISO 27001, GDPR, HIPAA)
Compile into a comparison table, then write a 600-word
analysis of where we have advantage and where we don't.
Save the table and analysis to a single doc I can share.Agent mode plans, browses, summarises, and writes — surfacing the plan up front so you can intervene. The deliverable is a doc, the conversation captures the reasoning, and you can come back tomorrow and ask follow-ups in the same thread. Custom GPT actions can be wired into the same task if your organisation has them.
Skip it if...
Your team is on the Anthropic side of the fence — the equivalent capability ships free with Cowork on paid Claude plans and the MCP server ecosystem matches better. Also skip if you need pure browser autonomy as the entire job; Operator is more focused on that shape.
OpenAI Operator
Managed agent
OpenAI Operator
OpenAI · ChatGPT Pro (originally); expanded since
What it does best
Operator’s entire reason to exist is browser autonomy. The Computer-Using Agent (CUA) model behind it is purpose-built to look at a screen, read the visual UI, click buttons, type into fields, and navigate websites that don’t have an API. That sounds prosaic, but it’s the capability that unlocks the long tail of tasks the rest of the agent ecosystem can’t touch: booking a restaurant on a janky OpenTable competitor, refilling a prescription on a clinic portal, checking a permit status on a state-government site. The browser is the universal API; Operator drives it the way you do, badly and slowly but without complaint.
Pick this if you...
- Have a specific repeating task on a website with no API and you don’t want to scrape
- Are on ChatGPT Pro and want the dedicated surface for browser sessions rather than agent mode in the main app
- Don’t mind handing over a browser session with stored credentials inside OpenAI’s sandbox (security-team caveat)
- Are okay with the highest failure rate of the four on edge cases (CAPTCHAs, 2FA, weird modals)
Getting started: first task in five minutes
Go to operator.chatgpt.com. If you’re on a supported plan, Operator opens a dedicated browser surface. Pick a website that requires no login first — comparison shopping is a classic — and try a goal that has clear success criteria:
Goal: Find a flight from SFO to JFK on Monday next week,
returning Thursday. I want:
- A direct flight if available
- Economy class
- Between $300 and $600 total
- Aisle seat if possible
Compare what's available on Google Flights, Kayak, and
two major airline sites. Surface the best three options
with a one-line trade-off for each. Don't book — just
present the comparison and let me decide.Operator will open each site, search, surface results, and present a comparison. The careful framing — “don’t book” — is deliberate. For irreversible actions like purchases, you want Operator to stop before pressing the final button and ask you to confirm. The browser is real; the actions are real; treat it like giving someone your logged-in browser to use for ten minutes.
Skip it if...
You don’t actually need browser autonomy. If the task has an API (Stripe, GitHub, Salesforce, anything modern), it’s faster and more reliable to use Cowork with an MCP server or ChatGPT Agent with a custom GPT action. Operator excels at the long tail of no-API tasks; for the head it’s the wrong shape.
Manus
Managed agent
Manus
Butterfly Effect · Free + paid (credit-based)
What it does best
Manus took the “hand the agent a goal and walk away” framing further than any of its competitors when it launched in March 2025, and that DNA still defines the product. The site tagline — “Hands On AI. Less structure, more intelligence.” — sums up the bet: Manus wants you to describe deliverables, not workflow. The workspace produces tangible artefacts (slides, websites, desktop apps, designs), the Browser Operator handles the browser-autonomy axis, and Wide Research handles broad information-gathering tasks. Of the four, Manus has the most opinionated UI about what “output” looks like — the others are chat-shaped and produce files; Manus is workspace-shaped and produces things.
Pick this if you...
- Want a deliverable-shaped output (a slide deck, a one-pager, a generated website) rather than a chat thread
- Are willing to run on credit metering and to monitor credit burn the way you’d monitor cloud spend
- Don’t need to integrate with an existing Anthropic or OpenAI investment
- Want to evaluate the most autonomous of the four on tasks you don’t mind handing to a relatively new vendor
- Are intrigued by the viral demo videos and want to see how the product holds up on your own work
Getting started: first task in five minutes
Sign up at manus.im. The free tier gives you enough credits to try one substantial task; the paid tiers (Starter, Pro, Team) buy more credits per month, with the Team plan adding SSO and shared workspace features. Try something deliverable-shaped:
Build me a one-page marketing site for a fake product
called "Halftime" — a punctual meeting timer that ends
calls at the 30-minute mark with a 60-second wind-down.
Requirements:
- Modern, single-page layout, no signup form, no email capture
- Hero section, three feature cards, FAQ, footer
- Subtle teal-on-navy aesthetic
- Hand it back to me as a deployable HTML/CSS file
When done, also generate three social-share images for
LinkedIn, Twitter, and a slack-card preview.Manus will plan, generate, render, and produce downloadable artefacts in its workspace. The experience is closer to dropping a brief on a designer than typing a prompt at a chatbot. Whether the output is good enough depends on the task; the value proposition is that there is an output and you didn’t have to assemble it yourself.
Skip it if...
You hate credit-metered pricing for cognitive reasons. Some teams find “is this credit worth it?” anxiety more taxing than the actual cost — Cowork and ChatGPT Agent fold their cost into a flat monthly bill, which removes that decision overhead. Also skip if you want deep integration with Anthropic or OpenAI ecosystems; Manus is independent and connects outward via its own connector model rather than native parity.
Safety, supervision, and consequences
These four products are different from the chat-shaped AI assistants you used in 2023. Managed agents take actions in the real world. They send emails to your customers. They book rooms on your corporate card. They commit to merge requests. They edit files on your local disk. They acknowledge incidents in PagerDuty. The blast radius of a confused agent is larger than the blast radius of a confused chatbot, by roughly the distance between “saying something wrong” and “doing something wrong.”
The four products take broadly similar but not identical approaches to that problem. Each has three lines of defence:
- Sandboxing. All four run the agent inside an isolated environment — a VM, a virtual computer, a browser session. That contains the blast radius from “your laptop” to “the sandbox.” What it doesn’t contain is the network egress — if the sandbox has internet, it can email, post, click “send,” and pay.
- Approval gates. Each product asks for explicit permission before irreversible actions, though the bar for what counts as “irreversible” varies. Cowork is explicit about asking before destructive file operations and outbound messages. ChatGPT Agent and Operator both pause at purchases and credential-bearing steps. Manus surfaces a plan-then-execute pattern with intervention points. Felix Rieseberg, the head of Cowork, has publicly described “approve-every-command” as unsustainable UX — the goal is approve the consequential moments and trust the routine ones. That trust calibration is hard.
- Audit and observability. All four log the agent’s actions. Whether those logs land in your SIEM, your identity-provider audit trail, or just the vendor’s dashboard depends on the plan tier. Enterprise plans on Cowork, ChatGPT Agents, and (increasingly) Manus surface these logs in a form suitable for security review. Free and entry-tier plans usually don’t.
Practical advice. If you’re evaluating managed agents for production use, run them in a sandboxed identity first — a service account that has only the permissions the agent strictly needs, scoped to a single project or workspace. Watch the first dozen runs end-to-end before you trust the approval-gate UX. Treat the agent’s credentials the way you treat any production secret: short-lived, scoped, rotated, audited. And budget for the failure mode where the agent did exactly what you asked but you didn’t mean to ask for it — the most common incident type in agent-driven workflows is the user-error variety, not the vendor-error variety.
For workloads where the agent must never hold long-lived credentials at all — financial services, healthcare, anything regulated — managed agents are the wrong shape. The right shape is self-hosted agent runtimes inside your own VPC, even though the cost in engineering time is higher. None of the four products solves this constraint and none of them pretends to.
Pitfalls to plan around
Operator hits CAPTCHAs and stops — you can’t always tell at a glance
Every browser-driving agent — Operator, Cowork, ChatGPT Agent in browser mode, Manus Browser Operator — will hit a CAPTCHA at some point. The polite ones surface the blocker and pause. The less polite ones loop silently. Watch the first few sessions in real time. Plan for human-in-the-loop on any site that takes credentials.
Cowork lock-in is more real than the marketing suggests
Cowork sessions live inside the Claude desktop app. Task histories, configured skills, MCP wiring — none of that is portable to a different agent runtime without rebuilding. The trade-off is real: Cowork is excellent if you stay on Claude, but the migration cost out is higher than the marketing copy implies.
Manus credit anxiety is a real friction
The single most-cited complaint about Manus isn’t output quality — it’s the feeling of watching a credit meter while the agent works. Some teams love the pay-for-what-you-use model; others find it cognitively expensive to decide each task on its own merits. Know which kind of team you are before you adopt.
ChatGPT Agent tool sprawl confuses the model
Agent mode has the broadest tool surface of the four, which is usually an advantage and occasionally a problem. With browser, code interpreter, file analysis, image generation, custom GPT actions, native connectors, and memory all available, the model sometimes picks the wrong tool for the job. Pin the tool you want in the prompt for tasks where correctness matters more than convenience.
Operator vs Agent mode confusion inside OpenAI
OpenAI ships two surfaces — Operator at operator.chatgpt.com and Agent mode inside the main ChatGPT app — that overlap materially. Knowing which one to use for which task isn’t obvious. Rule of thumb: if the task is browser-only and session-bound, Operator. If it’s anything else, Agent mode in the main app.
Long-running tasks fail silently more often than they should
All four products will, eventually, leave you with a task that looks like it’s running but has actually stalled. Cowork and Manus have the strongest async story but even they aren’t bulletproof. Set yourself a calendar reminder to check on multi-hour tasks. Don’t assume silence is progress.
The Manus “Meta acquisition” story is not closed
Meta announced an acquisition of Butterfly Effect in December 2025; China’s NDRC blocked the deal in April 2026. Manus remains independent. If your organisation has a policy on Meta vendors specifically, the answer today is “Manus is not Meta.” Recheck before you sign a contract — these deals sometimes get refiled.
Community signal
The verbatim community signal on this category is still thinner than the hype suggests — most of these products are less than 18 months old. We’re using only sources we can verify through public coverage and product pages.
Anthropic’s own framing of Cowork on the product page describes it as having Claude “handle execution while keeping users informed.” Felix Rieseberg, head of the Cowork product, has been the loudest public voice from Anthropic on the question of how much approval the user should have to give — his framing in the Latent Space podcast was that “approve every command” is not a real long-term UX, and that sandboxing plus VMs matter more than constant human approval. We covered the broader Cowork-and-Managed-Agents story in detail at Claude Cowork, Managed Agents & Claude Design.
For Manus the verifiable public signal is the product’s viral early-2025 launch trajectory: Wikipedia documents that the invitation-only beta demonstration—showing autonomous task completion like resume screening and stock analysis—“drew more than one million views within twenty hours” of going live, and that invitation codes were resold on Chinese platforms for between $7,000 and $13,800. Reception has been mixed since then; the early demo set expectations the product has spent two years either meeting or falling short of, depending on the workload.
On Operator, the most consistent public feedback in the months after launch was about what it couldn’t do rather than what it could — CAPTCHAs, two-factor flows, and visually complex UIs trip it up at materially higher rates than humans, and OpenAI’s later positioning of ChatGPT Agent as the broader agent surface (with Operator’s browser-control capability folded in) suggests the dedicated-destination model wasn’t where the product team wanted to stay. Treat Operator as a tactical tool for browser-only tasks; treat Agent mode as the strategic surface.
Frequently asked questions
What's the difference between a managed agent and a CLI agent like Claude Code or Aider?
A managed agent runs the agent loop on the vendor's infrastructure: you submit a goal, the vendor's sandbox executes tool calls, you receive deliverables. A CLI agent like Claude Code, Aider, or Goose runs locally — the loop, file edits, and shell commands all happen on your machine. The trade-off is cognitive: managed agents handle credential storage, sandbox isolation, and long-running execution for you but lock you into the vendor's runtime; CLI agents give you full control but expect you to supervise every step. Cowork, ChatGPT Agents, Operator, and Manus are all managed. For the CLI alternatives we've covered them at /blog/goose-vs-cline-vs-aider-vs-claude-code-vs-opencode-2026.
Is Operator still a separate product from ChatGPT Agents?
OpenAI shipped Operator as a standalone research preview in January 2025 (operator.chatgpt.com) and then introduced ChatGPT Agent in July 2025 as a more general agent mode inside ChatGPT itself. As of 2026 the two surfaces still coexist but OpenAI's product direction has been to fold browser-control capability into the broader Agent mode rather than keep Operator as a separate destination. If you're starting fresh today, ChatGPT Agent mode in the main app is usually the right entry point — Operator the destination is more useful for narrow, dedicated browser sessions.
How much does each of these cost in 2026?
Claude Cowork is bundled into paid Claude plans (Pro, Max, Team, Enterprise) — no separate fee. ChatGPT Agent mode is available on ChatGPT Pro at $200/mo with the highest message allowance; Plus and Team get more limited access. Operator was originally Pro-only ($200/mo) at launch and has expanded since. Manus runs on a credit-based system: a free tier with limited monthly credits, paid plans (Starter, Pro, Team) that buy more credits per month. Always confirm at the vendor's pricing page before you commit a budget — these change often.
Which of these can actually book a flight or fill out a form?
Operator was built specifically for that — its Computer-Using Agent (CUA) clicks, types, and navigates in a sandboxed browser. ChatGPT Agent inherits the same browser-control capability and adds a virtual computer with a terminal and code interpreter. Cowork can drive a desktop via Claude Computer Use plus a sandboxed Linux VM, so it covers similar ground. Manus also includes a Browser Operator and runs tasks in its own workspace. All four can technically book things; in practice all four also fail in nearly identical ways on CAPTCHAs, two-factor flows, and login walls. Plan for human-in-the-loop.
Can ChatGPT Agents talk to my custom GPTs?
Yes — that's one of the stronger integration arguments for ChatGPT Agent over the others. Custom GPT actions (the OpenAPI-described tools you've wired up for your GPT) are callable from inside an agent session, so any organisation that has already built around custom GPTs has an instant adoption ramp. Cowork uses MCP servers for the equivalent role; Operator stays mostly inside the browser sandbox; Manus has its own connector and tool model. Custom GPT compatibility is the single biggest reason teams already on ChatGPT pick Agent mode over the alternatives.
Is Manus actually owned by Meta now?
No. Meta announced an acquisition of Butterfly Effect (Manus's parent company) in December 2025 at a reported $2-3B valuation, but on April 27, 2026 China's National Development and Reform Commission blocked the deal and required the parties to withdraw. As of this writing Manus remains independent under Butterfly Effect, based in Singapore. Some marketing material online still reflects the pre-block period, which is why you may see conflicting claims — treat the deal as not closed.
Which is safest to give long-lived credentials to?
There's no clean winner — all four are closed-source SaaS holding tokens on their infrastructure. The relative-safety axes are: (1) credential scoping — Cowork and ChatGPT Agent both expose granular per-tool connector controls; Operator and Manus are more all-or-nothing on the browser sandbox side. (2) Audit logging — Anthropic and OpenAI's enterprise plans have SOC 2 Type II coverage and SIEM-friendly logs; Manus's enterprise story is newer. (3) Egress controls — all four sandbox network egress to some degree but only the enterprise tiers expose configurable egress allowlists. If credentials must never leave your environment, you don't want a managed agent — you want a self-hosted agent SDK.
Can I use these in Europe or only the US?
Cowork: rolled out in the same regions as Claude itself; broadly available globally on paid plans. ChatGPT Agent: available in most regions where ChatGPT is, with EU rollout following Pro/Plus availability. Operator: launched US-only in January 2025 and expanded to additional regions throughout 2025; check operator.chatgpt.com for current availability before you commit. Manus: global availability via manus.im, with a credit-based plan model that works in most regions. Always check the vendor's current region list at purchase time — these change with payment processors and local regulation.
What's the right one to start with if I want to evaluate the category?
If your team is already on Claude paid plans, Cowork is the lowest-friction starting point — it's included, no separate signup. If your team is on ChatGPT Pro/Team, ChatGPT Agent is similarly free-at-margin. Operator is worth a try only if browser control is the entire point. Manus is the most opinionated of the four, with a workspace-and-deliverables framing — it's best evaluated on a self-contained task you don't mind handing to a credit-metered SaaS. Pick whichever you already have a subscription for; run the same prompt across all four if you can; pick by output quality on tasks you actually do.
Are there open-source alternatives if I don't want a managed agent?
Yes — and the gap is closing. For agent loops you can self-host: LangGraph, AutoGen, smolagents, OpenHands, and the open-source Agents SDK (Anthropic's open library that underpins Cowork) all let you build the same shape of system on your own infrastructure. The cost is that you write your own sandboxing, persistence, and credential handling — which is the work the managed services charge for. For most enterprises the choice isn't open-vs-managed but managed-for-the-common-case and self-hosted for sensitive workloads. Our CLI agent comparison at /blog/goose-vs-cline-vs-aider-vs-claude-code-vs-opencode-2026 covers the developer side; for production agent runtimes the picture is still moving quickly.
Sources
Claude Cowork
- claude.com/cowork — product page, capabilities, framing
- anthropic.com/news — launch and enterprise updates
- /blog/claude-cowork-managed-agents-claude-design-2026-explained — deep dive companion piece
ChatGPT Agents
- chatgpt.com — agent mode in the main app
- help.openai.com — ChatGPT Agent help docs (Plus / Pro / Team / Enterprise availability)
- en.wikipedia.org/wiki/ChatGPT — Operator (Jan 2025) and Agent (Jul 2025) launch history
OpenAI Operator
- operator.chatgpt.com — dedicated browser-control surface
- openai.com/index/introducing-operator — launch post (Jan 2025)
Manus
- manus.im — product page, capabilities, tagline
- manus.im/pricing — free + paid plans, credit model
- en.wikipedia.org/wiki/Manus_(AI_agent) — Butterfly Effect history, March 2025 launch, Meta deal block
Related comparisons
- /blog/claude-cowork-managed-agents-claude-design-2026-explained — Cowork deep dive
- /blog/goose-vs-cline-vs-aider-vs-claude-code-vs-opencode-2026 — CLI agents (the unmanaged alternative)
- /blog/cursor-vs-windsurf-vs-antigravity-vs-kiro-2026 — IDE agents (the editor-shaped alternative)
Internal links
- /blog/what-is-mcp — MCP primer (Cowork uses MCP under the hood)
- /best-mcp-servers — curated server roundup
- /servers — browse the directory
Bottom line
Pick by the subscription you already pay for. Cowork on paid Claude. ChatGPT Agents on paid ChatGPT. Operator only when browser autonomy is the whole job. Manus when you want the most goal-then-walk-away of the four and you can live with credit metering. Run the same representative task across two of them before you commit a workflow.