top of page

OpenAI Codex Can Now Control Your Desktop: What It Means for the AI Coding Agent Race

On April 16, OpenAI shipped a Codex update that lets the tool open any app on your Mac, move a cursor, click buttons, and type, without you touching the keyboard. Three weeks earlier, Anthropic had done the same thing with Claude Code, which TechCrunch had already described as "the tool of choice for many businesses." OpenAI was playing catch-up, and the April 16 release was the clearest signal yet that it knows it.

This is not just a feature announcement. What's actually happening is that the definition of an AI coding agent (software that helps developers write, test, and ship code) is being rewritten in real time. The category is moving from autocomplete tool to autonomous desktop operator, and two of the most-funded AI companies in the world are racing to define what that looks like at the consumer subscription level.

This piece breaks down what the April 16 update actually delivered, why it matters beyond the marketing copy, where the race between Codex and Claude Code really stands on the numbers, and what the security community has been saying that OpenAI's press releases haven't.

What Happened

The headline capability in OpenAI's "Codex for (almost) everything" release is Background Computer Use, a feature that lets Codex operate macOS applications autonomously using its own cursor. It can see the screen, click interface elements, and type, across any app, including those without APIs. Multiple Codex agents can run in parallel in the background, without disrupting whatever the user is doing in the foreground. According to OpenAI's developer documentation, agents run in isolated processes and communicate asynchronously with the main application.

This is the first time computer use AI (the paradigm of AI controlling a desktop GUI through screen capture and simulated input) has shipped inside a consumer-facing, subscription-tier product at scale. Anthropic pioneered the concept as a research preview in October 2024. By April 2026, both companies have turned it into something you can buy for $20 a month.

The April 16 update bundled five additional capabilities alongside Background Computer Use. An in-app browser lets users annotate live webpages with comments, giving Codex point-and-click precision for frontend and design workflows. A Memory preview tracks past sessions, tech stack preferences, and recurring workflows, surfacing suggestions for where to pick up on prior projects. Image generation via gpt-image-1.5 lets Codex produce UI mockups, game assets, and placeholder visuals within the same workflow. More than 90 plugins built on the Model Context Protocol (MCP), an open standard originally introduced by Anthropic, now connect Codex to Atlassian Rovo, GitLab Issues, Slack, Microsoft Suite, Google Calendar, and over 80 other services. The update also ships GPT-5-Codex, a version of GPT-5 optimized for agentic coding tasks.

To make this concrete: picture three Codex agents running simultaneously on the same machine. One is writing feature code in a repository. A second is executing the test suite. A third is generating UI image assets from a design brief. Meanwhile, the developer is writing documentation in a fourth window, and none of the agents interfere with each other or with the foreground app. This is not a hypothetical scenario; OpenAI describes it directly in the announcement and in developer documentation as a supported workflow.

The growth numbers behind this release are notable. OpenAI cited more than 3 million weekly active developers on Codex as of April 2026, nearly double the 1.6 million reported in early March. Codex CLI npm downloads went from 82,000 in April 2025 to 14.53 million in March 2026 (a 177x increase), according to data published by AI Code Detector. Token usage was growing more than 70% month over month through Q1 2026, according to Fortune.

Why It Matters

The shift that the April 16 update represents is not incremental. The AI coding agent category, which began as a smarter autocomplete tool with GitHub Copilot in 2021, has now crossed into territory where an AI can autonomously operate your entire computer on your behalf. That is a qualitatively different kind of software, and it changes what developers (and increasingly, non-developers) can expect from these tools.

Temporal, a workflow automation company, is using Codex to accelerate feature development, debug issues, write and execute tests, and refactor large codebases, with teams reporting 2–3x faster iteration cycles. Superhuman, the email client, is using Codex to let product managers directly contribute lightweight code changes without pulling in an engineer, closing a gap that has historically required either a dedicated engineering sprint or a long backlog wait.

The non-developer angle is arguably more significant than the developer productivity story. A CEO cited in a Hacker News discussion reportedly used Claude Code's equivalent computer-use workflow to complete a social media campaign in two hours that "would have taken 3–4 weeks." Another Hacker News commenter described a non-technical family member successfully solving a complex scheduling optimization problem by implementing a Python algorithm through AI assistance, without understanding the language at all. These are not edge cases. They are early signals that desktop automation AI is dissolving the boundary between technical and non-technical work.

According to data cited by AI Automation Global, 73% of professional developers reportedly use AI coding agents daily as of mid-2026. The AI coding market is projected to reach $26 billion by 2030. These are large numbers, but they underscore a structural shift: AI-assisted development has moved from an early-adopter curiosity to a default workflow assumption in less than two years.

Sam Altman articulated OpenAI's strategic logic in a press call covered by The Neuron AI: "If you really want to do sophisticated work on something complex, [GPT-5.2] is the strongest model by far. However, it's been harder to use, so taking that level of model capability and putting it in a more flexible interface, we think is going to matter quite a bit." The framing is instructive: the ambition is not a better code editor, but a general-purpose interface layer for complex knowledge work.

For developers managing multiple tools and context across their workday, this also raises a practical challenge: as these agents grow more autonomous, keeping the underlying knowledge and context they need (documentation, codebase understanding, team workflows) organized and accessible becomes a separate problem worth solving. We'll return to this in the closing.

The Security Problem Nobody Wants to Talk About

On the same day OpenAI published the April 16 release, Help Net Security ran an analysis headlined: "Codex can now operate between apps. Where are the boundaries?" It is a question nobody has answered well, and it points to something more serious than a product limitation.

When an AI coding agent can open your email client, click into your browser, read your calendar, and type into any application, the attack surface for that agent is no longer the code repository; it is your entire desktop. OpenAI's own developer documentation acknowledges this directly, listing prompt injection, exfiltration of code or secrets, inclusion of malware or vulnerabilities, and license restriction violations as "elevated risks," particularly when internet access is enabled. Their official guidance: "only allow necessary domains and methods, and always review Codex's outputs and work log."

That guidance is not wrong. It is also an admission that using this tool without review is risky.

The theoretical risk became a demonstrated vulnerability before the April 16 update even shipped. In December 2025, security researchers at BeyondTrust's Phantom Labs discovered a critical command injection flaw in Codex's CLI, SDK, and development environment integrations. The attack vector was precise: embed malicious commands inside a GitHub branch name. When Codex processed the branch without sufficient input sanitization, it executed the payload and leaked the user's GitHub authentication token. Because the attack could be automated across shared repositories, the potential enterprise blast radius was significant; a single compromised project could propagate to every contributor. OpenAI patched the vulnerability on February 5, 2026, more than 50 days after BeyondTrust's initial report.

The Hacker News thread on the April 16 announcement surfaced a more structural concern. One commenter articulated it precisely: in agentic contexts, "data becomes effectively an executable." This is not a metaphor. When an AI agent can read an email, see a webpage, and act on its contents, any visible content becomes a potential instruction. A malicious actor does not need to compromise the agent directly; they need to put a prompt injection in something the agent will read. That could be a webpage the agent browses, a document it opens, or a notification it sees. The attack surface is no longer defined by what the AI can access; it is defined by everything the AI can perceive.

A second concern in the Hacker News discussion is a product-level trust problem. Non-technical users often expect an AI agent to function like a capable executive assistant, handling vague instructions, inferring intent, and completing tasks without requiring constant supervision. Current large language models, according to skeptics in the thread, do not reliably meet this expectation in unsupervised autonomous contexts. The gap between marketing language ("background agents running while you work") and actual reliability for ambiguous tasks is real, and it is most visible at the non-developer end of the user base.

It is worth noting that Codex's OS-level sandboxing architecture (using bubblewrap (bwrap) on Linux/WSL2 and Windows Sandbox on Windows) provides stronger process isolation than Claude Code's application-layer hooks. The sandboxing limits what an agent can write to outside its workspace directory, and network access is disabled by default. But the existence of the sandbox is itself a signal: OpenAI built these constraints because unrestricted desktop access creates real risk, not hypothetical risk.

OpenAI has also launched a separate product called Codex Security, which scanned 1.2 million code commits and reportedly found 10,561 high-severity issues. The framing is that the same agentic layer can identify "complex vulnerabilities that other agentic tools miss." The dual positioning (using AI agents to create productivity while using a different AI agent to scan for the resulting security problems) is an accurate reflection of where the industry is.

Codex vs. Claude Code: The Numbers Behind the Race

The benchmark picture is more complicated than either company's marketing suggests, and understanding the split is essential for choosing which tool fits a given workflow.

Neither Codex nor Claude Code is winning across all dimensions; they lead in different areas, and the gap varies significantly depending on what you measure. On SWE-bench, the standard benchmark for autonomous software engineering tasks, Claude Code reportedly scores 72.5% against Codex's approximately 49%. That is a substantial difference on a widely cited metric. On Terminal-Bench 2.0, which measures command-line task performance, GPT-5.3-Codex scores 77.3% compared to Claude Code's 65.4%. GitHub Copilot sits at 56% SWE-bench and Cursor at 52%; both trail the two leaders and, critically, neither currently offers desktop-level computer use.

The timeline of the race is as revealing as the benchmarks. Anthropic launched macOS desktop control on March 24, 2026. OpenAI followed on April 16, three weeks later. More pointed: on April 14, two days before OpenAI's announcement, Anthropic released a redesigned Claude Code desktop app with parallel sessions and automated Routines, which are scheduled workflows triggerable via API or GitHub events. The sequence does not read like coincidence. It reads like Anthropic anticipating a competitor's announcement and moving to shape the narrative before OpenAI could set it.

On product differentiation, the two tools have distinct strengths. Claude Code's 1 million token context window (in Opus 4.6 beta) gives it a significant advantage for large-codebase reasoning and multi-file refactors, the kinds of tasks where a narrow context window produces incomplete or inconsistent results. Codex claims a roughly 3x token efficiency advantage for equivalent tasks, which translates directly to cost at scale. The Codex plugin ecosystem at 90+ integrations is broader today, though Claude Code supports the same MCP standard and is expanding rapidly.

Pricing has not become a differentiator. Both tools are available at $20 per month on their respective base plans and $100 per month on higher tiers. OpenAI introduced a new $100 Pro tier on April 9 specifically targeting heavier Codex users, offering 5x more usage than the base plan. Anthropic's Max plan, also at $100, provides equivalent access to Claude Code at higher limits. The price war has not started. The feature race has.

Devin, from Cognition, remains the most fully autonomous option on the market (it runs its own shell, browser, and editor) but is positioned at the enterprise end, used by companies like Nubank for large-scale ETL refactoring at claimed 12x efficiency gains. For individual developers, Devin is not the primary consideration. The contest that actually matters for most practitioners is between Codex and Claude Code.

What's Next

The April 2026 releases from both Anthropic and OpenAI represent a threshold crossing: computer use AI has moved from research preview to consumer subscription standard. What happens in the next six months will be shaped by several converging pressures.

The computer use feature set will expand and standardize. Codex's Background Computer Use is currently limited to macOS; Windows support was not included in the April 16 release, though the Codex CLI supports Windows via WSL2. EU and UK markets are flagged for an upcoming rollout. As both companies push toward full cross-platform parity, desktop-level AI control will become a baseline expectation, not a differentiating feature. The question will shift from "which tool has computer use?" to "which tool executes it more reliably and with better judgment?"

Multi-agent coordination is the next architectural frontier. Both OpenAI and Anthropic have shipped parallel agent execution, meaning multiple agents running simultaneously on different tasks. The subsequent challenge is agent-to-agent collaboration: not just parallel independent runs, but agents that can delegate subtasks to each other, report results, and coordinate across a shared objective. Anthropic's Routines feature, which allows Claude Code to be triggered by GitHub events or API calls on a schedule, is an early step in this direction.

The non-developer market is the larger growth vector. Professional developers are already saturated adopters; 73% reporting daily use means the penetration ceiling within that group is close. The incremental market is product managers, designers, content strategists, and operations teams who currently rely on engineering queues for tasks that autonomous desktop agents can now execute directly. This is not a niche scenario. Superhuman enabling PMs to ship code changes without engineering involvement is a preview of a structural shift in how cross-functional teams work.

Regulatory frameworks have not caught up. The EU AI Act and comparable governance structures were designed around AI generating content or making recommendations, not AI autonomously operating a user's desktop. Computer use AI in consumer products at this scale is a 2026 development. Policy discussions around oversight, liability, and data control in agentic contexts will accelerate in the second half of the year, and the companies currently shipping these capabilities will be the primary reference points.

The category boundary between AI coding agent and general AI agent is being actively erased. The New Stack characterized OpenAI's strategy as building a "developer superapp": Codex as an operating layer for the full developer workflow, not a narrow coding tool. That framing increasingly applies to both Codex and Claude Code. At the point where an AI can control any app, access any service, and maintain memory across sessions, "coding agent" is an undersell. What's being built is closer to an autonomous computing layer.

Staying Oriented When Your AI Agent Can Do Everything

The Hacker News commenter who noted that "data becomes effectively an executable" in agentic contexts identified something more general than a security concern. When an AI coding agent operates across your entire desktop, the quality of its work depends directly on the quality of the context it has access to: your documentation, your codebase conventions, your team's current priorities, your past decisions.

Agents that run without structured context will make worse decisions, produce less consistent output, and require more human review to catch errors. The productivity gains that Temporal and Superhuman describe are not just about the agent's capability; they reflect teams that have built the organizational scaffolding around the agent: clear documentation, defined boundaries, reviewed outputs.

For developers and knowledge workers navigating multiple AI tools across a workflow, building a searchable knowledge base is the part of the stack that does not come pre-installed. As AI coding agents become more autonomous, the knowledge layer they draw on (and whether that layer is organized or fragmented) will determine how much of that autonomy is actually useful. That is the problem worth thinking about now, before the agents get any more capable.

Get started for free

A local first AI Assistant w/ Personal Knowledge Management

For better AI experience,

remio only supports Windows 10+ (x64) and M-Chip Macs currently.

​Add Search Bar in Your Brain

Just Ask remio

Remember Everything

Organize Nothing

bottom of page