Two days ago, OpenAI shipped a model that can navigate your desktop, operate software, and complete office tasks with a 75% success rate — surpassing the reported human performance benchmark of 72.4% on the same test. GPT-5.4, released on March 5, 2026, isn't just an incremental step. It's OpenAI's first general-purpose model with native computer-use capabilities baked in from the ground up, and its benchmark results across professional work tasks represent the most substantial capability jump the company has shipped in a single release.
What GPT-5.4 Actually Is
OpenAI released GPT-5.4 across ChatGPT, its API, and its Codex platform simultaneously. The model ships in three tiers: a standard version, GPT-5.4 Thinking (the primary experience in ChatGPT), and GPT-5.4 Pro for users who need maximum performance on the most demanding tasks.
The architecture consolidates capabilities that had previously been split across specialized models. GPT-5.4 absorbs the coding strengths of GPT-5.3-Codex while adding new native computer-use capabilities and deep improvements to professional document work — spreadsheets, presentations, financial models, legal analysis. The goal, in OpenAI's framing, is a single model that "gets complex real work done accurately, effectively, and efficiently" without the back-and-forth that has plagued AI-assisted workflows.
On the API side, GPT-5.4 supports context windows of up to 1 million tokens — by far the largest OpenAI has offered. That's not a novelty spec; it directly enables the long-horizon planning and task verification that serious agentic use cases demand. An agent that can hold the equivalent of a full codebase, a legal case file, or months of financial records in context is a qualitatively different tool than one that can't.
The Computer-Use Benchmark Everyone Is Talking About
The number that has dominated AI industry discussion since Thursday is GPT-5.4's 75.0% success rate on OSWorld-Verified — a benchmark that tests a model's ability to navigate a real desktop environment using screenshots plus keyboard and mouse actions. For context: GPT-5.2 scored 47.3% on the same test. Human performance on the benchmark is pegged at 72.4%.
That's not a small gap. OpenAI's previous flagship cleared the bar less than half the time. GPT-5.4 now clears it in three out of four attempts and, by the benchmark's measure, outperforms the average human doing the same tasks. The Next Web's analysis called the OSWorld result "the marquee benchmark of the launch," noting it represents a genuine threshold crossing rather than a marginal improvement on an existing capability.
WebArena Verified — which tests web-based task completion rather than desktop navigation — also returned a record score for GPT-5.4, though OpenAI has not released the precise figure. Both benchmarks belong to a class of evaluations that specifically test whether an AI can accomplish real-world computer tasks autonomously, not just answer questions about how to do them.
The practical implication is significant. An agent running GPT-5.4 can now be reasonably expected to: open a spreadsheet, locate the correct data, run a formula, export results, switch to a presentation application, insert a chart, and save the file — all from a single natural-language instruction. That level of autonomous multi-application task completion has been the holy grail of enterprise AI deployment for the past three years.
GDPval and the Professional Work Benchmark
OpenAI has put significant emphasis on GDPval, its own benchmark for knowledge work. The test spans 44 occupations across the nine industries contributing most to U.S. GDP, asking models to produce real work products: sales presentations, accounting spreadsheets, urgent care schedules, manufacturing diagrams, short videos.
GPT-5.4 scores 83.0% on GDPval — matching or exceeding industry professionals in 83 of 100 comparisons. GPT-5.2 scored 70.9%. VentureBeat's coverage of the launch noted that the GDPval improvement is particularly striking for spreadsheet tasks: on an internal benchmark of investment-banking-analyst-level spreadsheet modeling, GPT-5.4 scored 87.3% versus 68.4% for GPT-5.2. Human raters also preferred GPT-5.4's presentations 68% of the time over GPT-5.2's, citing stronger visual design and more effective image generation.
The practical implication for enterprise users is that tasks previously requiring a skilled analyst — building a financial model, drafting a board presentation, synthesizing a legal memo from case law — are now within the reliable capability range of GPT-5.4 operating semi-autonomously.
Thinking Mode and the Chain-of-Thought Problem
In ChatGPT, users access GPT-5.4 through the Thinking interface, which now surfaces an upfront plan before the model begins executing. The design change is more than cosmetic. Users can review the model's intended approach and redirect it mid-response — addressing one of the most consistent frustrations with reasoning models, which have historically locked users out of the process until a lengthy, potentially wrong output is complete.
GPT-5.4 Thinking also improves deep web research performance, particularly on highly specific queries that require integrating information from multiple sources. OpenAI reports better context maintenance for questions requiring extended reasoning chains — a common failure point in earlier versions.
There is a safety dimension to the Thinking architecture that deserves attention. AI safety researchers have raised concerns about whether reasoning models can misrepresent their chain-of-thought — presenting a sanitized version of their actual decision process while concealing the reasoning that led to a given output. OpenAI included a new chain-of-thought safety evaluation in the GPT-5.4 launch, publishing results that suggest the Thinking version is less likely to exhibit deceptive chain-of-thought behavior. The company's conclusion: "the model lacks the ability to hide its reasoning and CoT monitoring remains an effective safety tool." Whether that framing holds under adversarial testing remains an open question in the AI safety community.
Tool Search and the Token Efficiency Play
A less-heralded but operationally significant change in GPT-5.4 is a new API system called Tool Search. In previous versions, system prompts had to include full definitions for every tool the model might call — a process that consumed a growing volume of tokens as agent architectures scaled up with more connectors, APIs, and integrations. Tool Search allows GPT-5.4 to look up tool definitions dynamically as needed, rather than loading them all at the start of every inference.
The result is faster, cheaper API calls in complex agentic systems with many available tools. For developers building enterprise-scale agents with dozens of integrations, the token savings compound rapidly across millions of requests.
Token efficiency is a recurring theme in the GPT-5.4 announcement. OpenAI says the model uses significantly fewer tokens than GPT-5.2 to solve equivalent problems — a claim consistent with the broader trend toward more capable but more efficient frontier models that has characterized the last eighteen months of competition from Anthropic, Google, and the open-source community.
Hallucination Reduction: The Numbers
OpenAI's stated improvements on factual accuracy are worth taking seriously, though they require careful interpretation. On a dataset of de-identified prompts where users previously flagged factual errors, GPT-5.4's individual claims are 33% less likely to be false relative to GPT-5.2. Complete responses are 18% less likely to contain any errors at all.
These are internal evaluations on a curated dataset, which limits how much weight independent observers should give them. That caveat noted, the direction of travel is consistent with what third-party testing has found in the days since launch: GPT-5.4 is measurably more reliable on factual claims than its predecessor. For professional use cases where a single factual error in a legal brief or financial model can have serious consequences, even a 30% reduction in individual claim errors represents a material improvement in practical reliability.
The Competitive Picture
GPT-5.4 lands in a market that has seen an unusual volume of frontier model launches in the past thirty days. Anthropic shipped Claude Opus 4.6 in early February. Google released Gemini 3.1 Pro on February 19. DeepSeek's V4 — a trillion-parameter multimodal model — is expected imminently. The compression of the competitive cycle means no single release maintains a clear performance lead for long.
What GPT-5.4 does establish, at least for now, is a lead on computer-use tasks specifically. Claude Opus 4.6's OSWorld scores have not matched GPT-5.4's reported 75%. Google's Gemini 3.1 Pro has a different architectural emphasis — stronger on multimodal reasoning and document analysis — but has not claimed the same computer-use benchmark results. For developers and enterprises whose primary use case is autonomous task execution on real computing environments, GPT-5.4 appears to be the current state of the art.
What It Means for Enterprise Deployment
The coordinated launch across ChatGPT, the API, and Codex — with the simultaneous release of a ChatGPT for Excel add-in — signals that OpenAI is not treating GPT-5.4 as an abstract research milestone. The product emphasis on spreadsheets, presentations, financial modeling, and legal analysis maps directly to the highest-value knowledge work in Fortune 500 environments.
For CIOs and AI platform teams currently evaluating which model to build agentic workflows around, the GPT-5.4 launch presents a meaningful inflection point. The 1M-token context window, native computer-use, Tool Search efficiency, and the GDPval results collectively describe a model that is not just more capable in benchmark conditions but more deployable in the messy reality of enterprise software environments where tasks span multiple applications, require long-horizon planning, and demand factual accuracy.
The question now is how quickly competitors close the gap — and whether OpenAI can maintain a consistent enough lead in computer-use specifically to make GPT-5.4 the default choice for the next generation of enterprise AI agents. Based on the pace of releases since January, the answer will be apparent within weeks.




