GPT-5.4 Accidentally Leaked in OpenAI Codex Repository: What the Code Reveals
Developers spotted GPT-5.4 references twice in OpenAI's public Codex repo pull requests, revealing a 2-million-token context window and full-resolution image handling before the code was scrubbed.
Developers spotted GPT-5.4 references twice in OpenAI's public Codex repo pull requests, revealing a 2-million-token context window and full-resolution image handling before the code was scrubbed.
A Model Name That Was Not Supposed to Be Public
On February 28, 2026, a pull request appeared in OpenAI's public Codex GitHub repository containing something developers were not meant to see: a version cutoff reading "GPT-5.4 or newer" for an image handling feature. The reference was quickly scrubbed via force pushes and edits, but not before X user @scaling01 flagged it publicly. Then on March 2, a second Codex pull request surfaced with a direct function call using gpt-5.4 as the model argument, plus a slash command explicitly described as "toggle Fast mode for GPT-5.4." An OpenAI employee named Tibo compounded the situation by accidentally posting a screenshot on X that included GPT-5.4 in the model picker before deleting it.
These are not speculative rumors. They are code artifacts from OpenAI's own repositories, providing concrete evidence that GPT-5.4 exists internally and is actively being developed for production deployment.
What the Leaked Code Reveals
Two-Million-Token Context Window
The most significant technical detail extracted from the leaked code is a context window of 2 million tokens. For perspective, GPT-5.3 Codex operates with a 256,000-token context window. An eight-fold expansion to 2 million tokens would place GPT-5.4 in territory currently occupied only by Google's Gemini 3.1 Pro, which offers a 1-million-token window. If the 2-million-token figure holds, GPT-5.4 would set a new record among commercially available frontier models.
A context window of this size has practical implications for enterprise users. Legal firms could process entire case files in a single conversation. Software development teams could load complete codebases for analysis. Research teams could feed entire paper corpora for synthesis. The shift from hundreds of thousands to millions of tokens is not incremental—it fundamentally changes what tasks are feasible in a single model interaction.
Full-Resolution Image Handling
The leaked pull request adds a feature flag that preserves original PNG, JPEG, and WebP image bytes, passing them directly to the Responses API without compression. Current GPT models compress uploaded images before processing, which can degrade analysis quality for tasks requiring pixel-level precision—medical imaging, satellite imagery analysis, document OCR, and design review among them.
The code sets the minimum model version to (5, 4), meaning this feature explicitly requires GPT-5.4 or newer. This is not a general capability being backported; it is a GPT-5.4-specific enhancement.
Fast Mode Toggle
The slash command for "toggle Fast mode for GPT-5.4" suggests OpenAI is building a speed-optimized inference path, likely similar to the existing approach where models offer different speed-quality tradeoffs. This aligns with OpenAI's recent pattern of releasing both standard and optimized variants of each model generation.
Timeline and Release Probability
OpenAI has maintained an accelerating release cadence within the GPT-5 series:
| Model | Release Date |
|---|---|
| GPT-5 | August 2025 |
| GPT-5.1 | November 2025 |
| GPT-5.2 | January 2026 |
| GPT-5.3 Codex | February 5, 2026 |
The interval between releases has compressed from three months to roughly one month. If this pace continues, a March or early April release for GPT-5.4 is plausible.
Prediction markets on Manifold currently give GPT-5.4 a 55% probability of shipping before April 2026 and 74% before June. However, shipping both GPT-5.3 and GPT-5.4 within the same month would require unprecedented acceleration even by OpenAI's recent standards.
OpenAI has made no official announcement regarding GPT-5.4. The company has not commented on the leaked code references.
Competitive Context
The timing of these leaks is notable. OpenAI is navigating a period of significant public relations pressure following its Pentagon classified network deal, which has triggered the "Cancel ChatGPT" movement and driven measurable user migration to Anthropic's Claude. A major model release could serve as both a technical milestone and a strategic counter-narrative.
Meanwhile, the competitive landscape has intensified. Google's Gemini 3.1 Pro already offers a 1-million-token context window with strong multimodal performance. Anthropic's Claude Opus 4.6 has been drawing praise for coding and reasoning capabilities. A 2-million-token GPT-5.4 with enhanced vision capabilities would represent OpenAI's bid to reclaim technical leadership on multiple fronts simultaneously.
What This Means for Developers
For developers building on the OpenAI API, the leaked features suggest several preparation steps. Applications that currently chunk documents to fit within context limits may need architectural revisions to take advantage of million-scale windows. Image processing pipelines that pre-compress inputs before API submission may benefit from passing raw bytes directly. Teams evaluating multi-model strategies should factor in the possibility that GPT-5.4 could narrow or close capability gaps that currently favor competitors in specific domains.
The Fast mode toggle also suggests that OpenAI is continuing to invest in inference optimization, which could translate to lower latency and potentially lower costs for real-time applications.
Conclusion
The GPT-5.4 leak is not a roadmap slide or a conference teaser. It is production code referencing a specific model version with specific capabilities—a 2-million-token context window, full-resolution image handling, and a fast inference mode. Whether the model ships in March, April, or later, the code confirms that OpenAI is building the next iteration of its flagship model with substantial capability expansions. Developers and enterprise users should monitor OpenAI's official channels for an announcement that, based on the evidence, appears to be a matter of weeks rather than months away.
Pros
- A 2-million-token context window would set a new commercial frontier model record, enabling entirely new categories of single-session tasks
- Full-resolution image handling removes a significant limitation for vision-dependent professional workflows
- Fast mode toggle suggests users will have speed-quality flexibility without switching models
- Accelerating release cadence demonstrates OpenAI's engineering velocity and competitive responsiveness
- Code-level evidence provides more reliable insight than roadmap speculation or conference previews
Cons
- No official announcement from OpenAI; all information derives from leaked code that may not reflect final specifications
- A 2-million-token context window will likely carry significant cost implications for API users
- Shipping GPT-5.4 weeks after GPT-5.3 could fragment developer attention and create migration fatigue
- Enhanced capabilities do not address the trust and ethical concerns driving the current Cancel ChatGPT movement
References
Comments0
Key Features
GPT-5.4 was discovered twice in OpenAI's public Codex GitHub repository pull requests on February 28 and March 2, 2026. The leaked code reveals a 2-million-token context window (8x larger than GPT-5.3's 256K), full-resolution image handling that preserves original PNG/JPEG/WebP bytes without compression, and a Fast mode toggle for speed-optimized inference. Prediction markets give the model a 55% chance of launching before April 2026.
Key Insights
- Two separate Codex pull requests and an employee screenshot confirm GPT-5.4 exists as an internal OpenAI model, not speculation
- The 2-million-token context window would double Gemini 3.1 Pro's 1M-token record and represent an 8x expansion over GPT-5.3 Codex
- Full-resolution image handling eliminates compression artifacts, enabling pixel-level analysis for medical imaging, satellite data, and document OCR
- The Fast mode toggle indicates OpenAI is building speed-optimized inference paths, suggesting both standard and rapid variants at launch
- OpenAI's release cadence has compressed from 3-month to 1-month intervals within the GPT-5 series, making a March-April launch plausible
- Prediction markets assign 55% probability to a pre-April release and 74% to pre-June, reflecting high confidence in near-term availability
- The leak arrives amid the Cancel ChatGPT movement, creating strategic pressure for OpenAI to deliver a compelling technical counter-narrative
- Enterprise applications involving legal case files, full codebases, and research corpora would benefit most from million-scale context windows
Was this review helpful?
Share
Related AI Reviews
OpenAI Secures Pentagon Classified Network Deal Hours After Anthropic Blacklisted
OpenAI deploys AI models in the Pentagon's classified network with three red-line safeguards, filling the gap left by Anthropic's supply-chain-risk designation.
OpenAI Finalizes $110 Billion Funding Round at $730 Billion Valuation
OpenAI closes the largest private funding round in history with $110B from Amazon, Nvidia, and SoftBank, reaching a $730 billion valuation.
OpenAI Nears $100 Billion Funding Round at an $850 Billion Valuation
OpenAI is finalizing the first phase of a record-breaking $100B+ funding round with Amazon, SoftBank, Nvidia, and Microsoft, pushing its valuation past $850 billion.
OpenAI Launches Trusted Access for Cyber: A $10M Bet on AI-Powered Defense
OpenAI introduces Trusted Access for Cyber, an identity-based framework pairing GPT-5.3-Codex's high-capability cybersecurity skills with $10 million in API credits to accelerate defensive security operations.
