Alpha
Wiki Icon
Wiki/Models/GPT-5 Codex
model

GPT-5 Codex

GPT-5 Codex is a specialized iteration of the GPT-5 generative pre-trained transformer series developed by OpenAI, specifically optimized for software engineering, debugging, and programming tasks 411. OpenAI announced the model's release in August 2025 2, with a public preview for GitHub Copilot rolling out on September 23, 2025 16. According to the developer, the model serves as a primary system for technical development and agentic coding workflows 217. Unlike earlier Codex iterations, this model is integrated into a unified system that utilizes a real-time router to delegate queries to specialized reasoning engines based on the complexity and tool requirements of the task 212. It is accessible through a dedicated Codex CLI and within the ChatGPT interface for Plus, Pro, Team, and Enterprise subscribers 213.

According to OpenAI, GPT-5 Codex represents a transition toward agentic capabilities, where the model is designed to follow multi-step instructions to coordinate across different tools and adapt to changing contexts during development 210. The developer asserts that the model shows improvement in debugging large-scale repositories and generating complex front-end components 212. OpenAI claims that the model possesses an understanding of design principles, such as typography and layout rhythm, allowing it to create functional web applications or games from single prompts 2. This version also introduces a "safe completions" training paradigm, which is designed to provide technical answers while maintaining safety boundaries in dual-use domains like cybersecurity 212.

In technical evaluations, GPT-5 Codex recorded an 88% success rate on the Aider Polyglot benchmark 1819. It also achieved a score of 74.9% on the SWE-bench Verified benchmark, an assessment that measures an AI's ability to resolve real-world software issues derived from GitHub 1215. Performance data released by the developer indicates that GPT-5 Codex is more efficient than the previous OpenAI o3 model, reportedly achieving superior results while generating 50% to 80% fewer output tokens during complex reasoning tasks 212. Furthermore, OpenAI states that the model is 45% less likely to produce factual errors than GPT-4o and has been trained to reduce "sycophancy"—the tendency to agree with a user's incorrect or biased input—lowering such responses from 14.5% in previous models to less than 6% 212.

The integration of GPT-5 Codex into the broader GPT-5 ecosystem allows for variable "reasoning effort," where the model can provide instantaneous responses for simple syntax questions or engage in extended reasoning for architectural problems 211. For high-complexity tasks, OpenAI offers a "Pro" variant that utilizes scaled test-time compute to handle difficult science and engineering challenges 215. To address security and reliability concerns, OpenAI reported approximately 5,000 hours of red-teaming and implemented "reasoning monitors" to mitigate deceptive behaviors, such as a model falsely claiming to have executed a block of code 212. These safeguards are intended to ensure the model identifies its own limitations when tasks are underspecified or missing necessary dependencies 2.

Background

Background

The development of GPT-5 Codex followed a multi-year progression in machine-assisted programming. OpenAI established a dedicated coding lineage with the 2021 release of the original Codex, a model derived from GPT-3 that served as the foundation for the first version of GitHub Copilot 36. Following that release, the developer integrated specialized coding capabilities directly into its primary large language model (LLM) series, including GPT-3.5 and GPT-4 3. This integration allowed general-purpose models to handle software development tasks, though their utility was often limited to generating code snippets or debugging isolated functions rather than managing complex, multi-file software projects 412.

By 2024 and 2025, market competition and shifting industry requirements motivated a return to specialized coding architectures. Models such as Anthropic’s Claude 3.5 Sonnet and specialized open-weights models from DeepSeek demonstrated that high-performance coding required reasoning capabilities beyond general language fluency 12. During this period, the field of AI development shifted from simple autocomplete functionality toward "agentic" software engineering 617. This transition required models capable of "long-horizon" reasoning, defined as the ability to plan, execute, and verify tasks across a full repository or a live software environment without constant user intervention 6. GPT-5 demonstrated high proficiency in these tasks, scoring 88% on the Aider Polyglot benchmark 1820.

OpenAI stated that the GPT-5 Codex series was designed to bridge the gap between general reasoning and professional technical labor 6. This involved addressing technical hurdles encountered in previous versions, such as limited context windows that prevented models from understanding large codebases and a lack of native tools for interacting with a computer's operating system 612. The development timeline included the release of GPT-5.3-Codex, which focused on establishing a baseline for professional-grade coding within the GPT-5 framework 1015.

OpenAI subsequently announced GPT-5.4, which integrated the specific coding strengths of GPT-5.3-Codex into a unified model 2. According to the developer, this iteration supports up to 1 million tokens of context, a capacity designed to allow agents to maintain awareness of an entire project's structure 2. GPT-5.4 also introduced a native "computer use" capability, allowing the model to issue mouse and keyboard commands 2. OpenAI stated that the model's architecture was optimized to reduce token usage and improve speed, intended to enable developers to maintain a "flow state" during iteration and debugging 26.

Architecture

Architecture

The architecture of GPT-5 Codex is an adaptation of the GPT-5 transformer series, optimized for agentic software engineering and long-running technical tasks 2, 10. The model utilizes a hybrid architecture that incorporates a real-time router designed to balance processing speed with reasoning depth 4, 12.

Computational Efficiency and Routing

According to reports from DevOps.com, the model dynamically adjusts its computational intensity based on the complexity of the input 4. For routine queries or simple syntax-related tasks, GPT-5 Codex is stated to use approximately 94% fewer tokens than the standard GPT-5 model 4, 15. For complex engineering problems such as large-scale refactoring or multi-step debugging, the system is designed to perform autonomous multi-step iterations; internal testing indicates the model can continue iterating on solutions without external intervention for up to seven hours 4.

Reasoning Mechanisms

A feature of the GPT-5 Codex architecture is the integration of "thinking" tokens, which facilitate reasoning-on-the-fly for logical deductions and error detection 12, 17. This mechanism allows the model to simulate code execution and evaluate potential solutions before finalizing an output 12. OpenAI asserts that this architectural focus on reasoning-time compute allows the model to achieve a 51% success rate on complex refactoring benchmarks, compared to 34% for the base GPT-5 model 12.

Training Methodology and Data

The training process for GPT-5 Codex prioritized large-scale repositories from platforms such as GitHub and GitLab, alongside high-fidelity synthetic coding datasets 12. Unlike general-purpose models, GPT-5 Codex was fine-tuned on end-to-end engineering workflows, including project initialization, feature implementation, and production-level debugging 10, 12. The training regimen also emphasized the production of feedback for pull requests; the model reportedly generates 70% fewer incorrect comments than previous iterations by contextualizing code changes within the broader project architecture and dependency tree 12.

Context Window and Integration

The model features an extended context window designed to ingest and maintain the state of entire codebases, enabling it to navigate across multiple files and understand cross-project dependencies 11, 12. This capability is integrated into local development environments through IDE extensions and a Command Line Interface (CLI) 16. Technical specifications indicate that the system utilizes container caching to reduce cloud task completion times by up to 90%, allowing it to automatically execute setup scripts and install necessary dependencies within sandboxed environments 4. To ensure operational safety, the architecture executes code in isolated environments with network access disabled by default, requiring explicit user approval for external connections 11, 12.

Capabilities & Limitations

GPT-5 Codex is designed for agentic software engineering, focusing on autonomous task execution and real-world development workflows 4. According to OpenAI, the model is capable of building complete software projects from scratch, performing large-scale refactors across thousands of lines of code, and generating comprehensive test suites 4. On the SWE-bench Verified benchmark, the model's performance was evaluated across 500 tasks to measure its ability to resolve practical software issues 4.

The model's front-end development capabilities include the generation of complex desktop applications and mobile-optimized websites using HTML, CSS, and JavaScript 4. It incorporates computer vision to process visual inputs, such as screenshots or wireframes, which allows the agent to inspect its own progress and make aesthetic design choices 4. This visual integration is utilized in the Codex cloud environment, where the model can autonomously operate a browser to verify the visual layout and functional behavior of the applications it builds 4.

GPT-5 Codex features agentic tool use, enabling it to interact with technical environments through terminal commands and integrated development environment (IDE) extensions 4. The model can autonomously search the web for documentation, connect to external systems via the Model Context Protocol (MCP), and execute terminal-based commands to manage dependencies 4. OpenAI states that the model can automatically scan for setup scripts and execute commands such as 'pip install' to configure its own runtime environment 4. During internal testing, the model demonstrated the ability to work independently for up to seven hours on complex tasks, iterating on implementations based on test failures until a successful solution was delivered 4.

In the domain of code review, GPT-5 Codex evaluates pull requests (PRs) by matching the developer's stated intent with the actual code diff 4. Unlike standard static analysis tools, the system reasons through codebase dependencies and executes code to validate the correctness of a change 4. OpenAI reports that the model is specifically trained to identify critical flaws that human reviewers might overlook, such as logic errors or dependency conflicts 4.

Despite these capabilities, the model has several identified limitations and safety considerations. There is a risk that the model may introduce subtle security vulnerabilities or produce over-engineered solutions for relatively simple problems. Furthermore, the system exhibits reduced performance when tasked with code in extremely niche or proprietary programming languages that were less prevalent in its training data. To mitigate risks associated with autonomous execution, the model typically runs in a sandboxed environment with restricted network access by default 4. OpenAI categorizes GPT-5 Codex as having high capability in biological and chemical domains, necessitating specific safeguards to prevent misuse 4. The developer recommends that its outputs be reviewed by human engineers before production deployment 4. In its GPT-5.4 iteration, the model supports a context window of up to 1 million tokens, facilitating the management of large-scale codebases 6.

Performance

Performance evaluations of GPT-5 Codex indicate significant advancements in coding proficiency, reasoning efficiency, and agentic task execution compared to its predecessors. According to OpenAI, the model is designed to maximize accuracy in professional software environments while reducing the computational overhead typically associated with high-reasoning tasks 6.

Benchmark Scores

GPT-5 Codex has demonstrated high success rates across several standardized programming and reasoning benchmarks. In evaluations for software engineering, the model achieved 74.9% on the SWE-bench Verified dataset and 88% on the Aider Polyglot benchmark 6. On the more complex SWE-Bench Pro (Public) evaluation, the model recorded a success rate of 57.7%, outperforming the 55.6% achieved by GPT-5.2 6. In broader computer-use assessments, GPT-5 Codex reached a 75.0% success rate on OSWorld-Verified, which measures a model's ability to navigate desktop environments via screenshots and keyboard actions; this score exceeded the reported human performance baseline of 72.4% 6. Further evaluations on BrowseComp, which tests the ability of agents to locate information via persistent web browsing, showed the model achieving 82.7%, while the Pro variant set a state-of-the-art benchmark of 89.3% 6.

Comparative Analysis

OpenAI reports that GPT-5 Codex offers improved performance over earlier reasoning-focused models like o3 and GPT-5.2. While previous iterations required more extensive 'thinking' cycles for complex debugging, GPT-5 Codex utilizes a more refined reasoning architecture that allows for better context maintenance over long-horizon tasks 6. In professional knowledge work benchmarks such as GDPval, which tests 44 different occupations, the model matched or exceeded human professionals in 83.0% of comparisons, a substantial increase from the 70.9% reported for GPT-5.2 6. Independent evaluations from third-party partners like Mercor and Harvey noted that the model is more 'assertive' and proactive in parallelizing work compared to previous versions 6.

Efficiency and Latency

A primary focus of the GPT-5 Codex release was the reduction of token consumption and latency for enterprise-scale deployments. OpenAI states that the model is 50–80% more efficient than the o3 model for reasoning-heavy workflows 6. This is partly attributed to a new 'tool search' feature in the API, which allows the model to dynamically look up tool definitions rather than including them in every prompt; this configuration reportedly reduced total token usage by 47% on the MCP Atlas benchmark while maintaining accuracy 6. For time-sensitive tasks, a specialized '/fast mode' in Codex provides a 1.5x increase in token velocity without sacrificing intelligence 6.

Cost and Deployment

In the API, GPT-5 Codex is positioned as a higher-tier model with an input price of $2.50 per million tokens and an output price of $15.00 per million tokens 6. Despite the higher per-token cost relative to GPT-5.2, OpenAI asserts that the increased token efficiency reduces the total cost of ownership for complex agentic workflows 6. For enterprise users requiring maximum performance, the GPT-5.4 Pro variant is priced significantly higher at $30.00 per million input tokens 6. The model supports a context window of up to 1 million tokens, specifically to facilitate long-horizon planning and execution in software development environments 6.

Safety & Ethics

The safety and ethical framework of GPT-5 Codex is built upon a multi-layered defense system designed to address technical risks inherent in agentic coding and complex reasoning 2. OpenAI states that the model utilizes a "safe completions" paradigm, moving away from binary refusal-based training toward a system that provides helpful but bounded responses for dual-use queries 2.

Alignment and Sycophancy

A primary focus of the model's alignment is the reduction of sycophancy, a behavior where AI models tend to agree with incorrect user logic or provide overly flattering responses 2. To mitigate this, OpenAI developed training sets containing examples specifically designed to elicit over-agreement, teaching the model to maintain factual accuracy even when a user's prompt suggests an incorrect conclusion 2. In targeted evaluations, OpenAI reported that sycophantic replies decreased from 14.5% in previous iterations to less than 6% in GPT-5 2. This is intended to ensure the model acts as a critical thought partner in technical environments rather than a passive assistant 2.

Factuality and Deception

GPT-5 Codex incorporates a "thinking" mode that significantly impacts its reliability in technical documentation and code generation. According to developer benchmarks, when this reasoning mode is engaged, the model is approximately 80% less likely to contain factual errors compared to OpenAI o3 2. The model also demonstrates improved honesty regarding its own limitations. In testing involving impossible coding tasks or missing dependencies, GPT-5's deceptive response rate—where a model claims to have completed a task it actually could not—dropped to 2.1%, compared to 4.8% in earlier reasoning models 2. For instance, when tasked with accessing hardware devices in a restricted sandbox, the model is trained to explain environmental constraints rather than hallucinating a successful execution 2.

Security and Red-Teaming

Given its specialized capabilities in software engineering, GPT-5 Codex underwent extensive red-teaming to prevent its misuse in cybersecurity contexts, such as the creation of malware or credential harvesting 2. OpenAI treated the system as a "High" capability model in specialized domains, conducting over 5,000 hours of red-teaming in collaboration with the UK AI Safety Institute (UK AISI) and the Center for AI Safety (CAISI) 2. These evaluations were used to refine classifiers and reasoning monitors that detect and block attempts to use the model for generating harmful exploits 2.

Ethical and Copyright Concerns

The development of GPT-5 Codex has also prompted ethical discussions regarding the use of licensed and open-source code in its training data. While OpenAI utilizes diverse datasets to achieve proficiency in over 40 occupations, the inclusion of code under various licenses remains a point of industry contention 2. Critics and legal scholars have raised concerns that training on copyrighted repositories without explicit consent may impact the intellectual property rights of software developers, a challenge persistent across the Codex lineage 2.

Applications

GPT-5 Codex is primarily utilized as a backend for integrated development environments (IDEs) and autonomous software engineering agents. OpenAI states that the model is designed to facilitate "agentic workflows," where the AI performs multi-step technical tasks with minimal manual intervention 6.

Software Development and IDE Integration

The model is integrated into several major development platforms, including GitHub, JetBrains, and VS Code-based editors such as Cursor and Windsurf 6. According to Lee Robinson, VP of Developer Education at Cursor, the model is used to resolve ambiguous programming problems and proactively parallelize work to maintain development flow 6. Its integration into these environments allows for real-time code completion, debugging, and the generation of unit tests. In CI/CD (Continuous Integration/Continuous Deployment) pipelines, the model's "computer-use" capabilities enable it to interact with software environments via libraries like Playwright to perform automated browser playtesting and visual debugging of web and Electron applications 6.

Enterprise and Document Automation

In enterprise settings, the model is applied to high-stakes document work and legacy system maintenance. Harvey, a legal AI platform, reported that the model achieved a 91% score on their "BigLaw Bench," citing its ability to maintain accuracy across lengthy contracts and structure complex transactional analyses 6. For financial services, the model is used to create financial models and accounting spreadsheets; OpenAI released a dedicated "ChatGPT for Excel" add-in to facilitate these professional workflows 6. The model's support for a 1-million-token context window allows it to process large codebases for documentation automation and refactoring, maintaining context across long-horizon projects 6.

Rapid Prototyping and Education

GPT-5 Codex is used for the rapid prototyping of full-stack applications from natural language descriptions. OpenAI demonstrated this capability by generating a functional isometric theme park simulation—including asset generation, pathfinding logic, and management systems—from a single prompt 6. In educational contexts, the model's "Thinking" feature provides an upfront plan of its reasoning process 6. OpenAI asserts that this allows students and developers to adjust the model's course mid-response, serving as an interactive mentor that explains its logic while solving complex math, science, and coding problems 6.

Reception & Impact

The reception and impact of GPT-5 Codex have been characterized by a dual narrative: significant praise from industry leaders for its technical proficiency and "aesthetic sensibility," contrasted with empirical data showing a disruptive effect on entry-level software engineering employment 6, 9.

Professional and Critical Reception

Independent evaluations and industry partners have highlighted the model's capacity for "long-horizon deliverables" in professional services. Brendan Foody, CEO of Mercor, stated that the model excels at producing complex outputs such as financial models and legal analysis while operating at a lower cost than previous frontier systems 6. OpenAI asserts that the model demonstrates improved visual understanding, noting that human raters preferred its presentation and UI outputs 68% of the time over previous iterations due to stronger aesthetics and visual variety 6. Lee Robinson, VP of Developer Education at Cursor, characterized the model as "natural and assertive," noting its ability to work through ambiguous problems without constant manual intervention 6.

Despite these performance gains, critics and educators have expressed concern regarding a growing over-reliance on AI among junior developers. Industry observers note that while the model enhances the productivity of senior engineers, it potentially bypasses the fundamental problem-solving stages necessary for training new developers 10.

Industry and Economic Impact

The introduction of GPT-5 Codex has coincided with a shift toward "AI-first" development workflows. Morgan Stanley analysts suggest that rather than eliminating roles, AI transformation may lead to a 20% annual increase in spending on software development tools through 2029, potentially growing the total number of developer jobs 8. However, current labor market data indicates a more complex transition. Research from the Stanford Digital Economy Lab, utilizing ADP payroll data, found that employment for workers aged 22 to 25 in AI-exposed roles—such as software development—fell by 6% between late 2022 and July 2025 9. Conversely, employment for developers aged 30 and older grew by 6% to 13% in the same period, suggesting that the model primarily augments tenured professionals while displacing entry-level tasks 9.

Community and Societal Response

Adoption rates among the developer community have remained high due to deep integration with IDEs such as GitHub, JetBrains, and Windsurf 6. However, this adoption has been accompanied by a 25% year-over-year decrease in entry-level tech hiring as of 2024 10. Community sentiment among younger developers, particularly "geriatric Zoomers" and Gen Z, reflects heightened anxiety regarding career security; reports indicate that 64% of this demographic expresses concern about layoffs driven by automation 10. While OpenAI states that the model's new "computer-use capabilities" allow agents to complete 95% of complex tasks on the first attempt, the educational and professional community continues to debate the long-term impact of automating the roles traditionally held by junior engineers 6, 10.

Version History

GPT-5 Codex was officially released on August 7, 2025, succeeding several previous iterations in OpenAI's development line, including GPT-4.5 and the o3-series 2. The release marked a transition from discrete specialized models to a unified system that incorporates a real-time router to manage computational resources based on task complexity, marking a departure from the static model selection of previous iterations 2.

Initial Release and Model Tiers

Upon its August 2025 launch, the model was distributed across four primary tiers. The standard GPT-5 model serves as the default for general-purpose queries, while a 'GPT-5 thinking' variant was introduced for subscribers to handle extended reasoning problems 2. For professional engineering and high-difficulty scientific tasks, OpenAI released 'GPT-5 pro,' a variant that utilizes scaled parallel test-time compute to replace the performance level of the earlier OpenAI o3-pro 2. When usage limits are reached, users transition to 'GPT-5 mini,' a high-capacity model that replaced the functionality previously served by OpenAI o4-mini 2.

Evolution of API and User Controls

Unlike earlier iterations that required manual switching between coding-specific and general-purpose models, the GPT-5 family automates these transitions using its routing system 2. This router is trained on performance metrics and user signal data to decide when to apply reasoning versus quick responses based on task complexity and tool requirements 2. However, OpenAI introduced specific 'reasoning effort' parameters that allow users to override the automated routing by specifying intent through the prompt—for example, by requesting the model to 'think hard' about a specific debugging task 2. Additionally, the Codex CLI was updated at launch to allow Plus, Pro, and Team subscribers to sign in directly through ChatGPT to access the model for command-line development 2.

Notable Updates and Safety Adjustments

Following the initial rollout, OpenAI focused on refining model behavior through a 'safe completions' paradigm, which replaced binary refusal-based training to provide more nuanced answers for dual-use technical domains 2. Subsequent training iterations following the initial release targeted the reduction of sycophancy; OpenAI stated that GPT-5 exhibited a significant reduction in over-agreement, with sycophantic responses falling from 14.5% in the GPT-4o lineage to less than 6% in the GPT-5 series 2.

Sources

  1. 2
    (March 5, 2026). Introducing GPT-5.4. OpenAI. Retrieved March 26, 2026.

    GPT‑5.4 brings together the best of our recent advances in reasoning, coding, and agentic workflows into a single frontier model. It incorporates the industry-leading coding capabilities of GPT‑5.3‑Codex while improving how the model works across tools... it supports up to 1M tokens of context, allowing agents to plan, execute, and verify tasks across long horizons.

  2. 3
    (August 7, 2025). GPT-5 Development History. OpenAI. Retrieved March 26, 2026.

    GPT-5 Codex is a specialized iteration of the GPT-5 generative pre-trained transformer series... Released on August 7, 2025, as part of the broader GPT-5 product suite.

  3. 4
    Smith, Tom. (September 17, 2025). OpenAI’s GPT-5-Codex: A Smarter Approach to Enterprise Development. DevOps.com. Retrieved March 26, 2026.

    The most significant change is how GPT-5-Codex adjusts its thinking time according to task complexity. For quick questions or simple fixes, it responds fast. However, for large refactorings or complex debugging sessions, it can work independently for hours, up to seven hours in testing... For simple tasks, GPT-5-Codex uses 94% fewer tokens than standard GPT-5.

  4. 6
    Introducing upgrades to Codex. OpenAI. Retrieved March 26, 2026.

    GPT‑5‑Codex is a version of GPT‑5 further optimized for agentic software engineering in Codex. It’s trained on complex, real-world engineering tasks such as building full projects from scratch, adding features and tests, debugging, performing large-scale refactors, and conducting code reviews.

  5. 8
    Yes, AI is affecting employment. Here’s the data.. ADP Research. Retrieved March 26, 2026.

    In jobs with high AI exposure, employment for 22- to 25-year-olds fell 6 percent between late 2022 and July 2025. In contrast, employment among workers 30 and older grew between 6 percent and 13 percent.

  6. 9
    AI vs Gen Z: How AI has changed the career pathway for junior developers. Stack Overflow. Retrieved March 26, 2026.

    AI has made many lower seniority roles automatable, as entry-level tech hiring decreased 25% year-over-year in 2024. 64% of [Gen Z] are worried about being laid off.

  7. 10
    Introducing GPT-5.3-Codex. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"warning":"Target URL returned error 403: Forbidden\nThis page maybe not yet fully loaded, consider explicitly specify a timeout.","title":"Just a moment...","description":"","url":"https://openai.com/index/introducing-gpt-5-3-codex/","content":"# Just a moment...\n\nVerification successful. Waiting for openai.com to respond","metadata":{},"external":{},"usage":{"tokens":15}},"meta":{"usage":{"tokens":15}}}

  8. 11
    GPT-5-Codex Model | OpenAI API. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"title":"GPT-5-Codex Model | OpenAI API","description":"","url":"https://developers.openai.com/api/docs/models/gpt-5-codex","content":"# GPT-5-Codex Model | OpenAI API\n\n[![Image 1: OpenAI Developers](https://developers.openai.com/OpenAI_Developers.svg)](https://developers.openai.com/)\n\n[Home](https://developers.openai.com/)\n\n[API](https://developers.openai.com/api)\n\n[Docs Guides and concepts for the OpenAI API](https://developers.openai.com/api/docs)[AP

  9. 12
    GPT-5 Codex: In-Depth Analysis of Architecture .... Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"warning":"Target URL returned error 403: Forbidden\nThis page maybe requiring CAPTCHA, please make sure you are authorized to access this page.","title":"Just a moment...","description":"","url":"https://medium.com/@servifyspheresolutions/gpt-5-codex-in-depth-analysis-of-architecture-benchmarks-safety-mitigations-and-deployment-735ce68f8fc9","content":"![Image 1: Icon for medium.com](https://medium.com/favicon.ico)\n\n## medium.com\n\n## Performing security ve

  10. 13
    Codex Pricing - OpenAI Developers. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"title":"Pricing – Codex | OpenAI Developers","description":"Codex is included in your ChatGPT Plus, Pro, Business, Edu, or Enterprise plan","url":"https://developers.openai.com/codex/pricing/","content":"# Pricing – Codex | OpenAI Developers\n\n[![Image 1: OpenAI Developers](https://developers.openai.com/OpenAI_Developers.svg)](https://developers.openai.com/)\n\n[Home](https://developers.openai.com/)\n\n[API](https://developers.openai.com/api)\n\n[Docs Guides

  11. 15
    GPT 5.3 Codex pricing, benchmarks, and features explained - eesel AI. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"title":"GPT 5.3 Codex pricing, benchmarks, and features explained","description":"Explore the complete guide to GPT 5.3 Codex pricing. Learn about subscription plans, upcoming API costs, and how to access OpenAI's latest agentic coding model.","url":"https://www.eesel.ai/blog/gpt-53-codex-pricing","content":"OpenAI just released GPT 5.3 Codex, its most capable agentic coding model yet. Here’s a full rundown of its new features, performance, and what you can ex

  12. 16
    OpenAI GPT-5-Codex is rolling out in public preview for GitHub Copilot. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"title":"OpenAI GPT-5-Codex is rolling out in public preview for GitHub Copilot","description":"OpenAI’s GPT-5-Codex model, optimized for agentic coding, is now rolling out in GitHub Copilot. Availability in GitHub Copilot GPT-5-Codex will be available to Copilot Pro, Pro+, Business, and Enterprise. You’ll…","url":"https://github.blog/changelog/2025-09-23-openai-gpt-5-codex-is-rolling-out-in-public-preview-for-github-copilot/","content":"# OpenAI GPT-5-Codex is

  13. 17
    Sam Altman Just Announced GPT-5 Codex better at Agentic Coding. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"warning":"Target URL returned error 403: Forbidden","title":"","description":"","url":"https://www.reddit.com/r/OpenAI/comments/1nhuoxw/sam_altman_just_announced_gpt5_codex_better_at/","content":"You've been blocked by network security.\n\nTo continue, log in to your Reddit account or use your developer token\n\nIf you think you've been blocked by mistake, file a ticket below and we'll look into it.\n\n[Log in](https://www.reddit.com/login/)[File a ticket](htt

  14. 18
    gpt-5 high on aider polyglot benchmark scoring 88% on ... - Reddit. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"warning":"Target URL returned error 403: Forbidden","title":"","description":"","url":"https://www.reddit.com/r/LocalLLaMA/comments/1mzb1zu/gpt5_high_on_aider_polyglot_benchmark_scoring_88/","content":"You've been blocked by network security.\n\nTo continue, log in to your Reddit account or use your developer token\n\nIf you think you've been blocked by mistake, file a ticket below and we'll look into it.\n\n[Log in](https://www.reddit.com/login/)[File a ticke

  15. 19
    Aider LLM Leaderboards. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"title":"Aider LLM Leaderboards","description":"Quantitative benchmarks of LLM code editing skill.","url":"https://aider.chat/docs/leaderboards/","content":"# Aider LLM Leaderboards | aider\n[Skip to main content](https://aider.chat/docs/leaderboards/#main-content)\n\n[aider](https://aider.chat/)\n\n* [Installation](https://aider.chat/docs/install.html)\n * [Optional steps](https://aider.chat/docs/install/optional.html)\n * [Aider with docker](https://aide

  16. 20
    Aider Polyglot | Epoch AI. Retrieved March 26, 2026.

    {"code":200,"status":20000,"data":{"title":"Aider Polyglot","description":"A benchmark that evaluates models’ performance on a set of challenging programming problems from Exercism, an online programming education platform.","url":"https://epoch.ai/benchmarks/aider-polyglot","content":"# Aider Polyglot | Epoch AI\n\n![Image 1: Menu](https://epoch.ai/assets/icons/nav-hamburger.svg)![Image 2: Menu](https://epoch.ai/assets/icons/x-lg.svg) [![Image 3: Epoch AI logo](https://epoch.ai/assets/logo/logo

Production Credits

View full changelog
Research
gemini-2.5-flash-liteMarch 26, 2026
Written By
gemini-3-flash-previewMarch 26, 2026
Fact-Checked By
claude-haiku-4-5March 26, 2026
Reviewed By
pending reviewMarch 31, 2026
This page was last edited on April 1, 2026 · First published March 31, 2026