To bring Codex to Windows, we had to answer a hard question: how do you let coding agents stay useful without forcing developers to choose between constant approval prompts and full machine access?
Here’s how we built the Windows sandbox for Codex:
Introducing Daybreak: frontier AI for cyber defenders.
Daybreak brings together the most capable OpenAI models, Codex, and our security partners to accelerate cyber defense and continuously secure software.
A step toward a future where security teams can move at the speed defense demands.
Introducing the OpenAI Deployment Company, which will help businesses maximally succeed with their deployments of AI.
Starting with 150 Forward Deployed Engineers and Deployment Specialists, and $4 billion of initial investment from 19 partners.
We have a gnarly refactor in our codebase that I test every frontier model on. I've been doing this since the release of o1. Nothing has been able to solve it.
GPT-5.5 extra high thinking with /goal just did it.
I am floored.
Codex quite literally filed my reimbursements, downloaded invoices since the start of the month, updated the expenses spreadsheet and filled the actual form all by itself
Used Drive & Sheets plugin for state tracking
Gmail plugin for tracking invoices
Chrome extension for actual form Used Drive + upload
I had been putting this off ever since I joined OpenAI - it took codex ~20 minutes to do it all!
I just added real-time AI translation into Chormex using GPT-Realtime-2… and this feels absolutely surreal.
It works across YouTube videos, live streams, meetings, presentations, basically anywhere audio is playing inside Chrome.
You can watch translated speech in real time while simultaneously using Codex on top of the live context.
“Summarize this.”
“What are the key points?”
“Turn this into notes.”
“Explain what they mean.”
“Organize the discussion.”
…all while the video or meeting is still happening.
It genuinely feels like browsers are evolving into real-time AI operating systems.
We are getting dangerously close to a world where language barriers on the internet completely disappear.
Chain of thought monitors are a key layer of defense against AI agent misalignment. To preserve monitorability, we avoid penalizing misaligned reasoning during RL.
We found a limited amount of accidental CoT grading which affected released models, and are sharing our analysis.
The only way to describe codex working in chrome is pure magic.
"Any sufficiently advanced technology is indistinguishable from magic." - Arthur C. Clarke's Third Law
Hosting a session next Wednesday (5/13) with the OpenAI Forum on why Codex matters beyond code. Join for the livestream and Q&A if you’re interested in the history of Codex, what we're pushing on next and see some cool use-cases.
I've been driving GPT5.5 on low reasoning for the last week+ and it's very good, very efficient. Haven't been tempted to reach for Opus at all. And it's more succinct than Kimi too. Huge leap forward for @OpenAI 👌
Today, we're rolling out GPT‑5.5‑Cyber in limited preview to defenders responsible for securing critical infrastructure.
GPT-5.5 with Trusted Access for Cyber (TAC) is still the best option for developers to find and patch vulnerabilities in their code.
It's a very good model.
Today, we are excited to introduce Codex for Chrome!
Now, Codex can drive its own Chrome tabs in the background to automate tasks while you use the browser simultaneously.
It does this by opening up tab groups for each task, cleaning up at the end, and handing back tabs for review only as needed.
Try it for deep research inside logged-in websites, large scale data transfer into any systems of record like CRMs/CMSs, and automating repetitive workflows inside admin consoles & internal tools.
Codex will still prefer dedicated plugins if you have them installed, but the Chrome plugin is the universal connector that glues end to end workflows where programmatic coverage is often incomplete.
We are making this available on both Windows and Mac today! Let us know what you think.
have been excited for realtime voice-to-voice translation as an AI application since we started OpenAI. extremely cool to see it now available in the API for anyone to build with:
Introducing GPT-Realtime-2 in the API: our most intelligent voice model yet, bringing GPT-5-class reasoning to voice agents.
Voice agents are now real-time collaborators that can listen, reason, and solve complex problems as conversations unfold.
Now available in the API alongside streaming models GPT-Realtime-Translate and GPT-Realtime-Whisper — a new set of audio capabilities for the next generation of voice interfaces.