Finding signal on Twitter is more difficult than it used to be. We curate the best tweets on topics like AI, startups, and product development every weekday so you can focus on what matters.

OpenAI Codex Development and Interface Design Choices

OpenAI’s hottest app isn’t ChatGPT—it’s Codex. In the last few weeks alone, the Codex team shipped a desktop app, GPT-5.3 Codex (a new flagship model), and Spark, the fastest coding model I’ve ever used. Usage has grown fivefold since January and over a million people now use Codex weekly. Codex was also the app that OpenAI chose to run an ad for in the Super Bowl. I talked to Thibault (@thsottiaux), head of Codex, and Andrew (@ajambrosino), a member of technical staff who built the Codex app, for @every’s AI & I about what OpenAI is building and how they’re using it internally. We get into: - Why they built a GUI instead of a terminal. Terminals work for quick tasks, they say, but feel limiting when you’re running multiple agents in parallel. The IDE, meanwhile, overwhelms users—and the Codex team wants the AI to dynamically decide which tools to show you for a given task. - How they’re teaching the model to read between the lines. Codex is great at following instructions, but optimize too hard in that direction, and it starts taking you literally—like copying a typo directly into the code. The team obsesses over this tradeoff, and is also introducing “personalities,” modes users can toggle between that control how blunt or supportive the model feels. - How OpenAI uses its own coding agent. Codex lets you schedule prompts to run on a recurring basis, and the team has dozens of automations running at all times. For example, one scans for merge conflicts every couple of hours so code is always ready to ship, and another picks a random file from the codebase multiple times a day and hunts for bugs no one would've gone looking for. - Why speed is a dimension of intelligence. OpenAI’s newest model (Spark) is so fast that they actually slow it down so you can read the output. They see the speed enabling three things: staying super in the flow, replacing brittle developer tools with intelligent ones that can adapt on the fly, and redirecting the model mid-task— especially with voice—so coding starts to feel more and more like a conversation. - Code review is the next bottleneck. Models can generate code faster than ever, but someone still has to verify that it works. The team is exploring a future where the model proves its own fix works—retracing the click path a user would take, screenshotting the results, and attaching the evidence to a pull request. This is a must-watch for anyone who uses AI coding agents—and is curious about the future of programming. Watch below! Timestamps: Introduction: 00:01:27 OpenAI’s evolving bet on its coding agent: 00:05:27 The choice to invest in a GUI (over a terminal): 00:09:42 The AI workflows that the Codex team relies on to ship: 00:20:38 Teaching Codex how to read between the lines: 00:26:45 Building affordances for a lightening fast model: 00:28:45 Why speed is a dimension of intelligence: 00:33:15 Code review is the next bottleneck for coding agents: 00:36:30 How the Codex team positions against the competition: 00:41:24

Video thumbnail
View

Topics

Read the stories that matter.

Save hours a day in 5 minutes