Finding signal on Twitter is more difficult than it used to be. We curate the best tweets on topics like AI, startups, and product development every weekday so you can focus on what matters.
President & Co-Founder @OpenAI
Page 1 • Showing 16 tweets
seeing so much positive progress across each part of openai right now, very proud of the team
agents are up and productive, time for bed
with codex, feels like whatever can be imagined can be created
going to soon feel how inefficient it’s been to do work with a computer
vibe coding is the manifesting of vision, abstracted from the details (ideally the unnecessary ones!) of how that vision comes to be
AI is the refining of compute into brainpower
If you’re an infrastructure or security engineer, now is the best time to join OpenAI. It’s hard not to be inspired by what today’s coding tools are capable of, and we have line of sight to making them much better. While our core ML infrastructure problems remain much the same as always — training and inferencing models at scale, co-designing end-to-end for maximum effect, managing complexity and maintaining fast iteration — what it feels like to solve these problems in practice is changing fast. As the models have been improving, our ability to get value out of them is increasingly bottlenecked by thoughtfully designed infrastructure — whether figuring out how to manage agent cross-collaboration, having ergonomic sandboxes that let the agents complete end-to-end workflows securely, building tools/abstractions/observability/frameworks which allow the agents to move faster, and scaling supervision of the agents' work. Engineering is already different from a few months ago, and will continue to evolve. Having seen many generations of engineering and AI tools, I believe what is most important going forward will be skills like the following: strong understanding of your domain, ability to think through abstraction/architecture/design/how the pieces should fit together, and deep curiosity to explore what these models have to offer. If you’d like to help us build the future of AI, while using AI to get there, email me: [email protected]. Include a description of a surprising or creative way you’ve gotten value out of the models recently, and your contributions to any project in your career on which you’ve made a significant difference in its outcome. Feel free to include any other context that can help us understand how you operate and the problems you want to work on.
codex is so good at the toil — fixing merge conflicts, getting CI to green, rewriting between languages — it raises the ambition of what i even consider building
how did we ever write all that code by hand
Software development is undergoing a renaissance in front of our eyes. If you haven't used the tools recently, you likely are underestimating what you're missing. Since December, there's been a step function improvement in what tools like Codex can do. Some great engineers at OpenAI yesterday told me that their job has fundamentally changed since December. Prior to then, they could use Codex for unit tests; now it writes essentially all the code and does a great deal of their operations and debugging. Not everyone has yet made that leap, but it's usually because of factors besides the capability of the model. Every company faces the same opportunity now, and navigating it well — just like with cloud computing or the Internet — requires careful thought. This post shares how OpenAI is currently approaching retooling our teams towards agentic software development. We're still learning and iterating, but here's how we're thinking about it right now: As a first step, by March 31st, we're aiming that: (1) For any technical task, the tool of first resort for humans is interacting with an agent rather than using an editor or terminal. (2) The default way humans utilize agents is explicitly evaluated as safe, but also productive enough that most workflows do not need additional permissions. In order to get there, here's what we recommended to the team a few weeks ago: 1. Take the time to try out the tools. The tools do sell themselves — many people have had amazing experiences with 5.2 in Codex, after having churned from codex web a few months ago. But many people are also so busy they haven't had a chance to try Codex yet or got stuck thinking "is there any way it could do X" rather than just trying. - Designate an "agents captain" for your team — the primary person responsible for thinking about how agents can be brought into the teams' workflow. - Share experiences or questions in a few designated internal channels - Take a day for a company-wide Codex hackathon 2. Create skills and AGENTS[.md]. - Create and maintain an AGENTS[.md] for any project you work on; update the AGENTS[.md] whenever the agent does something wrong or struggles with a task. - Write skills for anything that you get Codex to do, and commit it to the skills directory in a shared repository 3. Inventory and make accessible any internal tools. - Maintain a list of tools that your team relies on, and make sure someone takes point on making it agent-accessible (such as via a CLI or MCP server). 4. Structure codebases to be agent-first. With the models changing so fast, this is still somewhat untrodden ground, and will require some exploration. - Write tests which are quick to run, and create high-quality interfaces between components. 5. Say no to slop. Managing AI generated code at scale is an emerging problem, and will require new processes and conventions to keep code quality high - Ensure that some human is accountable for any code that gets merged. As a code reviewer, maintain at least the same bar as you would for human-written code, and make sure the author understands what they're submitting. 6. Work on basic infra. There's a lot of room for everyone to build basic infrastructure, which can be guided by internal user feedback. The core tools are getting a lot better and more usable, but there's a lot of infrastructure that currently go around the tools, such as observability, tracking not just the committed code but the agent trajectories that led to them, and central management of the tools that agents are able to use. Overall, adopting tools like Codex is not just a technical but also a deep cultural change, with a lot of downstream implications to figure out. We encourage every manager to drive this with their team, and to think through other action items — for example, per item 5 above, what else can prevent a lot of "functionally-correct but poorly-maintainable code" from creeping into codebases.
taste is a new core skill
codex's shell-fu is incredible to behold and learn from
feels like such a wasted opportunity every moment your agents aren't running
gb200 has really been enabling us to do some amazing things