OpenAI's Next Grand Challenge

OpenAI has announced a sweeping new research ambition: building what it calls an AI researcher — a fully automated, agent-based system capable of independently tackling large, complex scientific problems. In an exclusive interview with MIT Technology Review, Chief Scientist Jakub Pachocki described the initiative as OpenAI's North Star for the coming years, representing a convergence of the company's work on reasoning models, coding agents, and interpretability into a unified long-horizon goal.

The timeline is concrete and near-term in ways that distinguish this announcement from the more diffuse AGI promises the industry has traded in for years. OpenAI plans to build an autonomous AI research intern — a system capable of independently working on specific research problems for days at a time — by September 2026. The full multi-agent AI researcher, capable of tackling problems too large or complex for humans to manage, is targeted for a 2028 debut.

Codex as the Blueprint

Pachocki pointed to OpenAI's existing Codex agent as both the evidence base and the early prototype for the more ambitious AI researcher vision. Codex, which OpenAI released in January, is an agent-based coding system that can autonomously generate, run, and debug code to complete complex programming tasks. It has been broadly adopted within OpenAI itself, with Pachocki noting that most of the company's technical staff now use Codex as a core part of their workflow.

The philosophical leap Pachocki is making is that if an AI system can autonomously solve complex coding problems — which require creative reasoning, decomposition of large tasks into subtasks, tracking of complex state over extended work sessions, and error correction — then the same capability architecture can be extended to scientific problem solving in domains like biology, chemistry, physics, and mathematics.

Our jobs are now totally different than they were even a year ago. Nobody really edits code all the time anymore. Instead, you manage a group of Codex agents, Pachocki told MIT Technology Review. The vision is that the same management relationship — human directing, AI executing — could eventually apply to research itself, with scientists directing AI agents that independently pursue experimental hypotheses, review literature, design analyses, and generate results.