An analysis finds that improvements in 'reasoning' AI models may soon slow down
'Reasoning' AI models show gains, but future improvements may slow due to computational and cost limitations, says Epoch AI analysis.

Epoch AI, a nonprofit organization focused on AI research, recently published a detailed analysis suggesting potential stagnation in the progression of 'reasoning' AI models. These models have shown spectacular improvements in areas such as mathematics and programming, highlighted by the success of OpenAI's o3 models. The analysis points out that although these models are significantly advantageous due to their advanced problem-solving capacities, their dependency on increased computing poses eventual limits to scaling. The use of reinforcement learning—a method where a model learns by receiving feedback—has enabled these gains, although further scalability might be challenged by computational and cost constraints.
Reasoning models are set apart by their training in two stages: an initial general training followed by reinforcement learning. According to Epoch AI, AI labs, including OpenAI, have recently increased the computation focused on this second stage. OpenAI notably invested around ten times more computation in its recent models. Dan Roberts, a researcher at OpenAI, revealed that future endeavors will emphasize reinforcement learning by assigning more computing resources than what was allocated for the basic training phase, yet this increased emphasis cannot extend indefinitely due to foreseeable constraints.
A key figure in this analysis, Josh You from Epoch AI, notes the staggering growth rate of reasoning models with performance surging four times annually and a tenfold increase in reinforcement-related improvements occurring every 3 to 5 months. Despite this momentum, he foresees that the breakthroughs from reasoning model training could align with the general AI frontier by 2026, hinting at the slow-down of advancements in reasoning AI. This potential convergence, if occurs, signals a plateau in the pace of reasoning model progress which may impact the expectations and investments ongoing in the AI sector.
The analysis constructs its arguments on several assumptions, partly grounded on statements from industry leaders. Nevertheless, it highlights additional challenges other than computational barriers, such as substantial overheads for dedicated research and operational expenses. The report further warns that persistent, substantial overheads may impose stricter scaling limits on reasoning models. Investing in scaling models without ensuring the complementary infrastructure for efficient, cost-effective computations will likely impede further developments.
The AI industry could be rightly concerned; despite immense investments in reasoning models, there are still prevalent weaknesses like hallucinations—a scenario where models produce inaccurate outputs. As these models are expensive to administer, confirming robust returns on investment becomes imperative. If the suggestions by Epoch AI hold true, companies might urgently need to rethink their approaches toward sustaining AI growth without unlimited foundational amendments.
Sources: TechCrunch, Epoch AI