When AI Experts Disagree: Insights from Scientific Discourse
Two of the leading voices in artificial intelligence recently gathered to discuss the current landscape of the field and its future trajectory. François Chollet, the creator of the highly-regarded Keras library and author of the ARC-AGI benchmark, which evaluates whether AI models have achieved “general” or human-level intelligence, joined forces with Dwarkesh Patel, a prominent podcaster in the AI space.
Chollet, known for his cautious approach to AI predictions, has been skeptical about the rapid advancements in the technology. Historically viewed as somewhat of a pessimist, he has nonetheless observed a shortening of his timelines regarding the achievement of artificial general intelligence (AGI), citing significant progress in addressing key challenges, particularly in models’ abilities to recall and apply prior knowledge.
Conversely, Patel’s own investigations have led him to adopt a more skeptical viewpoint regarding AI’s capacity to learn continuously or “on the job.” He expressed concern about the challenges inherent in integrating such capabilities into AI applications. “Humans are learning from their failures. They’re picking up small improvements and efficiencies as they work,” he noted, emphasizing a fundamental difference between human learning and AI model development.
Contrasting Views on AI Progress
The striking contrast between Chollet and Patel encapsulates a broader dilemma in the AI field: two highly knowledgeable individuals can arrive at divergent conclusions about the pace of AI advancements. For those outside of these expert circles, discerning who holds the more accurate view can seem daunting.
The Quest for Reliable Predictions
A promising avenue for reconciling these differing perspectives is being undertaken by the Forecasting Research Institute (FRI), which initiated the Existential Risk Persuasion Tournament (XPT) in 2022. Designed to produce high-quality forecasts about potential risks facing humanity in the coming century, the tournament surveyed subject matter experts and “superforecasters” — individuals with exceptional track records in prediction, but not necessarily experts in existential risks.
The contrast in outlooks was evident: experts were more likely to predict calamities from emerging threats such as AI, while generalists argued that the burden of proof should be on the experts to justify their fears regarding technologies that have yet to be realized. Despite this fundamental divide, the tournament required both groups to predict the upcoming pace of AI development over short, medium, and long-term horizons.
Evaluation of Predictions: Who Got It Right?
In a comprehensive analysis published recently, the authors, including Philip Tetlock and Josh Rosenberg, reviewed the predictions made by both groups about AI progress within the 2022-2025 timeframe. The hope was that, by assessing the relative accuracy of these forecasts, we could ascertain which perspectives warranted more credibility.
Disappointingly, the results did not provide a clear answer. Both experts and superforecasters underestimated the pace of AI evolution. For instance, while superforecasters anticipated an AI achieving excellence in the International Mathematical Olympiad by 2035, experts were slightly more optimistic, predicting it would occur in 2030. However, AI models surpassed these expectations much sooner, achieving this milestone in the summer of 2025.
On average, superforecasters assigned a mere 9.7% probability to the observed outcomes, while experts estimated a higher 24.6%. Though experts performed marginally better, the overall predictive accuracy between the two groups was statistically similar. Notably, the predictions about upcoming technological advancements didn’t correlate well with the differing perceptions of risk each group held regarding AI. Both parties proved themselves ineffectual at forecasting with high precision, a stark reminder of the unpredictable nature of technological advancement.
The Wisdom of Crowds
A compelling insight from the FRI study was the efficacy of aggregating forecasts. The median of collective predictions significantly outperformed individual forecasts, underscoring a phenomenon well-documented in predictive analytics: the wisdom of crowds.
Ezra Karger, an economist affiliated with the project, pointed out that previous research indicated little significant disagreement between groups when it came to the shorter-term implications of AI development. Both camps seemed to share a misjudgment regarding the pace of AI evolution, suggesting a collective underestimation of technological advancements.
Learning From the Past to Chart Our Future
The findings prompt a reevaluation of our understanding of AI and its potential risks. The rapid advancements observed in AI since 2022, particularly post-ChatGPT, have arguably altered the landscape of possibilities. History teaches us that failure to anticipate exponential trends can hinder our ability to address emerging threats effectively. The underestimation of AI’s development bucket could pose existential risks that merit serious consideration.
In conclusion, navigating the complex discussions surrounding AI progress and its potential implications necessitates a proactive approach, weighing different viewpoints and drawing insights from various fields. While the disagreement between experts and generalists may leave us with more questions than answers, the continual evolution of AI demands our diligent attention.
Image Credit: www.vox.com






