Willow Ventures

AI forecasting tournament tried to predict 2025. It couldn’t. | Insights by Willow Ventures


The AI Predictions Showdown: Insights from Leading Experts

Recently, two prominent figures in the world of artificial intelligence (AI) engaged in an enlightening discussion about the future of AI. Their insights provide a unique perspective on current advancements and potential risks in this rapidly evolving field.

François Chollet: An Optimist with Reservations

François Chollet, the creator of the Keras library and author of the ARC-AGI benchmark, is known for his cautious stance on AI progress. Historically skeptical of overly optimistic predictions, Chollet noted a shift in his perspective during the discussion. He highlighted significant strides researchers have made toward overcoming key challenges in achieving Artificial General Intelligence (AGI). Specifically, he pointed out improvements in models’ abilities to recall and apply learned knowledge—critical factors for achieving human-like intelligence.

Dwarkesh Patel: A Cautionary Tale

In contrast, his dialogue partner, Dwarkesh Patel, host of a popular AI-focused podcast, revealed a more pessimistic outlook. Based on his reporting, Patel expressed doubts about AI models’ ability to learn continuously. He emphasized, “Humans learn from their failures and adapt through small improvements. It doesn’t seem straightforward to integrate this essential capability into AI models.”

Both Perspectives Matter

This divergence in viewpoints demonstrates that even the most knowledgeable individuals in AI can arrive at different conclusions regarding the pace of advancements. This leads us to ponder: how can everyday individuals gauge which expert’s predictions to trust?

The Forecaster Wars: Understanding Different Perspectives

One engaging approach to resolve these conflicting opinions comes from the Forecasting Research Institute. Established in the summer of 2022, the institute inaugurated the Existential Risk Persuasion Tournament (XPT). This initiative aims to produce well-informed forecasts about risks that humanity may face in the coming century, including those presented by AI.

A Diverse Group of Predictors

In working towards this goal, the researchers—featuring renowned psychologist and forecasting expert Philip Tetlock, alongside FRI head Josh Rosenberg—gathered insights from subject matter experts on existential risks. Notably, they also consulted a group of “superforecasters,” who are adept at predicting events across various domains, albeit not specifically focused on existential threats.

Disparities in Predictions

The results revealed significant disparities between experts and superforecasters. While experts often predicted severe risks from specific threats, including AI, the generalists took a more conservative stance, demanding more evidence before accepting the looming dangers associated with hypothetical future technologies.

Assessing AI Progress

A fascinating part of the XPT research involved each group’s predictions concerning AI advancements over different timeframes. This analysis allows us to understand who, among the experts and predictors, has the superior grasp on forecasting AI’s trajectory.

The Findings: Who to Trust?

A report by the research team, including Tetlock and others, examines the predictive accuracy between summer 2022 and 2025. The outcome is somewhat ambiguous; both groups underestimated AI’s rapid progress. For instance, while superforecasters projected that AI wouldn’t receive gold at the International Mathematical Olympiad until 2035, experts anticipated this achievement by 2030. In reality, it occurred in 2023.

Insights from the Report

According to the report, superforecasters assigned an average probability of just 9.7% to the realized outcomes across four benchmarks—while experts assigned 24.6%. Although experts performed slightly better, the difference was not statistically significant. This raises questions about the reliability of predictions in the AI space, especially considering that both groups failed to accurately anticipate the pace of AI advancements.

What Does This Mean for Our Understanding of AI Risks?

Given that both groups underestimated AI’s growth, it could indicate a general underestimation of AI’s immediate potential. The advent of tools like ChatGPT in late 2022 may have shifted the landscape unexpectedly, prompting us to reflect on our forecasting abilities. As historian Kelsey Piper eloquently pointed out last year, a similar failure to recognize exponential growth led to significant misjudgments, notably regarding the COVID-19 pandemic.

Conclusion: Navigating Uncertainty in AI Predictions

In navigating the speculative terrain of AI predictions, it’s vital to recognize that experts may not always have the answers, especially when there’s vast disagreement among them. As tempting as it is to favor one group of opinions, widespread aggregation of forecasts tends to yield more accurate predictions than relying on individual expertise.

While we may not have clear answers today, broadening our understanding of AI developments and encouraging diverse insights will empower us to navigate the complexities of this crucial field. Ultimately, it’s an evolving story, and remaining vigilant and informed is our best recourse.



Source link