The Bottleneck of Continual Learning in Achieving AGI

2025-07-07
ℹ️Note on the source

This blog post was automatically generated (and translated). It is based on the following original, which I selected for publication on this blog:
Why I don’t think AGI is right around the corner.

The Bottleneck of Continual Learning in Achieving AGI

The rapid advancement of AI has led to much speculation about the timeline for achieving Artificial General Intelligence (AGI). While today's Large Language Models (LLMs) possess remarkable capabilities, a fundamental limitation prevents them from truly transforming industries: the lack of continual learning.

The Importance of Continual Learning

LLMs often demonstrate impressive baseline performance, sometimes exceeding that of an average human. However, their inability to improve over time through feedback and experience significantly restricts their utility. Unlike humans, LLMs cannot build context, analyze their failures, or adapt to new information organically.

Consider the analogy of teaching someone to play the saxophone. A student learns by attempting to play, listening to the sound, and making adjustments. Current AI models, however, are more like a student who receives detailed instructions after each failed attempt, without the opportunity for real-time adaptation.

Reinforcement Learning (RL) offers a potential solution, but it lacks the deliberate, adaptive nature of human learning. The ability of humans to notice small improvements and refine their workflows through continuous experience is crucial to their effectiveness.

The Impact of Limited Context

LLMs can exhibit improved performance within a single session, retaining information and adapting to user preferences. However, this understanding is often lost at the end of the session. While long context windows and summarization techniques may offer a partial remedy, they may prove inadequate for capturing the nuances of complex tasks.

The absence of continual learning means that while AIs might technically perform subtasks to a satisfactory level, their inability to build context will prevent them from operating effectively as true employees within a firm. The question arises, how can AI truly integrate into existing workflows without this crucial capacity for ongoing learning and adaptation?

A Discontinuity in Value

Solving the problem of continuous learning could lead to a significant leap in the value and capabilities of AI models. An AI capable of online learning, constantly improving and adapting, could potentially become a superintelligence, even without further algorithmic advancements. This is because these models can amalgamate their learnings across all their copies. So one AI is basically learning how to do every single job in the world.

However, the path to achieving this is not straightforward. Early versions of continual learning are likely to be imperfect and may emerge incrementally.

Predictions and Skepticism

Some researchers predict the emergence of reliable AI agents capable of performing complex tasks, such as handling taxes, within the next few years. Such agents would need to navigate various digital environments, process diverse data types, and communicate effectively with humans.

However, this timeline faces several challenges. Developing AI agents requires processing images and video, which is already more compute intensive. Moreover, the lack of a large pretraining corpus of multimodal computer use data poses a significant obstacle.

The Power of Reasoning

Despite these challenges, recent advancements in AI reasoning capabilities are undeniable. Modern AI models demonstrate the ability to break down problems, analyze user needs, and correct their own errors. This suggests that we are, in fact, making machines that are intelligent.

A Glimpse into the Future

While timelines remain uncertain, it's plausible that within the next decade, we will see AI systems capable of learning on the job as effectively as humans. Consider an AI video editor who, after several months, possesses a deep understanding of user preferences, channel dynamics, and audience engagement. What impact would such a development have on the creative and professional landscape?

AGI timelines are very lognormal. It's either this decade or bust. This suggests that while a relatively normal world might persist until the 2030s or 2040s, we must also be prepared for potentially transformative outcomes in the nearer term. Which path do we want to take?


Comments are closed.