Research

The 5 Existential Barriers to AGI : Why scaling isn't working anymore

Welcome to AI Weekly, today we're exploring the barriers and top 5 limits to having AGI with the current LLM-based models. A lot of hype was built in 2025 and looking at the markets we're starting to see cracks in the model of "scale".

We have hit a ceiling. The industry assumption was that if we just kept making models bigger, they would eventually solve everything. However, verified research from Anthropic, Apple, and Nature confirms that "brute-force" scaling has reached a point of diminishing returns.

Here are the five specific failure modes identified in the literature.

1. Bigger Models Are Getting Less Reliable

Anthropic Research: Reliability & Inverse Scaling We assumed size equaled smarts. But Anthropic’s research on "Inverse Scaling" reveals that while larger models handle simple tasks well, they can become more chaotic on complex ones. As the "chain of thought" grows longer, the model's error rate increases. They don't just fail; they hallucinate more confidently ("sycophancy"), making them unsafe for autonomous workflows.

2. The "Reasoning" is Fake

Apple Research: GSM-Symbolic (arXiv) Apple debunked the idea that LLMs are learning logic. In their GSM-Symbolic benchmark, they showed that changing a trivial variable in a math problem (like swapping the name "David" for "Clara") caused model accuracy to drop by up to 65%. This proves the models are relying on fragile pattern matching, not genuine reasoning.

3. We Are Running Out of Human Data

Nature: AI models collapse when trained on recursively generated data This is the "pollution" crisis. As the internet fills with AI-generated text, new models are forced to train on the output of older models. The Nature study proves this causes "Model Collapse": the models lose the "tails" of the data (nuance and creativity) and converge on a generic, low-quality average.

4. The Return on Investment Has Flatlined

PNAS: Scaling language model size yields diminishing returns The economics are breaking. A massive study published in PNAS found that "frontier" models (often 10x larger and more expensive) were statistically no more effective at persuasion than much smaller models. We are paying exponential costs for improvements that are virtually invisible in real-world utility.

5. The "Easy Wins" Are Gone

Ilya Sutskever: The Age of Scaling is Over (Transcript) Ilya Sutskever, the co-creator of ChatGPT, has publicly admitted that the "Age of Scaling"—the strategy of simply building bigger GPU clusters—is finished. The pre-training paradigm has plateaued, and the industry is now scrambling for entirely new architectures (like inference-time reasoning) to find the next jump in intelligence.

Artificial Intelligence Weekly