Aussie AI
AI Hitting the Wall?
-
Last Updated 28 February, 2025
-
by David Spuler, Ph.D.
Update Jan 2025: DeepSeek Ends the Wall Debate
The wall debate is over, but it's not OpenAI that provided the sledgehammer. The recent release of DeepSeek R1 by a Chinese startup has largely ended the wall debate. Their model beat OpenAI's o1 model on multiple benchmarks (but not all), and did so using only single-step inference (whereas o1 is multi-step inference). Hence, they used training advances to train a much smarter model than other single-step models.
Bye-bye, wall!!
The Nov 2024 Wall Debate
Recently, there's been a two-fold indication that AI progress is "plateauing" or "hitting a wall." The two main indicators are:
- Inference-based reasoning ("test time compute")
- Underwhelming progress in new models
The GPT "o1" model released in September 2024 wasn't a bigger, more heavily-trained model with trillions more weights. Instead, it's a model that improves intelligence by doing multiple steps of inference, rather than one smarter step in an uber-trained model. This algorithm for "multi-step reasoning" is known as "chain-of-thought" and uses repeated calls to process queries, before merging them together into the one final response.
Why does this change to multi-step inference for reasoning support the "wall" theory? Well, inference is a slow process when it runs, and "o1" is therefore slow for users — the line of logic goes that OpenAI wouldn't tolerate using this slow method if they could do it with one request to a bigger model. It almost seems like a kind of workaround.
Hence, wall.
Secondly, there are also rumors that the big players are having difficulty training much better next-gen models. In particular, there are indicators that the GPT-5 release is having trouble gaining capabilities compared to GPT-4. Instead of launching GPT-5 soon, we got "o1" with its multiple steps.
Obviously, training trillion-parameter models is a specialist field, and it's evolving fast, with literally billions of dollars in funding being applied there. But open source models seem to be keeping up with the leading commercial vendors (albeit, after a lag), which tends to indicate that there's only incremental progress in reasoning capabilities, and the commercial vendors don't have a huge "secret sauce" algorithmic advantage in training. Some of the constraints include:
- Shortage of new high-quality training data (text).
- Complexity of software algorithms to train ever-bigger LLMs.
- Sheer volume of training data needed for multimodal LLMs (audio, images, and video).
- Capital cost of GPUs to crunch all that.
- Apparent lack of a new algorithmic advance in one-shot reasoning.
- Fundamental limitations of the way that LLMs and Transformers work.
On the other hand, there's a lot of research happening in training and in making LLMs better at reasoning in general. Some of the newer areas include:
- Newer GPU hardware for training (e.g., Blackwell).
- Faster software training algorithms (optimizing both computations and inter-GPU network traffic).
- Resiliency improvements to training (both software and hardware).
- Synthetic training data and derivative data.
- Multi-step reasoning algorithms are smarter (if slower).
- Long context processing seems to be a solved problem now.
- Inference optimization research (makes each step of multi-step reasoning faster).
- Next-gen architectures beyond LLMs (e.g., SSMs, Mamba, Hyena, and hybrid versions).
Is there a wall? OpenAI CEO Sam Altman posted on X that "there is no wall." And there are certainly signs that many of the bigger players are still gearing up to use NVIDIA Blackwell GPUs for even bigger training runs. And there have been two multi-billion dollar fund raises in just the last month. So, the plateau may only be a temporary thing.
Research on the AI Progress Wall
Articles and papers on recent AI progress:
- Deirdre Bosa, Jasmine Wu, Dec 11 2024, The limits of intelligence — Why AI advancement could be slowing down, https://www.cnbc.com/2024/12/11/why-ai-advancement-could-be-slowing-down.html
- The Information, Nov 2024, OpenAI Shifts Strategy as Rate of GPT AI Improvement Slows https://www.theinformation.com/articles/openai-shifts-strategy-as-rate-of-gpt-ai-improvements-slows
- Bloomberg, Nov 2024, OpenAI, Google and Anthropic are Struggling to Build More Advanced AI, https://www.bloomberg.com/news/articles/2024-11-13/openai-google-and-anthropic-are-struggling-to-build-more-advanced-ai
- Gary Marcus, Nov 25, 2024, A new AI scaling law shell game? Scaling laws ain’t what they used to be, https://garymarcus.substack.com/p/a-new-ai-scaling-law-shell-game
- Kyle Orland, 13 Nov 2024, What if AI doesn’t just keep getting better forever? New reports highlight fears of diminishing returns for traditional LLM training. https://arstechnica.com/ai/2024/11/what-if-ai-doesnt-just-keep-getting-better-forever/
- Will Lockettm Nov 2024, Apple Calls BS On The AI Revolution, They aren’t late to the AI game; they are just the only sceptical big tech company. https://medium.com/predict/apple-calls-bullshit-on-the-ai-revolution-ae38fdf83392
- Sam Altman, Nov 14, 2024, there is no wall, https://x.com/sama/status/1856941766915641580
- Shirin Ghaffary, December 6, 2024, Tech CEOs Say It’s Getting Harder to Build Better AI Systems. The comments follow a renewed debate over whether AI is hitting a scaling wall. https://www.bloomberg.com/news/newsletters/2024-12-05/tech-ceos-say-it-s-getting-harder-to-build-better-ai-systems
- Maxwell Zeff, November 20, 2024, Current AI scaling laws are showing diminishing returns, forcing AI labs to change course, https://techcrunch.com/2024/11/20/ai-scaling-laws-are-showing-diminishing-returns-forcing-ai-labs-to-change-course/ ("at least 10 to 20x gains in model performance ...intelligent prompting, UX decisions, and passing context at the right time into the models...")
- Kylie Robison, Dec 14, 2024, OpenAI cofounder Ilya Sutskever says the way AI is built is about to change. “We’ve achieved peak data and there’ll be no more,” OpenAI’s former chief scientist told a crowd of AI researchers. https://www.theverge.com/2024/12/13/24320811/what-ilya-sutskever-sees-openai-model-data-training
- Joe Procopio, Dec 17, 2024, We’ve Hit The “AI Wall.” Here’s What That Means For the Tech Industry. https://ehandbook.com/weve-hit-the-ai-wall-here-s-what-that-means-for-the-tech-industry-97f543a68e77
- Lan Chu, Jan 2025, Is AI progress slowing down? https://levelup.gitconnected.com/is-ai-progress-slowing-down-69d4f1215e49
- Duncan Anderson, Jan 2025, The wall that wasn’t: Benchmark results for the latest AI models suggest that any “scaling wall” has already been breached and we’re on the path to AGI. https://medium.com/barnacle-labs/the-wall-that-wasnt-62c617f66ad4
- Jano le Roux, Jan 2025, Why AI’s Growth Will Hit A Wall Very Very Soon, https://medium.com/swlh/why-ais-growth-will-hit-a-wall-very-very-soon-f6c138b7cfcb
- Kyle Wiggers, January 27, 2025, Viral AI company DeepSeek releases new image model family, https://techcrunch.com/2025/01/27/viral-ai-company-deepseek-releases-new-image-model-family/
- Manish Singh, January 27, 2025, DeepSeek ‘punctures’ AI leaders’ spending plans, and what analysts are saying, https://techcrunch.com/2025/01/27/deepseek-punctures-tech-spending-plans-and-what-analysts-are-saying/
- Rafe Brena, Jan 31, 2025, AI Isn’t ‘Hitting A Wall.” Here Is Why: What does DeepSeek have to do with it? https://pub.towardsai.net/ai-isnt-hitting-a-wall-here-is-why-e75fe86e47f1
- Ahmed El-Kishky, Alexander Wei, Andre Saraiva, Borys Minaev, Daniel Selsam, David Dohan, Francis Song, Hunter Lightman, Ignasi Clavera, Jakub Pachocki, Jerry Tworek, Lorenz Kuhn, Lukasz Kaiser, Mark Chen, Max Schwarzer, Mostafa Rohaninejad, Nat McAleese, o3 contributors, Oleg Mürk, Rhythm Garg, Rui Shu, Szymon Sidor, Vineet Kosaraju, Wenda Zhou, 3 Feb 2025, Competitive Programming with Large Reasoning Models, https://arxiv.org/abs/2502.06807 (OpenAI's paper on o3 that has similar conclusions to what DeepSeek showed about Reinforcement Learning for reasoning models, namely that "scaling general-purpose reinforcement learning" still works.)
- Jeremy Kahn, February 26, 2025, The $19.6 billion pivot: How OpenAI’s 2-year struggle to launch GPT-5 revealed that its core AI strategy has stopped working, https://fortune.com/2025/02/25/what-happened-gpt-5-openai-orion-pivot-scaling-pre-training-llm-agi-reasoning/
More AI Research
Read more about: