A chill seems to be setting in over Wall Street. Tech billionaire Peter Thiel’s hedge fund recently sold its entire $100m (£76m) stake in Nvidia, the world’s most valuable chip company at the heart of the artificial intelligence (AI) boom.
Meanwhile, Michael Burry – famed for sounding the alarm before the 2008 financial crisis and Christian Bale’s depiction of him in the movie The Big Short – bet almost $200m (£152m) against the chipmaker.
Why would two investors of their calibre turn against a company whose share price has risen almost 15-fold in 5 years?
Partly because this isn’t just an Nvidia issue. The company makes the processors on which much of the AI industry rests – an industry already worth trillions of dollars that is almost single-handedly driving US economic growth.
But something in that growth story may be starting to fray. Many researchers and investors now suspect that AI’s astonishing momentum rests on a technical assumption that may not hold forever.
In other words, an AI bubble may be forming – and could easily be popped by a fatal flaw hiding in plain sight.
The big bet: bigger models = better AI
To understand what’s going on in the global economy right now, you first need to understand what AI, as we know it, actually is.
The current boom in AI technology has ridden on a wave known as ‘deep learning’, which is an approach to creating intelligent computer systems using ‘artificial neural networks’.
Neural networks aren’t new: the idea dates back to 1944, but only recently have they become big and fast enough to work well.
These systems are made up of interconnected nodes (artificial ‘neurons’) that process information and pass it to other nodes. Deep learning models stack multiple layers of these nodes, each layer extracting increasingly complex features from the data. The ‘deep’ refers to having many layers.
Without getting too bogged down in details, what this produces is models that are very good approximators, learning to predict what things should look like based on the patterns in their training data.
Large language models (LLMs) are the type of deep learning model most people are now familiar with, powering chatbots like OpenAI’s ChatGPT, Google’s Gemini and Anthropic’s Claude.
LLMs are trained on vast amounts of text so that they become very adept at predicting the next word in a sequence.
“They’re sort of like autocomplete on your phone,” says Gary Marcus, a leading voice in the AI sceptic community.
“You type something, and it guesses what’s going to come next.
“Basically, these are really sophisticated devices for making that prediction – looking at context, not just from the last few words like your phone might use, but using everything maybe in all the conversations that you’ve had going back some distance.”
Over time, a simple mantra took hold in Silicon Valley: make the models bigger and they’ll get better.
There are three levers to pull to achieve this.
- First, increase the model size. This entails adding more layers or nodes so the system learns far more parameters (the internal variables that encode knowledge).
- Second, increase the amount of training data. By feeding the model more examples, it can learn more patterns.
- Third, increase the amount of computing power, known in the industry as ‘compute’. This involves using more and faster chips during training, allowing the model to learn from the data more effectively.
Increasing these three factors in tandem led to what appeared to be a remarkably predictable rise in performance. Thus, the ‘scaling laws’ were born.
Much like Moore’s Law (which predicted that the number of transistors on a chip would double roughly every two years, enabling computers to shrink from room-sized to pocket-sized), AI companies assumed that simply scaling up models – making them larger, training on more data and using more compute – would deliver steady, almost guaranteed improvements in capability.

For several years, this held true. Applying these scaling laws turbo-charged AI development, and within half a decade, deep learning models went from quirky toys to systems that hundreds of millions of users rely on daily.
OpenAI’s successive GPT models are a prime example of the scaling mindset. GPT-3, released in 2020, contained 175 billion parameters, making it by far the largest model of its time.
Its 2023 successor GPT-4 is estimated to be 10 times larger, at roughly 1.8 trillion parameters. Training data has exploded as well: GPT-4 was reportedly trained on an astonishing 13 trillion tokens of text (a token is roughly 3/4 of a word). For comparison, the entire English Wikipedia contains only about 5 billion words – making it thousands of times smaller than GPT-4’s training material.
Other leading models from Anthropic, Google and Meta have all followed a similar pattern.
In theory, each 10-fold increase in model size and data was expected to yield new capabilities and better performance across tasks. And indeed, performance on many benchmarks has shot up.
For instance, GPT-4 achieved a score of 84.6 per cent on the Massive Multitask Language Understanding tests – a benchmark for AI systems that covers 57 topics – whereas GPT-3.5 scored 70 per cent.
That improvement closed the gap to human-level performance on many tasks. GPT-4, for example, could pass the bar exam and other professional tests that stumped its predecessors.
This progress fueled grand claims that an artificial general intelligence (AGI) – an AI that could do all your work better than you, drive your car, book your holidays and even make scientific breakthroughs – was on the horizon, and justified sky-high valuations for AI startups and suppliers.
The only problem? The scaling laws may not have been ‘laws’ at all.
The 3 biggest limits of today’s AI
“If I told you that my baby weighed 9 pounds at birth, and 18 months later it had doubled in weight,” Marcus posits, “that doesn’t mean it’s going to keep doubling and become a trillion-pound baby by the time it goes to college.”
What he means is that while the scaling laws looked like a real relationship at the time, and delivered impressive results to boot, there was no empirical evidence that they would hold forever.
Cracks are now showing, with bigger models not yielding proportional gains. Models may be tens of times larger than they were a couple of years ago, but they’re not 10 times smarter by most metrics.
All of this puts the AI frenzy in a different light. If simply throwing more data and computing power at the problem no longer yields dramatically better results, then the economic foundations of the AI boom start to wobble.
“The thing about these systems is they’re really just mimics – they don’t have a deep understanding of what they’re talking about,” Marcus says.
At their core, as Marcus puts it, today’s AI models are still “giant statistical machines” that learn correlations, not true comprehension. They predict outputs based on patterns in their training data, not by reasoning about the world the way humans do.
Unlike a calculator, which gives the right answer every time for the problems it’s built for, a neural network can never be 100 per cent correct 100 per cent of the time. It works more like a human brain, making its best guess based on patterns it has seen before.
This fundamental limitation leads to three well-known failures of AI models.
Read more:
1. Hallucinations
The over-generalisations or outright fabrications that even the latest state-of-the-art models produce are often euphemistically called ‘hallucinations’.
Most of us have encountered these by now. The AI confidently invents facts, cites nonexistent research or asserts something completely false.
Marcus uses the example of his friend Harry Shearer, an actor behind the voice of Mr Burns in The Simpsons, to explain. Shearer once found an AI-generated biography claiming he was British, which is wrong, as a quick check of Wikipedia would show.
Why did the model say that? Possibly because many other voice actors or comedians it read about in that category were British, so the pattern-matching statistical machine guessed that Shearer was too.
Fundamentally, the AI had no concept of who Harry Shearer actually is, or even what an actor or Britain is – it just regurgitated a likely-seeming correlation from its training data.
“They break everything into little pieces of information, and they learn the correlations between those bits of information,” Marcus says. In other words, there are no hard facts in deep learning models, only connections – and there never will be.

Empirically, newer models do hallucinate less than older ones, but they still do it quite frequently.
A 2024 study found ChatGPT-4 produced false information 28.6 per cent of the time, compared to a 39.6 per cent hallucination rate for GPT-3.5, in tests where factual accuracy was measured. According to OpenAI, the company has made “significant advances in reducing hallucinations” in its latest model, GPT-5, but they’re still commonplace.
In truth, no current AI model can be trusted to be consistently correct. We humans are still needed to sense-check what comes out, which doesn’t replace human expertise in a given subject, but necessitates it.
2. The ‘outlier problem’
Hallucinations are one issue; another is what happens when these models encounter situations outside the distribution of their training data. If an AI sees something genuinely new or weird – something that wasn’t well-represented in the billions of examples it ingested – it can completely break down.
Marcus calls this the ‘outlier problem’. He says, “There’s this infinite periphery around the centre of things that the systems haven’t been exposed to.”
Take self-driving cars. They can often reliably recognise other vehicles moving in familiar, orderly ways. But if they come across a lorry tipped on its side across two lanes – a shape they’ve barely, if ever, seen in training data – the system may fail to register it as a hazard at all. It doesn’t take much imagination to picture how costly such an error would be.
This is a major problem in terms of where the AI industry can go from here, and it’s why we haven’t seen lone AI scientist models winning Nobel Prizes for novel discoveries yet. Today’s deep learning models can remix human knowledge, but not extend it much beyond the frontier of what it’s seen.
3. Data limits
Where does all of this leave us? Well, AI models are now incredibly expensive to train and run, requiring not just vast amounts of data but enormous computational infrastructure.
And they’re literally running out of good data to learn from – so much so that companies are now scraping and transcribing everything (like YouTube video subtitles) just to get a bit more text to feed the beast.
“Everybody’s been using essentially 100 per cent of the internet for the last couple of years, and they’re not getting the same gains anymore,” Marcus says. “There isn’t 10 more internets to draw on.”
In fact, a 2024 analysis by the non-profit research institute Epoch AI estimated that at some point between 2028 to 2032, we may exhaust the supply of high-quality human text data to train on.
According to Elon Musk, who founded his own AI company, xAI, in 2023, that point may already have been reached. “The cumulative sum of human knowledge has been exhausted in AI training. That happened basically last year,” Musk said in an interview in January that was livestreamed on his social media platform X.

The cost of an AI revolution
All this talk of scale makes it sound like we understand how these systems truly work. We don’t. We know what their architecture looks like and how to make them perform tasks, but when it comes to the computations they’re doing internally to generate outputs, they’re essentially black boxes.
“We know how to build them, but we don’t know how to predict exactly what they’ll do,” Marcus says. “Fundamentally, the whole idea of using a black box where you just pour data in, like you would pour cranberries into a grinder, and expect cognition to come out of it, I think, is just a bad idea to start with.”
According to a survey by the Association for the Advancement of Artificial Intelligence, Marcus’ views are the consensus, with a comfortable majority of AI researchers agreeing that simply scaling current approaches won’t yield AGI.
Now, the economic underpinnings of this approach are beginning to show strain. The push for ever-larger models has an astronomical price tag. As early as last year, Anthropic CEO Dario Amodei predicted models could soon cost $10bn (£7.6bn)or more to train.
Training and using AI also takes a heavy environmental toll. While exact figures are difficult to ascertain, a 2021 preprint study by researchers from Google and the University of California, Berkeley, estimated that the training process alone for GPT-3 was 1,287 megawatt hours of electricity – enough to power 120 US homes for a year. That model was orders of magnitude smaller than those in use today; GPT-4, for example, was estimated to have needed 40 times that amount of power.
Using these models is costly, too. According to Goldman Sachs, a ChatGPT query needs nearly 10 times as much power as a typical Google search, and data centre power demand is projected to grow 160 per cent by 2030.
These hidden costs – the electric bills, the water for cooling servers, the supply chain for GPUs – are the less glamorous forces propelling (and potentially unravelling) the AI bubble.
Is this really a bubble?
This arms race has been a boon for chipmakers: Nvidia’s revenue surged from just over $20bn (£16bn) in 2022 to almost $130bn (£104bn) in the 12 months prior to August 2025. Its latest quarterly results and outlook were also positive, restoring at least some faith in the $4.5tr (£3.6tr) company.
And herein lies a curiosity, and a potential reason why it may not be time to stash your cash under the mattress just yet. Because, despite the eye-watering costliness of these systems, money is being made.
A recent bulletin, also from Goldman Sachs, cautiously touted that “we are not in a bubble… yet,” and the reason for this is that companies like Alphabet (Google’s parent company), Nvidia and Microsoft – all of which are at the core of the AI boom – are making hundreds of billions of dollars.
These are not the flimsy Pet.coms of the dot-com bubble in the late 1990s. They’re financial behemoths with money to invest in costly data centres and model training.
So while it might be true that numerous AI start-ups will go belly-up as the scaling train runs out of steam, the mega-companies propping up the global economy could hold firm.
There is also another way for these companies to stay profitable. Even AI companies that are losing money hand over fist could harness a treasure trove of lucrative data if they can capture enough market share.
Marcus calls this a move towards “surveillance capitalism” – the idea that our personal data becomes a raw material to be mined and sold. Think about your social media feeding you targeted adverts and selling your data elsewhere – the same techniques could be a cash cow for the AI industry.
He adds, “I think they’re definitely thinking about targeted ads and so forth. For personal data, they don’t have to solve the grand problems of artificial intelligence, they just have to get people to type stuff in – and they’re already doing that.”
A new way forward
If scaling current models won’t get us to the kind of transformational AI that many had predicted, what will?
One option, Marcus argues, is a return to an older idea that has been quietly waiting in the wings: neuro-symbolic AI. For the past half-century, AI research has largely split into two camps: those building neural networks and those developing symbolic systems.
Symbolic systems, as the name implies, manipulate symbols with formal logic.
“It’s called symbol manipulation because you have symbols that stand for things, like in algebra,” Marcus explains. “Classical computer programming is almost entirely made up of stuff like that, and neural networks don’t do that very well.”
He continues: “The classical stuff is really good at, for example, representing databases and ontologies. Like a robin is a bird, a bird is an animal, and concluding therefore that a robin is an animal. Classical AI techniques are perfect at that stuff. They never hallucinate.”
By combining the clear, rule-based logic of older AI with the pattern-spotting power of neural networks, Marcus thinks researchers could get much closer to true general intelligence. These hybrid systems would sidestep the rigid limits of traditional software while also reducing the errors and made-up answers that plague today’s models.
Some companies are already experimenting with this approach, most notably Google DeepMind.
Its AlphaFold2 system, which can accurately predict the 3D structure of proteins from their amino-acid sequence, has been widely hailed as one of the most important scientific breakthroughs of recent years. Crucially, it blends neural networks with elements of symbolic manipulation.

It is perhaps not surprising, then, that AlphaFold2 earned the 2024 Nobel Prize in Chemistry – a win Marcus has called “the first Nobel Prize for Neurosymbolic AI”. This wasn’t an AI system making discoveries unaided, but it was a major validation of the approach.
A neuro-symbolic strategy won’t, Marcus says, deliver AGI outright – but it could represent a significant leap forward.
And despite his pessimism about the current state of the field, he remains cautiously optimistic about what comes next.
“Will we have artificial general intelligence by 2027? I can say with absolute certainty, or nearly absolute certainty, no, we won’t.”
But, he adds, “I absolutely think a better artificial intelligence is possible.
“The tragedy of this era is that we’re spending so much money on one bet. That one bet, trillions of dollars, on the one bet is that scaling, adding more data and adding more compute will bring us to artificial general intelligence. I think there’s actually lots of evidence against that at this point.”
Read more:
