The AI Industry’s Scaling Obsession Is Headed for a Cliff

Date:

Share:


A new study from MIT suggests the biggest and most computationally intensive AI models may soon offer diminishing returns compared to smaller models. By mapping scaling laws against continued improvements in model efficiency, the researchers found that it could become harder to wring leaps in performance from giant models whereas efficiency gains could make models running on more modest hardware increasingly capable over the next decade.

“In the next five to 10 years, things are very likely to start narrowing,” says Neil Thompson, a computer scientist and professor at MIT involved in the study.

Leaps in efficiency, like those seen with DeepSeek’s remarkably low-cost model in January, have already served as a reality check for the AI industry, which is accustomed to burning massive amounts of compute.

As things stand, a frontier model from a company like OpenAI is currently much better than a model trained with a fraction of the compute from an academic lab. While the MIT team’s prediction might not hold if, for example, new training methods like reinforcement learning produce surprising new results, they suggest that big AI firms will have less of an edge in the future.

Hans Gundlach, a research scientist at MIT who led the analysis, became interested in the issue due to the unwieldy nature of running cutting edge models. Together with Thompson and Jayson Lynch, another research scientist at MIT, he mapped out the future performance of frontier models compared to those built with more modest computational means. Gundlach says the predicted trend is especially pronounced for the reasoning models that are now in vogue, which rely more on extra computation during inference.

Thompson says the results show the value of honing an algorithm as well as scaling up compute. “If you are spending a lot of money training these models, then you should absolutely be spending some of it trying to develop more efficient algorithms, because that can matter hugely,” he adds.

The study is particularly interesting given today’s AI infrastructure boom (or should we say “bubble”?)—which shows little sign of slowing down.

OpenAI and other US tech firms have signed hundred-billion-dollar deals to build AI infrastructure in the United States. “The world needs much more compute,” OpenAI’s president, Greg Brockman, proclaimed this week as he announced a partnership between OpenAI and Broadcom for custom AI chips.

A growing number of experts are questioning the soundness of these deals. Roughly 60 percent of the cost of building a data center goes toward GPUs, which tend to depreciate quickly. Partnerships between the major players also appear circular and opaque.



Source link

━ more like this

Waze expands speed bump, toll, and emergency vehicle alerts worldwide

It looks like Waze is finally ready to make good on some old promises. After a long stretch of silence, the Google-owned navigation...

Alder Lake is ending and here’s what it means for your current PC

It looks like the sun is finally setting on Intel’s 12th Gen Alder Lake series, signaling the end of the road for one...

1Password helps prevent your passwords from going to scam sites

Phishing scams are evolving fast, and AI-assisted sites are making fake login pages look more convincing than ever. To help users stay safe,...

You might actually be able to buy a Tesla robot in 2027

Tesla CEO Elon Musk has once again laid out an ambitious timeline for the company’s long-awaited humanoid robot, Optimus. Speaking at the World...

Your next road trip is booked: Forza Horizon 6 comes this May

After months of anticipation and speculation, and even a leaked release date, Playground Games and Xbox have finally given fans what they’ve been...
spot_img