option
Home
News
AI Scaling Breakthrough Questioned by Experts

AI Scaling Breakthrough Questioned by Experts

April 10, 2025
96

AI Scaling Breakthrough Questioned by Experts

There's been some buzz on social media about researchers discovering a new AI "scaling law," but experts are taking it with a grain of salt. AI scaling laws, which are more like informal guidelines, show how AI models get better as you throw more data and computing power at them. Up until about a year ago, the big trend was all about "pre-training" – basically, training bigger models on bigger datasets. That's still a thing, but now we've got two more scaling laws in the mix: post-training scaling, which is all about tweaking a model's behavior, and test-time scaling, which involves using more computing power during inference to boost a model's "reasoning" capabilities (think models like R1). Recently, researchers from Google and UC Berkeley dropped a paper that some folks online are calling a fourth law: "inference-time search." This method has the model spit out a bunch of possible answers to a query at the same time and then pick the best one. The researchers claim it can juice up the performance of an older model, like Google's Gemini 1.5 Pro, to beat OpenAI's o1-preview "reasoning" model on science and math benchmarks. Eric Zhao, a Google doctorate fellow and one of the paper's co-authors, shared on X that by just randomly sampling 200 responses and letting the model self-verify, Gemini 1.5 – which he jokingly called an "ancient early 2024 model" – could outdo o1-preview and even get close to o1. He pointed out that self-verification gets easier as you scale up, which is kind of counterintuitive but cool. But not everyone's convinced. Matthew Guzdial, an AI researcher and assistant professor at the University of Alberta, told TechCrunch that this approach works best when you've got a solid way to judge the answers. Most questions aren't that straightforward, though. He said, "If we can't write code to define what we want, we can't use [inference-time] search. For something like general language interaction, we can't do this... It's generally not a great approach to actually solving most problems." Zhao responded, saying their paper actually looks at cases where you don't have a clear way to judge the answers, and the model has to figure it out on its own. He argued that the gap between having a clear way to judge and not having one can shrink as you scale up. Mike Cook, a research fellow at King's College London, backed up Guzdial's view, saying that inference-time search doesn't really make the model's reasoning better. It's more like a workaround for the model's tendency to make confident mistakes. He pointed out that if your model messes up 5% of the time, checking 200 attempts should make those mistakes easier to spot. This news might be a bit of a downer for the AI industry, which is always on the hunt for ways to boost model "reasoning" without breaking the bank. As the paper's authors noted, reasoning models can rack up thousands of dollars in computing costs just to solve one math problem. Looks like the search for new scaling techniques is far from over. *Updated 3/20 5:12 a.m. Pacific: Added comments from study co-author Eric Zhao, who takes issue with an assessment by an independent researcher who critiqued the work.*
Related article
Authentic Focusing System Developed for Affordable Augmented Reality Authentic Focusing System Developed for Affordable Augmented Reality Revolutionizing Projection-Based Augmented RealityResearchers from the prestigious Institute of Electrical and Electronics Engineers (IEEE) have made a groundbreaking leap forward
How we’re using AI to help cities tackle extreme heat How we’re using AI to help cities tackle extreme heat It's looking like 2024 might just break the record for the hottest year yet, surpassing 2023. This trend is particularly tough on folks living in urban heat islands—those spots in cities where concrete and asphalt soak up the sun's rays and then radiate the heat right back out. These areas can warm
'Degraded' Synthetic Faces May Enhance Facial Recognition Technology 'Degraded' Synthetic Faces May Enhance Facial Recognition Technology Researchers at Michigan State University have come up with an innovative way to use synthetic faces for a noble cause—enhancing the accuracy of image recognition systems. Instead of contributing to the deepfakes phenomenon, these synthetic faces are designed to mimic the imperfections found in real-
Comments (35)
0/200
JustinJackson
JustinJackson April 11, 2025 at 12:00:00 AM GMT

The hype around this new AI scaling law is a bit overblown, if you ask me. Experts are skeptical, and I'm not surprised. It's interesting, but I'm not ready to bet the farm on it just yet. Anyone else feeling the same?

PatrickCarter
PatrickCarter April 11, 2025 at 12:00:00 AM GMT

この新しいAIスケーリング法の話題、少し大げさに感じます。専門家も懐疑的で、私も驚きません。興味深いですが、まだ全面的に信じるのは早いかなと思います。皆さんも同じ気持ちですか?

AnthonyPerez
AnthonyPerez April 11, 2025 at 12:00:00 AM GMT

이 새로운 AI 스케일링 법에 대한 과대광고는 조금 과장된 것 같아요. 전문가들도 회의적이고, 저도 놀랍지 않아요. 흥미롭긴 하지만, 아직 이것에 전부를 걸기에는 이릅니다. 다른 분들도 같은 생각이신가요?

JamesTaylor
JamesTaylor April 11, 2025 at 12:00:00 AM GMT

O hype em torno dessa nova lei de escalabilidade de IA está um pouco exagerado, se me perguntar. Os especialistas estão céticos e eu não fico surpreso. É interessante, mas ainda não estou pronto para apostar tudo nisso. Alguém mais sente o mesmo?

GeorgeEvans
GeorgeEvans April 11, 2025 at 12:00:00 AM GMT

El entusiasmo alrededor de esta nueva ley de escalabilidad de IA está un poco exagerado, si me lo preguntas. Los expertos son escépticos y no me sorprende. Es interesante, pero no estoy listo para apostar todo en esto todavía. ¿Alguien más siente lo mismo?

PaulHernández
PaulHernández April 11, 2025 at 12:00:00 AM GMT

Heard about this new AI scaling law? Sounds cool but honestly, I'm not convinced. It feels like every other week there's a new 'breakthrough' that fizzles out. Experts seem skeptical too, so I'm just gonna wait and see. Anyone else feel the same?

Back to Top
OR