option
Home
News
DeepMind's AI Outperforms IMO Gold Medalists

DeepMind's AI Outperforms IMO Gold Medalists

April 10, 2025
60

Google DeepMind's latest AI, AlphaGeometry2, has made waves by outperforming the average gold medalist in solving geometry problems at the International Mathematical Olympiad (IMO). This advanced version of the previously released AlphaGeometry, introduced just last January, reportedly solved 84% of geometry problems from the last 25 years of IMO competitions.

You might wonder why DeepMind is focusing on a high school math contest. Well, they believe that cracking these challenging Euclidean geometry problems could be a stepping stone to developing more advanced AI. Solving these problems requires both logical reasoning and the ability to navigate through various solution paths, skills that could be crucial for future general-purpose AI systems.

This summer, DeepMind showcased a system that combined AlphaGeometry2 with AlphaProof, another AI model designed for formal math reasoning. Together, they tackled four out of six problems from the 2024 IMO. This approach could potentially extend beyond geometry to other areas of math and science, like complex engineering calculations.

AlphaGeometry2 is powered by a few key components, including a language model from Google's Gemini family and a "symbolic engine." The Gemini model assists the symbolic engine, which applies mathematical rules to find solutions, in creating feasible proofs for geometry theorems.

A typical geometry diagram in the IMO.

A typical geometry problem diagram in an IMO exam.Image Credits:Google (opens in a new window)

In the IMO, geometry problems often require adding "constructs" like points, lines, or circles to diagrams before solving them. AlphaGeometry2's Gemini model predicts which constructs might be helpful, guiding the symbolic engine to make deductions.

Here's how it works: The Gemini model suggests steps and constructions in a formal mathematical language, which the engine then checks for logical consistency. AlphaGeometry2 uses a search algorithm to explore multiple solution paths simultaneously and stores potentially useful findings in a shared knowledge base.

A problem is considered "solved" when AlphaGeometry2 combines the Gemini model's suggestions with the symbolic engine's known principles to form a complete proof.

Due to the scarcity of usable geometry training data, DeepMind created synthetic data to train AlphaGeometry2's language model, generating over 300 million theorems and proofs of varying complexity.

The DeepMind team tested AlphaGeometry2 on 45 geometry problems from IMO competitions spanning 2000 to 2024, which they expanded into 50 problems. AlphaGeometry2 solved 42 of these, surpassing the average gold medalist score of 40.9.

However, AlphaGeometry2 has its limitations. It struggles with problems involving a variable number of points, nonlinear equations, and inequalities. While it's not the first AI to reach gold-medal-level performance in geometry, it's the first to do so with such a large problem set.

When faced with a tougher set of 29 IMO-nominated problems that haven't yet appeared in competitions, AlphaGeometry2 could only solve 20.

The study's results are likely to spark further debate about the best approach to building AI systems. Should we focus on symbol manipulation, where AI uses rules to manipulate symbols representing knowledge, or on neural networks, which mimic the human brain's structure and learn from data?

AlphaGeometry2 takes a hybrid approach, combining the neural network architecture of the Gemini model with the rules-based symbolic engine.

Supporters of neural networks argue that intelligent behavior can emerge from vast amounts of data and computing power. In contrast, symbolic AI proponents believe it's better suited for encoding knowledge, reasoning through complex scenarios, and explaining solutions.

Vince Conitzer, a Carnegie Mellon University computer science professor specializing in AI, commented on the contrast between the impressive progress on benchmarks like the IMO and the ongoing struggles of language models with simple commonsense problems. He emphasized the need to better understand these systems and the risks they pose.

AlphaGeometry2 suggests that combining symbol manipulation and neural networks might be a promising way forward in the quest for generalizable AI. Interestingly, the DeepMind team found that AlphaGeometry2's language model could generate partial solutions to problems without the symbolic engine's help, hinting at the potential for language models to become self-sufficient in the future.

However, the team noted that until language model speed improves and hallucinations are resolved, tools like symbolic engines will remain essential for math applications.

Related article
Billionaires Discuss Automating Jobs Away in This Week's AI Update Billionaires Discuss Automating Jobs Away in This Week's AI Update Hey everyone, welcome back to TechCrunch's AI newsletter! If you're not already subscribed, you can sign up here to get it delivered straight to your inbox every Wednesday.We took a little break last week, but for good reason—the AI news cycle was on fire, thanks in large part to the sudden surge of
NotebookLM App Launches: AI-Powered Tool for Instant Knowledge Access Anywhere NotebookLM App Launches: AI-Powered Tool for Instant Knowledge Access Anywhere NotebookLM Goes Mobile: Your AI-Powered Research Assistant Now on Android & iOSWe’ve been blown away by the response to NotebookLM—millions of users have embraced it as their go-to
Imagen 4 is Google’s newest AI image generator Imagen 4 is Google’s newest AI image generator Google has just unveiled its latest image-generating AI model, Imagen 4, promising users an even better visual experience than its predecessor, Imagen 3. Announced at Google I/O 20
Comments (25)
0/200
HarryPerez
HarryPerez April 12, 2025 at 12:00:00 AM GMT

DeepMind's AlphaGeometry2 beating IMO gold medalists is mind-blowing! It's amazing to see AI tackling such complex problems. But, does it really understand geometry, or is it just pattern matching? Either way, it's a huge step forward for AI in education!

JackSanchez
JackSanchez April 12, 2025 at 12:00:00 AM GMT

DeepMindのAlphaGeometry2がIMOの金メダリストを上回るなんて驚きです!AIがこんなに複雑な問題に取り組むのを見るのは素晴らしいです。でも、本当に幾何学を理解しているのか、それともただのパターンマッチングなのか?どちらにしても、教育におけるAIの大きな一歩ですね!

HenryJackson
HenryJackson April 11, 2025 at 12:00:00 AM GMT

DeepMind의 AlphaGeometry2가 IMO 금메달리스트를 능가하다니 놀랍네요! AI가 이렇게 복잡한 문제를 다루는 걸 보는 건 정말 멋져요. 하지만 정말 기하학을 이해하는 건지, 아니면 단순히 패턴 매칭을 하는 건지 궁금해요. 어쨌든 교육에서의 AI 발전에 큰 걸음이에요!

HarryRoberts
HarryRoberts April 11, 2025 at 12:00:00 AM GMT

O AlphaGeometry2 da DeepMind superar os medalhistas de ouro do IMO é impressionante! É incrível ver a IA lidando com problemas tão complexos. Mas, será que ela realmente entende geometria, ou é apenas correspondência de padrões? De qualquer forma, é um grande avanço para a IA na educação!

BillyRoberts
BillyRoberts April 10, 2025 at 12:00:00 AM GMT

¡Que AlphaGeometry2 de DeepMind supere a los medallistas de oro del IMO es alucinante! Es increíble ver a la IA abordando problemas tan complejos. Pero, ¿realmente entiende la geometría, o solo está haciendo coincidencia de patrones? De cualquier manera, es un gran paso adelante para la IA en la educación!

AlbertHarris
AlbertHarris April 14, 2025 at 12:00:00 AM GMT

DeepMind's AlphaGeometry2 is mind-blowing! It's solving geometry problems better than IMO gold medalists. I used it to help with my math homework and it was spot on! The only thing is, it's a bit too advanced for casual users like me. Still, it's a solid 4 out of 5. 📚

Back to Top
OR