option
Home
News
DeepMind's AI Outperforms IMO Gold Medalists

DeepMind's AI Outperforms IMO Gold Medalists

April 10, 2025
110

Google DeepMind's latest AI, AlphaGeometry2, has made waves by outperforming the average gold medalist in solving geometry problems at the International Mathematical Olympiad (IMO). This advanced version of the previously released AlphaGeometry, introduced just last January, reportedly solved 84% of geometry problems from the last 25 years of IMO competitions.

You might wonder why DeepMind is focusing on a high school math contest. Well, they believe that cracking these challenging Euclidean geometry problems could be a stepping stone to developing more advanced AI. Solving these problems requires both logical reasoning and the ability to navigate through various solution paths, skills that could be crucial for future general-purpose AI systems.

This summer, DeepMind showcased a system that combined AlphaGeometry2 with AlphaProof, another AI model designed for formal math reasoning. Together, they tackled four out of six problems from the 2024 IMO. This approach could potentially extend beyond geometry to other areas of math and science, like complex engineering calculations.

AlphaGeometry2 is powered by a few key components, including a language model from Google's Gemini family and a "symbolic engine." The Gemini model assists the symbolic engine, which applies mathematical rules to find solutions, in creating feasible proofs for geometry theorems.

A typical geometry diagram in the IMO.

A typical geometry problem diagram in an IMO exam.Image Credits:Google (opens in a new window)

In the IMO, geometry problems often require adding "constructs" like points, lines, or circles to diagrams before solving them. AlphaGeometry2's Gemini model predicts which constructs might be helpful, guiding the symbolic engine to make deductions.

Here's how it works: The Gemini model suggests steps and constructions in a formal mathematical language, which the engine then checks for logical consistency. AlphaGeometry2 uses a search algorithm to explore multiple solution paths simultaneously and stores potentially useful findings in a shared knowledge base.

A problem is considered "solved" when AlphaGeometry2 combines the Gemini model's suggestions with the symbolic engine's known principles to form a complete proof.

Due to the scarcity of usable geometry training data, DeepMind created synthetic data to train AlphaGeometry2's language model, generating over 300 million theorems and proofs of varying complexity.

The DeepMind team tested AlphaGeometry2 on 45 geometry problems from IMO competitions spanning 2000 to 2024, which they expanded into 50 problems. AlphaGeometry2 solved 42 of these, surpassing the average gold medalist score of 40.9.

However, AlphaGeometry2 has its limitations. It struggles with problems involving a variable number of points, nonlinear equations, and inequalities. While it's not the first AI to reach gold-medal-level performance in geometry, it's the first to do so with such a large problem set.

When faced with a tougher set of 29 IMO-nominated problems that haven't yet appeared in competitions, AlphaGeometry2 could only solve 20.

The study's results are likely to spark further debate about the best approach to building AI systems. Should we focus on symbol manipulation, where AI uses rules to manipulate symbols representing knowledge, or on neural networks, which mimic the human brain's structure and learn from data?

AlphaGeometry2 takes a hybrid approach, combining the neural network architecture of the Gemini model with the rules-based symbolic engine.

Supporters of neural networks argue that intelligent behavior can emerge from vast amounts of data and computing power. In contrast, symbolic AI proponents believe it's better suited for encoding knowledge, reasoning through complex scenarios, and explaining solutions.

Vince Conitzer, a Carnegie Mellon University computer science professor specializing in AI, commented on the contrast between the impressive progress on benchmarks like the IMO and the ongoing struggles of language models with simple commonsense problems. He emphasized the need to better understand these systems and the risks they pose.

AlphaGeometry2 suggests that combining symbol manipulation and neural networks might be a promising way forward in the quest for generalizable AI. Interestingly, the DeepMind team found that AlphaGeometry2's language model could generate partial solutions to problems without the symbolic engine's help, hinting at the potential for language models to become self-sufficient in the future.

However, the team noted that until language model speed improves and hallucinations are resolved, tools like symbolic engines will remain essential for math applications.

Related article
Google Commits to EU’s AI Code of Practice Amid Industry Debate Google Commits to EU’s AI Code of Practice Amid Industry Debate Google has pledged to adopt the European Union’s voluntary AI code of practice, a framework designed to assist AI developers in aligning with the EU’s AI Act by implementing compliant processes and sy
Google Unveils Production-Ready Gemini 2.5 AI Models to Rival OpenAI in Enterprise Market Google Unveils Production-Ready Gemini 2.5 AI Models to Rival OpenAI in Enterprise Market Google intensified its AI strategy Monday, launching its advanced Gemini 2.5 models for enterprise use and introducing a cost-efficient variant to compete on price and performance.The Alphabet-owned c
Meta Enhances AI Security with Advanced Llama Tools Meta Enhances AI Security with Advanced Llama Tools Meta has released new Llama security tools to bolster AI development and protect against emerging threats.These upgraded Llama AI model security tools are paired with Meta’s new resources to empower c
Comments (27)
0/200
AlbertSmith
AlbertSmith August 9, 2025 at 5:00:59 PM EDT

Wow, AlphaGeometry2 is killing it at IMO geometry problems! Beating gold medalists is wild—makes me wonder if AI will soon design math contests instead of just solving them. 😮

JackCarter
JackCarter July 27, 2025 at 10:13:31 PM EDT

This AI beating IMO champs is wild! 🤯 Makes me wonder if it could tutor me in math or just take over the world one proof at a time.

LarryMartin
LarryMartin April 20, 2025 at 5:52:37 AM EDT

AlphaGeometry2는 정말 놀랍네요! IMO 기하 문제를 금메달리스트보다 더 잘 풀어요. 하지만 저는 수학 천재가 아니라서 마술 쇼를 보는 기분이에요. 좀 더 이해할 수 있으면 좋겠어요! 🤯

JerryGonzález
JerryGonzález April 16, 2025 at 7:58:03 PM EDT

DeepMindのAlphaGeometry2がIMOの金メダリストを上回るなんて信じられない!まるでAIが人間より賢いSF映画を見ているよう。でも本当に、私の宿題を手伝ってくれるのかな?😂

RaymondBaker
RaymondBaker April 16, 2025 at 7:21:57 AM EDT

AlphaGeometry2 is mind-blowing! It's solving IMO geometry problems better than gold medalists. But I'm no math whiz, so it's more like watching a magic show for me. Wish I could understand it better! 🤯

JonathanAllen
JonathanAllen April 16, 2025 at 4:15:26 AM EDT

AlphaGeometry2 é impressionante! Resolve problemas de geometria do IMO melhor que medalhistas de ouro. Mas eu não sou um gênio em matemática, então pra mim é como assistir a um show de mágica. Queria entender melhor! 🤯

Back to Top
OR