At the 2025 International Mathematical Olympiad (IMO), held in Queensland, Australia, DeepMindās advanced AI system Gemini Deep Think achieved a historic milestone: it solved five of six world-class high school math problems in natural language, earning an official gold-medal-level performanceāthe first-ever time an AI has been officially graded at this level.
But how it reached this milestone, what it signifies for AIās future, and what comes next is a deeper story.

š From Silver to Gold: Evolution of Math AI
- 2024 achievement: DeepMind combined AlphaProof (formal proofs) and AlphaGeometry (neuro-symbolic geometry) to earn a silver-level score by solving four of six problems. These earlier models required human-guided formal language conversion.
- 2025 advancement: Gemini Deep Think handled all stepsāfrom problem understanding to full natural-language proofsāwithout formal language conversion, completing all work within the official 4.5-hour limit.
š§ What Powers Gemini Deep Think?
- Parallel-thinking search
Simultaneously explores multiple reasoning paths, evaluating each before converging on a solutionāmirroring how humans brainstorm. - Rich training data
Trained on thousands of curated IMO-style proofs, Gemini Deep Think develops both pattern recognition and structural reasoning. - Natural language fluency
Produces rigorous, human-readable solutionsāmarking a shift from symbolic logic to general conversational problem solving.
š Competition and Integrity in AI Math
While other AI labs also demonstrated high-level performance, DeepMind stood out for submitting its results directly to the IMO for formal evaluation. This official grading by human judges added credibility and transparency to the achievement.
š Beyond the IMO: Wider Implications
- Scientific collaboration: These capabilities could soon assist researchers with theorem generation, verification, and idea exploration.
- AI research frontier: The leap from silver-level formal logic to gold-level natural language reasoning is a major milestone toward broader artificial general intelligence.
- Benchmarking challenge: As AI solves existing test sets, the need grows for new, dynamic reasoning benchmarks that span multiple disciplines.
š§ Caveats and Limits
Despite the win, Gemini Deep Think is not flawless:
- It failed to solve the hardest problem on the testāone that stumped most human contestants.
- Natural-language reasoning excels in well-structured tasks but still struggles with ambiguous, open-ended real-world problems.
- The system requires enormous compute resources, raising questions about accessibility, efficiency, and environmental impact.
š ļø Whatās Next?
- Selective release: Gemini Deep Think will be available first to academic partners, with broader availability to follow through premium platforms.
- Research extensions: Its technology is expected to expand into areas like physics, computer science, and unsolved mathematical domains.
- Benchmark evolution: New challenges such as cross-domain problem-solving are being developed to keep AI models growing meaningfully.
š¤ Frequently Asked Questions
Q: Did Gemini Deep Think actually receive an official IMO gold medal?
A: Not an award per se, but its solutions were officially evaluated by IMO judges and earned a gold-medal-level score.
Q: How did this differ from DeepMindās 2024 entry?
A: In 2024, earlier models used formal logic and human assistance. This year, Gemini Deep Think completed all reasoning in natural language, independently.
Q: What is parallel-thinking?
A: Itās a reasoning strategy where the AI explores multiple solutions at once, evaluates them in parallel, and then chooses or blends the most effective paths.
Q: Will DeepMind release this model publicly?
A: Initially, access will be limited to research institutions, with future public release expected through select platforms.
Q: Does this mean AI can now solve any math problem?
A: No. It shows significant progress in structured problem-solving, but open-ended, abstract problems still pose major challenges.
š Final Thought
Gemini Deep Thinkās gold-standard success at the IMO marks a monumental leap in AIās reasoning capabilitiesāmoving from structured, formal logic to fluid, human-like problem solving. But while it unlocks new possibilities for scientific collaboration and advanced reasoning, vast challenges and real-world applicability still lie ahead.
AI can now think like a mathlete. Next, weāll see if it can think like a scientist.

Sources Google Deepmind


