OpenAI recently announced that its generative AI model had achieved a Gold medal performance in the International Mathematical Olympiad (IMO). Shortly thereafter, Demis Hassabis—CEO of Google DeepMind and Nobel laureate—announced that their model has achieved the same performance. Last year, their specialized model called AlphaProof and AlphaGeometry could only manage a Silver.

This isn’t just another benchmark. The IMO fiercely tests pre-university students’ mathematical creativity, problem-solving skills, and logical depth. It does not test routine calculations; instead, it demands original, often elegant, ideas. The problems are notoriously hard even for some mathematicians, let alone machines.

The exam typically consists of six problems spread over two days, and examiners score each contestant out of 42 points. Achieving a Gold usually requires 30+ points, depending on that year’s difficulty. India has had an impressive history at the IMO, with students consistently ranking in the top 20 countries globally and winning Golds and Silvers several times..

So, what does it mean when an AI model secures a Gold in such a competition? It’s exciting news for more than one reason. It signals that generative AI—originally embraced for tasks like writing emails, generating code snippets, or searching resources online—can now solve some of the hardest intellectual problems, which everyone once believed required deep human intuition and creativity.

This doesn’t mean AI has become sentient, though that misunderstanding persists. What it does mean is that these models are now robust tools capable of complex reasoning, mathematical insight, and strategic problem-solving—far beyond their initial design goals.

What makes this even more remarkable is that these AI systems are general-purpose large language models (LLMs). Unlike AlphaGo or AlphaFold, which were trained with the explicit purpose of mastering Go or protein folding, these models were not specialised for maths. These models—trained on vast amounts of general text and code rather than specialized mathematical datasets—nevertheless solved the Olympiad problems without any internet access or human intervention.

There have been concerns, of course. Notably, Fields Medalist Terence Tao cautioned that comparing an AI’s performance to a human’s isn’t exactly level playing field—these models, even unaided by internet access or human intervention, demonstrate far greater sustained capability than a single high‑school student solving Olympiad problems. But even he acknowledged that this is a major milestone.

Every major development in generative AI seems to divide people into two camps. One camp is enthusiastic, even ecstatic—predicting that machines are becoming superintelligent and that a technological doomsday may not be far. The other is deeply sceptical, calling it all a hype bubble bound to burst.

Rather than seeing this as a step toward the end of human supremacy, we can see it as a new phase in technological evolution. Humans have never been the strongest or fastest species, but we have always thrived by building tools—from hammers to computers—that enhance our capabilities.

Every wave of innovation, from home appliances to digital computing, has been met with resistance. AI is no different. And that resistance is not necessarily bad—it forces us to be careful, thoughtful, and inclusive in how we deploy these technologies.

But let’s not ignore the potential. If used wisely, AI can help us do things better, faster, and smarter. A machine that can solve IMO-level maths doesn’t just beat a benchmark—it opens new doors to what we might achieve next. @prateekdwv.