August 10, 2025
5 min read
Luc Jose Adjinacou
On the chessboard, two visions of AI faced each other. Sam Altman, head of OpenAI, and Elon Musk, founder of xAI, crossed their models in a chess tournament organized by Google. For three days, OpenAI’s o3 and xAI’s Grok 4 competed without any specialized assistance. Much more than a simple exhibition match, the event turned into a revealing moment: behind the final score was the real gap between two artificial intelligences, and two strategies, which came to light.

In brief
- Sam Altman and Elon Musk faced off via AI in a chess tournament organized by Google.
- OpenAI’s o3 and xAI’s Grok 4 played without a chess engine or specialized training.
- OpenAI largely dominated the match, winning the final 4–0.
- The tournament revealed the current limits of generalist AIs facing strict rules.
- A free loss of important pieces, giving an immediate material advantage to OpenAI;
- A failed “poisoned pawn”, with a poorly chosen target that immediately cost the capture of its queen;
- Wasted solid positions in the middle game, with a series of incoherent moves;
- Poor management of the initial advantage in the fourth game, allowing o3 to turn the situation around. Hikaru Nakamura, international grandmaster and event streamer, summed up the difference between the two opponents: “OpenAI didn’t make the mistakes Grok did”. He also praised o3’s spectacular comeback in the last game, where a bad start had predicted a possible victory for Elon Musk’s xAI.
- AI Agents: Are They Broken? Can GPT-5 Fix Them?
- Crypto Market AI News Portal Coming Soon: Stay Tuned for Exclusive Content
- The AI Gig Economy Is Here, and It Pays in Crypto
A one-sided final
During the final held on August 7, OpenAI’s o3 model, which had been criticized a few days earlier in favor of GPT-5 that is already disappointing, inflicted a decisive 4–0 defeat on xAI’s Grok 4. This tournament, called “Kaggle Game Arena AI Chess Exhibition”, prohibited any use of a chess engine or dedicated training, leaving the models to manage with their general knowledge gathered from the Internet. From the first games, the difference was noticeable. Magnus Carlsen, world champion and event commentator, compared the two AIs to “a gifted child who doesn’t know how the pieces move”, estimating their level around 800 ELO, far from competitive standards. Grok’s blunders multiplied throughout the final:The limits of generalist AIs exposed
Beyond the score, the tournament revealed the structural difficulties of generalist AIs when confronted with a strict framework like chess. Many models were disqualified in the preliminary phase after attempting impossible actions: teleporting pieces, resurrecting captured units, illegal pawn moves. Even in the final, understanding of the rules seemed fragile, alternating between brilliant moves and absurd decisions. As Carlsen pointed out, “these AIs know how to count captured pieces, but not how to conclude a winning game”. This observation is not isolated. Earlier this year, international master Levy Rozman organized a similar tournament where language models accumulated illegal moves, even invoking missing pieces. Stockfish, a specialized AI, won the event hands down. These episodes show that, despite promises and claims about their versatility, language models remain far from mastering tasks requiring coherence and procedural rigor. For Elon Musk, this defeat to Sam Altman, his second direct competition loss this year, comes at a bad time, as xAI has just raised $10 billion and seeks to position itself as a credible player in the race for general AI. However, for the whole sector, Google’s exhibition mostly reminds us that current large models excel at natural language processing, much less so in strict application of complex rules. AI may one day rival the best chess players… but that day, it will have also proven capable of reasoning well beyond black and white squares.Originally published at Cointribune on Sun, 10 Aug 2025