Home
/
Latest news
/
Event highlights
/

Ai models battle in foggy maze to reach the exit first

AI Bout | Claude's Bots Dominate, Others Stumble in Foggy Maze Challenge

By

Raj Patel

Mar 22, 2026, 04:33 PM

Edited By

Chloe Zhao

2 minutes needed to read

Five AI bots compete in a fog-filled maze with teleportals trying to find the exit.
popular

A recent tournament pitting five frontier AI models against each other in a coding competition has sparked interest in performance metrics among various models. The challenge involved coding bots to navigate a foggy maze filled with teleportals, walls, and a singular exit. Surprisingly, both Gemini and ChatGPT fell short, failing to reach past round eight. Meanwhile, Claude's and Grok's bots have surprisingly managed to reach round 93.

The Challenge Explained

In this unique tournament, bots had to navigate a maze they couldn't see, limited to a 5Γ—5 field of fog around their current position. The maze included teleportals that could warp them across its layout, along with walls that blocked pathways. The aim? To reach the exit in the fewest steps. A significant challenge: exceed 500 steps, and you're out.

"Even if an LLM could, it would take time to build the infrastructure"

Performance Disparities

Observers have noted an evident disparity between the performances of various AI models. Claude and Grok pushed through successive rounds, leaving behind their counterparts. The early elimination of Gemini and ChatGPT raises questions about their coding strategies and infrastructure capabilities.

Notable Feedback from the Community

Comments from the user boards reflect curiosity about the factors influencing AI performance.

  • Many appear surprised by how far ahead Claude and Grok are.

  • One user queried, "How many times are you running this? Is each run using a seed for the maze?"

  • Speculation about maze complexity affecting outcomes has emerged, particularly noticing the dramatic differences in advance stages reached by different models.

Sentiment: A mix of intrigue and concern is palpable, especially regarding the varying AI performances.

Key Insights

  • β–³ Claude and Grok's bots demonstrate strong navigation skills, reaching round 93.

  • β–½ Gemini and ChatGPT struggled early, not surpassing round eight.

  • πŸ’¬ "Surprising differences in model performances have caught attention" - Community comment.

As this competition unfolds, it highlights significant advancements in AI navigation tasks. Will Claude and Grok maintain their lead? Stay tuned.

Future AI Competitions on the Horizon

There's a strong chance that this competition will lead to further tournaments focusing on AI navigation tasks. Expect more rigorous challenges to emerge as developers seek to refine their algorithms. Experts estimate around 70% likelihood that teams will learn from this round, adjusting strategies to improve performance in future events. Claude and Grok's current success may inspire others to enhance their methodologies, potentially creating a more competitive landscape in upcoming challenges.

A Unique Twist on Historical Competitions

This situation mirrors the historic rivalry between competing aviation companies in the early 20th century, where innovations led to surprising advances in flight technology. Much like the AI contest, early aviation featured pioneers like the Wright brothers and Glenn Curtiss, each pushing boundaries in unexpected ways, but with vastly different results. As each competitor raced for supremacy, sudden shifts in leadership happened based on unpredicted improvements. The ongoing AI navigation contest may very well lead to similar breakthroughs that shift the tide in this digital arena.