A recent viral post on X claimed that Google's Gemini model outperformed Anthropic's Claude in the original Pokémon games. However, it was revealed that Gemini had a significant advantage due to a custom minimap designed by the developer, allowing it to navigate more effectively. This incident highlights the challenges in AI benchmarking, as custom implementations can skew results. Anthropic's Claude model also demonstrated varying scores on different benchmarks, illustrating the complexities of model comparisons. As AI benchmarks evolve, their reliability continues to be questioned.