It actually gets better! Just checked some other benchmarks, and on the LM Arena[0] the new Gemini 2.5 Pro is winning every category! Not only that, but it's also winning in all the subcategories as well. This is just nuts. To spell that out it means that Gemini 2.5 Pro is the best at:
- Vision
- WebDev
- Search
- Coding
- Math
- Creative Writing
- Instruction following
- Long queries
- Multi-turn
This feels like a GPT-4 moment. Google is moving so fast at this point that unless OpenAI is cooking up an absolutely insane upgrade with the o4 model, it's going to be outdated before they can even finish their training run. Google already has a reasoning model that is all around smarter, cheaper, and more multi-modal.
My money is on Google for winning the AI wars. They're winning on every single front. They have the best frontier model, the best cheap models, products that are already insanely popular they can integrate their models into, their own in-house chips and infrastructure, and boatloads of their own (not investors) cash because they're already a hugely profitable company.
[0]: Yeah I know there been some controversies but it's still a solid benchmark