On Tuesday, Google unveiled Gemini 2.5, a new family of AI reasoning models that pauses to “think” before answering a question.
To kick off the new family of models, Google is launching Gemini 2.5 Pro Experimental, a multimodal, reasoning AI model that the company claims is its most intelligent model yet. This model will be available on Tuesday in the company’s developer platform, Google AI Studio, as well as in the Gemini app for subscribers to the company’s $20-a-month AI plan, Gemini Advanced.
Moving forward, Google says all of its new AI models will have reasoning capabilities baked in.
Since OpenAI launched the first AI reasoning model in September 2024, o1, the tech industry has raced to match or exceed that model’s capabilities with their own. Today, Anthropic, DeepSeek, Google, and xAI all have AI reasoning models, which use extra computing power and time to fact-check and reason through problems before delivering an answer.
Reasoning techniques have helped AI models achieve new heights in math and coding tasks. Many in the tech world believe reasoning models will be a key component of AI agents, autonomous systems that can perform tasks largely san human intervention. However, these models are also more expensive.
Google has experimented with AI reasoning models before, previously releasing a “thinking” version of Gemini in December. But Gemini 2.5 represents the company’s most serious attempt yet at besting OpenAI’s o series of models.
Google claims that Gemini 2.5 Pro outperforms its previous frontier AI models, and some of the leading competing AI models, on several benchmarks. Specifically, Google says it designed Gemini 2.5 to excel at creating visually compelling web apps and agentic coding applications.
On an evaluation measuring code editing, called Aider Polyglot, Google says Gemini 2.5 Pro scores 68.6%, outperforming top AI models from OpenAI, Anthropic, and Chinese AI lab DeepSeek.
However, on another test measuring software dev abilities, SWE-bench Verified, Gemini 2.5 Pro scores 63.8%, outperforming OpenAI’s o3-mini and DeepSeek’s R1, but underperforming Anthropic’s Claude 3.7 Sonnet, which scored 70.3%.
On Humanity’s Last Exam, a multimodal test consisting of thousands of crowdsourced questions relating to mathematics, humanities, and the natural sciences, Google says Gemini 2.5 Pro scores 18.8%, performing better than most rival flagship models.
To start, Google says Gemini 2.5 Pro is shipping with a 1 million token context window, which means the AI model can take in roughly 750,000 words in a single go. That’s longer than the entire “Lord of The Rings” book series. And soon, Gemini 2.5 Pro will support double the input length (2 million tokens).
Google didn’t publish API pricing for Gemini 2.5 Pro. The company says it’ll share more in the coming weeks.
You Might Also Like
WhatsApp will now charge AI chatbots to operate in Italy
Meta announced Wednesday that it will charge developers for running chatbots on WhatsApp in regions where regulators are forcing the...
TechCrunch Mobility: RIP, Tesla Autopilot, and the NTSB investigates Waymo
Welcome back to TechCrunch Mobility — your central hub for news and insights on the future of transportation. To get...
Ring is adding a new content verification feature to videos
It’s going to be more difficult to fake or manipulate Ring camera video footage going forward. On Thursday, the Amazon-owned...
Here are the 55 US AI startups that raised $100M or more in 2025
The AI industry entered 2025 with strong momentum. There were 49 startups that raised funding rounds worth $100 million or more in...








