Google I/O 2025: Gemini 2.5 AI Models Upgraded With Deep Think Mode, Native Audio Output

Google is testing an enhanced reasoning mode called Gemini 2.5 Pro Deep Think that uses new reasoning techniques.

Advertisement
Written by Akash Dutta, Edited by Siddharth Suvarna | Updated: 21 May 2025 06:00 IST
Highlights
  • Gemini 2.5 Pro ranks on top of the WebDev Arena and LMArena leaderboards
  • Google is adding Native Audio Output to Live API with Gemini 2.5 models
  • Gemini API and SDK will now support Anthropic’s MCP

Gemini 2.5 Pro and Flash will now include thought summaries in the Gemini API and Vertex AI

Photo Credit: Google

Google showcased several new features for the Gemini 2.5 family of artificial intelligence (AI) models at the Google I/O 2025 on Tuesday. The Mountain View-based tech giant introduced an enhanced reasoning mode dubbed Deep Think, which is powered by the Gemini 2.5 Pro model. It also unveiled a new, natural and human-like speech called Native Audio Output, which will be available via the Live application programming interface (API). Additionally, the company is also bringing thought summaries and thinking budgets with the latest Gemini models for developers.

Gemini 2.5 Pro Ranks on top of the LMArena Leaderboard

In a blog post, the tech giant detailed all the new capabilities and features that it will be shipping to the Gemini 2.5 AI model series throughout the next few months. Earlier this month, Google released an updated version of the Gemini 2.5 Pro with improved coding capabilities. The updated model also ranked in the top position on the WebDev Arena and LMArena leaderboards.

Now, Google is improving the AI model further with the Deep Think mode. The new reasoning mode allows Gemini 2.5 Pro to consider multiple hypotheses before responding. The company says it uses a different research technique compared to the Thinking versions of the older models.

Advertisement

Based on internal testing, the tech giant shared the reasoning mode's benchmark scores across different parameters. Notably, the Gemini 2.5 Pro Deep Think is claimed to score 49.4 percent on the 2025 UAMO, one of the toughest mathematics benchmark tests. It also scores competitively on LiveCodeBench v6 and MMMU.

Deep Think is currently under testing, and Google says it is conducting safety evaluations and getting input from safety experts. Currently, the reasoning mode is only available to trusted testers via the Gemini API. There is no word on its release date.

Google also announced adding new capabilities to the Gemini 2.5 Flash model, which was released just a month ago. The company said the AI model's key benchmarks for reasoning, multimodality, code and long context have been improved. Additionally, it is also more efficient and uses 20-30 percent fewer tokens, the company claimed.

Advertisement

This new version of Gemini 2.5 Flash is currently available in preview to developers via Google AI Studio. Enterprises can access it via the Vertex AI platform, and individuals can find it in the Gemini app. Notably, the model will be widely available for production in June.

Developers accessing the Live API will now get a new feature with the Gemini 2.5 series of AI models. The company is introducing a preview version of Native Audio Output, which can generate speech in a more expressive and human-like manner. Google said the feature allows users to control the tone, accent, and style of speech generated.

Advertisement

The early version of the capability comes with three features. First is Affective Dialogue, where the AI model can detect emotions in the user's voice and respond accordingly. The second is Proactive Audio, which enables the model to ignore background conversations and only respond when it is spoken to. And finally, Thinking, which lets the speech generation leverage Gemini's thinking capabilities to verbally answer complex queries.

Apart from this, the 2.5 Pro and Flash models in the Gemini API and in Vertex AI will also show thought summaries. These are essentially the model's raw thought process, which were previously only visible in Gemini's reasoning models. Now, Google will show a detailed summary including headers, key details and information about model actions with every response.

Advertisement

In the coming weeks, developers will also be able to use thinking budgets with the Gemini 2.5 Pro. This will allow them to decide how many tokens a model consumes before responding. Finally, Project Mariner's Computer Use agentic function will also be added to the API and in Vertex AI soon.

 

Catch the latest from the Consumer Electronics Show on Gadgets 360, at our CES 2026 hub.

Advertisement

Related Stories

Popular Mobile Brands
  1. Wheel of Fortune India on OTT: When, Where to Watch Akshay Kumar's Game Show
  2. Sirai OTT Release: When, Where to Watch the Tamil Courtroom Drama Online
  3. Ustaad Bhagat Singh OTT Release: When, Where to Watch the Telugu Action Drama
  4. Hackers Steal Hundreds of Gigabytes of Data from European Space Agency
  1. Is Space Sticky? New Study Challenges Standard Dark Energy Theory
  2. Sirai OTT Release: When, Where to Watch the Tamil Courtroom Drama Online
  3. Wheel of Fortune India OTT Release: When, Where to Watch Akshay Kumar-Hosted Global Game Show
  4. NASA Confirms Expedition 74 Will Continue ISS Work After Crew-11 Exit
  5. European Space Agency Hit by Cyberattacks, Hundreds of Gigabytes of Data Stolen by Hackers
  6. Ustaad Bhagat Singh OTT Release: When, Where to Watch Harish Shankar's Telugu Action Drama Film
  7. Bha Bha Ba is Now Streaming: All You Need to Know About This Malayalam Comedy Thriller Film
  8. World’s Biggest Alien Search Enters Final Stage With 100 Mystery Signals
  9. NASA Pulls Out Artemis II Rocket to Launch Pad Ahead of Historic Moon Mission
  10. Shambhala OTT Release: When, Where to Watch the Telugu Supernatural Horror Film
Gadgets 360 is available in
Download Our Apps
Available in Hindi
© Copyright Red Pixels Ventures Limited 2026. All rights reserved.