top of page

Gemini Live Thinking Mode Found in Google App 17.2

Gemini Live Thinking Mode Found in Google App 17.2

Code hidden within the latest Google App 17.2 update reveals that Google is finally addressing the biggest complaints surrounding its voice AI. The teardown points to the imminent launch of a dedicated Gemini Live Thinking Mode, alongside a suite of experimental features housed under a new "Labs" setting.

For users who have spent the last few months wrestling with the current version of Gemini Live—powered by the faster but often shallower Gemini 2.5 Flash—this is the pivot point. The leak confirms Google is moving toward a model that pauses to process complex queries rather than spitting out the first hallmark of a hallucination.

Here is the breakdown of what was found, why the community is demanding these specific changes, and how the new Labs feature works.

Why Users Need a Slower, Smarter AI

Why Users Need a Slower, Smarter AI

Before dissecting the code, we need to address the user experience gap that Gemini Live Thinking Mode intends to fill. The current iteration of Gemini Live is fast, but that speed has come at a cost.

Feedback from early adopters has been consistent: the current model often feels "brain dead" during nuanced conversations. The system prioritizes latency over logic, leading to surface-level answers that fall apart under scrutiny. But the frustration goes deeper than just bad answers.

The Interruption Problem

The most vocal complaint involves the aggressive voice activity detection (VAD). Users attempting to have a fluid conversation find themselves cut off the moment they pause to breathe or think.

The system assumes silence equals task completion. It triggers a response before the user has finished their thought. This creates a disjointed loop where users feel rushed. The demand is clear: people want a "manual send" button or a toggle that forces the AI to wait. While the leaks in version 17.2 don't explicitly confirm a manual button, the introduction of a Thinking Mode implies a fundamental shift in how the AI manages time and pacing.

Exploring the New "Labs" in Google App 17.2

Exploring the New "Labs" in Google App 17.2

Google appears to be adopting a strategy similar to other tech giants by burying these advanced, potentially unstable features behind a "Labs" toggle. This allows them to push Gemini Live Thinking Mode to power users without disrupting the mainstream experience.

According to the file teardown, the Labs section will host two primary categories:

  1. Live Thinking Mode

  2. Live Experimental Features

This segmentation suggests that Google acknowledges the trade-off between speed and intelligence. They aren't replacing the fast, chatty version of Gemini; they are offering a separate lane for heavy lifting.

How Gemini Live Thinking Mode Works

The code strings describe Gemini Live Thinking Mode as a state where the AI "takes time to think" to provide "more detailed responses."

This is likely the mobile implementation of the reasoning capabilities found in the Gemini 3 Pro model, which launched back in November 2025. Until now, mobile users were stuck with the lightweight Flash models to save on battery and reduce latency.

By enabling this mode, you are essentially telling the app that you don't mind a 5-to-10-second delay if it means the agent can perform chain-of-thought processing. For coding questions, complex travel planning, or philosophical debates, this added latency is a feature, not a bug.

However, some users are skeptical. There is a fear that "Thinking Mode" might simply act as a buffer for additional safety layers and PR-friendly filtering, rather than genuine cognitive processing. If the delay only results in a more sanitized, "corporate safe" response, adoption will plummet. The value proposition here relies entirely on the quality of the insight gained during that pause.

Project Astra and Vision Capabilities

Perhaps the most exciting addition found alongside Gemini Live Thinking Mode is the reference to vision. The experimental flags include functionality for "responding when it sees something."

This is the long-awaited arrival of Project Astra’s real-time multimodal capabilities. Instead of taking a static photo and asking for analysis, this feature implies a continuous video stream where Gemini watches what the camera sees and comments in real-time.

For hardware troubleshooting or navigating a new city, this transforms the phone from a passive screen into an active observer. It also ties into the "Agent" concept, where the AI isn't just a chatbot but an interface that understands your physical environment.

Multimodal Memory and Context

Multimodal Memory and Context

The "Experimental Features" section also references "multimodal memory." This is distinct from standard history. Standard history is a log of text. Multimodal memory means the AI remembers images you showed it last week, a sound clip you analyzed yesterday, and the context of a voice conversation you had an hour ago.

Currently, context windows on mobile are limited. You often have to re-upload images or remind the bot of previous constraints. If Gemini Live Thinking Mode can access a persistent, cross-media memory bank, it solves the continuity problem that plagues long-term AI interactions.

The Problem of Noise and Reliability

The Reddit threads dissecting this leak highlighted another critical area: noise cancellation. The code indicates "better noise handling" is part of this update.

For a voice-first feature, Gemini Live has historically struggled in environments that aren't a quiet home office. Using it in a car or a busy café often resulted in the AI transcribing background chatter as user input. Improved noise gating is less "sexy" than a Thinking Mode, but it is technically required for the product to be usable in the real world.

When Will We See Gemini Live Thinking Mode?

The presence of these strings in Google App 17.2 suggests the code is already on devices. The server-side switch hasn't been flipped yet.

Historically, Google releases these features in staggered rollouts, often prioritizing US English users and AI Premium subscribers. The community sentiment is strongly against this regional locking. Users in Europe and Asia are already voicing frustration that features announced in 2025 are still absent in their regions in early 2026.

Developer Access

There is also a push for these capabilities to open up in AI Studio. Developers want to build applications that utilize the "Thinking" endpoint without the Gemini app wrapper. If the architecture supports it, we might see third-party apps utilizing Gemini Live Thinking Mode for specialized tasks like medical diagnosis assistance or legal document review on mobile.

Assessing the Upgrade Path

Assessing the Upgrade Path

This update represents a maturity milestone for mobile AI. We are moving past the "wow factor" of instant voice generation and entering the utility phase.

Users don't just want an AI that talks back instantly; they want an AI that understands when to shut up, when to listen, and when to think. The success of Gemini Live Thinking Mode won't depend on how smart the model is—we know Gemini 3 is capable. It will depend on the implementation.

If Google can solve the interruption issue and prove that the "thinking" time produces tangible value, this will define the standard for mobile agents in 2026. If it remains a rude, interruption-prone chatbot that just takes longer to answer, the toggle will likely stay off.

FAQ

What is the difference between Gemini Live Thinking Mode and the standard mode?

Standard mode prioritizes speed, likely using the Gemini 2.5 Flash model for instant replies. Thinking Mode allows the AI to pause and process complex logic, likely leveraging Gemini 3 Pro for deeper reasoning and more detailed answers.

Will Gemini Live Thinking Mode stop the AI from interrupting me?

While not explicitly confirmed as a "do not interrupt" setting, the move to a thinking-based architecture suggests a change in pacing. However, users are specifically requesting a manual "send" button to fully prevent premature interruptions.

How do I enable the new Gemini Live features?

These features will appear under a new "Labs" or "Opt-in" section within the Google App settings. You will need to be on version 17.2 or higher, and the features must be activated via a server-side update from Google.

Does the new update include vision capabilities?

Yes. The code references "responding when it sees something," which aligns with Project Astra's technology. This allows the AI to analyze real-time video feeds from your camera rather than just static images.

Is Gemini Live Thinking Mode free to use?

The leak does not specify pricing, but advanced reasoning features are often bundled with the Google One AI Premium subscription. Basic functionality may be free, but the full "Thinking" capability could be gated.

Get started for free

A local first AI Assistant w/ Personal Knowledge Management

For better AI experience,

remio only supports Windows 10+ (x64) and M-Chip Macs currently.

​Add Search Bar in Your Brain

Just Ask remio

Remember Everything

Organize Nothing

bottom of page