Google launches Search Live globally with real-time camera and voice search
Google is expanding Search Live globally to users in more than 200 countries, enabling real-time voice and camera search through the Google app and Lens. The feature, powered by Gemini 3.1 Flash Live—a new multilingual audio and video model—allows users to point their phone camera at objects and ask questions with instant spoken responses.
Google Expands Search Live to 200+ Countries
Google is rolling out its Search Live feature globally, making real-time voice and camera-based search available to users in more than 200 countries. The feature is now accessible through the Google app on both Android and iOS, as well as through Google Lens.
How Search Live Works
Search Live enables two primary interaction modes:
Voice Search: Users can ask questions aloud and receive spoken answers paired with relevant web links.
Camera Search: With the camera active, users can point their phone at physical objects and ask contextual questions. Google cites assembling furniture as an example use case—a user could point their camera at a shelf and ask for assembly instructions.
Powered by Gemini 3.1 Flash Live
The feature runs on Google's new Gemini 3.1 Flash Live model, a multilingual audio and voice model designed to enable more natural conversational interactions. The model processes real-time audio and visual input, eliminating the need for users to photograph objects and wait for processing—searches occur while the camera is actively pointed at the subject.
Specific details about the model's capabilities, latency, or technical specifications have not yet been disclosed by Google.
Integration and Accessibility
Search Live is integrated into the AI mode within the Google app, positioning it as a core search interface rather than a standalone feature. The global rollout marks a significant expansion from previous limited availability, though Google has not specified whether all regions receive identical functionality or if certain features vary by location.
What This Means
Google is attempting to shift mobile search from text-based queries to multimodal interactions where users can leverage their device's camera and microphone as primary input methods. This represents a strategic response to AI-powered search competition and aligns with Google's broader efforts to integrate Gemini capabilities across its product suite. The real-time processing requirement suggests significant advances in inference speed—Search Live must provide responses while the camera remains pointed at objects, not after capturing static images. The feature's availability across 200+ countries indicates Google's confidence in Gemini 3.1 Flash Live's multilingual capabilities, though performance variations across languages and regions remain unknown.
Related Articles
Google's Gemini API Agent Skill boosts coding task success from 28% to 97%
Google has released an Agent Skill for the Gemini API that provides language models with current information about their own APIs, SDKs, and best practices. Testing across 117 coding tasks showed Gemini 3.1 Pro's success rate jumped from 28.2% to 96.6%, though older 2.5-series models showed minimal improvement.
Google lets users import chat histories and personal data into Gemini from competing chatbots
Google announced 'switching tools' that allow users to transfer chat histories and personal information directly into Gemini from competing chatbots like ChatGPT and Claude. The feature uses a prompt-based system for importing memories and accepts zip file uploads for chat logs. This move targets Gemini's lagging consumer adoption, which sits at 750 million monthly active users compared to ChatGPT's 900 million weekly active users.
Google adds Import Memory and Chat History tools to Gemini to reduce AI switching friction
Google is launching "Import Memory" and "Import Chat History" features in Gemini that let users transfer their preferences and past conversations from other AI chatbots. The tools use copy-paste prompts and .zip file uploads to help users avoid retraining a new AI model on their preferences.
Apple gains full Gemini access, uses distillation to build lightweight on-device models
Apple has secured full access to Google's Gemini models within its data centers and is using knowledge distillation to generate training data for smaller, on-device AI models. The approach allows Apple to create lightweight versions that replicate Gemini's reasoning patterns while running directly on Apple devices, requiring significantly less processing power.
Comments
Loading...