Google Redefines Search: New AI “Multisearch” Blends Camera, Voice, and Text for Hyper-Personalized Discovery
Google Redefines Search: New AI “Multisearch” Blends Camera, Voice, and Text for Hyper-Personalized Discovery
MOUNTAIN VIEW, CALIFORNIA – Google has unveiled its most significant evolution of search in years, launching a new AI-powered feature globally called “Multisearch.” The groundbreaking technology allows users to combine images, voice commands, and text into a single, seamless query, fundamentally changing how people can find information and interact with the world around them.
Announced at a special AI event, the feature is now rolling out to the Google app on both Android and iOS. With Multisearch, a user can, for example, point their phone’s camera at a piece of furniture they like, and then verbally ask, “Where can I find this in a different color?” or type “and under ₹15,000.” Google’s advanced AI models, including its Multitask Unified Model (MUM), will process all three inputs—the image, the voice query, and the text—to deliver a highly contextual and accurate set of results.
The technology extends beyond simple shopping. During the keynote, Google demonstrated a user taking a photo of a complex car engine part and asking, “What is this, and how do I replace it?” The AI correctly identified the part, provided its name, and surfaced a step-by-step video tutorial for the replacement, all from one fluid query.
“We are moving beyond a text box to a more intuitive, natural way of exploring and understanding,” said Priya Sharma, Google’s Vice President of Search, during the presentation. “The world is not just text; it’s images, sounds, and the questions we have about them. Multisearch is designed to bridge that gap, making search a true extension of our curiosity.”
Industry analysts are calling this a pivotal moment in the “AI assistant wars,” positioning Google to compete more directly with visually-driven discovery platforms like Pinterest and TikTok. By integrating its powerful computer vision (from Google Lens) with its best-in-class natural language processing, Google is creating a tool that doesn’t just answer questions, but helps users solve real-world problems. The company stated that Multisearch is the first step in a broader initiative to make its services more “ambient and multi-modal,” hinting at future integrations into smart glasses and other wearable technology.