Today : Apr 21, 2025
Technology
08 April 2025

Google Expands AI Mode With Multimodal Search Features

The new update allows users to upload images and ask complex questions for richer search results.

Google LLC is making significant strides in enhancing its search capabilities with the introduction of multimodal features in its AI Mode. This update allows users to upload images and ask complex questions about them, marking a notable evolution in how users interact with the platform. Initially rolled out to Google One AI Premium subscribers in March 2025, the feature is now expanding to millions of users enrolled in Google Labs in the U.S. This expansion comes on the heels of overwhelmingly positive feedback from early adopters, who praised the feature for its intuitive design and fast response times.

AI Mode, which builds on Google’s extensive work in visual search, is designed to handle more nuanced queries than traditional search options. Users have reported that queries made in AI Mode are, on average, twice as long as those typically entered into standard Google Search. This suggests that users are increasingly relying on AI Mode for more exploratory tasks, such as product comparisons, how-to guides, and travel planning.

The key enhancement in AI Mode is its new multimodal understanding, powered by Google Lens and a custom version of the Gemini AI model. With this capability, users can take a photo or upload an image, ask a question regarding what they see, and receive comprehensive answers enriched with links for further exploration. For instance, if a user snaps a photo of their bookshelf and queries, "If I enjoyed these, what are some similar books I might like?" AI Mode can identify each book in the image and provide tailored recommendations based on the titles recognized.

According to Google, the AI Mode can grasp the entire context of an image, including how objects relate to one another, their materials, colors, shapes, and arrangements. This depth of understanding is achieved through a technique known as "query fan-out," where the system runs multiple searches about the image and the objects within it. As a result, users receive more nuanced and contextually relevant responses than they would from traditional search methods.

Robby Stein, Google Search Vice President of Product, emphasized the importance of this update, stating, "AI Mode builds on years of work in the area of visual search and takes it a step further." This sentiment reflects the company’s commitment to enhancing user experience and keeping pace with the growing demand for advanced AI capabilities in search.

Google’s decision to expand AI Mode comes as a direct response to the rising popularity of generative AI search applications, such as Perplexity and OpenAI’s ChatGPT Search, which have gained traction for their ability to handle complex queries. By integrating multimodal capabilities into AI Mode, Google aims to provide users with a more comprehensive and interactive search experience.

The rollout of AI Mode is being closely monitored, as Google continues to collect feedback from users in Labs. This ongoing testing phase will allow the company to refine the feature further before a more general release. Users interested in exploring AI Mode can sign up through the Google app available on both Android and iOS devices.

As the digital landscape evolves, Google’s AI Mode represents a significant step forward in search technology, merging visual and textual inquiries to create a more engaging user experience. The ability to ask complex questions about images could redefine how users approach information gathering, making it easier to find relevant content quickly and efficiently.

In summary, Google’s introduction of multimodal capabilities in AI Mode is a game-changer for search technology. By allowing users to interact with images and receive detailed, context-rich responses, Google is setting a new standard for how search engines can facilitate deeper understanding and exploration. As this feature continues to develop, it will be interesting to see how users adapt to and leverage these advanced capabilities in their everyday search activities.