Google is hitting hard at MWC 2025 with a major announcement for its artificial intelligence: Gemini Live will gain vision. Thanks to this update, users will be able to interact with the chatbot by showing it their environment via their smartphone camera. An advancement that opens the door to new practical applications.
A small revolution for Gemini Live: the integration of vision
By the end of March, Gemini Live will receive a major update that will bring the ability to “see.”
Concretely, Pixel and Galaxy smartphone users subscribed to Gemini Advanced ($24/month) will be able to activate their phone’s camera to show the AI what’s around them.
What will this be used for?
- • Get advice on organizing a space (choice of colors, storage, etc.)
- • Identify objects or foods to suggest recipe ideas
- • Get help assembling a piece of furniture or solving a math problem
This innovation is based on Project Astra, unveiled by Google at Google I/O 2024, which aims to transform AI into an intelligent assistant that responds to real-time vision.
Use cases also for digital professionals?
Here are some examples of concrete use cases of such innovation for professionals:
- • E-commerce and retail : rapid identification of products thanks to vision, assistance in creating more optimized merchandising or advice on the layout of a landing page from a conversion point of view.
- • Marketing and content creation : real-time analysis of visuals to suggest improvements to layout or colors.
- • Education and training : interactive explanation of complex concepts requiring knowledge of a visual context.
- • International working made easy : instant translations and seamless interactions between multilingual collaborators.
Google’s ambition doesn’t stop there.
Ultimately, Project Astra could become Gemini’s primary voice assistant, with possible integration into smart glasses, paving the way for more natural and intuitive interaction.
Gemini 2.0 Flash: a faster and more powerful model
In addition to the vision, Google will roll out Gemini 2.0 Flash, a new multimodal language model (LLM), designed to improve:
- • Language comprehension
- • Multilingual interactions : more than 45 languages supported
- • Conversation fluidity : ability to switch between languages without changing settings
- • Better recognition of accents and dialects
Why is it important?
- • This makes it easier to access for international users.
- • Interactions become more natural, even with accent variations.
- • Gemini fits best into a multilingual professional or personal setting.
When will there be wider adoption?
For now, these new features remain reserved for paying subscribers and Pixel and Galaxy users.
But this technological advancement should soon be available to all Gemini users, even on iOS.
Towards an ever more interactive and useful AI
With this update, Google positions Gemini as an ultra-intelligent assistant, capable of interpreting text, sound and images.
An advance that brings AI a little closer to the ideal personal assistant model, ready to respond to our needs in real time and with greater precision.