alt- Gemini Live enables camera lens interaction
Google, recently at the Mobile World Congress 2025, revealed that the long-awaited feature of Gemini AI which had been vaguely touted for an entire year would soon be launching. The being known as Gemini Live which has been conversing would soon be gaining the ability to view live video and share screens, whose capabilities were shown earlier under Project Astra.
With the addition of video functions to Gemini, users will not be required to explain visually what things are; they can just show it to the AI. Currently, Google’s multimodal AI can do text, images, and all kinds of documents, but video input is unpredictable; at times, it makes summaries from YouTube, and at other times it fails for no good reason.
In March, Android’s Gemini app will be getting an exciting upgrade with respect to video. Users can activate their cameras in order to stream video directly to Gemini Live, or even share their screens in real-time for them to be able to ask Gemini questions about what it sees.
Keeping up with Google’s plethora of artificial intelligence initiatives is an undeniable challenge, though the 2024 Google I/O program lumped together as this event was a tribute to all things Gemini AI. The demonstration of Astra had a strong impact; it showed the world a more straightforward interaction with AI.
The original video, which can be seen down below, illustrated how Gemini Live will answer inquiries in real-time as a user moves their phone around a room. It provided insights into code displayed on a computer screen, helped explain some of the ways speakers operate, and interpreted a network diagram from a whiteboard. It even recalled where early in the video the user had left their glasses.However, it is unknown if the version of this technology that reaches Gemini will have the same features demonstrated during Google’s tightly managed I/O show.