Google’s Gemini update that can tell you live what it sees through your camera is now rolling out

Google’s Gemini update that can tell you live what it sees through your camera is now rolling out

[ad_1]

Google has started publishing an innovative AI video in Gemini in an actual time, allowing the platform to explain the visible input from the user device or smartphone camera and provide contextual answers. The confirmation of this publishing comes almost a year after the initial shows of "Project Asra", which is the basic technology that works with these capabilities During Google I/O 2024, reports from The Reddit user initially showed the appearance of this feature on the Xiaomi phone. This user later shared a video that showcases the new Gemini's ability to analyze the screen content. This screen reading function is one of the two main features that Google announced in early March to Gueeni Advanced with the Google One Ai Premium plan later in the same month.
The second feature that has been announced by the smartphone camera is used to process a direct video summary, allowing Gemini to understand questions about the user's circumference and answer them.
Video mini image
This update follows the last introduction to Gemini's The fabric feature, which helps users in writing and coding tasks, and add podcast summarization tools. This update besides Gemini Live Video really shows the current Google progress when it comes to AI's assistant technology compared to Apple, Samsung and even Amazon.
Video mini image

Of course, this does not seem to include every aspect of Project ASTRA as it was tried last year and can be seen in the video above. The full illustration of the assistant memory preserved for the elements you saw across your camera moments, and later told you where you saw that - or even gives you the option to draw on the elements to put more information on the specified area.

However, for me, it seems that we are not far from the future of the artificial intelligence assistant that was shown last year. The rate at which Gueini is progressing in the very promising feature of the collection, and I am very excited to find out how Google will work on it at the Google I/O conference for this year.


[ad_2]
Download

Name is the most famous version in the series of publisher
Publisher
Genre News & Magazines
Version
Update March 24, 2025
Get it On Google Play
Rate this post
Download
Rate this post


Google has started publishing an innovative AI video in Gemini in an actual time, allowing the platform to explain the visible input from the user device or smartphone camera and provide contextual answers. The confirmation of this publishing comes almost a year after the initial shows of “Project Asra”, which is the basic technology that works with these capabilities During Google I/O 2024, reports from The Reddit user initially showed the appearance of this feature on the Xiaomi phone. This user later shared a video that showcases the new Gemini’s ability to analyze the screen content. This screen reading function is one of the two main features that Google announced in early March to Gueeni Advanced with the Google One Ai Premium plan later in the same month.

The second feature that has been announced by the smartphone camera is used to process a direct video summary, allowing Gemini to understand questions about the user’s circumference and answer them.
Video mini image

This update follows the last introduction to Gemini’s The fabric feature, which helps users in writing and coding tasks, and add podcast summarization tools. This update besides Gemini Live Video really shows the current Google progress when it comes to AI’s assistant technology compared to Apple, Samsung and even Amazon.
Video mini image

Of course, this does not seem to include every aspect of Project ASTRA as it was tried last year and can be seen in the video above. The full illustration of the assistant memory preserved for the elements you saw across your camera moments, and later told you where you saw that – or even gives you the option to draw on the elements to put more information on the specified area.

However, for me, it seems that we are not far from the future of the artificial intelligence assistant that was shown last year. The rate at which Gueini is progressing in the very promising feature of the collection, and I am very excited to find out how Google will work on it at the Google I/O conference for this year.



Download

 
Report

You are now ready to download for free. Here are some notes:

  • Please check our installation guide.
  • To check the CPU and GPU of Android device, please use CPU-Z app
Rate this post

Leave a Comment

Your email address will not be published. Required fields are marked *