Google During the I/O event, Hassabis showed off Astra’s power with a demo video, in which the powerful AI model was scanning the surrounding environment through the camera and answering questions in realtime. With this, questions can be asked by pointing out things and Astra can give information about that object. For example, it is as if you opened the camera and drew a circle or arrow towards any object visible on the screen and asked ‘What is this called?’ and immediately you get a voice from the smartphone in which you describe that object. All necessary information is given about it. This is just an example. Astra can answer many complex and humorous questions in realtime.
Hassabis emphasized during the I/O the importance of AI agents that not only communicate but also execute tasks on behalf of users. Hassabis believes that the future of AI lies not just in fancy technology, but in practical use. He talked about AI agents, who do not just talk but perform many tasks on your behalf. He believes that in the future there will be different types of agents, from simple assistants to more advanced assistants, depending on your needs and your situation.
He says that Astra’s development Google Made possible by improvements to Gemini 1.5 Pro’s language model. Over the past six months, the team worked on making Astra faster and more responsive. This involved not only refining the model but also making sure everything ran smoothly on a larger scale.
Astra unveiled Gemini at this year’s I/O many announcements Is one of. Google also highlighted the 1.5 Flash AI model with Gemini 1.5 Pro, which has been developed to perform tasks like summarization and captioning faster. Additionally, Veo has also been announced, which can generate videos from text prompts. There’s also the Gemini Nano, a model designed to be used locally on a device like your phone.<!–
–>