Mountain View, California – Google’s ambitious Project Astra, focused on developing AI applications and “agents” for real-time, multimodal understanding, will not release any technology until at least next year.
During Google’s Q3 earnings call on Tuesday, CEO Sundar Pichai announced that the company is in the process of creating experiences where AI can visually perceive and comprehend its surroundings. Project Astra exemplifies this vision, with Google aiming to make such experiences available as early as 2025.
Unveiled at Google’s I/O developer conference in May 2024, Project Astra involves a range of technologies, including smartphone apps that can identify objects and provide related information, as well as AI assistants capable of executing tasks on behalf of users.
In a demonstration video shown at I/O, Google showcased a Project Astra prototype answering queries about objects visible through a smartphone camera, such as identifying a neighborhood or a part of a bicycle. However, recent reports suggest that plans to launch a consumer-oriented agent application by December may be delayed unless it is separate from Project Astra.
Anthropic, a company that recently developed a powerful generative AI model capable of controlling applications and web browsers on a PC, serves as an example of the challenges in building AI agents. Despite its advancements, Anthropic has encountered difficulties performing basic tasks efficiently.
The development of Project Astra reflects Google’s commitment to advancing AI technology to create more intuitive and interactive experiences for users. As the company navigates the complexities of AI development, the timeline for releasing these innovative experiences continues to evolve. Can Google achieve its vision for a future where AI seamlessly interacts with the world around us? Only time will tell.