Google introduced Gemini 2.0 on Wednesday and marked a significant step forward in artificial intelligence. The new model, is described by CEO Sundar Pichai as the start of “a new agentic era.” And aims to transform how AI understands and interacts with the world. Gemini 2.0 enhances the ability of AI to process information, think through multiple steps. And take action based on its findings.
Pichai emphasized that the model will make information far more useful, bringing Google closer to its goal of creating a universal assistant. The launch sparked a 4% increase in Google‘s stock, what follows a previous rise of 3.5% due to the company’s groundbreaking quantum chip announcement.
AI Agents and the Race for Dominance
AI “agents” are the latest trend in Silicon Valley. These digital assistants are designed to assess environments, make decisions. Also, take action to achieve specific goals. Despite the high costs associated with developing such technology, major tech companies are rushing to introduce more powerful AI models. These, fueled by the success of ChatGPT in 2022.
Gemini 2.0 is currently available to developers and select testers. It has plans to integrate it into Google‘s core products, whose includes Search and the Gemini platform.
Google: Advanced Technology and Future Prospects
Gemini 2.0 is powered by Google‘s proprietary sixth-generation Tensor Processing Units (TPUs), known as Trillium. Google emphasized that Trillium processors are used exclusively for training and running Gemini 2.0, setting the company apart from competitor. For example, Nvidia, which currently dominates AI chip manufacturing.
Millions of developers are already building applications using Gemini, which has been incorporated into seven Google products, each serving over two billion users. The first release from the 2.0 series, called Flash, will improve performance by handling a range of input types—text, images, video, and audio. Plus, providie outputs such as generated images and speech.
A chat-only version of Flash is already available to users, while testers have access to a multimodal version that can interpret both images and surroundings. Google is also experimenting with an AI product that can interact with software, websites, and other tools like a human user.
This feature is similar to offerings from competitors like OpenAI and Anthropic. Additionally, Google teased a new version of Project Astra. Which smartphone assistant that responds not only to voice commands but also to images. This could compete with Apple’s Siri and further enhance the user experience.
For more political, business, technological and cultural information visit all of our Mundo Ejecutivo platforms.