A new AI model for the agent era

A note from Google and Alphabet CEO Sundar Pichai:
Information is at the core of human development. That’s why we’ve focused for more than 26 years on our mission to organize the world’s information and make it accessible and accessible. And this is why we continue to push the boundaries of AI to organize that information in each input and make it accessible through any output, so that it can be useful for you.
That’s our view when we introduced Gemini 1.0 last December. The first model built to be natively multimodal, Gemini 1.0 and 1.5 bring major improvements with multimodality and high context to understand information in text, video, images, audio and code, and process many more on this.
Today millions of developers are building with Gemini. And it helps us reimagine all of our products — including all 7 of them with 2 billion users — and to create new ones. NotebookLM a great example of what multimodality and high context can do for people, and why so many love it.
Over the past year, we’ve invested in creating more agent models, meaning they can better understand the world around you, think several steps ahead, and act on your behalf, with your guidance.
Today we are excited to launch our next season of models created for this new agent era: introducing Gemini 2.0, our most capable model yet. With new advances in multimodality – such as native image and audio output – and using native tools, it will enable us to build new AI agents that will bring us closer to our vision. of a universal assistant.
We’ve got 2.0 into the hands of developers and trusted testers today. And we’re working quickly to embed it in our products, leading Gemini and Search. Starting today our Gemini 2.0 Flash experimental model will be available to all Gemini users. We also launched a new feature called Deep Researchwhich uses advanced reasoning and high context capabilities to act as a research assistant, exploring complex topics and compiling reports for you. It is available in Gemini Advanced now.
No product has been more transformed by AI than Search. Our AI Insights have now reached 1 billion people, enabling them to ask entirely new kinds of questions — quickly becoming one of our most popular Search features ever. As a next step, we bring the advanced reasoning capabilities of Gemini 2.0 to AI Overviews to address more complex topics and multi-step questions, including advanced math equations, multimodal questions and coding. We started limited testing this week and will roll it out more widely early next year. And we will continue to bring AI Generals to more countries and languages next year.
2.0 advances are backed by decades-long investment in our diverse full-stack approach to AI innovation. It’s built on custom hardware like Trillium, our sixth-generation TPU. TPUs run 100% of Gemini 2.0 training and inference, and now Trillium usually available to customers so they can build it too.
If Gemini 1.0 is about organizing and understanding information, Gemini 2.0 is about making it more useful. I can’t wait to see what the next season brings.
– Follower
https://storage.googleapis.com/gweb-uniblog-publish-prod/images/blog_gemini_hero_thumbnail.width-1300.png
2024-12-11 15:33:54