
Google Debuts an Updated Gemini 2.5 Pro AI Model Ahead of I/O
An anonymous reader quotes a report from TechCrunch: Google on Tuesday announced the launch of Gemini 2.5 Pro Preview (I/O edition), an updated version of its flagship Gemini 2.5 Pro AI model that the company claims tops a number of widely used benchmarks. Gemini 2.5 Pro Preview (I/O edition) is available via the Gemini API and Google's Vertex AI and AI Studio platforms, and is priced the same as the Gemini 2.5 Pro model it effectively replaces. It's also in Google's Gemini chatbot app for the web and for mobile devices.
The model's release comes ahead of Google's annual I/O developer conference (hence the "I/O edition" designation), where Google is expected to unveil a host of models, as well as AI-powered tools and platforms. [...] According to Google, Gemini 2.5 Pro Preview (I/O edition) has "significantly" improved capabilities for coding and building interactive web apps. The model is also better at tasks like code transformation -- that is, modifying a piece of code to achieve a specific goal -- and code editing, the company says. Google says the Gemini 2.5 Pro Preview (I/O edition) leads the WebDev Arena Leaderboard, a benchmark measuring a model's ability to create aesthetically pleasing and functional web apps. It also achieved a score of 84.8% on VideoMME, a popular benchmark designed to evaluate the video analysis capabilities of multi-modal large language models.
The model's release comes ahead of Google's annual I/O developer conference (hence the "I/O edition" designation), where Google is expected to unveil a host of models, as well as AI-powered tools and platforms. [...] According to Google, Gemini 2.5 Pro Preview (I/O edition) has "significantly" improved capabilities for coding and building interactive web apps. The model is also better at tasks like code transformation -- that is, modifying a piece of code to achieve a specific goal -- and code editing, the company says. Google says the Gemini 2.5 Pro Preview (I/O edition) leads the WebDev Arena Leaderboard, a benchmark measuring a model's ability to create aesthetically pleasing and functional web apps. It also achieved a score of 84.8% on VideoMME, a popular benchmark designed to evaluate the video analysis capabilities of multi-modal large language models.