
Google Gemini Set to Revolutionize AI Studio with Real-Time Upgrades and Enhanced Deep Research Features
Get ready for a major shake-up in the world of AI! Google is poised to unleash significant upgrades to its Gemini models and AI Studio, potentially unveiling groundbreaking features at the upcoming Google I/O 2025. Whispers of these enhancements are already circulating, hinting at a future where real-time multimodal processing and seamless coding workflows become the norm.
Over the weekend, eagle-eyed users noticed a subtle update within AI Studio, flagged with a simple “new” label. This seemingly minor change redirects to the Stream Realtime section, suggesting an impending upgrade to Flash 2.5. This upgrade will likely significantly enhance Gemini's ability to process real-time image, video, and potentially audio inputs. This is expected to be announced during Google I/O.

This move could pave the way for end-to-end input-output pipelines, transforming raw visual input into spoken responses in real-time. A separate backend tweak now allows for simultaneous web search and code execution, enabling greater autonomy and context-switching within the system.
Adding fuel to the fire, Logan Kilpatrick recently sparked a discussion on X about integrating an engineering agent into AI Studio. This move is very similar to OpenAI’s Codex agent within ChatGPT. The fact that Google is unifying model capabilities with deployability via Cloud Run is a sign that they are working towards a smooth coding and deployment flow. Whether OpenAI will also follow suit remains to be seen.
But that's not all! According to TestingCatalog, Google has even more in store for its AI users, an all-in-one product might be comming to Gemini soon, that contains a **Video Overviews** option. With Deep Research, users will soon be able to upload their own documents and images as referencing material so that the AI will scan the web for you and prepare a comprehensive report with key findings. This is very similar to a core function of NotebookLM.

The blog looked at Illuminate, a Google AI experiment that lets you turn content into AI-generated audio discussions. The new UI features experimental controls, such as an Edit button, caption toggles, and a cover image generation tool. However, the most interesting new Illuminate feature is a section called Sparks that’s in Early Preview with the following description:
Imagine any question could be instantly transformed into a short video, 100% AI-generated.
This feature would work by prompting the AI a question and Illuminate will create a video complete with audio commentary that addresses the prompt. These clips are between one and three minutes long and might use a video generation tool like Veo 3, potentially paired with a next-gen multimodal version of Gemini.
Google's recent strides in AI integration show the company's commitment to improving its AI offerings, and making them more accessible. By allowing users to study smarter, and by integrating AI tools into their current workflow. These advancements are poised to redefine how we interact with AI, offering a glimpse into the future of real-time media analysis, coding workflows, and information consumption.
Will Google's I/O 2025 deliver on these exciting promises? What features are you most looking forward to? Share your thoughts and predictions in the comments below!