Google has revealed that users can use Audio Overview to turn documents, slides, and even Deep Research reports into easy-to-listen-to podcasts. The firm said it will allow users to select a new “Generate Audio Overview” option to listen to an audio overview based on their research.
Google also upgraded Deep Research with Gemini 2.0 Flash Thinking Experimental to enhance Gemini’s capabilities across all research stages. The tech firm said that the new upgrade will help Gemini to create higher-quality, multi-page reports that will be more detailed and insightful.
Google lets users make AI podcasts from Gemini Deep Research
Google announced that its Gemini app will allow users to generate audio overviews based on Deep Research. The firm said that users can turn in-depth reports generated by Gemini into a conversational podcast featuring two AI “hosts.”
“Think of it as helping you go from zero to understanding a subject deeply.”
-Aarush Selvan, Gemini Senior Product Manager.
The tech firm released its NotebookLM Audio Overview feature last September and said that it can turn documents into engaging audio discussions. The company highlighted that two hosts can start up lively “deep dive” discussions based on users’ sources, “make connections between topics, and banter back and forth.”
See also Elon Musk's xAI and Nvidia is joining BlackRock and Microsoft for AI infrastructure development
The firm said the upgrade was meant to build on NotebookLM’s existing feature that helped users interact with all their notes, transcripts, and other documents. Google acknowledged that the feature was only an audio version since the app already used Google’s Gemini AI model to help summarize research.
Google has now introduced the same feature for Deep Research, which can be accessed when users toggle in the right-hand corner that has a few different options, including one to generate an audio overview. The firm also shared that users will be able to multitask while listening, even on the go, and it believes that “it’s great if you absorb information better that way.”
Lead product for Google AI Studio Logan Kilpatrick mentioned that Deep Research blew him away because he was able to browse 145 different sites based on his query.
Google also introduced Audio Overview to the Gemini app for free users and advanced subscribers earlier this week. The tech company mentioned that the feature will be more helpful for Deep Research, its “agentic” AI feature. Google said that the feature will let users call upon Gemini to explore a specific topic from the web and generate a detailed report based on its findings. According to the firm, users can select a new “Generate Audio Overview” option after Gemini finishes generating a report. The feature will also allow users to listen to an audio overview based on the research.
Google also mentioned that Audio Overview had some limitations where it could take several minutes to generate a podcast-like discussion. The company also said that the feature could only generate podcast-like discussion in English, with more languages to come at a later date. The firm highlighted that Audio Overview will be accessible on the web and in the Gemini mobile app, and users can easily share or download them to listen on the go.
See also Nvidia CEO dismisses DeepSeek AI cost-cutting fears
Google introduces Gemini 2.0 Flash Thinking Experimental mode
The tech firm also revealed Gemini’s latest feature called “Canvas”, which refines users’ documents and code and Audio Overview and transforms them into engaging podcast-style discussions. Google CEO Sundar Pichai said that users can create a doc or coding project with Canvas because “it spins up prototypes super fast, and you can edit with Gemini in real-time.”
Google also introduced Gemini 2.0 Flash Thinking Experimental mode on March 17, which can analyze users’ search history to better understand their interests and provide more helpful answers. The firm said that the model was trained to break down prompts into a series of steps to strengthen its reasoning capabilities and deliver better responses.
Google also highlighted that Gemini Advanced users will have access to a 1M token context window with 2.0 Flash Thinking Experimental. The tech company said that it will enable users to solve more complex problems than ever before by experimenting and analyzing large amounts of information.
Cryptopolitan Academy: Tired of market swings? Learn how DeFi can help you build steady passive income. Register Now