Afleveringen
-
Explore the synergy between long context models and Retrieval Augmented Generation (RAG) in this episode of Release Notes. Join Google DeepMind's Nikolay Savinov as he discusses the importance of large context windows, how they enable Al agents, and what's next in the field.
Chapters:
0:52 Introduction & defining tokens
5:27 Context window importance
9:53 RAG vs. Long Context
14:19 Scaling beyond 2 million tokens
18:41 Long context improvements since 1.5 Pro release
23:26 Difficulty of attending to the whole context
28:37 Evaluating long context: beyond needle-in-a-haystack
33:41 Integrating long context research
34:57 Reasoning and long outputs
40:54 Tips for using long context
48:51 The future of long context: near-perfect recall and cost reduction
54:42 The role of infrastructure
56:15 Long-context and agents -
Tulsee Doshi, Head of Product for Gemini Models joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini 2.5 Pro experimental launch. Gemini 2.5 is a well-rounded, multimodal thinking model, designed to tackle increasingly complex problems. From enhanced reasoning to advanced coding, Gemini 2.5 can create impressive web applications and agentic code applications. Learn about the process of building Gemini 2.5 Pro experimental, the improvements made across the stack, and what’s next for Gemini 2.5.
Chapters:
0:00 - Introduction
1:05 - Gemini 2.5 launch overview
3:19 - Academic evals vs. vibe checks
6:19 - The jump to 2.5
7:51 - Coordinating cross-stack improvements
11:48 - Role of pre/post-training vs. test-time compute
13:21 - Shipping Gemini 2.5
15:29 - Embedded safety process
17:28 - Multimodal reasoning with Gemini 2.5
18:55 - Benchmark deep dive
22:07 - What’s next for Gemini
24:49 - Dynamic thinking in Gemini 2.5
25:37 - The team effort behind the launchResources:
Gemini → https://goo.gle/41Yf72bGemini 2.5 blog post → https://goo.gle/441SHiVExample of Gemini’s 2.5 Pro’s game design skills → https://goo.gle/43vxkq1Demo: Gemini 2.5 Pro Experimental in Google AI Studio → https://goo.gle/4c5RbhE -
Zijn er afleveringen die ontbreken?
-
Dave Citron, Senior Director Product Management, joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini updates and demos. Learn more about Canvas for collaborative content creation, enhanced Deep Research with Thinking Models and Audio Overview and a new personalization feature.
0:00 - Introduction
0:59 - Recent Gemini app launches
2:00 - Introducing Canvas
5:12 - Canvas in action
8:46 - More Canvas examples
12:02 - Enhanced capabilities with Thinking Models
15:12 - Deep Research in action
20:27 - The future of agentic experiences
22:12 Deep Research and Audio Overviews
24:11 - Personalization in Gemini app
27:50 - Personalization in action
29:58 - How personalization works: user data and privacy
32:30 -The future of personalization -
Jack Rae, Principal Scientist at Google DeepMind, joins host Logan Kilpatrick for an in-depth discussion on the development of Google’s thinking models. Learn more about practical applications of thinking models, the impact of increased 'thinking time' on model performance and the key role of long context.
01:14 - Defining Thinking Models
03:40 - Use Cases for Thinking Models
07:52 - Thinking Time Improves Answers
09:57 - Rapid Thinking Progress
20:11 - Long Context Is Key
27:41 - Tools for Thinking Models
29:44 - Incorporating Developer Feedback
35:11 - The Strawberry Counting Problem
39:15 - Thinking Model Development Timeline
42:30 - Towards a GA Thinking Model
49:24 - Thinking Models Powering AI Agents
54:14 - The Future of AI Model Evals -
Tulsee Doshi, Gemini model product lead, joins host Logan Kilpatrick to go behind the scenes of Gemini 2.0, taking a deep dive into the model's multimodal capabilities and native tool use, and Google's approach to shipping experimental models.
Watch on YouTube: https://www.youtube.com/watch?v=L7dw799vu5o
Chapters:
Meet Tulsee Doshi
Gemini's Progress Over the Past Year
Introducing Gemini 2.0
Shipping Experimental Models
Gemini 2.0’s Native Tool Use
Function Calling
Multimodal Agents
Rapid Fire Questions -
Logan Kilpatrick sits down with Emanuel Taropa, a key figure in the development of Gemini to delve into the cutting edge of AI. Taropa provides insights into the technical challenges and triumphs of building and deploying large language models, focusing on the recent release of the Flash 8B Gemini model.
Their conversation covers everything from the intricacies of model architecture and training to the practical challenges of shipping AI models at scale, and even speculates on the future of AI.