TED Talks Daily: "The Next Computer? Your Glasses" by Shahram Izadi
Introduction to the Computing Revolution
In the April 17, 2025 episode of TED Talks Daily titled "The Next Computer? Your Glasses," host Elise Hu introduces Shahram Izadi, a pioneer in augmented reality (AR) and extended reality (XR) at Google. Izadi delves into the transformative potential of merging AI with wearable technology, envisioning a future where our glasses and headsets act as personal AI assistants, seamlessly integrating digital information with our physical world.
Shahram Izadi’s Journey and Vision
Izadi begins by reflecting on his 25-year journey in augmented reality, highlighting the early days of his PhD research where he aimed to fuse the real world with computing experiences. He acknowledges the parallel advancements in AI that have enabled computers to reason about the world, citing, “For our entire lives, we've been living for a computing revolution with leaps forward in compute, connectivity, mobile technologies, and now AI” (00:50).
He emphasizes the convergence of AI and XR as the next phase of this revolution, stating, “AI and XR are converging, unlocking radical new ways to interact with technology on your terms” (03:15). This synergy promises to make computers more lightweight, personal, and contextually aware, moving beyond traditional screens to immersive, conversational interfaces.
Integrating AI with XR: Android XR and Gemini
Izadi introduces Android XR, an operating system developed in collaboration with Samsung, which integrates XR hardware with Gemini, Google's AI assistant. This platform supports a wide range of devices, from glasses to headsets, with the flexibility to adapt to future form factors. He explains, “With Android XR, we’re building an operating system that brings XR hardware together with Gemini, our AI assistant to augment human intelligence” (03:45).
Demonstrations of Smart Glasses
To illustrate the capabilities of this technology, Izadi invites his colleague Nishtha to demonstrate smart glasses equipped with Gemini AI. The glasses feature a miniaturized camera, microphones, speakers, and a high-resolution lens display. During the live demo, Nishtha interacts with Gemini AI, showcasing its ability to:
-
Visual and Auditory Integration: Gemini can see and hear the environment, responding to visual and auditory inputs in real-time. Nishtha asks Gemini to write a haiku based on her current view, and the AI responds promptly: “Faces all aglow. Eager minds await the words” (06:51).
-
Memory and Contextual Awareness: The AI remembers past interactions without explicit prompts. When Nishtha inadvertently looks away, she asks Gemini about a book she briefly saw, and the AI accurately identifies it as "Atomic Habits by James Clear" (07:37).
-
Multilingual Capabilities: Gemini seamlessly translates signs and conversations into different languages, including Farsi and Hindi, demonstrating its advanced language processing and translation abilities (08:42).
-
Actionable Responses: The AI can perform tasks like summarizing complex diagrams and initiating actions such as playing music or navigating to specific locations. For instance, Nishtha requests Gemini to navigate her to Lighthouse Park, and the AI provides precise directions (10:13).
Izadi remarks, “This AI is inherently multimodal, so it's using visual data and natural language to digest complex queries in real time” (07:19), underscoring the sophisticated integration of AI with visual and contextual data.
Showcasing AI-Integrated Headsets
Transitioning from glasses to headsets, Izadi presents the Project Muhan headset developed by Samsung. He introduces his colleague Max, who demonstrates how the headset, powered by Gemini AI, enhances user interaction:
-
Hands-Free Control: Max controls the system using his eyes, hands, and voice, with AI assisting in managing digital content seamlessly. When Max requests, “Hey, Gemini, can you bring up my trip planner for me, please?” the AI promptly opens the trip planner (12:28).
-
Interactive Navigation and Information Retrieval: During the demo, Max asks Gemini to navigate to Cape Town, and the AI organizes his digital workspace while providing informative insights about locations like Table Mountain (13:02; 13:41).
-
Immersive Experiences: The headset allows users to explore virtual environments, such as viewing 360-degree snowboarding videos with real-time assistance from Gemini. Max engages in playful interactions, asking Gemini to narrate a video in a horror movie style, showcasing the AI’s versatility and creativity (15:38).
Izadi highlights, “The AI is taking Max's request, figuring out how best to answer it, opening up the Maps app, and from there, he can actually explore anywhere in the world in this 3D view” (13:27), emphasizing the AI’s ability to integrate and act upon user commands fluidly.
Augmenting Human Intelligence
Concluding his talk, Izadi envisions a future where AI and XR not only augment our reality but also enhance our intelligence. He states, “A world where helpful AI will converge with lightweight XR devices will become increasingly more wearable, giving us instant access to information” (17:12). This integration aims to create personalized, conversational interfaces that work on users’ terms, significantly enriching daily interactions with technology.
Key Takeaways and Future Outlook
-
Seamless Integration: The fusion of AI with AR and XR devices creates intuitive, immersive interfaces that blend digital and physical worlds.
-
Contextual Awareness: AI assistants like Gemini can understand and respond to complex, multimodal inputs, providing personalized and contextually relevant assistance.
-
Flexible and Wearable Technology: Lightweight, wearable devices such as glasses and headsets are becoming the next generation of computing platforms, offering versatile functionalities without the constraints of traditional screens.
-
Enhanced Human Intelligence: By augmenting our cognitive abilities, AI and XR technologies empower users to navigate, learn, and interact with their environment more effectively.
Izadi concludes with an inspiring vision of the future, where AI-driven XR devices not only enhance our reality but also our inherent intelligence, paving the way for unprecedented advancements in how we interact with technology and the world around us.
Notable Quotes:
-
“For our entire lives, we've been living for a computing revolution with leaps forward in compute, connectivity, mobile technologies, and now AI.” (00:50)
-
“AI and XR are converging, unlocking radical new ways to interact with technology on your terms.” (03:15)
-
“This AI is inherently multimodal, so it's using visual data and natural language to digest complex queries in real time.” (07:19)
-
“The AI is taking Max's request, figuring out how best to answer it, opening up the Maps app, and from there, he can actually explore anywhere in the world in this 3D view.” (13:27)
-
“A world where helpful AI will converge with lightweight XR devices will become increasingly more wearable, giving us instant access to information.” (17:12)
Conclusion
Shahram Izadi's presentation on TED Talks Daily offers a visionary glimpse into the future of computing, where AI and XR technologies converge to create intelligent, wearable devices that seamlessly integrate with our daily lives. Through engaging demonstrations and insightful discussions, Izadi illustrates how these advancements will not only transform our interaction with technology but also enhance our cognitive capabilities, paving the way for a new era of augmented intelligence.