Épisodes

  • Training General Robots for Any Task: Physical Intelligence’s Karol Hausman and Tobi Springenberg
    Jan 6 2026
    Physical Intelligence’s Karol Hausman and Tobi Springenberg believe that robotics has been held back not by hardware limitations, but by an intelligence bottleneck that foundation models can solve. Their end-to-end learning approach combines vision, language, and action into models like π0 and π*0.6, enabling robots to learn generalizable behaviors rather than task-specific programs. The team prioritizes real-world deployment and uses RL from experience to push beyond what imitation learning alone can achieve. Their philosophy—that a single general-purpose model can handle diverse physical tasks across different robot embodiments—represents a fundamental shift in how we think about building intelligent machines for the physical world. Hosted by Alfred Lin and Sonya Huang, Sequoia Capital
    Voir plus Voir moins
    1 h et 2 min
  • Why the Next AI Revolution Will Happen Off-Screen: Samsara CEO Sanjit Biswas
    Dec 16 2025
    Sanjit Biswas is one of the rare founders who has scaled AI in the physical world – first with Meraki, and now with Samsara, a $20B+ public company with sensors deployed across millions of vehicles and job sites. Capturing 90 billion miles of driving data each year, Samsara operates at a scale matched only by a small handful of companies. Sanjit discusses why physical AI is fundamentally different from cloud-based AI, from running inference on two- to ten-watt edge devices to managing the messy diversity of real-world data—weather, road conditions, and the long tail of human behavior. He also shares how advances in foundation models unlock new capabilities like video reasoning, why distributed compute at the edge still beats centralized data centers for many autonomy workloads, and how AI is beginning to coach frontline workers—not just detect risk, but recognize good driving and improve fuel efficiency. Sanjit also explains why connectivity, sensors, and compute were the original “why now” for Samsara, and how those compounding curves will reshape logistics, field service, construction, and every asset-heavy industry. Hosted by Sonya Huang and Pat Grady, Sequoia Capital
    Voir plus Voir moins
    38 min
  • The Rise of Generative Media: fal's Bet on Video, Infrastructure, and Speed
    Dec 10 2025
    fal is building the infrastructure layer for the generative media boom. In this episode, founders Gorkem Yurtseven, Burkay Gur, and Head of Engineering Batuhan Taskaya explain why video models present a completely different optimization problem than LLMs, one that is compute-bound, architecturally volatile, and changing every 30 days. They discuss how fal's tracing compiler, custom kernels, and globally distributed GPU fleet enable them to run more than 600 image and video models simultaneously, often faster than the labs that trained them. The team also shares what they’re seeing from the demand side: AI-native studios, personalized education, programmatic advertising, and early engagement from Hollywood. They argue that generative video is following a trajectory similar to early CGI—initial skepticism giving way to a new medium with its own workflows, aesthetics, and economic models.Hosted by Sonya Huang, Sequoia Capital
    Voir plus Voir moins
    1 h et 2 min
  • Why IDEs Won't Die in the Age of AI Coding: Zed Founder Nathan Sobo
    Dec 2 2025
    Nathan Sobo has spent nearly two decades pursuing one goal: building an IDE that combines the power of full-featured tools like JetBrains with the responsiveness of lightweight editors like Vim. After hitting the performance ceiling with web-based Atom, he founded Zed and rebuilt from scratch in Rust with GPU-accelerated rendering. Now with 170,000 active developers, Zed is positioned at the intersection of human and AI collaboration. Nathan discusses the Agent Client Protocol that makes Zed "Switzerland" for different AI coding agents, and his vision for fine-grained edit tracking that enables permanent, contextual conversations anchored directly to code—a collaborative layer that asynchronous git-based workflows can't provide. Nathan argues that despite terminal-based AI coding tools visual interfaces for code aren't going anywhere, and that source code is a language designed for humans to read, not just machines to execute. Hosted by Sonya Huang and Pat Grady, Sequoia Capital
    Voir plus Voir moins
    40 min
  • How End-to-End Learning Created Autonomous Driving 2.0: Wayve CEO Alex Kendall
    Nov 18 2025
    Alex Kendall founded Wayve in 2017 with a contrarian vision: replace the hand-engineered autonomous vehicle stack with end-to-end deep learning. While AV 1.0 companies relied on HD maps, LiDAR retrofits, and city-by-city deployments, Wayve built a generalization-first approach that can adapt to new vehicles and cities in weeks. Alex explains how world models enable reasoning in complex scenarios, why partnering with automotive OEMs creates a path to scale beyond robo-taxis, and how language integration opens up new product possibilities. From driving in 500 cities to deploying with manufacturers like Nissan, Wayve demonstrates how the same AI breakthroughs powering LLMs are transforming the physical economy. Hosted by: Pat Grady and Sonya Huang
    Voir plus Voir moins
    42 min
  • How Google’s Nano Banana Achieved Breakthrough Character Consistency
    Nov 11 2025
    When Google launched Nano Banana, it instantly became a global phenomenon, introducing an image model that finally made it possible for people to see themselves in AI-generated worlds. In this episode, Nicole Brichtova and Hansa Srinivasan, the product and engineering leads behind Nano Banana, share the story behind the model’s creation and what it means for the future of visual AI. Nicole and Hansa discuss how they achieved breakthrough character consistency, why human evaluation remains critical for models that aim to feel right, and how “fun” became a gateway to utility. They explain the craft behind Gemini’s multimodal design, the obsession with data quality that powered Nano Banana’s realism, and how user creativity continues to push the technology in unexpected directions—from personal storytelling to education and professional design. The conversation explores what comes next in visual AI, why accessibility and imagination must evolve together, and how the tools we build can help people capture not just reality but possibility. Hosted by: Stephanie Zhan and Pat Grady, Sequoia Capital
    Voir plus Voir moins
    44 min
  • OpenAI Sora 2 Team: How Generative Video Will Unlock Creativity and World Models
    Nov 6 2025
    The OpenAI Sora 2 team (Bill Peebles, Thomas Dimson, Rohan Sahai) discuss how they compressed filmmaking from months to days, enabling anyone to create compelling video. Bill, who invented the diffusion transformer that powers Sora and most video generation models, explains how space-time tokens enable object permanence and physics understanding in AI-generated video, and why Sora 2 represents a leap for video. Thomas and Rohan share how they're intentionally designing the Sora product against mindless scrolling, optimizing for creative inspiration, and building the infrastructure for IP holders to participate in a new creator economy. The conversation goes beyond video generation into the team’s vision for world simulators that could one day run scientific experiments, their perspective on co-evolving society alongside technology, and how digital simulations in alternate realities may become the future of knowledge work. Hosted by: Konstantine Buhler and Sonya Huang, Sequoia Capital
    Voir plus Voir moins
    1 h
  • Nvidia CTO Michael Kagan: Scaling Beyond Moore's Law to Million-GPU Clusters
    Oct 28 2025
    Recorded live at Sequoia’s Europe100 event: Michael Kagan, co-founder of Mellanox and CTO of Nvidia, explains how the $7 billion Mellanox acquisition helped transform Nvidia from a chip company into the architect of AI infrastructure. Kagan breaks down the technical challenges of scaling from single GPUs to 100K and eventually million-GPU data centers. He reveals why network performance—not just compute power—determines AI system efficiency. He discusses the shift from training to inference workloads, and his vision for AI as humanity's "spaceship of the mind," and why he thinks AI may help us discover laws of physics we haven’t yet imagined. Hosted by Sonya Huang and Pat Grady
    Voir plus Voir moins
    42 min
adbl_web_global_use_to_activate_DT_webcro_1694_expandible_banner_T1