Powered by RND
PodcastsBusinessTraining Data

Training Data

Sequoia Capital
Training Data
Latest episode

Available Episodes

5 of 55
  • Mapping the Mind of a Neural Net: Goodfire’s Eric Ho on the Future of Interpretability
    Eric Ho is building Goodfire to solve one of AI’s most critical challenges: understanding what’s actually happening inside neural networks. His team is developing techniques to understand, audit and edit neural networks at the feature level. Eric discusses breakthrough results in resolving superposition through sparse autoencoders, successful model editing demonstrations and real-world applications in genomics with Arc Institute's DNA foundation models. He argues that interpretability will be critical as AI systems become more powerful and take on mission-critical roles in society. Hosted by Sonya Huang and Roelof Botha, Sequoia Capital Mentioned in this episode: Mech interp: Mechanistic interpretability, list of important papers here Phineas Gage: 19th century railway engineer who lost most of his brain’s left frontal lobe in an accident. Became a famous case study in neuroscience. Human Genome Project: Effort from 1990-2003 to generate the first sequence of the human genome which accelerated the study of human biology Emergent Misalignment: Narrow finetuning can produce broadly misaligned LLMs Zoom In: An Introduction to Circuits: First important mechanistic interpretability paper from OpenAI in 2020 Superposition: Concept from physics applied to interpretability that allows neural networks to simulate larger networks (e.g. more concepts than neurons) Apollo Research: AI safety company that designs AI model evaluations and conducts interpretability research Towards Monosemanticity: Decomposing Language Models With Dictionary Learning. 2023 Anthropic paper that uses a sparse autoencoder to extract interpretable features; followed by Scaling Monosemanticity Under the Hood of a Reasoning Model: 2025 Goodfire paper that interprets DeepSeek’s reasoning model R1 Auto-interpretability: The ability to use LLMs to automatically write explanations for the behavior of neurons in LLMs Interpreting Evo 2: Arc Institute's Next-Generation Genomic Foundation Model. (see episode with Arc co-founder Patrick Hsu) Paint with Ember: Canvas interface from Goodfire that lets you steer an LLM’s visual output  in real time (paper here) Model diffing: Interpreting how a model differs from checkpoint to checkpoint during finetuning Feature steering: The ability to change the style of LLM output by up or down weighting features (e.g. talking like a pirate vs factual information about the Andromeda Galaxy) Weight based interpretability: Method for directly decomposing neural network parameters into mechanistic components, instead of using features The Urgency of Interpretability: Essay by Anthropic founder Dario Amodei On the Biology of a Large Language Model: Goodfire collaboration with Anthropic
    --------  
    47:07
  • ElevenLabs’ Mati Staniszewski: Why Voice Will Be the Fundamental Interface for Tech
    Mati Staniszewski, co-founder and CEO of ElevenLabs, explains how staying laser-focused on audio innovation has allowed his company to thrive despite the push into multimodality from foundation models. From a high school friendship in Poland to building one of the fastest-growing AI companies, Mati shares how ElevenLabs transformed text-to-speech with contextual understanding and emotional delivery. He discusses the company's viral moments (from Harry Potter by Balenciaga to powering Darth Vader in Fortnite), and explains how ElevenLabs is creating the infrastructure for voice agents and real-time translation that could eliminate language barriers worldwide. Hosted by: Pat Grady, Sequoia Capital Mentioned in this episode: Attention Is All You Need: The original Transformers paper Tortoise-tts: Open source text to speech model that was a starting point for ElevenLabs (which now maintains a v2) Harry Potter by Balenciaga: ElevenLabs’ first big viral moment from 2023 The first AI that can laugh: 2022 blog post backing up ElevenLab’s claim of laughter (it got better in v3) Darth Vader's voice in Fortnite: ElevenLabs used actual voice clips provided by James Earl Jones before he died Lex Fridman interviews Prime Minister Modi: ElevenLabs enabled Fridman to speak in Hindi and Modi to speak in English. Time Person of the Year 2024: ElevenLabs-powered experiment with “conversational journalism” Iconic Voices: Richard Feynman, Deepak Chopra, Maya Angelou and more available in ElevenLabs reader app SIP trunking: a method of delivering voice, video, and other unified communications over the internet using the Session Initiation Protocol (SIP) Genesys: Leading enterprise CX platform for agentic AI Hitchhiker’s Guide to the Galaxy: Comedy/science-fiction series by Douglas Adams that contains the concept of the Babel Fish instantaneous translator, cited by Mati FYI: communication and productivity app for creatives that Mati uses, founded by will.i.am Lovable: prototyping app that Mati loves
    --------  
    59:53
  • From DevOps ‘Heart Attacks’ to AI-Powered Diagnostics With Traversal’s AI Agents
    Anish Agarwal and Raj Agrawal, co-founders of Traversal, are transforming how enterprises handle critical system failures. Their AI agents can perform root cause analysis in 2-4 minutes instead of the hours typically spent by teams of engineers scrambling in Slack channels. Drawing from their academic research in causal inference and gene regulatory networks, they’ve built agents that systematically traverse complex dependency maps to identify the smoking gun logs and problematic code changes. As AI-generated code becomes more prevalent, Traversal addresses a growing challenge: debugging systems where humans didn’t write the original code, making AI-powered troubleshooting essential for maintaining reliable software at scale. Hosted by Sonya Huang and Bogomil Balkansky, Sequoia Capital Mentioned in this episode: SRE: Site reliability engineering. The function within engineering teams that monitors and improves the availability and performance of software systems and services. Golden signals: four key metrics used by Site Reliability Engineers (SREs) to monitor the health and performance of IT systems: latency, traffic, errors and saturation. MELT data:  Metrics, events, log, and traces. A framework for observability. The Bitter Lesson: Another mention of Nobel Prize  winner Rich Sutton’s influential post.
    --------  
    40:32
  • The Breakthroughs Needed for AGI Have Already Been Made: OpenAI Former Research Head Bob McGrew
    As OpenAI's former Head of Research, Bob McGrew witnessed the company's evolution from GPT-3’s breakthrough to today's reasoning models. He argues that there are three legs of the stool for AGI—Transformers, scaled pre-training, and reasoning—and that the fundamentals that will shape the next decade-plus are already in place. He thinks 2025 will be defined by reasoning while pre-training hits diminishing returns. Bob discusses why the agent economy will price services at compute costs due to near-infinite supply, fundamentally disrupting industries like law and medicine, and how his children use ChatGPT to spark curiosity and agency. From robotics breakthroughs to managing brilliant researchers, Bob offers a unique perspective on AI’s trajectory and where startups can still find defensible opportunities. Hosted by: Stephanie Zhan and Sonya Huang, Sequoia Capital Mentioned in this episode:  Solving Rubik’s Cube with a robot hand: OpenAI’s original robotics research Computer Use and Operator: Anthropic and OpenAI reasoning breakthroughs that originated with OpenAi researchers Skild and Physical Intelligence: Robotics-oriented companies Bob sees as well-positioned now Distyl: AI company founded by ex-Palintir alums to create enterprise workflows driven by proprietary data Member of the technical staff: Title at OpenAI designed to break down barriers between AI researchers and engineers Howie.ai: Scheduling app that Bob uses
    --------  
    48:51
  • OpenAI Codex Team: From Coding Autocomplete to Asynchronous Autonomous Agents
    Hanson Wang and Alexander Embiricos from OpenAI's Codex team discuss their latest AI coding agent that works independently in its own environment for up to 30 minutes, generating full pull requests from simple task descriptions. They explain how they trained the model beyond competitive programming to match real-world software engineering needs, the shift from pairing with AI to delegating to autonomous agents, and their vision for a future where the majority of code is written by agents working on their own computers. The conversation covers the technical challenges of long-running inference, the importance of creating realistic training environments, and how developers are already using Codex to fix bugs and implement features at OpenAI. Hosted by Sonya Huang and Lauren Reeder, Sequoia Capital  Mentioned in this episode:  The Culture: Sci-Fi series by Iain Banks portraying an optimistic view of AI The Bitter Lesson: Influential paper by Rich Sutton on the importance of scale as a strategic unlock for AI.
    --------  
    37:44

More Business podcasts

About Training Data

Join us as we train our neural nets on the theme of the century: AI. Sonya Huang, Pat Grady and more Sequoia Capital partners host conversations with leading AI builders and researchers to ask critical questions and develop a deeper understanding of the evolving technologies—and their implications for technology, business and society. The content of this podcast does not constitute investment advice, an offer to provide investment advisory services, or an offer to sell or solicitation of an offer to buy an interest in any investment fund.
Podcast website

Listen to Training Data, Working Hard with Grace Beverley and many other podcasts from around the world with the radio.net app

Get the free radio.net app

  • Stations and podcasts to bookmark
  • Stream via Wi-Fi or Bluetooth
  • Supports Carplay & Android Auto
  • Many other app features

Training Data: Podcasts in Family

Social
v7.20.2 | © 2007-2025 radio.de GmbH
Generated: 7/9/2025 - 12:16:01 PM