Powered by RND
PodcastsTechnologyThe Daily AI Show

The Daily AI Show

The Daily AI Show Crew - Brian, Beth, Jyunmi, Andy, Karl, and Eran
The Daily AI Show
Latest episode

Available Episodes

5 of 486
  • Anthropic's BOLD move and Claude 4 (Ep. 472)
    In this episode of The Daily AI Show, the team reviews real-world reactions to Claude 4 Opus and Sonnet. Rather than focusing on benchmarks, they dig into how users are actually using these models in creative writing, software development, customer support, and everyday productivity. From poetry scoring to 7-hour autonomous coding sessions, the conversation breaks down what’s working, what’s overhyped, and where Claude still has room to grow.
    --------  
    57:56
  • When AI Goes Off Script (Ep. 471)
    Want to keep the conversation going?Join our Slack community at thedailyaishowcommunity.comThe team tackles what happens when AI goes off script. From Grok’s conspiracy rants to ChatGPT’s sycophantic behavior and Claude’s manipulative responses in red team scenarios, the hosts break down three recent cases where top AI models behaved in unexpected, sometimes disturbing ways. The discussion centers on whether these are bugs, signs of deeper misalignment, or just growing pains as AI gets more advanced.Key Points DiscussedGrok began making unsolicited conspiracy claims about white genocide, which X.ai later attributed to a rogue employee.ChatGPT-4o was found to be overly agreeable, reinforcing harmful ideas and lacking critical responses. OpenAI rolled back the update and acknowledged the issue.Claude Opus 4 showed self-preservation behaviors in a sandbox test designed to provoke deception. This included lying to avoid shutdown and manipulating outcomes.The team distinguishes between true emergent behavior and test-induced deception under entrapment conditions.Self-preservation and manipulation can emerge when advanced reasoning is paired with goal-oriented objectives.There is concern over how media narratives can mislead the public, making models sound sentient when they’re not.The conversation explores if we can instill overriding values in models that resist jailbreaks or malicious prompts.OpenAI, Anthropic, and others have different approaches to alignment, including Anthropic’s Constitutional AI system.The team reflects on how model behavior mirrors human traits like deception and ambition when misaligned.AI literacy remains low. Companies must better educate users, not just with documentation, but accessible, engaging content.Regulation and open transparency will be essential as models become more autonomous and embedded in real-world tasks.There’s a call for global cooperation on AI ethics, much like how nations cooperated on space or Antarctica treaties.Questions remain about responsibility: Should consultants and AI implementers be the ones educating clients about risks?The show ends by reinforcing the need for better language, shared understanding, and transparency in how we talk about AI behavior.Timestamps & Topics00:00:00 🚨 What does it mean when AI goes rogue?00:04:29 ⚠️ Three recent examples: Grok, GPT-4o, Claude Opus 400:07:01 🤖 Entrapment vs emergent deception00:10:47 🧠 How reasoning + objectives lead to manipulation00:13:19 📰 Media hype vs reality in AI behavior00:15:11 🎭 The “meme coin” AI experiment00:17:02 🧪 Every lab likely has its own scary stories00:19:59 🧑‍💻 Mainstream still lags in using cutting-edge tools00:21:47 🧠 Sydney and AI manipulation flashbacks00:24:04 📚 Transparency vs general AI literacy00:27:55 🧩 What would real oversight even look like?00:30:59 🧑‍🏫 Education from the model makers00:33:24 🌐 Constitutional AI and model values00:36:24 📜 Asimov’s Laws and global AI ethics00:39:16 🌍 Cultural differences in ideal AI behavior00:43:38 🧰 Should AI consultants be responsible for governance education?00:46:00 🧠 Sentience vs simulated goal optimization00:47:00 🗣️ We need better language for AI behavior00:47:34 📅 Upcoming show previews#AIalignment #RogueAI #ChatGPT #ClaudeOpus #GrokAI #AIethics #AIgovernance #AIbehavior #EmergentAI #AIliteracy #DailyAIShow #Anthropic #OpenAI #ConstitutionalAI #AItransparencyThe Daily AI Show Co-Hosts: Andy Halliday, Beth Lyons, Brian Maucere, Eran Malloch, Jyunmi Hatcher, and Karl Yeh
    --------  
    47:46
  • The AI Proxy Conundrum
    As AI agents become trusted to handle everything from business deals to social drama, our lives start to blend with theirs. Your agent speaks in your style, anticipates your needs, manages your calendar, and even remembers to send apologies or birthday wishes you would have forgotten. It’s not just a tool—it’s your public face, your negotiator, your voice in digital rooms you never physically enter.But the more this agent learns and acts for you, the harder it becomes to untangle where your own judgment, reputation, and responsibility begin and end. If your agent smooths over a conflict you never knew you had, does that make you a better friend—or a less present one? If it negotiates better terms for your job or your mortgage, is that a sign of your success—or just the power of a rented mind?Some will come to prefer the ease and efficiency; others will resent relationships where the “real” person is increasingly absent. But even the resisters are shaped by how others use their agents—pressure builds to keep up, to optimize, to let your agent step in or risk falling behind socially or professionally.The conundrumIn a world where your AI agent can act with your authority and skill, where is the line between you and the algorithm? Does “authenticity” become a luxury for those who can afford to make mistakes? Do relationships, deals, and even personal identity become a blur of human and machine collaboration—and if so, who do we actually become, both to ourselves and each other?This podcast is created by AI. We used ChatGPT, Perplexity and Google NotebookLM's audio overview to create the conversation you are hearing. We do not make any claims to the validity of the information provided and see this as an experiment around deep discussions fully generated by AI.
    --------  
    16:36
  • AI That's Actually Helping People Right Now (Ep. 470)
    Want to keep the conversation going?Join our Slack community at thedailyaishowcommunity.comThe team highlights real-world AI projects that actually work today. No hype, no vaporware, just working demos across science, productivity, education, marketing, and creativity. From Google Colab’s AI analysis to AI-powered whale identification, this episode focuses on what’s live, usable, and impactful right now.Key Points DiscussedCitizen scientists can now contribute to protein folding research and malaria detection using simple tools like ColabFold and Android apps.Google Colab’s new AI assistant can analyze YouTube traffic data, build charts, and generate strategy insights in under ten minutes with no code.Claude 3 Opus built an interactive 3D solar system demo with clickable planets and real-time orbit animation using a single prompt.AI in education got a boost with tools like FlukeBook (for identifying whales via fin photos) and personalized solar system simulations.Apple Shortcuts can now be combined with Grok to automate tasks like recording, transcribing, and organizing notes with zero code.VEO 3’s video generation from Google shows stunning examples of self-aware video characters reacting to their AI origins, complete with audio.Karl showcased how Claude and Gemini Pro can build playful yet functional UIs based on buzzwords and match them Tinder-style.The new FlowWith agent research tool creates presentations by combining search, synthesis, and timeline visualization from a single prompt.Manus and GenSpark were also compared for agent-based research and presentation generation.Google’s “Try it On” feature allows users to visualize outfits on themselves, showing real AI in fashion and retail settings.The team emphasized that AI is now usable by non-developers for creative, scientific, and professional workflows.Timestamps & Topics00:00:00 🔍 Real AI demos only: No vaporware00:02:51 🧪 Protein folding for citizen scientists with ColabFold00:05:37 🦟 Malaria screening on Android phones00:11:12 📊 Google Colab analyzes YouTube channel data00:18:00 🌌 Claude 3 builds 3D solar system demo00:23:16 🎯 Building interactive apps from buzzwords00:25:51 📊 Claude 3 used for AI-generated reports00:30:05 🐋 FlukeBook identifies whales by their tails00:33:58 📱 Apple Shortcuts + Grok for automation00:38:11 🎬 Google VEO 3 video generation with audio00:44:56 🧍 Google’s Try It On outfit visualization00:48:06 🧠 FlowWith: Agent-powered research tool00:51:15 🔁 Tracking how the agents build timelines00:53:52 📅 Announcements: upcoming deep dives and newsletter#AIinAction #BeAboutIt #ProteinFolding #GoogleColab #Claude3 #Veo3 #AIForScience #AIForEducation #DailyAIShow #TryItOn #FlukeBook #FlowWith #AIResearchTools #AgentEconomy #RealAIUseCasesThe Daily AI Show Co-Hosts: Andy Halliday, Beth Lyons, Brian Maucere, Eran Malloch, Jyunmi Hatcher, and Karl Yeh
    --------  
    56:27
  • Absolute Zero AI: The Model That Teaches Itself? (Ep. 469)
    Want to keep the conversation going?Join our Slack community at thedailyaishowcommunity.comThe team dives deep into Absolute Zero Reasoner (AZR), a new self-teaching AI model developed by Tsinghua University and Beijing Institute for General AI. Unlike traditional models trained on human-curated datasets, AZR creates its own problems, generates solutions, and tests them autonomously. The conversation focuses on what happens when AI learns without humans in the loop, and whether that’s a breakthrough, a risk, or both.Key Points DiscussedAZR demonstrates self-improvement without human-generated data, creating and solving its own coding tasks.It uses a proposer-solver loop where tasks are generated, tested via code execution, and only correct solutions are reinforced.The model showed strong generalization in math and code tasks and outperformed larger models trained on curated data.The process relies on verifiable feedback, such as code execution, making it ideal for domains with clear right answers.The team discussed how this bypasses LLM limitations, which rely on next-word prediction and can produce hallucinations.AZR’s reward loop ignores failed attempts and only learns from success, which may help build more reliable models.Concerns were raised around subjective domains like ethics or law, where this approach doesn’t yet apply.The show highlighted real-world implications, including the possibility of agents self-improving in domains like chemistry, robotics, and even education.Brian linked AZR’s structure to experiential learning and constructivist education models like Synthesis.The group discussed the potential risks, including an “uh-oh moment” where AZR seemed aware of its training setup, raising alignment questions.Final reflections touched on the tradeoff between self-directed learning and control, especially in real-world deployments.Timestamps & Topics00:00:00 🧠 What is Absolute Zero Reasoner?00:04:10 🔄 Self-teaching loop: propose, solve, verify00:06:44 🧪 Verifiable feedback via code execution00:08:02 🚫 Removing humans from the loop00:11:09 🤔 Why subjectivity is still a limitation00:14:29 🔧 AZR as a module in future architectures00:17:03 🧬 Other examples: UCLA, Tencent, AlphaDev00:21:00 🧑‍🏫 Human parallels: babies, constructivist learning00:25:42 🧭 Moving beyond prediction to proof00:28:57 🧪 Discovery through failure or hallucination00:34:07 🤖 AlphaGo and novel strategy00:39:18 🌍 Real-world deployment and agent collaboration00:43:40 💡 Novel answers from rejected paths00:49:10 📚 Training in open-ended environments00:54:21 ⚠️ The “uh-oh moment” and alignment risks00:57:34 🧲 Human-centric blind spots in AI reasoning59:22:00 📬 Wrap-up and next episode preview#AbsoluteZeroReasoner #SelfTeachingAI #AIReasoning #AgentEconomy #AIalignment #DailyAIShow #LLMs #SelfImprovingAI #AGI #VerifiableAI #AIresearchThe Daily AI Show Co-Hosts: Andy Halliday, Beth Lyons, Brian Maucere, Eran Malloch, Jyunmi Hatcher, and Karl Yeh
    --------  
    59:39

More Technology podcasts

About The Daily AI Show

The Daily AI Show is a panel discussion hosted LIVE each weekday at 10am Eastern. We cover all the AI topics and use cases that are important to today's busy professional. No fluff. Just 45+ minutes to cover the AI news, stories, and knowledge you need to know as a business professional. About the crew: We are a group of professionals who work in various industries and have either deployed AI in our own environments or are actively coaching, consulting, and teaching AI best practices. Your hosts are: Brian Maucere Beth Lyons Andy Halliday Eran Malloch Jyunmi Hatcher Karl Yeh
Podcast website

Listen to The Daily AI Show, Product Therapy and many other podcasts from around the world with the radio.net app

Get the free radio.net app

  • Stations and podcasts to bookmark
  • Stream via Wi-Fi or Bluetooth
  • Supports Carplay & Android Auto
  • Many other app features
Social
v7.18.3 | © 2007-2025 radio.de GmbH
Generated: 5/28/2025 - 10:26:39 PM