Artwork
iconShare
 
Manage episode 507233897 series 3690669
Content provided by Mr. Dew. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Mr. Dew or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://staging.podcastplayer.com/legal.

In the grand finale of "All Things LLM," hosts Alex and Ben look ahead to the bleeding edge—and reflect on the ultimate question for AI: can we ever truly understand how these models think?

Inside this episode:

  • The rise of reasoning models: Discover why the next leap for AI isn’t just bigger models, but smarter thinking. Explore how OpenAI’s o1 and DeepSeek-R1 represent a paradigm shift, moving from brute-force “pre-train and scale” to dynamic, inference-time reasoning. Learn how these new models, designed for test-time compute, “think longer” to tackle complex challenges in math, code, and logic.
  • How reasoning emerges: Uncover the latest approaches—like inference-time scaling, majority voting, and the power of reinforcement learning—that let models break down problems step by step, creating explicit “chains of thought” and more reliable answers.
  • Interpretability and the black box: Go deep into the science of Mechanistic Interpretability (MI). Find out how tools like classifier probes, activation patching, and sparse auto-encoders (SAEs) are helping researchers reverse-engineer the inner workings of LLMs, from Golden Gate Bridge neurons to features for deception, coding errors, and more.
  • Ongoing debates: What’s the endgame for interpretability? Can we achieve a complete, human-understandable model, or is it as hard as explaining the brain? What’s the path to building both powerful and truly safe AI?

Perfect for listeners searching for:

  • Reasoning models vs. LLMs
  • Test-time compute and chain-of-thought
  • Mechanistic Interpretability (MI) in AI
  • Opening the black box of AI
  • Sparse auto-encoders and activation patching
  • Scaling laws beyond pre-training
  • AI safety and alignment
  • DeepSeek, OpenAI o1, Claude 3 research

Wrap-up:
Join us for a rich, forward-looking discussion at the intersection of science, engineering, and philosophy—where progress is rapid, safety is paramount, and interpretability is the new frontier. Whether you’re a developer, researcher, or lifelong learner, this episode brings you full circle on the state and future of LLMs.

Thank you for listening and sharing this journey with us. Stay tuned to "All Things LLM" for more breakthroughs, debates, and discoveries on the evolving landscape of artificial intelligence!

All Things LLM is a production of MTN Holdings, LLC. © 2025. All rights reserved.
For more insights, resources, and show updates, visit allthingsllm.com.
For business inquiries, partnerships, or feedback, contact: [email protected]

The views and opinions expressed in this episode are those of the hosts and guests, and do not necessarily reflect the official policy or position of MTN Holdings, LLC.

Unauthorized reproduction or distribution of this podcast, in whole or in part, without written permission is strictly prohibited.
Thank you for listening and supporting the advancement of transparent, accessible AI education.

  continue reading

15 episodes