Episodes

  • AI Agents at Work: Scaffold Required
    Dec 3 2025

    We review four clips from the Dwarkesh Patel Podcast with Satya Nadella, Microsoft's CEO. I highly recommend Dwarkesh’s show—technical & nerdy, but excellent.

    Satya talks about scaffolding—the software wrapped around AI models to make them actually work.

    So we speak with someone building that scaffolding: Neil McKechnie runs two AI-first startups as a CTO.

    He discusses how he orchestrates up to twelve different language models—GPT-5, Claude, Gemini, Llama, Mistral, Cohere, Perplexity.

    We discuss what it actually takes to build production systems with LLMs today—and what that reveals about the agent future we’re being pitched.

    Dwarkesh's Podcast:

    https://www.youtube.com/@DwarkeshPatel




    To stay in touch, sign up for our newsletter at https://www.superprompt.fm

    Show More Show Less
    40 mins
  • Whose Agent Is It Anyways?
    Nov 7 2025

    Description:

    AI agents from OpenAI, Google, and Anthropic promise to act on your behalf—booking flights, handling tasks, making decisions. What kind of agency do these systems actually have? And whose interests are they serving?

    Enterprise AI agents are already deployed in customer support, code generation, and task automation. Consumer agents—ChatGPT Agent Mode, personal task assistants—face a wider gap between marketing promises and actual capabilities.

    The alignment problem: agents need access to your calendar, email, and personal preferences to help you effectively. But the agent that knows you well enough to serve you is also positioned to steer you. When you delegate decisions to an agent, who decides what success looks like?

    To stay in touch, sign up for our newsletter at https://www.superprompt.fm

    Show More Show Less
    16 mins
  • AI Safety: Constitutional AI vs Human Feedback
    Jun 17 2024

    With great power comes great responsibility. How do leading AI companies implement safety and ethics as language models scale? OpenAI uses Model Spec combined with RLHF (Reinforcement Learning from Human Feedback). Anthropic uses Constitutional AI. The technical approaches to maximizing usefulness while minimizing harm. Solo episode on AI alignment.

    REFERENCE

    OpenAI Model Spec

    https://cdn.openai.com/spec/model-spec-2024-05-08.html#overview

    Anthropic Constitutional AI

    https://www.anthropic.com/news/claudes-constitution



    To stay in touch, sign up for our newsletter at https://www.superprompt.fm

    Show More Show Less
    17 mins
  • Open Source LLMs: How Open Is "Open"?
    Jun 10 2024

    Notable open source large language models from Meta, French AI company Mistral (valued at $2B), Microsoft, and Apple. Not all open source models are equally open—the restrictions and licensing constraints you need to know before deploying one. Solo episode.

    To stay in touch, sign up for our newsletter at https://www.superprompt.fm

    Show More Show Less
    13 mins
  • Open Source AI: The Safety Debate
    Jun 3 2024

    Why enterprises and entrepreneurs choose open source LLMs like Meta's Llama—cost-effectiveness, control, privacy, and security. The safety and ethics debate: which poses greater risk to humanity, open source or proprietary AI models? Both? Neither? Solo episode on open source LLMs.

    To stay in touch, sign up for our newsletter at https://www.superprompt.fm

    Show More Show Less
    16 mins
  • LLM Benchmarks: How to Know Which AI Is Better
    May 27 2024

    Beyond ChatGPT and Gemini: Anthropic's Claude and the $4 billion Amazon investment. How AI industry benchmarks work, including LMSYS Arena Elo and MMLU (Measuring Massive Multitask Language Understanding). How benchmarks are constructed, what they measure, and how to use them to evaluate LLMs. Solo episode.

    Anthropic's Claude
    https://claude.ai [Note: I am not sponsored by Anthropic]

    LMSYS Leaderboard
    https://chat.lmsys.org/?leaderboard

    To stay in touch, sign up for our newsletter at https://www.superprompt.fm

    Show More Show Less
    11 mins
  • Multimodal AI: When ChatGPT Learned to See
    May 20 2024

    Recent updates from Google and OpenAI feature multimodal capabilities—AI that processes multiple input types simultaneously. Why multimodal models outperform single-modality systems, demonstrated through a hypothetical chatCAT that helps owners understand their cats. Solo episode on multimodal architecture.

    To stay in touch, sign up for our newsletter at https://www.superprompt.fm

    Show More Show Less
    10 mins
  • Google Gemini: Three Models, One Strategy
    May 13 2024

    Google's Gemini family of multimodal AI models compared to OpenAI equivalents. What Nano, Pro, and Ultra each do, how they compare to GPT-3.5, GPT-4, and ChatGPT, and what "multimodal" means in practice. Solo episode on Google's LLM strategy.

    https://gemini.google.com. (Note: I am not sponsored by Google.)

    To stay in touch, sign up for our newsletter at https://www.superprompt.fm

    Show More Show Less
    7 mins