Vanishing Gradients

Episode 48: HOW TO BENCHMARK AGI WITH GREG KAMRADT


Listen Later

If we want to make progress toward AGI, we need a clear definition of intelligence—and a way to measure it.

In this episode, Hugo talks with Greg Kamradt, President of the ARC Prize Foundation, about ARC-AGI: a benchmark built on Francois Chollet’s definition of intelligence as “the efficiency at which you learn new things.” Unlike most evals that focus on memorization or task completion, ARC is designed to measure generalization—and expose where today’s top models fall short.

They discuss:

🧠 Why we still lack a shared definition of intelligence
🧪 How ARC tasks force models to learn novel skills at test time
📉 Why GPT-4-class models still underperform on ARC
🔎 The limits of traditional benchmarks like MMLU and Big-Bench
⚙️ What the OpenAI O₃ results reveal—and what they don’t
💡 Why generalization and efficiency, not raw capability, are key to AGI

Greg also shares what he’s seeing in the wild: how startups and independent researchers are using ARC as a North Star, how benchmarks shape the frontier, and why the ARC team believes we’ll know we’ve reached AGI when humans can no longer write tasks that models can’t solve.

This conversation is about evaluation—not hype. If you care about where AI is really headed, this one’s worth your time.

LINKS

  • ARC Prize -- What is ARC-AGI?
  • On the Measure of Intelligence by François Chollet
  • Greg Kamradt on Twitter
  • Hugo's High Signal Podcast with Fei-Fei Li
  • Vanishing Gradients YouTube Channel
  • Upcoming Events on Luma
  • Hugo's recent newsletter about upcoming events and more!
  • Watch the podcast here on YouTube!
  • 🎓 Want to go deeper?

    Check out Hugo's course: Building LLM Applications for Data Scientists and Software Engineers.
    Learn how to design, test, and deploy production-grade LLM systems — with observability, feedback loops, and structure built in.
    This isn’t about vibes or fragile agents. It’s about making LLMs reliable, testable, and actually useful.

    Includes over $800 in compute credits and guest lectures from experts at DeepMind, Moderna, and more.

    Cohort starts July 8 — Use this link for a 10% discount

    ...more
    View all episodesView all episodes
    Download on the App Store

    Vanishing GradientsBy Hugo Bowne-Anderson

    • 5
    • 5
    • 5
    • 5
    • 5

    5

    11 ratings


    More shows like Vanishing Gradients

    View all
    a16z Podcast by Andreessen Horowitz

    a16z Podcast

    1,032 Listeners

    Data Skeptic by Kyle Polich

    Data Skeptic

    480 Listeners

    The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) by Sam Charrington

    The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

    441 Listeners

    Super Data Science: ML & AI Podcast with Jon Krohn by Jon Krohn

    Super Data Science: ML & AI Podcast with Jon Krohn

    298 Listeners

    NVIDIA AI Podcast by NVIDIA

    NVIDIA AI Podcast

    322 Listeners

    DataFramed by DataCamp

    DataFramed

    267 Listeners

    Practical AI by Practical AI LLC

    Practical AI

    192 Listeners

    Google DeepMind: The Podcast by Hannah Fry

    Google DeepMind: The Podcast

    198 Listeners

    Machine Learning Street Talk (MLST) by Machine Learning Street Talk (MLST)

    Machine Learning Street Talk (MLST)

    88 Listeners

    Dwarkesh Podcast by Dwarkesh Patel

    Dwarkesh Podcast

    408 Listeners

    No Priors: Artificial Intelligence | Technology | Startups by Conviction

    No Priors: Artificial Intelligence | Technology | Startups

    121 Listeners

    Latent Space: The AI Engineer Podcast by swyx + Alessio

    Latent Space: The AI Engineer Podcast

    75 Listeners

    AI + a16z by a16z

    AI + a16z

    31 Listeners

    High Signal: Data Science | Career | AI by Delphina

    High Signal: Data Science | Career | AI

    4 Listeners

    OpenAI Podcast by OpenAI

    OpenAI Podcast

    28 Listeners