
Sign up to save your podcasts
Or


Send us a text
You might be using AI models in pathology without even knowing if they’re giving you reliable results.
 
Let that sink in for a second—because today, we’re fixing that.
In this episode, I walk you through the real statistics that power—and sometimes fail—AI in digital pathology. It's episode 4 of our AI series, and we’re demystifying the metrics behind both generative and non-generative AI. Why does this matter? Because accuracy isn't enough. And not every model metric tells you the whole story.
If you’ve ever been impressed by a model’s "99% accuracy," you need to hear why that might actually be a red flag. I share personal stories (yes, including my early days in Germany when I didn’t even know what a "training set" was), and we break down confusing metrics like perplexity, SSIM, FID, and BLEU scores—so you can truly understand what your models are doing and how to evaluate them correctly.
Together, we’ll uncover how model evaluation works for:
Whether you're a pathologist, a scientist, or someone leading a digital transformation team—you need this knowledge to avoid misleading data, flawed models, and missed opportunities.
🕒 EPISODE HIGHLIGHTS WITH TIMESTAMPS
📘 RESOURCE FROM THIS EPISODE:
🔗 Read the full paper discussed in this episode:
 "Statistics of generative and non-generative artificial intelligence models in medicine"
💬 Final Thoughts
Statistical literacy isn’t optional anymore—especially in digital pathology. AI isn’t just a buzzword; it’s a tool, and if we want to lead this field forward, we must understand the systems we rely on. This episode will help you become not just a user, but a better steward of AI.
🎙️ Tune in now and let's keep trailblazing—together.
Support the show
Get the "Digital Pathology 101" FREE E-book and join us!
 By Aleksandra Zuraw, DVM, PhD
By Aleksandra Zuraw, DVM, PhD5
77 ratings
Send us a text
You might be using AI models in pathology without even knowing if they’re giving you reliable results.
 
Let that sink in for a second—because today, we’re fixing that.
In this episode, I walk you through the real statistics that power—and sometimes fail—AI in digital pathology. It's episode 4 of our AI series, and we’re demystifying the metrics behind both generative and non-generative AI. Why does this matter? Because accuracy isn't enough. And not every model metric tells you the whole story.
If you’ve ever been impressed by a model’s "99% accuracy," you need to hear why that might actually be a red flag. I share personal stories (yes, including my early days in Germany when I didn’t even know what a "training set" was), and we break down confusing metrics like perplexity, SSIM, FID, and BLEU scores—so you can truly understand what your models are doing and how to evaluate them correctly.
Together, we’ll uncover how model evaluation works for:
Whether you're a pathologist, a scientist, or someone leading a digital transformation team—you need this knowledge to avoid misleading data, flawed models, and missed opportunities.
🕒 EPISODE HIGHLIGHTS WITH TIMESTAMPS
📘 RESOURCE FROM THIS EPISODE:
🔗 Read the full paper discussed in this episode:
 "Statistics of generative and non-generative artificial intelligence models in medicine"
💬 Final Thoughts
Statistical literacy isn’t optional anymore—especially in digital pathology. AI isn’t just a buzzword; it’s a tool, and if we want to lead this field forward, we must understand the systems we rely on. This episode will help you become not just a user, but a better steward of AI.
🎙️ Tune in now and let's keep trailblazing—together.
Support the show
Get the "Digital Pathology 101" FREE E-book and join us!

26,175 Listeners

3,844 Listeners

318 Listeners

496 Listeners

764 Listeners

302 Listeners

769 Listeners

233 Listeners

324 Listeners

85 Listeners

59,392 Listeners

16 Listeners

34 Listeners

4 Listeners

19 Listeners