
Sign up to save your podcasts
Or


Send us a text
You might be using AI models in pathology without even knowing if they’re giving you reliable results.
Let that sink in for a second—because today, we’re fixing that.
In this episode, I walk you through the real statistics that power—and sometimes fail—AI in digital pathology. It's episode 4 of our AI series, and we’re demystifying the metrics behind both generative and non-generative AI. Why does this matter? Because accuracy isn't enough. And not every model metric tells you the whole story.
If you’ve ever been impressed by a model’s "99% accuracy," you need to hear why that might actually be a red flag. I share personal stories (yes, including my early days in Germany when I didn’t even know what a "training set" was), and we break down confusing metrics like perplexity, SSIM, FID, and BLEU scores—so you can truly understand what your models are doing and how to evaluate them correctly.
Together, we’ll uncover how model evaluation works for:
Whether you're a pathologist, a scientist, or someone leading a digital transformation team—you need this knowledge to avoid misleading data, flawed models, and missed opportunities.
🕒 EPISODE HIGHLIGHTS WITH TIMESTAMPS
📘 RESOURCE FROM THIS EPISODE:
🔗 Read the full paper discussed in this episode:
"Statistics of generative and non-generative artificial intelligence models in medicine"
💬 Final Thoughts
Statistical literacy isn’t optional anymore—especially in digital pathology. AI isn’t just a buzzword; it’s a tool, and if we want to lead this field forward, we must understand the systems we rely on. This episode will help you become not just a user, but a better steward of AI.
🎙️ Tune in now and let's keep trailblazing—together.
Support the show
Get the "Digital Pathology 101" FREE E-book and join us!
By Aleksandra Zuraw, DVM, PhD5
77 ratings
Send us a text
You might be using AI models in pathology without even knowing if they’re giving you reliable results.
Let that sink in for a second—because today, we’re fixing that.
In this episode, I walk you through the real statistics that power—and sometimes fail—AI in digital pathology. It's episode 4 of our AI series, and we’re demystifying the metrics behind both generative and non-generative AI. Why does this matter? Because accuracy isn't enough. And not every model metric tells you the whole story.
If you’ve ever been impressed by a model’s "99% accuracy," you need to hear why that might actually be a red flag. I share personal stories (yes, including my early days in Germany when I didn’t even know what a "training set" was), and we break down confusing metrics like perplexity, SSIM, FID, and BLEU scores—so you can truly understand what your models are doing and how to evaluate them correctly.
Together, we’ll uncover how model evaluation works for:
Whether you're a pathologist, a scientist, or someone leading a digital transformation team—you need this knowledge to avoid misleading data, flawed models, and missed opportunities.
🕒 EPISODE HIGHLIGHTS WITH TIMESTAMPS
📘 RESOURCE FROM THIS EPISODE:
🔗 Read the full paper discussed in this episode:
"Statistics of generative and non-generative artificial intelligence models in medicine"
💬 Final Thoughts
Statistical literacy isn’t optional anymore—especially in digital pathology. AI isn’t just a buzzword; it’s a tool, and if we want to lead this field forward, we must understand the systems we rely on. This episode will help you become not just a user, but a better steward of AI.
🎙️ Tune in now and let's keep trailblazing—together.
Support the show
Get the "Digital Pathology 101" FREE E-book and join us!

229,277 Listeners

323 Listeners

126 Listeners

498 Listeners

310 Listeners

302 Listeners

499 Listeners

125 Listeners

320 Listeners

20 Listeners

95 Listeners

369 Listeners

16 Listeners

148 Listeners

4 Listeners