
Sign up to save your podcasts
Or


To trust something, you need to understand it. And, to understand something, someone often has to explain it. When it comes to AI, explainability can be a real challenge (definitionally, a "black box" is unexplainable)! With AI getting new levels of press and prominence thanks to the explosion of generative AI platforms, the need for explainability continues to grow. But, it's just as important in more conventional situations. Dr. Janet Bastiman, the Chief Data Scientist at Napier, joined Moe and Tim to, well, explain the topic! For complete show notes, including links to items mentioned in this episode and a transcript of the show, visit the show page.
By Michael Helbling, Moe Kiss, Tim Wilson, Val Kroll, and Julie Hoyer4.8
166166 ratings
To trust something, you need to understand it. And, to understand something, someone often has to explain it. When it comes to AI, explainability can be a real challenge (definitionally, a "black box" is unexplainable)! With AI getting new levels of press and prominence thanks to the explosion of generative AI platforms, the need for explainability continues to grow. But, it's just as important in more conventional situations. Dr. Janet Bastiman, the Chief Data Scientist at Napier, joined Moe and Tim to, well, explain the topic! For complete show notes, including links to items mentioned in this episode and a transcript of the show, visit the show page.

1,940 Listeners

1,085 Listeners

168 Listeners

302 Listeners

144 Listeners

3,986 Listeners

226 Listeners

269 Listeners

210 Listeners

135 Listeners

94 Listeners

225 Listeners

607 Listeners

58 Listeners

96 Listeners