Linear Digressions

SHAP: Shapley Values in Machine Learning

05.13.2018 - By Ben Jaffe and Katie MalonePlay

Download our free app to listen on your phone

Download on the App StoreGet it on Google Play

Shapley values in machine learning are an interesting and useful enough innovation that we figured hey, why not do a two-parter? Our last episode focused on explaining what Shapley values are: they define a way of assigning credit for outcomes across several contributors, originally to understand how impactful different actors are in building coalitions (hence the game theory background) but now they're being cross-purposed for quantifying feature importance in machine learning models. This episode centers on the computational details that allow Shapley values to be approximated quickly, and a new package called SHAP that makes all this innovation accessible.

More episodes from Linear Digressions