26.1 AI Podcast

Dr. Rachael Tatman: Data Scientist at Kaggle, Computational Sociolinguistics


Listen Later

This week’s episode may be our most cerebral to date, probably thanks to hosting our first University of Washington Husky, Ph.D. graduate. Dr. Rachael Tatman shares snippets of her experience at Kaggle, stochastic approaches to ML models, errors in ML models, understanding prediction, and importance of reproducibility.
A big takeaway references Dr. Tatman’s PyCon talk, “Put down the deep learning: When not to use neural networks and what to do instead.” We explore how many businesses benefit from a simple linear regression model instead of investing millions of dollars of compute time to a deep learning approach.
Included with our conversation with Dr. Tatman is a discussion of linguistics and will Don ever get accurate translations of his Korean friends’ tweets in real-time. So far, our conclusion is that we’re far away from the singularity.
Episode seven of 26.1 AI Podcast delivers lots of value for business leaders contemplating their future AI strategy.
...more
View all episodesView all episodes
Download on the App Store

26.1 AI PodcastBy Brian Ray; Don Sheu

  • 5
  • 5
  • 5
  • 5
  • 5

5

9 ratings