Machine Learning Street Talk (MLST)

Download our free app to listen on your phone

Download on the App StoreGet it on Google Play

Today Yannic Lightspeed Kilcher and I spoke with Alex Stenlake about Kernel Methods. What is a kernel? Do you remember those weird kernel things which everyone obsessed about before deep learning? What about Representer theorem and reproducible kernel hilbert spaces? SVMs and kernel ridge regression? Remember them?! Hope you enjoy the conversation!

00:00:00 Tim Intro

00:01:35 Yannic clever insight from this discussion 

00:03:25 Street talk and Alex intro 

00:05:06 How kernels are taught

00:09:20 Computational tractability

00:10:32 Maths 

00:11:50 What is a kernel? 

00:19:39 Kernel latent expansion 

00:23:57 Overfitting 

00:24:50 Hilbert spaces 

00:30:20 Compare to DL

00:31:18 Back to hilbert spaces

00:45:19 Computational tractability 2

00:52:23 Curse of dimensionality

00:55:01 RBF: infinite taylor series

00:57:20 Margin/SVM 

01:00:07 KRR/dual

01:03:26 Complexity compute kernels vs deep learning

01:05:03 Good for small problems? vs deep learning)

01:07:50 Whats special about the RBF kernel

01:11:06 Another DL comparison

01:14:01 Representer theorem

01:20:05 Relation to back prop

01:25:10 Connection with NLP/transformers

01:27:31 Where else kernels good

01:34:34 Deep learning vs dual kernel methods

01:33:29 Thoughts on AI

01:34:35 Outro

More episodes from Machine Learning Street Talk (MLST)