cover of episode Kernels!

Kernels!

2020/9/18
logo of podcast Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)

Frequently requested episodes will be transcribed first

Shownotes Transcript

Today Yannic Lightspeed Kilcher and I spoke with Alex Stenlake about Kernel Methods. What is a kernel? Do you remember those weird kernel things which everyone obsessed about before deep learning? What about Representer theorem and reproducible kernel hilbert spaces? SVMs and kernel ridge regression? Remember them?! Hope you enjoy the conversation!

00:00:00 Tim Intro

00:01:35 Yannic clever insight from this discussion 

00:03:25 Street talk and Alex intro 

00:05:06 How kernels are taught

00:09:20 Computational tractability

00:10:32 Maths 

00:11:50 What is a kernel? 

00:19:39 Kernel latent expansion 

00:23:57 Overfitting 

00:24:50 Hilbert spaces 

00:30:20 Compare to DL

00:31:18 Back to hilbert spaces

00:45:19 Computational tractability 2

00:52:23 Curse of dimensionality

00:55:01 RBF: infinite taylor series

00:57:20 Margin/SVM 

01:00:07 KRR/dual

01:03:26 Complexity compute kernels vs deep learning

01:05:03 Good for small problems? vs deep learning)

01:07:50 Whats special about the RBF kernel

01:11:06 Another DL comparison

01:14:01 Representer theorem

01:20:05 Relation to back prop

01:25:10 Connection with NLP/transformers

01:27:31 Where else kernels good

01:34:34 Deep learning vs dual kernel methods

01:33:29 Thoughts on AI

01:34:35 Outro