The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Rethinking Model Size: Train Large, Then Compress with Joseph Gonzalez - #378

05.25.2020 - By Sam CharringtonPlay

Download our free app to listen on your phone

Download on the App StoreGet it on Google Play

Today we’re joined by Joseph Gonzalez, Assistant Professor in the EECS department at UC Berkeley.

In our conversation, we explore Joseph’s paper “Train Large, Then Compress: Rethinking Model Size for Efficient Training and Inference of Transformers,” which looks at compute-efficient training strategies for models. We discuss the two main problems being solved; 1) How can we rapidly iterate on variations in architecture? And 2) If we make models bigger, is it really improving any efficiency?

More episodes from The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)