
Sign up to save your podcasts
Or


Why isn't Superman's suit Kryptonite-proof? This question reveals how large language models are trained. We break down transformers (the T in GPT), self-attention mechanisms, and the inference process—using Superman to explain why GPT-3 can generate coherent answers to questions it's never seen before. Solo episode on LLM architecture.
To stay in touch, sign up for our newsletter at https://www.superprompt.fm
By Tony Wan5
1717 ratings
Why isn't Superman's suit Kryptonite-proof? This question reveals how large language models are trained. We break down transformers (the T in GPT), self-attention mechanisms, and the inference process—using Superman to explain why GPT-3 can generate coherent answers to questions it's never seen before. Solo episode on LLM architecture.
To stay in touch, sign up for our newsletter at https://www.superprompt.fm

32,005 Listeners

26,217 Listeners

4,067 Listeners

1,092 Listeners

303 Listeners

334 Listeners

207 Listeners

9,927 Listeners

2,068 Listeners

197 Listeners

633 Listeners

616 Listeners

165 Listeners

466 Listeners

166 Listeners