
Sign up to save your podcasts
Or


Hey guys, in this episode I finally explain the Transformers network architecture! The paper Attention is all you need proposed the Transformer network and it was groundbreaking for firstly NLP field and now for all the Deep Learning fields. In the episode I explain the attention, the self-attention and the multi-head attention mechanisms for both Transformers encoder and Decoder, and also the positional encoding. Go listen to this episode because it's probably my best technical episode!
Original paper: https://arxiv.org/pdf/1706.03762.pdf
Self-attention GitHub code: https://github.com/filipelauar/projects/blob/main/self_attention.ipynb
Youtube video explaining the architecture: https://www.youtube.com/watch?v=TQQlZhbC5ps
Nice blog post with code 1: http://peterbloem.nl/blog/transformers
Nice blog post with code 2: https://nlp.seas.harvard.edu/2018/04/03/attention.html
Instagram: https://www.instagram.com/podcast.lifewithai/
Linkedin: https://www.linkedin.com/company/life-with-ai
By Filipe Lauar5
22 ratings
Hey guys, in this episode I finally explain the Transformers network architecture! The paper Attention is all you need proposed the Transformer network and it was groundbreaking for firstly NLP field and now for all the Deep Learning fields. In the episode I explain the attention, the self-attention and the multi-head attention mechanisms for both Transformers encoder and Decoder, and also the positional encoding. Go listen to this episode because it's probably my best technical episode!
Original paper: https://arxiv.org/pdf/1706.03762.pdf
Self-attention GitHub code: https://github.com/filipelauar/projects/blob/main/self_attention.ipynb
Youtube video explaining the architecture: https://www.youtube.com/watch?v=TQQlZhbC5ps
Nice blog post with code 1: http://peterbloem.nl/blog/transformers
Nice blog post with code 2: https://nlp.seas.harvard.edu/2018/04/03/attention.html
Instagram: https://www.instagram.com/podcast.lifewithai/
Linkedin: https://www.linkedin.com/company/life-with-ai