#48 - Attention is all you need. Understanding Transformers.
Life with AI - A podcast by Filipe Lauar - Thursdays
Categories:
Hey guys, in this episode I finally explain the Transformers network architecture! The paper Attention is all you need proposed the Transformer network and it was groundbreaking for firstly NLP field and now for all the Deep Learning fields. In the episode I explain the attention, the self-attention and the multi-head attention mechanisms for both Transformers encoder and Decoder, and also the positional encoding. Go listen to this episode because it's probably my best technical episode! Original paper: https://arxiv.org/pdf/1706.03762.pdf Self-attention GitHub code: https://github.com/filipelauar/projects/blob/main/self_attention.ipynb Youtube video explaining the architecture: https://www.youtube.com/watch?v=TQQlZhbC5ps Nice blog post with code 1: http://peterbloem.nl/blog/transformers Nice blog post with code 2: https://nlp.seas.harvard.edu/2018/04/03/attention.html Instagram: https://www.instagram.com/podcast.lifewithai/ Linkedin: https://www.linkedin.com/company/life-with-ai