Attention, Attention, Attention
I have wanted to take some time and learn about transformers for a while. They seems to be revolutionizing every major field within AI, so I could no longer afford to stay away.
The other day I decided to start from the beginning and went all in into the seminal paperĀ Attention is all you need.
The paper is quite interesting, but I failed to grasp the intuition behind the Qs, Ks, and Vs in there. I went looking elsewhere for a good explanation of what each of those was, but most articles just weren't clear enough.
Alas, at last, I found what I was looking for. The following YouTube series of 4 videos goes in depth about how the attention layer works and I recommend it to anyone trying to get started with transformers.
Update: 9th March, 2023 #
I have found another great resource for learning about transformers. This YouTube video provides a clear and thorough explanation of the attention mechanism and how it is used in transformer models. I highly recommend it to anyone looking to learn more about this topic.
- Next: Down with Covid