A nicely illustrated breakdown of how the Transformer (“Attention is All You Need”) neural network model and how it works. I’m impressed with results I’ve seen from it so far. (as a product it’s the Tensor2Tensor code) – self-attention means it continually looks at it itself while it is processing for contextual clues at various ranges within itself. http://jalammar.github.io/illustrated-transformer/

A nicely illustrated breakdown of how the Transformer (“Attention is All You Need”) neural network model and how it works.
 
I’m impressed with results I’ve seen from it so far. (as a product it’s the Tensor2Tensor code) – self-attention means it continually looks at it itself while it is processing for contextual clues at various ranges within itself.
 
http://jalammar.github.io/illustrated-transformer/
====

Leave a comment

Your email address will not be published. Required fields are marked *


+ 7 = twelve

Leave a Reply