positional encoding

scroll ↓ to Resources

Note

  • Adds information about the position of each token in the sequence to help the transformer understand word order, because attention is position-invariant
  • see decoder architecture
  • in simplest form, position encoding vectors are added to input sequences

Rotary Position Embeddings (RoPE)

Resources


table file.inlinks, file.outlinks from [[]] and !outgoing([[]])  AND -"Changelog"