Self-Attention

Concept

A variant of attention where a model attends to different positions of a single sequence to compute a representation of the same sequence, central to the Transformer architecture.

Mentioned in 2 videos

Save the 2 videos on Self-Attention to your own pod.

Sign up free to keep building your knowledge base on Self-Attention as more episodes are added.

Get Started Free