Rotary Positional Embeddings

Concept

Embeddings used in GPT OSS that encode token position directly into the attention mechanism to support longer contexts.

Mentioned in 1 video