Rotary Positional Embeddings
Concept
Embeddings used in GPT OSS that encode token position directly into the attention mechanism to support longer contexts.
Mentioned in 1 video
Embeddings used in GPT OSS that encode token position directly into the attention mechanism to support longer contexts.