Tensor Parallelism
Concept
A distributed training technique that splits model tensors across multiple devices.
Mentioned in 2 videos
Videos Mentioning Tensor Parallelism

A Comprehensive Overview of Large Language Models - Latent Space Paper Club
Latent Space
A distributed training technique that splits model tensors across multiple devices.

Stanford CS336 Language Modeling from Scratch | Spring 2026 | Lecture 7: Parallelism
Stanford Online
A distributed training technique where model layers are split across devices, requiring more communication but enabling larger models.