Jamba

Software / AppMentioned in 1 video

A hybrid Mixture-of-Experts (MoE) model trained by AI2, identified as a state-of-the-art non-Transformer architecture.