Jamba
Software / App
A hybrid Mixture-of-Experts (MoE) model trained by AI2, identified as a state-of-the-art non-Transformer architecture.
Mentioned in 1 video
A hybrid Mixture-of-Experts (MoE) model trained by AI2, identified as a state-of-the-art non-Transformer architecture.