Jamba
Software / AppMentioned in 1 video
A hybrid Mixture-of-Experts (MoE) model trained by AI2, identified as a state-of-the-art non-Transformer architecture.
A hybrid Mixture-of-Experts (MoE) model trained by AI2, identified as a state-of-the-art non-Transformer architecture.