Jamba

Software / App

A hybrid Mixture-of-Experts (MoE) model trained by AI2, identified as a state-of-the-art non-Transformer architecture.

Mentioned in 1 video