LLaMA CPP

Software / App

A framework enabling on-device inference for small models.

Mentioned in 3 videos