Introducing the Official
Llama API

Straight from Meta. Accelerated by Groq.

The fastest way to run the world’s most trusted openly available models with no tradeoffs.
Served directly on the most efficient inference chip.
and start building!
Not a wrapper. Not a copy. It’s the real thing, served directly from Meta and accelerated by Groq's purpose-built inference hardware
Llama 4 and more, available instantly with zero setup.