Fast LLM inference, OpenAI-compatible. Simple to integrate, easy to scale. Start building in minutes.
curl https://api.groq.com/openai/v1/chat/completions -s \
-H "Content-Type: application/json" \
-H "Authorization: Bearer $GROQ_API_KEY" \
-d '{
"model": "llama-3.3-70b-versatile",
"messages": [{
"role": "user",
"content": "Explain the importance of fast language models"
}]
}'
Essential resources to accelerate your development and maximize productivity
We’re adding new models all the time and will let you know when a new one comes online. See full details on our Models page.