Fast LLM inference, OpenAI-compatible. Simple to integrate, easy to scale. Start building in minutes.
import OpenAI from "openai";
const client = new OpenAI({
apiKey: process.env.GROQ_API_KEY,
baseURL: "https://api.groq.com/openai/v1",
});
const response = await client.responses.create({
model: "openai/gpt-oss-20b",
input: "Explain the importance of fast language models",
});
console.log(response.output_text);
from openai import OpenAI
import os
client = OpenAI(
api_key=os.environ.get("GROQ_API_KEY"),
base_url="https://api.groq.com/openai/v1",
)
response = client.responses.create(
input="Explain the importance of fast language models",
model="openai/gpt-oss-20b",
)
print(response.output_text)
curl -X POST https://api.groq.com/openai/v1/responses \
-H "Authorization: Bearer $GROQ_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "openai/gpt-oss-20b",
"input": "Explain the importance of fast language models"
}'