FEATURE | VALUE |
---|---|
Context Window (Tokens) | 8,192 |
Max Output Tokens | - |
Max File Size | - |
Token Generation Speed | 765 tps |
Input Token Price | $0.2/1M tokens |
Output Token Price | $0.2/1M tokens |
Tool Use | |
JSON Mode | |
Image Support |
Unlock the full potential of content moderation with Llama-Guard-3-8B - optimized for exceptional performance on Groq hardware now:
pip install groq
from groq import Groq
client = Groq()
completion = client.chat.completions.create(
model="llama-guard-3-8b",
messages=[
{
"role": "user",
"content": "Explain why fast inference is critical for reasoning models"
}
]
)
print(completion.choices[0].message.content)