FEATURE | VALUE |
---|---|
Context Window (Tokens) | 128K tokens (with a maximum of 5 image inputs) |
Max Output Tokens | N/A |
Max File Size | N/A |
Token Generation Speed | N/A |
Input Token Price | 0.11 per million tokens |
Output Token Price | 0.34 per million tokens |
Tool Use | |
JSON Mode | |
Image Support |
Experience the capabilities of meta-llama/llama-4-scout-17b-16e-instruct
on Groq:
pip install groq
from groq import Groq
client = Groq()
completion = client.chat.completions.create(
model="meta-llama/llama-4-scout-17b-16e-instruct",
messages=[
{
"role": "user",
"content": "Explain why fast inference is critical for reasoning models"
}
]
)
print(completion.choices[0].message.content)