FEATURE | VALUE |
---|---|
Context Window (Tokens) | 1M tokens (currently limited to 128K) |
Max Output Tokens | N/A but recommended image input up to 5 only for highest accuracy |
Max File Size | N/A |
Token Generation Speed | N/A |
Input Token Price | 0.20 per million tokens |
Output Token Price | 0.60 per million tokens |
Tool Use | |
JSON Mode | |
Image Support |
Experience the capabilities of meta-llama/llama-4-maverick-17b-128e-instruct
on Groq:
pip install groq
1from groq import Groq
2client = Groq()
3completion = client.chat.completions.create(
4 model="meta-llama/llama-4-maverick-17b-128e-instruct",
5 messages=[
6 {
7 "role": "user",
8 "content": "Explain why fast inference is critical for reasoning models"
9 }
10 ]
11)
12print(completion.choices[0].message.content)