Compound-beta-mini leverages Llama 3.3 70B to automatically use tools. Unlike compound-beta, it can only use one tool per request.
FEATURE | VALUE |
---|---|
Context Window (Tokens) | 128K |
Max Output Tokens | 8,192 |
Max File Size | N/A |
Token Generation Speed | 275 tokens per second |
Input Token Price | Varies. Pricing is based on input tokens to underlying models (while in preview) |
Output Token Price | Varies. Pricing is based on output tokens from underlying models (while in preview) |
Tool Use | |
Agentic Tooling | |
JSON Mode | |
Image Support |
Automatically access up-to-date information from the web using the built-in web search tool powered by Tavily.
Execute Python code automatically using the code execution tool powered by E2B.
Experience the capabilities of compound-beta-mini
on Groq:
pip install groq
1from groq import Groq
2client = Groq()
3completion = client.chat.completions.create(
4 model="compound-beta-mini",
5 messages=[
6 {
7 "role": "user",
8 "content": "Explain why fast inference is critical for reasoning models"
9 }
10 ]
11)
12print(completion.choices[0].message.content)