Skip to main content

Groq

Integrate Groq models through the Adaline Proxy to automatically capture telemetry — requests, responses, token usage, latency, and costs — with minimal code changes. Groq uses an OpenAI-compatible API.

Supported Models

Chat Models
ModelDescription
openai/gpt-oss-120bOpenAI GPT OSS 120B on Groq LPU
openai/gpt-oss-20bOpenAI GPT OSS 20B on Groq LPU
openai/gpt-oss-safeguard-20bOpenAI GPT OSS Safeguard 20B on Groq LPU
moonshotai/kimi-k2-instructMoonshot Kimi K2
moonshotai/kimi-k2-instruct-0905Moonshot Kimi K2 September 2025 snapshot
meta-llama/llama-4-maverick-17b-128e-instructLlama 4 Maverick
meta-llama/llama-4-scout-17b-16e-instructLlama 4 Scout
meta-llama/llama-guard-4-12bLlama Guard 4 (safety)
qwen/qwen3-32bQwen 3 32B
deepseek-r1-distill-llama-70bDeepSeek R1 Distill 70B
llama-3.3-70b-versatileLlama 3.3 70B
llama-3.1-8b-instantLlama 3.1 8B, ultra-fast inference
gemma2-9b-itGoogle Gemma 2 9B

Proxy Base URL

https://gateway.adaline.ai/v1/groq/

Prerequisites

  1. A Groq API key
  2. An Adaline API key, project ID, and prompt ID

Chat Completions

Complete Chat

from groq import Groq

client = Groq(
    api_key="your-groq-api-key",
    base_url="https://gateway.adaline.ai/v1/groq/"
)

headers = {
    "adaline-api-key": "your-adaline-api-key",
    "adaline-project-id": "your-project-id",
    "adaline-prompt-id": "your-prompt-id"
}

response = client.chat.completions.create(
    model="llama3-8b-8192",
    messages=[
        {"role": "system", "content": "You are a helpful assistant."},
        {"role": "user", "content": "What is the speed of light?"}
    ],
    extra_headers=headers
)

print(response.choices[0].message.content)

Stream Chat

from groq import Groq

client = Groq(
    api_key="your-groq-api-key",
    base_url="https://gateway.adaline.ai/v1/groq/"
)

headers = {
    "adaline-api-key": "your-adaline-api-key",
    "adaline-project-id": "your-project-id",
    "adaline-prompt-id": "your-prompt-id"
}

stream = client.chat.completions.create(
    model="llama3-8b-8192",
    messages=[
        {"role": "system", "content": "You are a helpful assistant."},
        {"role": "user", "content": "Explain how neural networks work."}
    ],
    stream=True,
    extra_headers=headers
)

for chunk in stream:
    if chunk.choices[0].delta.content is not None:
        print(chunk.choices[0].delta.content, end="")

Next Steps


Back to Integrations

Browse all integrations