Together AI
Integrate Together AI open-source models through the Adaline Proxy to automatically capture telemetry — requests, responses, token usage, latency, and costs — with minimal code changes.Supported Models
Together AI accepts any model name available on the Together AI platform. Popular models include: Chat Models| Model | Description |
|---|---|
meta-llama/Llama-3.1-405B-Instruct-Turbo | Llama 3.1 405B |
meta-llama/Llama-3.1-70B-Instruct-Turbo | Llama 3.1 70B |
meta-llama/Llama-3.1-8B-Instruct-Turbo | Llama 3.1 8B |
mistralai/Mixtral-8x7B-Instruct-v0.1 | Mixtral 8x7B MoE |
Qwen/Qwen2-72B-Instruct | Qwen 2 72B |
deepseek-ai/DeepSeek-R1 | DeepSeek R1 reasoning model |
Together AI uses a flexible model routing system — any model available on the Together AI models page can be used.
Proxy Base URL
Prerequisites
- A Together AI API key
- An Adaline API key, project ID, and prompt ID
Chat Completions
Complete Chat
Together AI’s Python SDK uses
supplied_headers for persistent headers set at client initialization, and extra_headers for per-request headers.Stream Chat
Next Steps
- Multi-Step Workflows — RAG pipelines, multi-step generation, and conversational agents
- Headers Reference — Complete header documentation
Back to Integrations
Browse all integrations