Skip to main content
The Iterate pillar is where product and engineering teams come together to design, develop, and refine prompts in a collaborative environment. It covers the entire prompt creation lifecycle — from selecting a model and composing messages through adding dynamic variables and testing in a live playground. The workflow is: configure your model, compose your prompt with roles and content, wire up variables and tools, then run and iterate in the Playground until you are satisfied with the output. The iterate pillar in Adaline

Model parameters

Every prompt starts with a model. The Editor displays all supported LLMs based on the providers configured in your workspace — OpenAI, Anthropic, Google, and others. Once you select a model, you can fine-tune generation settings like temperature, max tokens, top-p, and frequency/presence penalties to control how the model behaves. You can also configure a response format to constrain the model’s output to free-form text, a JSON object, or a strict JSON schema. Model settings in Adaline Use Parameters in Prompts covers model selection, all generation settings, and response format configuration in detail.

Prompt composition

Prompts in Adaline are built from messages. Each message has a role — system, user, assistant, or tool — and a content block that holds the actual payload. The role determines how the model treats the message, while the content block holds what is sent — whether that is text, images, PDFs, or a combination.
Content typeWhat it supports
TextThe foundational content type. Write instructions, add /* comments */ for team annotations (stripped before sending), and embed {{variables}} for dynamic inputs.
ImagesAttach images via upload, URL, or image variables for vision-capable models. Each message can contain multiple images.
PDFsAttach PDF documents as context for analysis, summarization, or extraction tasks.
Composing prompts in the Adaline Editor Once you are comfortable with basic prompt composition, Multi-shot Prompting covers how to teach the model a specific output format by providing example input/output pairs using the user and assistant roles.

Variables

Variables make your prompts dynamic and reusable. Define a variable by wrapping its name in double curly braces — {{variable_name}} — anywhere in your message text, and it automatically appears in the Variable Editor where you can set its value. Beyond static text and image variables, Adaline supports two advanced variable types that unlock powerful workflows:
  • API variables — Fetch live data from external HTTP endpoints at runtime. Configure an API source with a URL, method, headers, and body, and the response is injected directly into your prompt context when the prompt executes.
  • Prompt variables — Chain prompts together by using the output of one prompt as the input to another. This enables modular, agent-like workflows where each step can be independently authored, tested, and refined.
The Variable Editor in Adaline Use Variables in Prompts covers variable creation, naming rules, all variable types, and the Variable Editor interface.

Tools and MCP

Tool configuration in Adaline Tools let your LLM interact with external services, databases, and APIs during a conversation. When the model determines it needs external data or actions, it generates a structured tool call request that can be executed to fetch results and continue the conversation. You define tools using JSON schemas and optionally configure HTTP backends for automatic execution. For standardized access to external data sources and services, Adaline also supports the Model Context Protocol (MCP). Connect to remote MCP servers and their tools become available to the model alongside any custom tools you have defined — no additional backend code required. Use Tools in Prompts covers tool definition, JSON schemas, HTTP backend configuration, and the full tool calling workflow.

Playground

The Adaline Playground The Playground is your interactive testing sandbox. It lets you execute prompts with specific inputs, view model responses in real time, and iterate rapidly without leaving the Editor. You can add messages interactively, switch between models, and compare outputs side by side. The Playground supports the full testing lifecycle:
  • Run prompts — Execute your prompt, add follow-up messages, and switch between models to compare how different LLMs handle the same input.
  • View past runs — Access a versioned history of every playground execution. Restore any previous state to recover a good result or compare outputs across configurations.
  • Test tool calls — Handle tool call responses manually or enable auto tool calls for fully automated multi-turn conversations with tool-equipped prompts and MCP servers.
  • Link datasets — Connect datasets to your prompt to cycle through structured variable samples at scale. Each row in the dataset represents a different test case — select a row and run to test your prompt with that specific combination.

Use Parameters in Prompts

Select an LLM and fine-tune its generation settings.

Use Roles in Prompts

Structure prompts with role-based messages.

Use Variables in Prompts

Create dynamic, reusable prompt templates.

Run Prompts in Playground

Test your prompts interactively with real inputs.