Skip to main content

Models

Every agent runs on a model. The platform supports multiple LLM providers — Anthropic Claude, OpenAI GPT, Google Gemini, and others. You configure which models are available, and each agent picks one.

Why model choice matters

Models differ in:

  • Quality — how good the responses are
  • Speed — how fast a response comes back
  • Cost — how much each response costs
  • Tool use — how reliably the model picks and calls tools
  • Context size — how much input the model can handle in one go

A simple FAQ agent doesn't need the most expensive model. A complex multi-step analyst agent probably does. The platform lets you mix.

Configuring models

Go to Settings → Models. You'll see all models the tenant has access to, grouped by provider (OpenAI, Anthropic, etc.) with their status (Connected / Disconnected) and a default model marker.

AI Models page: default model card on top, OpenAI and Anthropic providers each with per-model toggles and Set Default actions

To enable a new model:

  1. Click Add Model next to the provider
  2. Enter the Model ID (the exact identifier used by the provider's API, e.g. claude-opus-4-7, gpt-5, gemini-3-pro)
  3. Set a friendly Display Name and optional description
  4. Save

Add Custom Model dialog: provider picker, Model ID, Display Name, Description

Once saved, the model appears in the model dropdown when you create or edit agents.

Switching an agent's model

Open the agent, change the model in the dropdown, save. The next run uses the new model. Behaviour can shift noticeably — re-test in the sandbox before publishing the change.

Model-specific features

Some models support features others don't:

  • Tool use — most modern models support it; quality varies
  • Vision — accepting images as input
  • Long context — 200k+ token windows
  • Streaming — token-by-token output (used by public chat for responsiveness)
  • Reasoning — exposing the model's chain of thought

The model picker shows which features each model supports. Pick the one that fits your agent's job.

Default model

Set a tenant default. New agents start with that model. Individual agents can still override.

Cost guard rails

Set spend caps per model and per tenant. When the cap is reached, agents using that model pause until you raise the limit or switch models.

This prevents runaway costs from a misbehaving agent or a webhook that fires too often.

Bring-your-own-key vs platform-managed

Two ways to get model access:

  • Bring your own key — you provide an API key from the provider. You pay them directly.
  • Platform-managed — the platform provides the model and bills you for usage. Simpler, sometimes more expensive.

You can mix per provider: BYO key for Anthropic, platform-managed for OpenAI.