Skip to content

Admin Dashboard: Models

The Models page lets you configure LLM providers and endpoints. Each model entry defines how the engine connects to an LLM backend — you can have multiple models from different providers and assign each agent its own model.

ProviderDescription
ollamaLocal model inference via Ollama. Free, private, no API key needed. Requires Ollama installed on the host.
openai_compatibleAny API that follows OpenAI chat completions format. Works with OpenAI, DeepInfra, Together, Groq, vLLM, LiteLLM.
anthropicNative Anthropic API. Supports Claude models with automatic message format conversion.
azure_openaiAzure-hosted OpenAI models. Uses deployment-based URLs and api-key header auth. Requires api_version.
googleGoogle Gemini models via the generateContent API. Uses x-goog-api-key header.
openrouterOpenRouter aggregator. Access 100+ models via a single API key. Base URL is preset automatically.
deepseekDeepSeek models (e.g. DeepSeek V3.2). OpenAI-compatible with preset base URL.
mistralMistral AI models (e.g. Mistral Medium 3). OpenAI-compatible with preset base URL.
xaixAI Grok models (e.g. Grok 4.1). OpenAI-compatible with preset base URL.
zaiZ.ai GLM models (e.g. GLM-5). OpenAI-compatible with preset base URL.

Click “Add Model” and fill in the fields:

  • Display name — a human-readable name used in the agent configuration dropdown.
  • Provider — select from the supported providers above.
  • Model name — the model identifier as expected by the provider API (e.g., llama3.2, claude-sonnet-4-20250514). For Azure OpenAI, this is the deployment name.
  • Base URL — custom endpoint URL. Required for Ollama and third-party providers. Leave empty for providers with preset URLs (OpenAI, Anthropic, OpenRouter, DeepSeek, Mistral, xAI, Z.ai). For Azure OpenAI, this is your Azure resource URL (e.g., https://my-company.openai.azure.com).
  • API Key — provider API key. Not needed for Ollama. Use the ${VAR} syntax when configuring via YAML.
  • API Version — (Azure OpenAI only) the API version string, e.g. 2024-10-21.
# Ollama (local, no API key needed)
models:
llama-local:
provider: ollama
model_name: llama3.2
base_url: "http://localhost:11434/v1"
# OpenAI-compatible (DeepInfra)
models:
qwen-3-32b:
provider: openai_compatible
model_name: Qwen/Qwen3-32B
base_url: "https://api.deepinfra.com/v1/openai"
api_key: ${DEEPINFRA_API_KEY}
# Anthropic
models:
claude-sonnet-4:
provider: anthropic
model_name: claude-sonnet-4-20250514
api_key: ${ANTHROPIC_API_KEY}
# Azure OpenAI (deployment-based)
models:
gpt4-azure:
provider: azure_openai
base_url: "https://my-company.openai.azure.com"
model_name: "gpt-4o-deploy"
api_version: "2024-10-21"
api_key: ${AZURE_OPENAI_KEY}
# Google Gemini
models:
gemini-pro:
provider: google
model_name: "gemini-3.1-pro"
api_key: ${GOOGLE_API_KEY}
# OpenRouter (base_url is preset automatically)
models:
openrouter-claude:
provider: openrouter
model_name: "anthropic/claude-sonnet-4-20250514"
api_key: ${OPENROUTER_API_KEY}
# DeepSeek (preset base_url)
models:
deepseek-v3:
provider: deepseek
model_name: "deepseek-chat"
api_key: ${DEEPSEEK_API_KEY}
# Mistral (preset base_url)
models:
mistral-medium:
provider: mistral
model_name: "mistral-medium-3"
api_key: ${MISTRAL_API_KEY}
# xAI (preset base_url)
models:
grok:
provider: xai
model_name: "grok-4.1"
api_key: ${XAI_API_KEY}
# Z.ai / GLM (preset base_url)
models:
glm-5:
provider: zai
model_name: "glm-5"
api_key: ${ZAI_API_KEY}