Skip to main content
AgentOS provides access to 47 models across 5 performance tiers, spanning frontier reasoning models to cost-effective local options.

Model Tiers

Most capable models for complex reasoning, research, and advanced tasks
ModelProviderContextPrice (per 1M tokens)Features
claude-opus-4-6Anthropic200K15/15 / 75Tools, Vision, 32K output
o3OpenAI200K10/10 / 40Reasoning, 100K output
gemini-2.5-proGoogle1M1.25/1.25 / 10Vision, Code exec, 65K output
grok-3xAI131K3/3 / 15Tools, Vision
samba-llama-3.1-405bSambaNova4K5/5 / 10405B parameters

Model Aliases

AgentOS provides convenient aliases for quick model selection:
const ALIASES = {
  // Tier shortcuts
  "best": "claude-opus-4-6",
  "frontier": "claude-opus-4-6",
  "smart": "claude-sonnet-4-6",
  "fast": "claude-haiku-4-5",
  "cheap": "gpt-4o-mini",
  
  // Claude shortcuts
  "opus": "claude-opus-4-6",
  "sonnet": "claude-sonnet-4-6",
  "haiku": "claude-haiku-4-5",
  
  // OpenAI shortcuts
  "gpt4": "gpt-4o",
  "gpt4o": "gpt-4o",
  "gpt41": "gpt-4.1",
  "o3": "o3",
  "o4": "o4-mini",
  
  // Google shortcuts
  "flash": "gemini-2.5-flash",
  "pro": "gemini-2.5-pro",
  "gemini": "gemini-2.5-flash",
  
  // Other providers
  "deepseek": "deepseek-chat",
  "ds": "deepseek-chat",
  "r1": "deepseek-reasoner",
  "llama": "llama-3.3-70b",
  "grok": "grok-2",
  "grok3": "grok-3",
  "mistral": "mistral-large",
  "sonar": "sonar-pro",
  "command": "command-a",
  "jamba": "jamba-1.5-large",
  "qwen": "qwen-max",
  "glm": "glm-4-plus",
  "kimi": "moonshot-v1-128k",
  "ernie": "ernie-4.0-turbo",
  "bedrock": "bedrock-claude-sonnet",
  "nova": "bedrock-nova-pro",
  "copilot": "copilot-gpt-4o",
}

Using Aliases

# CLI usage
agentos message default "Hello" --model sonnet
agentos models describe haiku

# Or via API
const result = await trigger('catalog::resolve', { model: 'opus' });
// Returns full model entry for claude-opus-4-6

Model Capabilities

Tool Use (Function Calling)

Models that support structured tool invocation:
  • All Claude models (Opus, Sonnet, Haiku)
  • All OpenAI models (GPT-4o, o3, o4-mini, GPT-4.1)
  • All Gemini models (Pro, Flash)
  • DeepSeek (Chat, Reasoner)
  • Llama 3.3 70B (all providers)
  • Cohere Command series
  • xAI Grok series
  • Mistral Large
  • AI21 Jamba series
  • All Chinese models (Qwen, GLM, Moonshot, ERNIE)
  • AWS Bedrock models

Vision Support

Models that can process images:
  • Claude Opus 4.6, Sonnet 4.6, Haiku 4.5
  • GPT-4o, GPT-4.1, o3, o4-mini, GPT-4o mini
  • Gemini 2.5 Pro, Gemini 2.5 Flash
  • Grok-2, Grok-3
  • AWS Bedrock Claude, Nova Pro
  • GitHub Copilot GPT-4o

Long Context (>100K tokens)

Models with extended context windows:
  • 1M tokens: Gemini 2.5 (Flash, Pro), GPT-4.1, Qwen Turbo
  • 300K tokens: AWS Bedrock Nova Pro
  • 256K tokens: Cohere Command series, AI21 Jamba, MiniMax ABAB 7
  • 200K tokens: All Claude models, OpenAI o3/o4-mini, Perplexity Sonar, Bedrock Claude
  • 128K+: Most other modern models

Pricing Comparison

Prices shown as input / output per 1 million tokens (USD).

Best Value Models

ModelInputOutputUse Case
qwen-turbo$0.05$0.15High-volume, 1M context
deepseek-chat$0.14$0.28General purpose
gpt-4o-mini$0.15$0.6Fast OpenAI option
gemini-2.5-flash$0.15$0.6Google ecosystem
jamba-1.5-mini$0.2$0.4256K context
hf-llama-3.3-70b$0.36$0.36Open source
hf-mistral-7bFreeFreeDevelopment/testing

Most Expensive (Frontier)

ModelInputOutputJustification
claude-opus-4-6$15$75Most capable reasoning
o3$10$40Advanced reasoning
gemini-2.5-pro$1.25$101M context + code exec
samba-llama-3.1-405b$5$10405B parameters

Model Selection Guide

By Use Case

// Use fast tier
model: "claude-haiku-4-5" // or "gpt-4o-mini" or "gemini-2.5-flash"

By Budget

  • Under $0.50/1M tokens: Qwen Turbo, DeepSeek Chat, HF models, local (free)
  • 0.500.50-2/1M: Most balanced tier models
  • 22-5/1M: Smart tier models
  • Over $5/1M: Frontier models (use sparingly)

CLI Usage

# List all models
agentos models list

# Filter by tier
agentos models list --tier smart

# Filter by provider
agentos models list --provider anthropic

# Filter by capability
agentos models list --tools
agentos models list --vision

# Describe a model
agentos models describe claude-sonnet-4-6

# Output:
# Model: claude-sonnet-4-6
# Provider: Anthropic
# Tier: smart
# Context: 200,000 tokens
# Max Output: 16,000 tokens
# Pricing: $3.00 / $15.00 per 1M tokens
# Capabilities: tools, vision

Programmatic Access

import { trigger } from 'iii-sdk';

// List all models
const models = await trigger('catalog::models', {});

// Filter by tier
const smartModels = await trigger('catalog::models', { tier: 'smart' });

// Filter by provider
const anthropicModels = await trigger('catalog::models', { provider: 'anthropic' });

// Resolve model or alias
const model = await trigger('catalog::resolve', { model: 'sonnet' });
// Returns full ModelEntry with pricing, capabilities, etc.

// List providers
const providers = await trigger('catalog::providers', {});
// Returns provider configs with availability status

// Get aliases
const aliases = await trigger('catalog::aliases', {});

HTTP API

# List models
curl http://localhost:3111/api/models

# List providers
curl http://localhost:3111/api/providers

# Get aliases
curl http://localhost:3111/api/models/aliases

# Test provider
curl -X POST http://localhost:3111/api/providers/anthropic/test

Model Metadata

Each model entry includes:
interface ModelEntry {
  id: string;                    // Model identifier
  provider: string;              // Provider name
  name: string;                  // Display name
  tier: "frontier" | "smart" | "balanced" | "fast" | "local";
  contextWindow: number;         // Max input tokens
  maxOutput: number;             // Max output tokens
  inputPrice: number;            // $ per 1M input tokens
  outputPrice: number;           // $ per 1M output tokens
  supportsTools: boolean;        // Function calling
  supportsVision: boolean;       // Image input
  local: boolean;                // Self-hosted
}

Next Steps

Provider Details

Learn about each provider’s setup

Routing Logic

Understand automatic model selection

Build docs developers (and LLMs) love