Skip to main content
Glass supports multiple language model providers, giving you flexibility in choosing the best models for your workflow. This guide covers setup and configuration for each provider.

Supported Providers

Anthropic

Claude 4.6 Sonnet, Opus, and other Claude models

OpenAI

GPT-4, GPT-5, o1, o3, and other OpenAI models

Ollama

Local models for privacy and offline usage

OpenAI-Compatible

Any OpenAI-compatible API endpoint

Copilot

GitHub Copilot Chat integration

Google Gemini

Gemini models via Agent Servers

Anthropic Claude

Set up Anthropic’s Claude models:

Get API Key

1

Create Account

2

Generate Key

Navigate to Settings → API Keys → Create Key
3

Copy Key

Save your API key (starts with sk-ant-)

Configure in Glass

1

Open Settings

cmd-, or ctrl-,
2

Navigate to AI

Settings → AI → Providers → Anthropic
3

Paste Key

Enter your API key and save

Available Models

ModelContextBest For
claude-4.6-sonnet200kGeneral use, balanced performance
claude-4.6-opus200kComplex reasoning, highest quality
claude-4.6-sonnet-thinking200kTasks requiring reasoning steps

Custom Models

Add custom Anthropic models:
{
  "language_models": {
    "anthropic": {
      "available_models": [
        {
          "name": "claude-custom-model",
          "display_name": "Claude Custom",
          "max_tokens": 200000,
          "max_output_tokens": 8192,
          "cache_configuration": {
            "max_cache_anchors": 4,
            "should_speculate": true,
            "min_total_token": 1024
          }
        }
      ]
    }
  }
}

OpenAI

Configure OpenAI models:

Get API Key

1

Create Account

2

Add Payment

Add payment method (required for API access)
3

Generate Key

Navigate to API Keys → Create new secret key
4

Copy Key

Save your API key (starts with sk-)
OpenAI API access requires a paid account with credits. GitHub Copilot subscription does not include API access.

Configure in Glass

1

Open Settings

cmd-, or ctrl-,
2

Navigate to AI

Settings → AI → Providers → OpenAI
3

Paste Key

Enter your API key and save

Available Models

ModelContextBest For
gpt-5200kLatest, best quality
gpt-5-mini200kFast, cost-effective
o3128kAdvanced reasoning
o3-mini128kFast reasoning

Custom Models

{
  "language_models": {
    "openai": {
      "available_models": [
        {
          "name": "gpt-4-custom",
          "display_name": "GPT-4 Custom",
          "max_tokens": 128000,
          "max_output_tokens": 4096,
          "capabilities": {
            "chat_completions": true
          }
        }
      ]
    }
  }
}

Ollama

Run models locally with Ollama:

Setup Ollama

1

Install Ollama

Download from ollama.ai:
# macOS
brew install ollama

# Linux
curl -fsSL https://ollama.ai/install.sh | sh

# Windows
# Download installer from ollama.ai
2

Start Ollama

ollama serve
Or use system service (Linux/macOS):
systemctl start ollama
3

Pull Models

Download code models:
# For agents
ollama pull qwen2.5-coder:32b

# For completions
ollama pull codellama:7b

# For chat
ollama pull llama3.3:70b
4

Configure Glass

{
  "language_models": {
    "ollama": {
      "api_url": "http://localhost:11434"
    }
  },
  "agent": {
    "default_model": {
      "provider": "ollama",
      "model": "qwen2.5-coder:32b"
    }
  }
}
ModelSizeRAMQuality
qwen2.5-coder:32b19GB32GBBest
deepseek-coder-v2:16b9GB16GBGreat
codellama:13b7GB12GBGood
mistral:7b4GB8GBFast

Model Selection

Ollama models appear in the model picker:
Only models that are currently pulled will appear in Glass.

OpenAI-Compatible APIs

Connect to any OpenAI-compatible endpoint:

Configuration

{
  "language_models": {
    "openai_compatible": {
      "api_url": "https://your-endpoint.com/v1",
      "api_key": "your-key",
      "available_models": [
        {
          "provider": "custom",
          "name": "your-model",
          "max_tokens": 8192
        }
      ]
    }
  }
}

Compatible Services

LM Studio

Local inference with UI:
{
  "api_url": "http://localhost:1234/v1"
}

Together AI

Hosted inference:
{
  "api_url": "https://api.together.xyz/v1",
  "api_key": "your-together-key"
}

Azure OpenAI

Azure deployment:
{
  "api_url": "https://your-resource.openai.azure.com/openai/deployments/your-deployment/chat/completions?api-version=2024-02-15-preview",
  "api_key": "your-azure-key"
}

Replicate

Model hosting:
{
  "api_url": "https://api.replicate.com/v1",
  "api_key": "your-replicate-token"
}

GitHub Copilot

Use GitHub Copilot Chat in Glass:

Requirements

  • Active GitHub Copilot subscription
  • GitHub account

Setup

1

Sign In

cmd-shift-p → “Sign in to GitHub”
2

Authorize Copilot

Follow the GitHub authentication flow
3

Enable

Copilot Chat is automatically enabled after sign-in
GitHub Copilot uses its own completions system separate from Glass AI features.

Google Gemini

Access Gemini via Agent Servers:

Configuration

{
  "agent_servers": {
    "gemini": {
      "command": "npx",
      "args": ["-y", "@modelcontextprotocol/server-gemini"],
      "env": {
        "GEMINI_API_KEY": "your-gemini-key"
      }
    }
  }
}

Get API Key

1

Create Account

2

Generate Key

Navigate to Get API Key → Create API Key
3

Configure

Add key to agent server config

Provider Selection

Choose the right provider for your needs:

Decision Matrix

PriorityRecommended ProviderWhy
Best qualityAnthropic ClaudeIndustry-leading models
Latest modelsOpenAICutting-edge capabilities
PrivacyOllamaLocal, offline, private
CostOllamaFree, unlimited
SpeedOpenAI GPT-4oFastest cloud models
CustomizationOpenAI-compatibleSelf-hosted, flexible

Multi-Provider Setup

Use multiple providers simultaneously:
{
  "agent": {
    "default_model": {
      "provider": "anthropic",
      "model": "claude-4.6-sonnet"
    },
    "inline_assistant_model": {
      "provider": "openai",
      "model": "gpt-4o-mini"
    },
    "thread_summary_model": {
      "provider": "ollama",
      "model": "llama3.3:70b"
    }
  },
  "edit_predictions": {
    "provider": "ollama",
    "ollama": {
      "model": "codellama:7b"
    }
  }
}

Troubleshooting

  • Verify key is correct (check for typos)
  • Ensure key has not expired
  • Check account has credits/subscription
  • Try regenerating the key
  • Verify API URL is correct
  • Check provider is authenticated
  • Verify API endpoint is reachable
  • Ensure provider is enabled in settings
  • Try refreshing Glass (cmd-r / ctrl-r)
  • Verify Ollama is running: ollama list
  • Check API URL: http://localhost:11434
  • Ensure models are pulled
  • Check firewall/network settings
  • Try restarting Ollama service
  • Check your API usage/quota
  • Wait and retry
  • Consider upgrading plan
  • Switch to different provider temporarily

Next Steps

Build docs developers (and LLMs) love