Skip to main content
NoteWise routes all LLM calls through LiteLLM, which means any model string that LiteLLM supports will work — not just the providers listed here. The table below covers the eight natively supported providers with first-class API key handling.
The default model is gemini/gemini-2.5-flash. Google Gemini has a free tier, so it’s the easiest way to get started without a credit card. Grab a key at aistudio.google.com.

Supported providers

ProviderEnv varExample model string
Google GeminiGEMINI_API_KEYgemini/gemini-2.5-flash
OpenAIOPENAI_API_KEYgpt-4o
AnthropicANTHROPIC_API_KEYclaude-3-5-sonnet-20241022
GroqGROQ_API_KEYgroq/llama3-70b-8192
xAIXAI_API_KEYxai/grok-2
MistralMISTRAL_API_KEYmistral/mistral-large-latest
CohereCOHERE_API_KEYcommand-r-plus
DeepSeekDEEPSEEK_API_KEYdeepseek/deepseek-chat

Setting your API key

Overriding the model per run

Use --model (or -m) to change the model for a single invocation without touching your config:
notewise process "URL" --model claude-3-5-sonnet-20241022
notewise process "URL" --model gpt-4o
notewise process "URL" --model groq/llama3-70b-8192
The corresponding API key must already be set in the environment or config.env--model only changes which model is used, not which key is read.

LiteLLM model string format

LiteLLM uses a provider/model-name convention for most providers. The prefix tells LiteLLM which API endpoint and authentication header to use:
gemini/gemini-2.5-flash      ← Google Gemini
groq/llama3-70b-8192         ← Groq
xai/grok-2                   ← xAI
mistral/mistral-large-latest ← Mistral
deepseek/deepseek-chat        ← DeepSeek
Some providers (OpenAI, Anthropic, Cohere) use unprefixed model names:
gpt-4o
claude-3-5-sonnet-20241022
command-r-plus
NoteWise detects the provider from the model string prefix (or the model name itself for OpenAI/Anthropic/Cohere) and determines which API key env var to validate. If you use a model string that doesn’t match any known pattern — for example, a locally-hosted Ollama model — NoteWise will still pass it straight to LiteLLM; no API key check is performed.

Provider details

Default provider. Gemini 2.5 Flash offers a generous free tier and is fast enough for real-time note generation on most videos.
SettingValue
Env varGEMINI_API_KEY
Default modelgemini/gemini-2.5-flash
Key sourceaistudio.google.com/app/apikey
Other available Gemini models (from the setup wizard):
gemini/gemini-2.5-flash
gemini/gemini-2.5-flash-lite
gemini/gemini-2.5-pro
SettingValue
Env varOPENAI_API_KEY
Example modelgpt-4o
Key sourceplatform.openai.com/api-keys
OpenAI reasoning models (o1, o3, o4 series) are also supported. Note that reasoning models do not accept a temperature parameter — NoteWise passes the configured temperature for all models, so you may see a warning from LiteLLM if you use a reasoning model with a non-default temperature.Available models include:
gpt-4o-mini
gpt-4o
o3-mini
SettingValue
Env varANTHROPIC_API_KEY
Example modelclaude-3-5-sonnet-20241022
Key sourceconsole.anthropic.com/settings/keys
Claude models accept max_tokens and temperature as normal. If MAX_TOKENS is not set in config, NoteWise lets LiteLLM use the model’s default.Available models include:
claude-haiku-4-5-20251001
claude-sonnet-4-5-20250929
claude-4-opus-20250514
SettingValue
Env varGROQ_API_KEY
Example modelgroq/llama3-70b-8192
Key sourceconsole.groq.com/keys
Groq’s inference API is very fast and well-suited to long transcripts. Model strings must include the groq/ prefix.Available models include:
groq/llama-3.1-8b-instant
groq/llama-3.3-70b-versatile
groq/meta-llama/llama-4-scout-17b-16e-instruct
SettingValue
Env varXAI_API_KEY
Example modelxai/grok-2
Key sourceconsole.x.ai
Model strings must include the xai/ prefix.Available models include:
xai/grok-3
xai/grok-3-mini-latest
xai/grok-4-0709
SettingValue
Env varMISTRAL_API_KEY
Example modelmistral/mistral-large-latest
Key sourceconsole.mistral.ai/api-keys
Model strings must include the mistral/ prefix.Available models include:
mistral/mistral-small-latest
mistral/mistral-medium-latest
mistral/mistral-large-latest
SettingValue
Env varCOHERE_API_KEY
Example modelcommand-r-plus
Key sourcedashboard.cohere.com/api-keys
Cohere models use unprefixed names (no cohere/ prefix required).Available models include:
command-a-03-2025
command-r-plus-08-2024
command-r-08-2024
SettingValue
Env varDEEPSEEK_API_KEY
Example modeldeepseek/deepseek-chat
Key sourceplatform.deepseek.com/api_keys
Model strings must include the deepseek/ prefix.Available models include:
deepseek/deepseek-chat
deepseek/deepseek-v3
deepseek/deepseek-reasoner

Generation parameters

Two parameters affect LLM output quality and length, and can be set globally in config or overridden per run:
ParameterConfig keyCLI flagDefaultRange
TemperatureTEMPERATURE--temperature / -t0.70.01.0
Max tokensMAX_TOKENS--max-tokens / -kmodel defaultprovider-dependent
TEMPERATURE controls how creative or deterministic the output is. Lower values (closer to 0.0) produce more consistent, factual notes. Higher values (closer to 1.0) produce more varied phrasing. MAX_TOKENS caps the length of each LLM response. If not set, NoteWise lets LiteLLM use the model’s built-in default. For long chapters, a low MAX_TOKENS may cause notes to be truncated — leave this unset unless you have a specific reason to cap output length.
Some providers (such as OpenAI’s o-series reasoning models) do not support the temperature parameter. Sending it will cause an API error. If you use those models, set TEMPERATURE=1.0 in your config as a workaround, or check the provider’s documentation for the correct value.

Using any LiteLLM-compatible model

Because NoteWise passes model strings directly to LiteLLM, you can use any model that LiteLLM supports beyond the eight providers above — including locally-hosted models via Ollama:
notewise process "URL" --model ollama/llama3
For models outside the native provider list, set the required API key or base URL in the environment according to LiteLLM’s documentation. NoteWise will not validate the key or warn if it is missing for unknown providers.

Build docs developers (and LLMs) love