ask
Send a question to the LLM provider with automatic memory context injection. Usage:llm authorization
Pipeline Flow:
- Memory search: Query memory for relevant context (hybrid FTS5 + vector)
- Context injection: Add top 5 memory results to prompt
- LLM call: Send to configured provider (Anthropic, OpenAI, etc.)
- Response: Display LLM output with usage metrics
Free-Form Text (Implicit LLM)
Any input that doesn’t match a command is automatically sent to the LLM pipeline. Example:ask Explain vector search.
Security: Requires llm authorization (same as ask)
providers
List configured LLM providers and their online status. Usage:system:providers authorization
Offline Mode:
- online: API reachable, ready for requests
- offline: Network error, invalid API key, or service unavailable
Provider Selection
OneClaw supports 6 LLM providers:| Provider | Models | Use Case |
|---|---|---|
| Anthropic | claude-sonnet-4, claude-opus-4 | Production, high-quality reasoning |
| OpenAI | gpt-4o, gpt-4o-mini | General purpose, function calling |
| DeepSeek | deepseek-chat | Cost-effective alternative |
| Groq | llama-3.3-70b, mixtral-8x7b | Ultra-low latency inference |
| gemini-2.0-flash | Multimodal, large context | |
| Ollama | llama3.2, mistral, qwen | Local/offline, privacy-focused |
Configuration
Primary provider inconfig/oneclaw.toml:
Automatic Fallback Chain
When primary provider fails, OneClaw automatically tries fallback providers: Example config:- Try Anthropic (primary)
- If fails → Try OpenAI (fallback 1)
- If fails → Try Ollama (fallback 2)
- If all fail → Offline mode response
Complexity Analysis
OneClaw automatically analyzes query complexity to optimize LLM usage: Simple queries: Short, direct questions- “What is edge AI?”
- “List the layers”
- “Compare the benefits of edge AI vs cloud AI for industrial IoT”
- “Based on my stored sensor data, what patterns do you see?”
- Affects model selection in multi-model setups
- Adjusts context window size
- Logged in metrics for analysis
Context Manager
The orchestrator enriches prompts with:- System prompt: Role definition (“You are OneClaw, a helpful AI assistant…”)
- Memory context: Top 5 relevant memory entries
- User message: Original query
Performance Metrics
Track LLM performance viametrics command:
- Calls: Total LLM API requests
- Failed: Network errors, rate limits, invalid responses
- Tokens: Total input + output tokens consumed
- Latency: Average response time per call