Available Modes
The CLI supports six primary modes, each accessed via themode parameter:
autopilot
Self-testing mode that evaluates temporal chains of varying lengths
train
Train entities with historical context and populate knowledge states
temporal_train
Train entities across temporal chains with causal evolution
evaluate
Run evaluation metrics on trained entities
interactive
Natural language query REPL for exploring simulation results
models
Manage and select LLM models
Autopilot Mode
Autopilot mode runs automated tests of temporal chains with different lengths to evaluate system performance:- Tests temporal chains of configurable lengths (default: 3, 5, 7 timepoints)
- Runs temporal training on each chain
- Computes aggregate evaluation metrics
- Generates cost and performance reports
- Checks for causal chain violations
- Temporal coherence scores
- Knowledge consistency scores
- Causal chain violation counts
- Per-template cost and token usage
- Summary reports in JSON and Markdown
Train Mode
Trains entities using historical contexts from templates:- Rich historical context from predefined templates
- Graph-based relationship modeling
- Knowledge state population via LLM
- Exposure event tracking
- Resolution level management
founding_fathers_1789- Constitutional inauguration- Additional contexts in
entity_templates.py
Temporal Train Mode
Trains entities across a temporal chain with causal propagation:- Builds causal temporal chains
- Propagates knowledge states forward
- Tracks knowledge growth per timepoint
- Records exposure events for each learning moment
- Validates temporal causality
Evaluate Mode
Runs evaluation metrics on all entities in the database:- Temporal Coherence - Consistency across timepoints
- Knowledge Consistency - Information conservation compliance
- Biological Plausibility - Constraint enforcement
- Per-entity metric scores
- Resolution level distribution
- Aggregate statistics
- Cost tracking
Interactive Mode
Natural language query REPL for exploring simulation data:- Parse natural language queries
- Retrieve entity knowledge states
- Synthesize contextual responses
- Track query costs
- Show simulation statistics
Models Mode
Manage LLM model selection and testing:- List available Llama models from OpenRouter
- View detailed model information (context length, pricing)
- Switch between models
- Test model connectivity
- Refresh model catalog
Branch Mode
Explore counterfactual “what-if” scenarios:- Interactive branching explorer
- Counterfactual query analysis
- Alternate timeline generation
Configuration
All modes use Hydra configuration fromconf/config.yaml:
Command-Line Overrides
Override any config value from the command line:Cost Tracking
All modes track LLM API costs and token usage:reports/ directory with:
- Total cost breakdown
- Token usage statistics
- Per-operation metrics
- Timestamp and configuration
Next Steps
Run Command
Execute simulations with ./run.sh
Training
Learn about training modes
Interactive Queries
Query your simulation data
Evaluation
Understand evaluation metrics

