Mars Mission Portal: Backward Reasoning from Disaster
The Ares III Mars Mission Portal template demonstrates PORTAL mode temporal reasoning—working backward from a catastrophic outcome to discover its root causes. This flagship example shows how schedule pressure, institutional culture, and cascading technical debt can create an inevitable failure trajectory.Template:
mars_mission_portal | Mode: PORTAL | Cost: ~0.40 (full)The Scenario
The Ares III crewed Mars mission loses contact during orbital insertion in March 2031. Last telemetry shows cascading systems failures in life support and communications. The mission was celebrated as humanity’s greatest achievement until silence fell. PORTAL mode works backward from this disaster through 5 years (or 10 in full mode) to January 2026, exploring how present-day decisions create future outcomes.Cast of Characters
Sarah Okafor
Mission CommanderExperienced astronaut, politically pressured by NASA leadership to maintain schedule. Mediates between engineering rigor and institutional pressure.Final state: Valence +0.47, Arousal 0.57, Energy 124.4
Lin Zhang
Systems EngineerDetected ALSS anomalies during testing but was overruled by schedule pressure. Data-driven, escalating urgency.Final state: Valence -0.20, Arousal 0.94, Energy 116.7
Raj Mehta
Flight EngineerBrilliant systems analyst, conflict-averse personality. Detects problems early but hesitates to escalate.Final state: Valence -0.20, Arousal 0.81, Energy 123.7
Thomas Webb
Mission DirectorPrioritized schedule and budget over safety margins. Made key decisions to reduce crew size and accept simplified life support.Final state: Valence -0.17, Arousal 0.78, Energy 118.4
Running the Simulation
Output Artifacts
The simulation generates rich, queryable outputs across multiple formats:- Dialogs
- Causal Graph
- Entity States
- Training Data
6 conversations, 78 dialog turns (quick mode)Each timepoint produces a multi-party dialog with per-character generation:
- Voice distinctiveness: 0.91-0.97 across all entity pairs
- Independent LLM calls per character with persona-derived parameters
- Fourth Wall context: back layer shapes voice, front layer provides content
Key Insights from Example Run
The Causal Chain Reveals Structural Failure
Schedule pressure emerges as the dominant institutional failure mode:- 2026: Lin detects O2 generator pressure fluctuations and overheating
- 2027: Lin finds 30% failure probability; proposes redundant system with existing components
- 2028: O2 generator at 92% efficiency; Lin pushes for redesign; Webb demands schedule adherence
- 2029: Lin discovers 30% CO2 scrubbing efficiency loss; Webb dismisses for schedule
- 2030: Raj patches around comm anomalies; Webb reduces bandwidth over Lin’s objections
- 2031: Mission failure
Emotional Arc Analysis
Lin Zhang has the highest arousal (0.94) in the cast—the most activated character. Her negative valence (-0.20) combined with near-maximum arousal drives repeated confrontations with Webb, citing specific data:- 30% CO2 scrubbing efficiency loss
- 30% failure probability
- 10% pressure variance
- 5-degree hourly temperature spikes
ADPRS Waveform Gating
24 evaluations, 8 divergent (33.33%). Sarah Okafor and Thomas Webb both hit the φ ceiling (1.0) at timepoints tp_000 through tp_003, mapping them totrained band, but they actually resolve at dialog. This is soft budget mode working as intended—spending enough to maintain quality without escalating to the highest tier.
Template Configuration
Here’s the actual template JSON (excerpt):Mechanisms in Action
M17: Portal Backward Inference
M17: Portal Backward Inference
Generates 3 candidate antecedent states per step, scores each with 405B judge model (direct state scoring, no mini forward-simulations), selects the most coherent backward chain.
M3: Knowledge Provenance
M3: Knowledge Provenance
Every fact an entity knows has a tracked exposure event with source, timestamp, and confidence score.Example exposure chain:In PORTAL mode, front-layer knowledge is filtered by causal ancestry so characters only reference information from timepoints upstream of their position.
M11: Per-Character Dialog Synthesis
M11: Per-Character Dialog Synthesis
Each conversation generated through LangGraph steering loop:
- Steering agent selects next speaker based on narrative goals
- Character agent generates ONE turn using PersonaParams derived from tensor state (arousal → temperature, energy → max_tokens)
- Quality gate evaluates after each turn
- Loop continues until steering agent ends dialog
M13: Relationship Tracking
M13: Relationship Tracking
Relationships evolve across timepoints with trust and alignment values:
Cost Comparison: Quick vs. Full Mode
| Metric | Full Run | Quick Run | Change |
|---|---|---|---|
| Cost | $0.49 | $0.18 | -63% |
| LLM calls | 912 | 479 | -47% |
| Tokens | 900K | 318K | -65% |
| Backward steps | 10 | 5 | -50% |
| Timepoints | 11 | 6 | -45% |
| Dialogs / turns | 11 / 143 | 6 / 78 | -45% |
| Training examples | 40 | 20 | -50% |
| Duration | ~103 min | ~44 min | -57% |
--portal-quick flag is designed for fast demos and iteration while maintaining the same per-step quality.
Next Steps
Run Convergence Testing
Export to Formats
Try Other Templates
Explore the full-mechanism showcase with branching survival strategies
Create Custom Scenario
Learn to build your own PORTAL scenario from scratch
Model Licensing: This example run used Llama models. Meta’s license restricts using Llama outputs to train non-Llama models. For training data generation, use DeepSeek (MIT) or Mistral (Apache 2.0):
./run.sh run --model deepseek/deepseek-r1 mars_mission_portal
