Skill Commands
Skill commands provide specialized capabilities beyond the core Double Diamond workflow. Each skill focuses on a specific domain expertise./octo:debate - AI Debate Hub
Structured three-way debates between Claude, Gemini, and Codex.
Syntax
What It Does
Orchestrates a structured debate with:- Three participants: Claude (moderator), Gemini, Codex
- Multiple rounds: Opening, rebuttals, synthesis
- Consensus building: Final recommendation with confidence score
- Adversarial mode: Red team vs blue team critique
Debate Structure
- Round 1 - Opening Statements
- Each AI presents their position
- Initial arguments and evidence
- Round 2 - Rebuttals
- Respond to other perspectives
- Address counterarguments
- Round 3 - Synthesis
- Claude moderates consensus
- Final recommendation with reasoning
Interactive Questions
Before debate, you’ll be asked:- Debate style: Collaborative vs Adversarial
- Depth: Quick (2 rounds) vs Deep (3+ rounds)
- Decision urgency: High stakes vs exploratory
Examples
When to Use
Use debate for:
- Comparing technology options
- Architecture decisions with trade-offs
- Security approach evaluation
- Adversarial code review
- High-stakes technical choices
Output
Natural Language Triggers
Auto-activates when you say:- “should”, “vs”, “or”, “compare”
- “versus”, “decide”, “which is better”
- “debate”, “argue for/against”
/octo:review - Code Review
Expert code review with comprehensive quality assessment.
Syntax
What Gets Reviewed
Code Quality
Code Quality
- Design patterns and architecture
- Code complexity (cyclomatic)
- Maintainability and readability
- Naming conventions
- Code duplication
Security
Security
- OWASP Top 10 vulnerabilities
- Authentication/authorization flaws
- Input validation
- SQL injection and XSS risks
- Sensitive data exposure
Performance
Performance
- Algorithm efficiency
- Database query optimization
- Memory usage
- Caching opportunities
- Scalability issues
Best Practices
Best Practices
- Industry standards
- Framework conventions
- Error handling
- Logging and monitoring
- Test coverage
Interactive Questions
Before review, you’ll be asked:- Goal: Pre-commit / Security focus / Performance / Architecture
- Priority concerns: Security / Performance / Maintainability / Testing
- Audience: Just me / Team review / Production release / External audit
Review Types
- Quick Review
- Full Review
- Security Focus
Pre-commit checks - Fast validation before committing
- Surface-level checks (5-10 sec)
- Critical issues only
- Best for small changes
Examples
Output Format
/octo:security - Security Audit
OWASP compliance and vulnerability detection.
Syntax
What Gets Audited
OWASP Top 10
- Injection flaws
- Broken authentication
- Sensitive data exposure
- XML external entities
- Broken access control
- Security misconfiguration
- XSS vulnerabilities
- Insecure deserialization
- Known vulnerable components
- Insufficient logging
Authentication & Auth
- Password storage (hashing/salting)
- Session management
- Token security (JWT/OAuth)
- Authorization logic
- Multi-factor authentication
Input Validation
- SQL injection prevention
- XSS protection
- Command injection
- Path traversal
- LDAP/XML injection
Data Protection
- Encryption at rest/transit
- Cryptographic implementations
- Key management
- PII handling
- GDPR/HIPAA compliance
Interactive Questions
- Threat model: Standard web app / High-value target / Compliance-driven / API-focused
- Compliance requirements: None / OWASP / GDPR/HIPAA/PCI / SOC2/ISO27001
- Risk tolerance: Strict zero-trust / Balanced / Pragmatic / Development-only
Examples
Output Format
/octo:tdd - Test-Driven Development
Red-green-refactor discipline with multi-AI test generation.
Syntax
TDD Workflow
What You Get
- Test-First: Failing tests written before implementation
- Minimal Code: Only enough code to pass tests
- Refactor: Clean up with confidence (tests protect you)
- Coverage: High test coverage by design
- Regression Protection: Catch breaks early
Interactive Questions
- Coverage goal: Critical paths / Standard ~80% / Comprehensive >90% / Mutation testing
- Test style: Unit tests / Integration / E2E / Mix of all
- Complexity: Simple CRUD / Moderate logic / Complex algorithms / Distributed systems
Examples
TDD Cycle Example
When to Use TDD
Use TDD for:
- Critical business logic
- Complex algorithms
- Features with clear requirements
- When you need high confidence
- Legacy code refactoring
- Prototypes and spikes
- UI/UX experimentation
- Unclear requirements (use
/octo:discoverfirst)
/octo:factory - Dark Factory Mode
Spec-in, software-out autonomous pipeline.
Syntax
What It Does
7-phase autonomous pipeline:1. Parse Spec
Validates NLSpec format and extracts:
- Satisfaction target (0.80-0.99)
- Complexity estimate
- Behaviors and constraints
2. Generate Scenarios
Multi-provider scenario generation:
- Codex: Technical scenarios
- Gemini: User scenarios
- Claude: Edge cases
3. Split Holdout
80/20 train/test split:
- 80% used for implementation
- 20% held back for blind validation
4. Embrace Workflow
Full 4-phase implementation:
- Discover → Define → Develop → Deliver
- Fully autonomous (no phase approval)
5. Holdout Tests
Blind evaluation:
- Test implementation against withheld scenarios
- Measure actual vs expected behavior
6. Score Satisfaction
Weighted scoring:
- Behavior coverage: 40%
- Constraint adherence: 20%
- Holdout pass rate: 25%
- Code quality: 15%
Interactive Questions
- Spec path: Where is the NLSpec file?
- Satisfaction target: Use spec default or override? (0.80-0.99)
- Cost confirmation: Proceed with ~$0.50-2.00 cost? (~20-30 agent calls)
Options
Path to NLSpec file defining the feature
Percentage of scenarios for blind validation (0.20-0.30)
Number of retry attempts on FAIL verdict
Non-interactive mode for automation pipelines
Examples
Output Structure
When to Use Factory
Use factory for:
- Features with clear specifications
- Autonomous development pipelines
- CI/CD integration
- When you have a complete NLSpec
- Spec-driven development
- Simple bug fixes
- Exploratory coding
- Unclear requirements (use
/octo:planfirst) - Tasks without specifications
Cost & Duration
- Cost: ~$0.50-2.00 per run (20-30 agent calls)
- Duration: 15-30 minutes depending on complexity
- Retries: Auto-retry on FAIL (up to max-retries)
/octo:prd - PRD Generation
AI-optimized Product Requirements Document with 100-point scoring.
Syntax
What You Get
Comprehensive PRD with:- Executive Summary - Vision and key value proposition
- Problem Statement - Quantified by user segment
- Goals & Metrics - SMART goals with P0/P1/P2 priorities
- Non-Goals - Explicit scope boundaries
- User Personas - 2-3 specific personas with needs
- Functional Requirements - FR-001 format with acceptance criteria
- Implementation Phases - Dependency-ordered rollout
- Risks & Mitigations - Identified risks with mitigation plans
Interactive Questions
Phase 0 clarification (mandatory):- Target Users: Who will use this? (developers/end-users/admins/agencies)
- Core Problem: What pain point does this solve? Metrics?
- Success Criteria: How will you measure success? KPIs?
- Constraints: Technical, budget, timeline, platform constraints?
- Existing Context: Greenfield or integrating with existing systems?
Scoring Framework (100 points)
| Category | Points | Criteria |
|---|---|---|
| AI-Specific Optimization | 25 | Structured for AI consumption, clear acceptance criteria |
| Traditional PRD Core | 25 | Problem statement, goals, requirements clarity |
| Implementation Clarity | 30 | Phasing, dependencies, technical feasibility |
| Completeness | 20 | All sections present, personas defined, risks identified |
Examples
Output Example
/octo:claw - OpenClaw Administration
Manage OpenClaw gateway instances across platforms.
Syntax
What It Manages
Gateway Lifecycle
- Start/stop/restart gateway
- Health checks and diagnostics
- Daemon installation
- Version updates and rollback
5 Platforms
- macOS: launchd service
- Ubuntu/Debian: systemd service
- Docker: compose orchestration
- OCI (ARM): ARM-optimized containers
- Proxmox: LXC containers
6 Channels
- Telegram
- Discord
- Slack
- Signal
- iMessage
Security
- Security audit and hardening
- Firewall configuration
- Tailscale VPN setup
- Credential management
- SSL/TLS configuration
Methodology
Every claw action follows:- DETECT - Identify platform (never assume OS)
- DIAGNOSE - Non-destructive checks before changes
- EXECUTE - Platform-specific commands
- VERIFY - Confirm the change took effect
Examples
Platform-Specific Commands
- macOS
- Ubuntu/Debian
- Docker
When to Use Claw
- Managing OpenClaw gateway instances
- Platform-specific administration tasks
- Channel configuration (WhatsApp, Telegram, etc.)
- Security hardening and VPN setup
- Troubleshooting gateway issues
- Multi-platform deployments
/octo:claw is specifically for OpenClaw gateway administration. For general system commands, use /octo:setup or /octo:doctor.Planning & Orchestration Skills
/octo:plan - Strategic Planning
Create execution plans without running them.
See Workflow Commands - Plan for details.
/octo:parallel - Team of Teams
Decompose work into parallel packages.
/octo:multi - Force Multi-Provider
Manual override for parallel execution.
/octo:spec - NLSpec Authoring
Write structured natural language specifications.
Skill Comparison
| Skill | Multi-AI | Duration | Cost | Use Case |
|---|---|---|---|---|
/octo:debate | Yes | 5-10 min | ~$0.08-0.20 | Compare options, adversarial review |
/octo:review | Yes | 3-8 min | ~$0.05-0.15 | Code quality assessment |
/octo:security | Yes | 3-8 min | ~$0.05-0.15 | OWASP audit, vulnerability scan |
/octo:tdd | Yes | 10-20 min | ~$0.15-0.40 | Test-first implementation |
/octo:factory | Yes | 15-30 min | ~$0.50-2.00 | Spec-to-software pipeline |
/octo:prd | Yes | 5-10 min | ~$0.08-0.20 | Product requirements |
/octo:claw | No | 2-5 min | Free | OpenClaw administration |
Next Steps
Try a Debate
Compare options with multi-AI perspectives
Code Review
Get comprehensive quality assessment
Workflow Commands
Learn the core Double Diamond phases
