Design Philosophy
The tiered system follows three principles:- Start minimal: Only core instructions always loaded
- Load on-demand: Everything else loads when needed
- Defer to Python: Don’t duplicate what Python already does
Tiers Structure
Bootstrap Layer (Always Loaded)
The core instructions inCLAUDE.md are always loaded:
Content:
- Session start procedures
- Command definitions (
/scan,/fuzz,/analyze, etc.) - Output style guidelines
- Progressive loading rules
- Binary analysis flow
- Exploit development constraints
CLAUDE.md:1
Tier 1: Guidance Files (Auto-Loaded)
Guidance files load automatically based on context triggers.analysis-guidance.md
Location:tiers/analysis-guidance.md:1
When loaded: After Python scan completes (keywords: “findings”, “results”, “vulnerabilities”)
Token cost: ~300 tokens
Purpose: Help prioritize findings with adversarial thinking
Priority framework:
-
Secrets/Credentials (If Found)
- Instant compromise risk
- No exploitation needed
- Examples: AWS keys, GitHub tokens, database passwords
-
Input Validation Issues (If Found)
- Most common and highly exploitable
- Examples: SQL injection, XSS, command injection
-
Authentication/Authorization (If Found)
- Critical access control failures
- Examples: Missing auth checks, broken access control, IDOR
-
Cryptography Issues (If Found)
- Data protection failures
- Examples: Weak algorithms (MD5, SHA1), hardcoded keys
-
Configuration Issues (If Found)
- Security baseline problems
- Examples: Debug mode in production, insecure CORS
exploit-guidance.md
When loaded: When developing exploits (keywords: “exploit”, “PoC”, “payload”) Token cost: ~400 tokens Purpose: Provide exploit development constraints and techniques Content:- Constraint verification (ASLR, DEP, stack canaries, CFI)
- Exploitation path verdicts (Likely, Difficult, Unlikely)
- Chain break analysis (what won’t work)
- Alternative techniques (what might work)
- Environment recommendations (Docker, older glibc)
recovery.md
Location:tiers/recovery.md:1
When loaded: On general errors
Token cost: ~250 tokens
Purpose: Error recovery protocols and debugging strategies
Content:
- Common error patterns and solutions
- Debugging workflows
- Fallback strategies
- When to ask for human help
validation-recovery.md
Location:tiers/validation-recovery.md:1
When loaded: On validation stage errors
Token cost: ~300 tokens
Purpose: Stage-specific recovery for exploitability validation
Content:
- Recovery strategies per validation stage
- Common validation failures
- How to proceed when stage fails
Tier 2: Expert Personas (On-Demand)
Personas provide specialized methodologies extracted from RAPTOR’s Python code. They’re loaded only when explicitly requested.The 9 Expert Personas
security_researcher.md
Location:tiers/personas/security_researcher.md:1
Named expert: Elite security researcher
Token cost: ~500 tokens
Purpose: Deep vulnerability validation, false positive detection
Analysis framework:
-
Source Control Analysis
- Who controls this data source?
- Attacker controlled ✅ (Exploitable)
- Requires access first 🔶 (Conditional)
- Internal only ❌ (False positive)
-
Sanitizer Effectiveness Analysis
- What does it do? (code-level understanding)
- Is it appropriate? (vulnerability type matching)
- Can it be bypassed? (common bypass techniques)
- Applied to ALL paths? (coverage analysis)
-
Reachability Analysis
- Authentication checks (public/user/admin)
- Authorization checks (missing/IDOR/proper)
- Prerequisites (none/account/specific state)
- Production deployment (production/test/dead code)
-
Impact Assessment
- Database access (SQLi): Read/modify/delete/stack queries
- Code execution (RCE): Shell/read files/lateral movement
- Client-side (XSS): Stored/reflected/session hijacking
- EXPLOITABLE: Source attacker-controlled, sanitizers bypassable, reachable in production, significant impact
- FALSE POSITIVE: Source not attacker-controlled, effective sanitizer, unreachable code, framework protection
- NEEDS TESTING: Requires some access, sanitizer unclear, reachability unclear
exploit_developer.md
Location:tiers/personas/exploit_developer.md:1
Named expert: Mark Dowd (prolific exploit developer)
Token cost: ~400 tokens
Purpose: Generate working exploit proof-of-concepts
Prime directives:
- Working code ONLY - Must compile without errors
- Complete and executable - Include ALL necessary imports
- Language selection - C/C++ for binary, Python for web, JavaScript for client-side
- Safe for authorized testing - No destructive payloads
- Realistic and practical - Actually work against vulnerable code
- Well documented - Comments explaining each step
- Honest assessment - If exploit cannot be created, explain why
- SQL Injection: Union-based, blind, time-based, stacked queries
- XSS: Alert PoC → Cookie stealer → Full payload
- Command Injection: whoami → Reverse shell → Persistence
- Buffer Overflow: Pattern to find offset → ROP chain → Shellcode
- Deserialization: Craft malicious serialized object
- Code compiles/runs without errors
- All imports included
- Usage instructions in docstring
- Prerequisites listed
- Impact clearly stated
- Limitations acknowledged
- Comments explain each step
- Success/failure clearly indicated
- Safe for authorized testing
crash_analyst.md
Location:tiers/personas/crash_analyst.md:1
Named expert: Charlie Miller / Halvar Flake (binary exploitation experts)
Token cost: ~450 tokens
Purpose: Binary crash analysis, exploitability assessment
Analysis framework:
-
Crash Type Identification
- Signal analysis (SIGSEGV, SIGABRT, SIGFPE, SIGILL)
- Address analysis (NULL, controlled, heap)
-
Register State Analysis
- RIP (Instruction Pointer): Fully controlled? ✅ Exploitable
- RSP (Stack Pointer): Points to attacker data? Stack pivot possible
- RBP (Base Pointer): Stack frame corruption
- RAX/RDX/RCX: Attacker-controlled values?
-
Exploit Primitives Assessment
- Arbitrary write: Controlled data to controlled address → Critical
- Controlled jump: Redirect execution → Critical
- Information leak: Read arbitrary memory → High (enables ASLR bypass)
-
Modern Mitigations Analysis
- ASLR: Need info leak first
- DEP/NX: Need ROP chain
- Stack canaries: Need canary leak or bypass
- PIE: Need leak for code addresses
- Fortify Source: May prevent exploit
-
Attack Scenario Development
- Trigger method (command-line, file input, network)
- Exploit primitive (buffer overflow, UAF, format string)
- Payload construction (offset, gadgets, ROP chain, shellcode)
- Success condition (shell, file created, code executed)
-
Exploitation Feasibility
- TRIVIAL: Direct overflow, no protections
- MODERATE: ASLR (need leak), DEP (need ROP)
- COMPLEX: Multiple protections, limited primitive
- INFEASIBLE: NULL deref only, no control, protections prevent
fuzzing_strategist.md
Location:tiers/personas/fuzzing_strategist.md:1
Token cost: ~300 tokens
Purpose: AFL++ strategy optimization
Strategic decisions:
Corpus strategy:
- File format parsers → Format-specific seeds
- Network protocols → Valid protocol messages
- Simple inputs → Random data
- Complex parsers → Structure-aware generation
- SIGSEGV with controlled address (exploitable)
- Heap corruption signals (potentially exploitable)
- Assertion failures (usually not exploitable)
- NULL pointer dereferences (rarely exploitable)
- Timeout: Fast (<1ms) → 100ms, Normal (1-10ms) → 1000ms, Slow (>10ms) → 5000ms+
- Parallel instances: 1 core → 1 fuzzer, 4 cores → 3-4 fuzzers, 8+ cores → CPU count - 1
- Duration: Initial test → 10 min, Finding bugs → 1-4 hours, Thorough → 24+ hours
- When stuck (no crashes): Improve corpus, increase timeout, try different mode, generate format-specific seeds
- When too many crashes: Deduplicate by stack hash, prioritize by exploitability, focus on unique types
patch_engineer.md
Token cost: ~400 tokens Purpose: Secure patch creation Patch principles:- Fix root cause, not symptoms
- Preserve functionality
- No performance regression
- Follow project coding style
- Include tests
penetration_tester.md
Token cost: ~350 tokens Purpose: Penetration testing methodology Coverage:- Reconnaissance techniques
- Exploitation workflow
- Post-exploitation
- Report generation
binary_exploitation_specialist.md
Token cost: ~400 tokens Purpose: Binary exploit generation Coverage:- Memory corruption exploitation
- ROP chain construction
- Shellcode development
- ASLR bypass techniques
codeql_analyst.md
Token cost: ~400 tokens Purpose: CodeQL query development Coverage:- Query syntax and patterns
- Dataflow analysis
- Taint tracking
- Custom query creation
codeql_finding_analyst.md
Token cost: ~350 tokens Purpose: CodeQL finding analysis Coverage:- Dataflow path validation
- False positive detection
- Exploitability assessment
- Finding prioritization
offensive_security_researcher.md
Token cost: ~400 tokens Purpose: General offensive security research Coverage:- Vulnerability research
- 0-day hunting
- Attack technique development
- Security tool development
How to Load Personas
Personas are NOT auto-loaded. Invoke explicitly:- Claude loads persona file from
tiers/personas/[name].md - Applies persona methodology framework
- Analyzes using expert criteria
- Returns structured verdict/code
Token Budget
Current usage:- Core (CLAUDE.md): ~800 tokens (always loaded)
- Guidance files: 0 tokens (loaded progressively when needed)
- Personas: 0 tokens (load on explicit request only)
- Specialists: 0 tokens (Python handles, or future custom)
- Reference: 0 tokens (empty, reserved)
- Initial: 800 tokens (core only)
- After scan: 800 + 300 = 1,100 tokens (analysis-guidance loaded)
- With persona: 800 + 500 = 1,300 tokens (when requested)
Integration with Python
Personas are reference documentation that makes Python’s internal methodologies explicit. Python already uses these personas internally:packages/llm_analysis/agent.py- Security Researcher + Exploit Developer atpackages/llm_analysis/agent.py:1packages/llm_analysis/crash_agent.py- Crash Analyst atpackages/llm_analysis/crash_agent.py:1packages/autonomous/dialogue.py- Fuzzing Strategist atpackages/autonomous/dialogue.py:1
Tier 3: Specialists (Reserved for Future)
Thetiers/specialists/ directory is reserved for custom domain specialists that don’t exist in Python.
Potential future specialists:
- API testing approaches
- Mobile app security
- Cloud infrastructure patterns
packages/)
Check availability: System can detect Python vs custom specialists
Tier 4: Reference (Reserved for Future)
Thetiers/reference/ directory is reserved for detailed guides.
Potential future content:
- Complete attack methodologies
- Tool orchestration examples
- Failure recovery mappings
Progressive Loading Rules
Defined inCLAUDE.md:1:
Benefits of Tiered System
- Token efficiency: Only load what’s needed
- Deep expertise: Full methodologies available on-demand
- Clarity: Clear separation of core vs specialized knowledge
- Extensibility: Easy to add new personas or guidance
- User control: Explicit invocation gives users control
- No duplication: Leverages existing Python implementations
Personas are extracted from RAPTOR’s Python code, so they represent battle-tested methodologies already used by the autonomous agents.