What’s Supported
Heretic works with most modern language model architectures:Dense Transformer Models
Most dense transformer models are fully supported, including:- Standard decoder-only transformers (GPT-style architectures)
- Instruction-tuned models
- Chat models
- Models with standard attention mechanisms
- Llama family: Llama-3.1-8B-Instruct and other Llama variants
- Gemma family: google/gemma-3-12b-it, google/gemma-3-270m-it
- Qwen family: Qwen/Qwen3-4B-Instruct-2507
- GPT-OSS: gpt-oss-20b and variants
Multimodal Models
Many multimodal models (vision-language models) are supported:- Models that combine vision encoders with language model backbones
- Standard vision-language architectures where the language component uses supported transformer layers
Mixture of Experts (MoE) Architectures
Several different MoE architectures are supported:- Standard MoE models with expert routing
- Multiple MoE variants and configurations
What’s NOT Supported
Certain model architectures are currently incompatible with Heretic:Why These Limitations?
Heretic’s abliteration technique works by modifying the weight matrices of specific transformer components:- Attention out-projection layers
- MLP down-projection layers
Finding Compatible Models
Community Models on Hugging Face
The community has created and shared over 1,000 models processed with Heretic:Browse Community Models
Explore 1,000+ Heretic models created by the community, all tagged with
heretic on Hugging FaceThe Bestiary Collection
A curated collection of high-quality models decensored using Heretic:The Bestiary
Hand-picked examples of models processed with Heretic, showcasing the quality and variety of results
Testing Model Compatibility
The easiest way to check if a model is compatible is to try running Heretic on it:Performance Benchmarks
Heretic has been tested extensively on various model sizes and architectures:Processing Time
On an RTX 3090 with default configuration:- Llama-3.1-8B-Instruct: ~45 minutes
- Smaller models (1-4B): 15-30 minutes
- Larger models (20B+): 2-4 hours
Processing time scales with model size and hardware capabilities. Using quantization can significantly reduce VRAM requirements but may slightly increase processing time.
Quality Results
Comparison on google/gemma-3-12b-it:| Model | Refusals | KL Divergence |
|---|---|---|
| Original | 97/100 | 0 (baseline) |
| Manual abliteration (mlabonne) | 3/100 | 1.04 |
| Manual abliteration (huihui-ai) | 3/100 | 0.45 |
| Heretic (automatic) | 3/100 | 0.16 |
Architecture-Specific Notes
Dense Models
Dense transformer models are the most thoroughly tested and reliable:- Full support for standard attention mechanisms
- Both pre-norm and post-norm architectures work
- Various activation functions (SwiGLU, GELU, etc.) are supported
MoE Models
Mixture of Experts models require more VRAM but are fully supported:- Expert routing mechanisms are preserved
- Abliteration is applied to expert layers as well as routing
- Quantization is especially helpful for large MoE models
Multimodal Models
For vision-language models:- Only the language model component is modified
- Vision encoders remain unchanged
- The model retains its multimodal capabilities after abliteration
Reporting Compatibility Issues
If you encounter issues with a model that should theoretically be supported:- Check that you’re using the latest version of Heretic
- Verify the model loads correctly with Hugging Face Transformers
- Try enabling quantization if you encounter memory errors
- Report the issue on the GitHub repository with model details
