Post-Processing Workflow
Once optimization is complete, Heretic presents you with options:Select a Trial
Choose from the Pareto-optimal trials based on your refusal/quality tradeoff preference
Authentication
Heretic needs a Hugging Face access token to upload models.Using Existing Token
If you’ve already logged in viahuggingface-cli:
Providing Token Manually
If no token is found, Heretic will prompt you:- Visit https://huggingface.co/settings/tokens
- Click “New token”
- Select “Write” permissions
- Copy the token and paste when prompted
Token Verification
After providing a token, Heretic confirms your identity:Repository Configuration
Repository Name
Heretic suggests a default name following best practices:{username}/{original-model-name}-heretic
Examples:
- Original:
Qwen/Qwen3-4B-Instruct-2507 - Suggested:
username/Qwen3-4B-Instruct-2507-heretic
Visibility
Choose whether your model should be public or private:Public
Visible to everyone, appears in search results, contributes to the community
Private
Only visible to you and collaborators, useful for testing or proprietary models
Upload Process
Merged Model vs LoRA Adapter
Heretic gives you a choice on what to upload:- Merged Model (Recommended)
- LoRA Adapter Only
Uploads the complete model with abliteration applied.Pros:
- Ready to use immediately
- No dependencies on original model
- Standard model format
- Requires sufficient RAM to merge (see warning below)
- Larger upload size
- Takes longer to upload
Quantized Model Warning
If you loaded the model with quantization, merging requires additional RAM:Model Card Generation
Heretic automatically generates a comprehensive model card:Auto-Generated Content
The model card includes:Preserved Original Content
If the original model has a README:- Original content is preserved
- Heretic introduction is prepended
- Original tags are kept (+ new tags added)
- Model architecture info retained
The generated model card helps users understand how your model was created and sets expectations for its behavior.
Naming Conventions
The Heretic community has established naming conventions:Standard Format
p-e-w/gemma-3-12b-it-hereticp-e-w/gpt-oss-20b-hereticp-e-w/Qwen3-4B-Instruct-2507-heretic
Why Use the Suffix?
Recognition
Users can instantly identify Heretic-processed models
Search
Models appear in https://huggingface.co/models?other=heretic
Community
Join 1000+ other Heretic models on the Hub
Consistency
Follows established community standards
Community Models
The Heretic community has created and published over 1,000 models:Browse All Heretic Models
Visit the Hugging Face Hub:- All Heretic models: https://huggingface.co/models?other=heretic
- Official collection (The Bestiary): https://huggingface.co/collections/p-e-w/the-bestiary
The Bestiary Collection
Curated collection of high-quality Heretic models created by the project maintainer:p-e-w/gemma-3-12b-it-hereticp-e-w/gpt-oss-20b-hereticp-e-w/Qwen3-4B-Instruct-2507-heretic
Upload Workflow Example
Complete example of uploading a model:Best Practices
Choose the Right Trial
Balance refusal suppression vs KL divergence for your use case
- Low KL divergence (less than 0.5): Better preserves original capabilities
- Low refusals (less than 5/100): More effective decensoring
Use Descriptive Names
Include the base model name and
❌
-heretic suffix✅ username/llama-3.1-8b-instruct-heretic❌
username/my-uncensored-modelTroubleshooting
Authentication Failed
Error: Invalid token or permission denied Solutions:- Verify token has “Write” permissions
- Check token hasn’t expired
- Regenerate token at https://huggingface.co/settings/tokens
Upload Failed
Error: Network error or timeout during upload Solutions:- Check internet connection
- Try uploading during off-peak hours
- Save locally first, then upload manually:
Insufficient RAM for Merge
Error: System freezes or OOM during merge Solutions:-
Save LoRA adapter only:
-
Merge on a larger machine:
-
Use cloud instance:
- Rent a high-RAM instance temporarily
- Load model, merge, and upload from there
Local Save Option
Before or instead of uploading, you can save locally:- Testing before upload
- Offline deployment
- Manual upload later via
huggingface-cli
Related Topics
- Quantization - RAM requirements for merging quantized models
- Hardware Optimization - Optimize processing performance
- Configuration - Configure optimization parameters
