Multiple Flux LoRAs
You can load multiple LoRA adapters for Flux models by providing arrays in thelora_config parameter.
Example
- Two LoRAs are loaded: an anime style and a realistic photography style
- Each LoRA has a scale of 0.6, allowing them to blend together
- Both are loaded from safetensors format (
from_pt: "true")
Multiple SDXL LoRAs
SDXL supports loading multiple LoRAs from both local paths and HuggingFace Hub.Example: blueprint and papercut styles
- Loading one LoRA from a local file path (blueprint style)
- Loading another LoRA from HuggingFace Hub (papercut style)
- Using different scale values (0.8 and 0.7) for fine-tuned blending
- Combining both effects in the generation
Hyper SDXL LoRA
Hyper-SDXL enables ultra-fast inference in just 2-4 steps using specialized LoRA adapters from ByteDance.Features
- Generate high-quality images in 2-4 inference steps
- Compatible with custom base models
- Supports DDIM scheduler with trailing timestep spacing
- Can disable classifier-free guidance for 1-2 step generation
Example: 2-step generation
Key parameters for Hyper-SDXL
num_inference_steps: Set to 2 or 4 depending on the LoRA variantdo_classifier_free_guidance: Set toFalsefor 1-2 steps,Truefor 4+ stepsdiffusion_scheduler_config: Must useFlaxDDIMSchedulerwithtrailingtimestep spacingscale: Typically 0.7 for optimal results
Configuration tips
Balancing multiple LoRAs
- Start with equal scales (e.g., 0.5-0.7 each) and adjust based on results
- Lower scales produce more subtle effects
- Higher scales strengthen the LoRA’s influence but may introduce artifacts
- Total combined scale doesn’t need to equal 1.0
LoRA sources
You can mix and match LoRAs from:- Local file paths (safetensors or checkpoint files)
- HuggingFace Hub repositories
- Different LoRA formats (standard LoRA, LoCON, etc.)
Performance considerations
- Loading multiple LoRAs increases memory usage slightly
- Inference speed is generally unaffected as LoRAs are merged during initialization
- All LoRAs are merged using device-side JIT compilation for efficiency
Related resources
- LoRA inference - Single LoRA loading and configuration
- Checkpointing - Save models with LoRA weights