Core Concepts
The Text-to-Image Pipeline
- Text Encoding: Your prompt is converted to embeddings via CLIP
- Latent Generation: Start with random noise in latent space
- Denoising: The model iteratively removes noise guided by your prompt
- Decoding: VAE converts the latent to a visible image
Key Nodes
- CheckpointLoaderSimple: Loads MODEL, CLIP, and VAE
- CLIPTextEncode: Converts text to conditioning
- EmptyLatentImage: Creates the initial noise tensor
- KSampler: Performs the denoising process
- VAEDecode: Converts latent to pixels
- SaveImage: Saves the final output
Basic Workflow
A serene mountain landscape at sunset, dramatic lighting,
high detail, photorealistic, 8k resolution
Advanced Prompting
Emphasis Syntax
Control the importance of specific words:() is 1.1.
Dynamic Prompts
Use wildcards for variation:Comments
C-style comments work in prompts:Textual Inversion
Place embeddings inmodels/embeddings/ and reference:
Parameter Guide
Seed
- Purpose: Controls randomness
- Same seed + settings = identical output
- Tip: Lock seed when iterating on prompts
Steps
- Range: 15-50 typically
- 15-20: Fast preview quality
- 25-35: Production quality
- 40+: Diminishing returns
CFG Scale
- Range: 1-20
- 1-5: Creative, loose interpretation
- 7-9: Balanced (recommended)
- 10-15: Strict adherence to prompt
- 15+: May oversaturate or distort
Sampler Selection
Fast samplers:euler: Simple, fast, good for previewseuler_ancestral: Adds variation between steps
dpm++ 2m karras: Excellent quality/speed balancedpm++ sde karras: High quality, slowerdpm++ 2m sde karras: Best quality, slowest
ddim: Deterministic, good for img2imguni_pc: Fast convergence, experimental
Scheduler
- normal: Linear noise schedule
- karras: Better for DPM++ samplers
- exponential: Smoother transitions
- sgm_uniform: For SD3/SDXL
Workflow Example: High-Quality Portrait
Batch Generation
Generate multiple variations:- Set batch_size in EmptyLatentImage to 4 (or higher)
- Keep seed the same for subtle variations
- Or use different seeds for diverse results
LoRA Integration
Enhance your model with LoRAs:- Add LoraLoader node
- Connect MODEL and CLIP from checkpoint
- Select LoRA file
- Set strength_model (0.5-1.0 typical)
- Set strength_clip (0.5-1.0 typical)
- Connect LoraLoader outputs to KSampler
Troubleshooting
Black/blank images
- Check VAE compatibility with your model
- Try loading a separate VAE with VAELoader
- Reduce CFG scale
Low quality results
- Increase steps (try 30-40)
- Use better sampler (dpm++ 2m karras)
- Improve prompt detail and specificity
- Check if model supports your resolution
Out of memory
- Reduce image dimensions
- Lower batch size
- Use tiled VAE for very large images
- Enable CPU offloading:
--lowvramflag
Next Steps
- Explore Image-to-Image for variations
- Learn Inpainting for editing
- Try ControlNet for precise control
- Experiment with different models and LoRAs