Overview
The Training Configuration page (/training) provides a comprehensive form for setting up training hyperparameters, optimizers, learning rate schedules, and callbacks. Configurations are saved to a Training Library for reuse across experiments.
Page Layout
The page uses a sidebar + main form layout:- Left Sidebar (25%): List of saved training configurations
- Main Form (75%): Hyperparameter configuration interface
Training Library Sidebar
Saved Configs List
Displays all saved training configurations as a vertical list. Each entry shows:- Name: User-defined configuration name
- Selection indicator: ● (filled) if selected, ○ (empty) if not
- Delete button: 🗑️ icon to permanently remove
- Click name: Load configuration into form
- Click delete: Remove from library (confirmation not required)
New Config Button
+ New Config (bottom of sidebar)- Clears all form fields
- Resets to default values
- Deselects current configuration
The sidebar processes actions BEFORE form widgets render, ensuring loaded values appear correctly in the form.
Training Configuration Form
Config Name Input
Enter a descriptive name for this training configuration:- Example: “Adam_Default”, “SGD_LR0.01”, “FineTune_Light”
- Required before saving
- Used in library list and experiment selection
Optimizer Configuration
Optimizer Selection
Choose optimization algorithm from dropdown:- Adam (Default)
- AdamW
- SGD with Momentum
- RMSprop
Adaptive Moment Estimation
- Adaptive learning rates per parameter
- Combines momentum and RMSprop
- Fast convergence, low memory overhead
- Best for: Most scenarios, especially with limited tuning
Learning Rate
Slider: 0.0001 - 0.01 (4 decimal precision) Default values by optimizer:- Adam/AdamW/RMSprop: 0.001
- SGD: 0.01 (SGD typically needs higher LR)
Learning Rate Schedule
Adjust learning rate during training to improve convergence.Strategy Selection
- Constant (Default)
- ReduceLROnPlateau
- Cosine Annealing
Fixed learning rate throughout training.
- No adjustments
- Simplest approach
- Best for: Initial experiments, short training runs
Learning rate schedules are applied automatically during training. No additional configuration needed.
Training Parameters
Max Epochs
Slider: 10 - 200- Default: 100
- Maximum number of training epochs
- Training may stop earlier with Early Stopping
Batch Size
Dropdown: 16, 32, 64, 128- Default: 32
- Number of samples per gradient update
- Smaller (16): More gradient updates, slower, better generalization, less GPU memory
- Larger (128): Fewer updates, faster, may generalize worse, more GPU memory
Shuffle Data
Checkbox: Shuffle training data each epoch- Default: Enabled
- Randomizes sample order to prevent order bias
- Always leave enabled unless you have specific reasons
Regularization
Techniques to prevent overfitting by constraining model complexity.L2 Weight Decay
Checkbox: Enable L2 regularization- Adds penalty for large weights to loss function
- Encourages smaller, more distributed weights
- Recommended for most scenarios
- Default: 0.0001
- Regularization strength
- Higher = stronger penalty
L2 weight decay is integrated into optimizers (AdamW uses decoupled weight decay for better performance).
Class Imbalance Handling
Select method to handle imbalanced classes during training.- Auto Class Weights (Recommended)
- Focal Loss
- None
Automatically compute class weights inversely proportional to frequency.
- Loss function scales per-class losses by weight
- Classes with fewer samples get higher weights
- No data manipulation
weight[c] = n_samples / (n_classes * n_samples_c)Best for: Most imbalanced datasetsCallbacks
Automated actions during training.Early Stopping
Checkbox: Enable early stopping- Default: Enabled
- Stops training when validation metric stops improving
- Prevents overfitting and wasted compute
- Default: 10
- Number of epochs with no improvement before stopping
- Higher patience = more tolerance for plateaus
How it works: Monitors validation loss. If loss doesn’t improve for
patience epochs, training stops and the best model weights are restored.Model Checkpointing
Checkbox: Enable model checkpointing- Default: Enabled
- Saves best model weights during training
- Allows recovery of best model even if training continues past optimal point
- Val Loss (default): Save when validation loss decreases
- Val Accuracy: Save when validation accuracy increases
Checkpoints are saved to
repo/models/checkpoints/{experiment_id}/best_model.ptConfiguration Summary
The form builds a configuration dictionary:Saving Configurations
Save Buttons
Two buttons appear at the bottom:- Save as New
- Update
Primary button (when not editing)
- Saves as new entry in library
- Generates unique ID
- Clears form after save
- Shows success toast
- Config name must not be empty
- All fields must have valid values
- ✅ Success message: “Config '' saved!”
- Library updates with new/updated entry
- Form remains for further edits or clears (Save as New)
Example Configurations
Quick Experiment (Fast Training)
Quick Experiment (Fast Training)
Standard Training (Balanced)
Standard Training (Balanced)
Transfer Learning Fine-Tuning
Transfer Learning Fine-Tuning
Transformer Training (High Capacity)
Transformer Training (High Capacity)
Severe Imbalance Handling
Severe Imbalance Handling
Tips & Best Practices
Next Steps
After saving your training configuration:Training Monitor
Compose experiments and start training with real-time monitoring