Before you begin
You need Python 3.8+ and pip installed on your system.Train your first model
Install dependencies
Install the required packages:This installs NumPy, pandas, matplotlib, psutil, and other core dependencies.
Verify your environment
Run the environment verification script to confirm everything is set up correctly:You should see output showing your Python version, platform, and package status:
The dataset status may show files as unavailable—that’s expected for synthetic mode.
Run your first training workflow
Execute the full workflow with synthetic data:This command:
- Trains a neural network with the baseline configuration
- Runs benchmarks to measure latency and memory
- Performs statistical analysis over 5 repeated runs
The baseline experiment uses synthetic data by default, so no dataset download is required.
Review your results
After training completes, you’ll find:
- Checkpoints: Saved model weights in
experiments/logs/ - Metrics: Training loss, validation accuracy, and convergence data
- Benchmarks: Latency measurements, memory footprint, and throughput statistics
- Reports: Statistical summaries and artifacts in
artifacts/
- Parameter count vs. memory footprint
- Inference latency across different batch sizes
- Accuracy under various precision constraints
What you just built
You trained a feed-forward neural network that exposes system-level trade-offs:- Parameter count vs. memory: See how model size affects RAM usage
- Precision vs. accuracy: Compare float32, float16, and int8 representations
- Batch size vs. latency: Understand throughput vs. per-sample inference time
Next steps
Train with real data
Download Fashion-MNIST and train on production-like data
Run specific stages
Execute training, benchmarking, or analysis independently
Customize experiments
Modify layer sizes, precision, and hardware constraints
Understand the architecture
Learn how the framework is structured
Train with Fashion-MNIST
To train on real data instead of synthetic samples:Stage-specific runs
You can run individual stages of the workflow independently:Understanding the workflow modes
Therun_workflow.py script supports three modes:
| Mode | What it does | Use case |
|---|---|---|
train | Trains the model and saves checkpoints | Iterate on model architecture |
benchmark | Runs inference benchmarks and statistical analysis | Profile performance changes |
full | Executes train → benchmark → analysis | End-to-end experiment |
Troubleshooting
Import errors after installation
Import errors after installation
Ensure you’re in the correct directory and Python can find the task modules:The scripts automatically add the task directory to your Python path.
Out of memory during training
Out of memory during training
Reduce the batch size in your experiment configuration or use a smaller model architecture. The framework will show memory allocation spikes during forward/backward passes.
Dataset integrity check fails
Dataset integrity check fails
If you’re using real data, re-download the Fashion-MNIST files:For synthetic mode, no dataset files are required.
Key concepts
Before diving deeper, understand these core concepts:- Synthetic mode: Uses generated random data for fast, deterministic testing
- Experiment configs: Pre-defined settings in
config.pythat control architecture and training - Reproducibility: All experiments use explicit seeds and fixed dependency versions
- Hardware constraints: Simulated precision and memory limits without specialized hardware