Prerequisites
- A Runpod account with billing set up
- An SSH key configured in the Runpod Settings tab
Setup
Create a Runpod account and add an SSH key
Sign up at console.runpod.io. Go to Settings and add your SSH public key so you can connect to pods from your terminal.
Launch a 1xH100 pod using the official template
Deploy using the official Parameter Golf template:Launch template on Runpod
- Select a 1xH100 GPU for initial experiments
- Enable SSH terminal access
- Leave all other settings at their defaults
- Click Deploy
pip install.Clone the repository on your remote machine
SSH into your pod once it’s running. You’ll land in
/workspace/.Download FineWeb training data
Download the cached 1024-token FineWeb export. This defaults to the full validation split plus 80 training shards (8B tokens):To download a smaller subset while iterating, pass
--train-shards N:Scaling to 8xH100s
To run a full leaderboard-eligible submission, changenproc_per_node to match the number of GPUs on your pod:
Environment variables
| Variable | Default | Description |
|---|---|---|
RUN_ID | random UUID | Name for this run’s log directory |
DATA_PATH | ./data/datasets/fineweb10B_sp1024 | Path to dataset shards |
TOKENIZER_PATH | ./data/tokenizers/fineweb_1024_bpe.model | Path to tokenizer model |
VOCAB_SIZE | 1024 | Vocabulary size, must match tokenizer |
MAX_WALLCLOCK_SECONDS | 600 | Hard stop in seconds; set to 0 to disable |
VAL_LOSS_EVERY | 1000 | Print validation loss every N steps |
TRAIN_LOG_EVERY | 200 | Print training loss every N steps |
Understanding the log output
During training, the script printstrain_loss at every TRAIN_LOG_EVERY steps. At the end, it prints:
val_loss— cross-entropy loss on the validation setval_bpb— bits per byte, the leaderboard metricfinal_int8_zlib_roundtriplines — compressed model size in bytes
By default,
train_gpt.py enforces a 10-minute wallclock cap (MAX_WALLCLOCK_SECONDS=600). To run longer, set MAX_WALLCLOCK_SECONDS=0 to disable the cap or pass an explicit second count.Next steps
Data setup
Download the full 10B token dataset or customize tokenizer variants.
Submission requirements
Learn what files and logs are required for a valid leaderboard submission.
