We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 14ab650 commit c82a24dCopy full SHA for c82a24d
docs/TRAINING.md
@@ -62,7 +62,7 @@ Links to the datasets:
62
We trained with four A100 GPUs, which took around 30 hours.
63
64
```
65
-OMP_NUM_THREADS=4 torchrun --master_port 25357 --nproc_per_node=4 train.py exp_id=[some unique id] model=[small/base] data=[base/with-mose/mega]
+OMP_NUM_THREADS=4 torchrun --master_port 25357 --nproc_per_node=4 cutie/train.py exp_id=[some unique id] model=[small/base] data=[base/with-mose/mega]
66
67
68
- Change `nproc_per_node` to change the number of GPUs.
0 commit comments