Multi-GPU Training
Cold Start Training with Multiple GPUs
#!/bin/bash
#SBATCH -p GPU-shared
#SBATCH --output=/path/to/your/log/file.txt
#SBATCH -N 1
#SBATCH -t 18:00:00
#SBATCH --gpus=v100-32:4
module load openmpi/4.0.5-gcc10.2.0
singularity exec --nv /path/to/your/isciml.sif mpirun -np 4 isciml train \
--sample_folder /path/to/your/samples/*Adj*Data \
--target_folder /path/to/your/targets/*Sus*Models \
--n_gpus 4 \
--batch_size 25 \
--max_epochs 12 \
--save_top_k 4 \
--strategy ddp_find_unused_parameters_true \
--n_blocks 5 \
--start_filters 128Warm Start Training
Important Notes
Last updated