Hello. I’m using a machine with 4 gpus controlled ...
# neural-forecast
j
Hello. I’m using a machine with 4 gpus controlled via slurm jobs (it’s an hpc) When trying toto allocate large models I’m getting OOM errors on torch cuda memory because it’s trying to allocate everything to the first gpu. I see the four devices in the torch printouts. Is there a way to use multiple gpus on a single node? Best