Mikkel
mikkelisk.bsky.social
Mikkel
@mikkelisk.bsky.social
I don't know what the worker threads you see are. Are you sure the code was written with multiple GPUs in mind? If so, you will likely see some reference to DDP somewhere pytorch.org/docs/stable/...
Distributed Data Parallel — PyTorch 2.5 documentation
pytorch.org
December 6, 2024 at 4:15 PM
It's unclear what your issue is. Do the other 7 GPUs have some utilization or zero utilization?
Are you talking about 8 dataloader worker threads or the number you get from torch.get_num_threads()/torch.get_num_interop_threads()?
December 6, 2024 at 3:57 PM
No, but it would be very nice to keep up with you here, rather than over there going forwards:)
November 18, 2024 at 12:38 PM