I want to train my model through DistributedDataParallel on a single machine that has 8 GPUs. But I want to train my model on four specified GPUs with device IDs 4, 5, 6, 7.
How to specify the GPU device ID for DistributedDataParallel?
I think the world size will be 4 for this case, but what should be the rank in this case?