I want to train the model, but training on 1 GPU is too slow. So I want to use 4 GPUs. Is distributed data parallel possible? If so, could you please explain how to set for ddp? I really appreciate any help you can provide.
I want to train the model, but training on 1 GPU is too slow.
So I want to use 4 GPUs.
Is distributed data parallel possible? If so, could you please explain how to set for ddp?
I really appreciate any help you can provide.