I have a program that using distributeddataparallel with pytorch to perform muti-gpu parallelism. The script can run correctly by CUDA_VISIBLE_DEVICES=0,1 python ddp_test.
CUDA_VISIBLE_DEVICES=0,1 python ddp_test.