Single-GPU Setup: If you’re not using multiple GPUs for distributed training, remove the distributed setup code (dist.init_process_group) to avoid unnecessary complexity.
GPU Conflicts: Ensure that GPU 0 is not being used by any other processes. You can set CUDA_VISIBLE_DEVICES='1' to explicitly restrict the use of GPU 1.