### Discussed in https://github.com/PyTorchLightning/pytorch-lightning/discussions/8019 <div type='discussions-op-text'> <sup>Originally posted by **thomas-happify** June 17, 2021</sup> ``` seed_everything(43, worker=True) Trainer(gpus=1, accumulate_grad_batches=16, accelerator='ddp', plugin='deepspeed', max_epoch=20) Trainer(gpus=2, accumulate_grad_batches=8, accelerator='ddp', plugin='deepspeed', max_epoch=20) ``` shouldn't these two args have similar training results? when I use `gpus=2` and `accumulate_grad_batches=8`, the model can't converge.</div>