File tree Expand file tree Collapse file tree 1 file changed +2
-3
lines changed
pytorch_lightning/accelerators Expand file tree Collapse file tree 1 file changed +2
-3
lines changed Original file line number Diff line number Diff line change 4444from pytorch_lightning .utilities .exceptions import MisconfigurationException
4545
4646try :
47- import torch_xla
4847 import torch_xla .core .xla_model as xm
4948except ImportError :
5049 XLA_AVAILABLE = False
@@ -241,8 +240,8 @@ def select_training_type_plugin(self):
241240 use_ddp_cpu_spawn = self .use_ddp and self .distributed_backend == "ddp_cpu"
242241 use_ddp_cpu_torch_elastic = use_ddp_cpu_spawn and self .is_using_torchelastic
243242 use_ddp_cpu_slurm = use_ddp_cpu_spawn and self .is_slurm_managing_tasks
244- use_ddp_sharded = self .distributed_backend == "ddp_sharded"
245- use_ddp_sharded_spawn = self .distributed_backend == "ddp_sharded_spawn"
243+ # use_ddp_sharded = self.distributed_backend == "ddp_sharded"
244+ # use_ddp_sharded_spawn = self.distributed_backend == "ddp_sharded_spawn"
246245
247246 if self .on_tpu :
248247 ddp_plugin_cls = TPUSpawnPlugin
You can’t perform that action at this time.
0 commit comments