Skip to content

Commit 2e005a8

Browse files
committed
make it a method
1 parent 9d2692d commit 2e005a8

File tree

2 files changed

+9
-10
lines changed

2 files changed

+9
-10
lines changed

pytorch_lightning/trainer/connectors/accelerator_connector.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -701,15 +701,15 @@ def select_training_type_plugin(self) -> TrainingTypePlugin:
701701
cluster_environment=self.select_cluster_environment(), parallel_devices=self.parallel_devices
702702
)
703703
elif self.use_ddp:
704-
use_slurm_ddp = self.use_ddp and self._is_slurm_managing_tasks
704+
use_slurm_ddp = self.use_ddp and self._is_slurm_managing_tasks()
705705
use_torchelastic_ddp = self.use_ddp and TorchElasticEnvironment.is_using_torchelastic()
706706
use_kubeflow_ddp = self.use_ddp and KubeflowEnvironment.is_using_kubeflow()
707707
use_ddp_spawn = self._distrib_type == _StrategyType.DDP_SPAWN
708708
use_ddp_cpu_spawn = use_ddp_spawn and self.use_cpu
709709
use_tpu_spawn = self.use_tpu and self._distrib_type == _StrategyType.TPU_SPAWN
710710
use_ddp_cpu_torch_elastic = use_ddp_cpu_spawn and TorchElasticEnvironment.is_using_torchelastic()
711711
use_ddp_cpu_kubeflow = use_ddp_cpu_spawn and KubeflowEnvironment.is_using_kubeflow()
712-
use_ddp_cpu_slurm = use_ddp_cpu_spawn and self._is_slurm_managing_tasks
712+
use_ddp_cpu_slurm = use_ddp_cpu_spawn and self._is_slurm_managing_tasks()
713713
use_ddp_sharded = self._distrib_type == _StrategyType.DDP_SHARDED
714714
use_ddp_sharded_spawn = self._distrib_type == _StrategyType.DDP_SHARDED_SPAWN
715715
use_ddp_fully_sharded = self._distrib_type == _StrategyType.DDP_FULLY_SHARDED
@@ -805,7 +805,7 @@ def select_accelerator(self) -> Accelerator:
805805
def select_cluster_environment(self) -> ClusterEnvironment:
806806
if self._cluster_environment is not None:
807807
return self._cluster_environment
808-
if self._is_slurm_managing_tasks:
808+
if self._is_slurm_managing_tasks():
809809
env = SLURMEnvironment()
810810
rank_zero_info("Multiprocessing is handled by SLURM.")
811811
elif TorchElasticEnvironment.is_using_torchelastic():
@@ -998,7 +998,6 @@ def _set_distrib_type_if_training_type_plugin_passed(self):
998998
if self._training_type_plugin is not None:
999999
self._distrib_type = getattr(self._training_type_plugin, "distributed_backend", None)
10001000

1001-
@property
10021001
def _is_slurm_managing_tasks(self) -> bool:
10031002
"""Returns whether we let SLURM manage the processes or not.
10041003

tests/accelerators/test_accelerator_connector.py

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -102,7 +102,7 @@ def test_accelerator_choice_ddp_spawn(cuda_available_mock, device_count_mock):
102102
def test_accelerator_choice_ddp_slurm(setup_distributed_mock):
103103
class CB(Callback):
104104
def on_fit_start(self, trainer, pl_module):
105-
assert trainer._accelerator_connector._is_slurm_managing_tasks
105+
assert trainer._accelerator_connector._is_slurm_managing_tasks()
106106
assert isinstance(trainer.accelerator, GPUAccelerator)
107107
assert isinstance(trainer.training_type_plugin, DDPPlugin)
108108
assert isinstance(trainer.training_type_plugin.cluster_environment, SLURMEnvironment)
@@ -134,7 +134,7 @@ def on_fit_start(self, trainer, pl_module):
134134
def test_accelerator_choice_ddp2_slurm(device_count_mock, setup_distributed_mock):
135135
class CB(Callback):
136136
def on_fit_start(self, trainer, pl_module):
137-
assert trainer._accelerator_connector._is_slurm_managing_tasks
137+
assert trainer._accelerator_connector._is_slurm_managing_tasks()
138138
assert isinstance(trainer.accelerator, GPUAccelerator)
139139
assert isinstance(trainer.training_type_plugin, DDP2Plugin)
140140
assert isinstance(trainer.training_type_plugin.cluster_environment, SLURMEnvironment)
@@ -309,7 +309,7 @@ def on_fit_start(self, trainer, pl_module):
309309
def test_accelerator_choice_ddp_cpu_slurm(device_count_mock, setup_distributed_mock):
310310
class CB(Callback):
311311
def on_fit_start(self, trainer, pl_module):
312-
assert trainer._accelerator_connector._is_slurm_managing_tasks
312+
assert trainer._accelerator_connector._is_slurm_managing_tasks()
313313
assert isinstance(trainer.accelerator, CPUAccelerator)
314314
assert isinstance(trainer.training_type_plugin, DDPPlugin)
315315
assert isinstance(trainer.training_type_plugin.cluster_environment, SLURMEnvironment)
@@ -778,7 +778,7 @@ def test_strategy_choice_ddp_spawn(cuda_available_mock, device_count_mock):
778778
def test_strategy_choice_ddp_slurm(setup_distributed_mock, strategy):
779779
class CB(Callback):
780780
def on_fit_start(self, trainer, pl_module):
781-
assert trainer._accelerator_connector._is_slurm_managing_tasks
781+
assert trainer._accelerator_connector._is_slurm_managing_tasks()
782782
assert isinstance(trainer.accelerator, GPUAccelerator)
783783
assert isinstance(trainer.training_type_plugin, DDPPlugin)
784784
assert isinstance(trainer.training_type_plugin.cluster_environment, SLURMEnvironment)
@@ -811,7 +811,7 @@ def on_fit_start(self, trainer, pl_module):
811811
def test_strategy_choice_ddp2_slurm(device_count_mock, setup_distributed_mock, strategy):
812812
class CB(Callback):
813813
def on_fit_start(self, trainer, pl_module):
814-
assert trainer._accelerator_connector._is_slurm_managing_tasks
814+
assert trainer._accelerator_connector._is_slurm_managing_tasks()
815815
assert isinstance(trainer.accelerator, GPUAccelerator)
816816
assert isinstance(trainer.training_type_plugin, DDP2Plugin)
817817
assert isinstance(trainer.training_type_plugin.cluster_environment, SLURMEnvironment)
@@ -987,7 +987,7 @@ def on_fit_start(self, trainer, pl_module):
987987
def test_strategy_choice_ddp_cpu_slurm(device_count_mock, setup_distributed_mock, strategy):
988988
class CB(Callback):
989989
def on_fit_start(self, trainer, pl_module):
990-
assert trainer._accelerator_connector._is_slurm_managing_tasks
990+
assert trainer._accelerator_connector._is_slurm_managing_tasks()
991991
assert isinstance(trainer.accelerator, CPUAccelerator)
992992
assert isinstance(trainer.training_type_plugin, DDPPlugin)
993993
assert isinstance(trainer.training_type_plugin.cluster_environment, SLURMEnvironment)

0 commit comments

Comments
 (0)