@@ -424,7 +424,7 @@ def __init__(
424424 gpu_ids , tpu_cores = self ._parse_devices (gpus , auto_select_gpus , tpu_cores )
425425
426426 # init connectors
427- self .data_connector = DataConnector (self , multiple_trainloader_mode )
427+ self ._data_connector = DataConnector (self , multiple_trainloader_mode )
428428 self .optimizer_connector = OptimizerConnector (self )
429429
430430 self .accelerator_connector = AcceleratorConnector (
@@ -514,7 +514,7 @@ def __init__(
514514 self .optimizer_connector .on_trainer_init ()
515515
516516 # init data flags
517- self .data_connector .on_trainer_init (
517+ self ._data_connector .on_trainer_init (
518518 check_val_every_n_epoch ,
519519 reload_dataloaders_every_n_epochs ,
520520 reload_dataloaders_every_epoch ,
@@ -663,7 +663,7 @@ def _fit_impl(
663663 )
664664
665665 # links data to the trainer
666- self .data_connector .attach_data (
666+ self ._data_connector .attach_data (
667667 model , train_dataloaders = train_dataloaders , val_dataloaders = val_dataloaders , datamodule = datamodule
668668 )
669669
@@ -747,7 +747,7 @@ def _validate_impl(
747747 )
748748
749749 # links data to the trainer
750- self .data_connector .attach_data (model , val_dataloaders = dataloaders , datamodule = datamodule )
750+ self ._data_connector .attach_data (model , val_dataloaders = dataloaders , datamodule = datamodule )
751751
752752 self .validated_ckpt_path = self .__set_ckpt_path (
753753 ckpt_path , model_provided = model_provided , model_connected = self .lightning_module is not None
@@ -837,7 +837,7 @@ def _test_impl(
837837 )
838838
839839 # links data to the trainer
840- self .data_connector .attach_data (model , test_dataloaders = dataloaders , datamodule = datamodule )
840+ self ._data_connector .attach_data (model , test_dataloaders = dataloaders , datamodule = datamodule )
841841
842842 self .tested_ckpt_path = self .__set_ckpt_path (
843843 ckpt_path , model_provided = model_provided , model_connected = self .lightning_module is not None
@@ -921,7 +921,7 @@ def _predict_impl(
921921 )
922922
923923 # links data to the trainer
924- self .data_connector .attach_data (model , predict_dataloaders = dataloaders , datamodule = datamodule )
924+ self ._data_connector .attach_data (model , predict_dataloaders = dataloaders , datamodule = datamodule )
925925
926926 self .predicted_ckpt_path = self .__set_ckpt_path (
927927 ckpt_path , model_provided = model_provided , model_connected = self .lightning_module is not None
@@ -985,7 +985,7 @@ def tune(
985985 )
986986
987987 # links data to the trainer
988- self .data_connector .attach_data (
988+ self ._data_connector .attach_data (
989989 model , train_dataloaders = train_dataloaders , val_dataloaders = val_dataloaders , datamodule = datamodule
990990 )
991991
@@ -1027,7 +1027,7 @@ def _run(self, model: "pl.LightningModule") -> Optional[Union[_EVALUATE_OUTPUT,
10271027 self .training_type_plugin .connect (model )
10281028
10291029 # hook
1030- self .data_connector .prepare_data ()
1030+ self ._data_connector .prepare_data ()
10311031 self .callback_connector ._attach_model_callbacks ()
10321032
10331033 if self ._ckpt_path and not self .training_type_plugin .restore_checkpoint_after_pre_dispatch :
@@ -1171,7 +1171,7 @@ def _post_dispatch(self):
11711171 # these `teardown` calls are here instead of in `_call_teardown_hook` since they are internal teardowns
11721172 # which need to happen before.
11731173 self .accelerator .teardown ()
1174- self .data_connector .teardown ()
1174+ self ._data_connector .teardown ()
11751175 self ._active_loop .teardown ()
11761176 self .logger_connector .teardown ()
11771177
@@ -1258,7 +1258,7 @@ def _run_predict(self) -> Optional[_PREDICT_OUTPUT]:
12581258 return self .predict_loop .run ()
12591259
12601260 def _run_sanity_check (self , ref_model ):
1261- using_val_step = self .data_connector ._val_dataloader_source .is_defined () and is_overridden (
1261+ using_val_step = self ._data_connector ._val_dataloader_source .is_defined () and is_overridden (
12621262 "validation_step" , ref_model
12631263 )
12641264 should_sanity_check = using_val_step and self .num_sanity_val_steps > 0 and self .limit_val_batches > 0
0 commit comments