Skip to content

Commit bc23966

Browse files
committed
flake8
1 parent 6b23877 commit bc23966

File tree

3 files changed

+7
-5
lines changed

3 files changed

+7
-5
lines changed

pytorch_lightning/core/memory.py

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -182,7 +182,7 @@ def __init__(self, model, mode: str = MODE_DEFAULT):
182182
self._model = model
183183
self._mode = mode
184184
self._layer_summary = self.summarize()
185-
self._precision_megabytes = (self._model.precision / 8.0) * 1e-6 # 1 byte -> 8 bits
185+
self._precision_megabytes = (self._model.precision / 8.0) * 1e-6 # 1 byte -> 8 bits
186186

187187
@property
188188
def named_modules(self) -> List[Tuple[str, nn.Module]]:
@@ -389,9 +389,11 @@ def get_gpu_memory_map() -> Dict[str, int]:
389389
}
390390
return gpu_memory_map
391391

392+
392393
def get_formatted_model_size(total_model_size: float) -> float:
393394
return f"{total_model_size:,.3f}"
394395

396+
395397
def get_human_readable_count(number: int) -> str:
396398
"""
397399
Abbreviates an integer number with K, M, B, T for thousands, millions,

pytorch_lightning/loggers/wandb.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -29,9 +29,8 @@
2929
_WANDB_AVAILABLE = _module_available("wandb")
3030

3131
try:
32-
from wandb.wandb_run import Run
33-
3432
import wandb
33+
from wandb.wandb_run import Run
3534
except ImportError:
3635
# needed for test mocks, these tests shall be updated
3736
wandb, Run = None, None

tests/core/test_memory.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -40,8 +40,8 @@ class PreCalculatedModel(BoringModel):
4040

4141
def __init__(self, precision: int = 32):
4242
super().__init__()
43-
self.layer = nn.Linear(32, 1000, bias=False) # 32K params
44-
self.layer1 = nn.Linear(1000, 218, bias=False) # 218K params
43+
self.layer = nn.Linear(32, 1000, bias=False) # 32K params
44+
self.layer1 = nn.Linear(1000, 218, bias=False) # 218K params
4545

4646
# calculate model size based on precision.
4747
self.pre_calculated_model_size = 1.0 / (32 / precision)
@@ -50,6 +50,7 @@ def forward(self, x):
5050
x = self.layer(x)
5151
return self.layer1(x)
5252

53+
5354
class UnorderedModel(LightningModule):
5455
""" A model in which the layers not defined in order of execution """
5556

0 commit comments

Comments
 (0)