You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"""Takes a name, target device, serialized TensorRT engine, and binding names / order and constructs
133
133
a PyTorch ``torch.nn.Module`` around it. Uses TensorRT Python APIs to run the engine
@@ -141,7 +141,7 @@ def __init__(
141
141
name (str): Name for module
142
142
settings (torch_tensorrt.dynamo.CompilationSettings): Settings used to compile engine, assumes engine was built with default compilation settings if object not passed
143
143
weight_name_map (dict): Mapping of engine weight name to state_dict weight name
144
-
engine_is_dds (bool): Whether the engine is Data Dependent Shape
144
+
requires_output_allocator (bool): Whether the engine requires an output allocator
logger.warning(f"Moved all input Tensors to cuda:{device_id}")
680
680
681
-
ifself.engine_is_dds:
681
+
ifself.requires_output_allocator:
682
682
ifself.cudagraphs_enabled:
683
683
raiseRuntimeError(
684
-
"The module is Data-Dependent Shape (DDS). It has to be handled by OutputAllocator which is not compatible with CUDA Graphs. Please disable CUDA Graphs."
684
+
"This module requires OutputAllocator which is not compatible with CUDA Graphs. Please disable CUDA Graphs."
685
685
)
686
-
logger.debug(
687
-
"The module is Data-Dependent Shape (DDS). Using output allocator."
0 commit comments