-
Notifications
You must be signed in to change notification settings - Fork 372
Closed
Description
Might be a copy paste error:
TensorRT/examples/dynamo/vgg16_fp8_ptq.py
Line 239 in 52ba6f1
out = model(data) |
above has this code:
trt_model = torchtrt.dynamo.compile(
exp_program,
inputs=[input_tensor],
enabled_precisions={torch.float8_e4m3fn},
min_block_size=1,
debug=False,
)
and then a note saying
# Inference compiled Torch-TensorRT model over the testing dataset
but then it looks like inference is done with the original model?
out = model(data)