We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 0bdde92 commit f47b773Copy full SHA for f47b773
tutorials/quantize_vit/run_vit_b_quant.py
@@ -36,6 +36,9 @@
36
if not TORCH_VERSION_AT_LEAST_2_5:
37
unwrap_tensor_subclass(model)
38
39
+# temporary workaround to recover the perf with quantized model under torch.compile
40
+torch.backends.mha.set_fastpath_enabled(False)
41
+
42
model = torch.compile(model, mode='max-autotune')
43
44
# Must run with no_grad when optimizing for inference
0 commit comments