We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b6cf8e5 commit 1b2d668Copy full SHA for 1b2d668
py/torch_tensorrt/dynamo/conversion/impl/quantize.py
@@ -45,7 +45,6 @@ def quantize(
45
Adds quantize and dequantize ops (QDQ) which quantize to INT8 or FP8 based
46
on the output_type set and dequantizes them back.
47
"""
48
-
49
with unset_fake_temporarily():
50
if isinstance(input_tensor, (torch.Tensor, TRTTensor)):
51
if input_tensor.dtype not in (
@@ -118,8 +117,6 @@ def quantize(
118
117
if not isinstance(input_tensor, TRTTensor):
119
input_tensor = get_trt_tensor(ctx, input_tensor, name + "_quantize_input")
120
121
- quantize_layer = ctx.net.add_quantize(input_tensor, scale, dtype)
122
123
# Add Q node
124
quantize_layer = ctx.net.add_quantize(input_tensor, scale, dtype)
125
if axis is not None:
0 commit comments