Fix automatic scripting for quantized models (#18089)
This commit is contained in:
parent
3d79bd1ac5
commit
77584d301e
@ -176,6 +176,8 @@ class TorchScriptPythonDecoder (Decoder):
|
||||
break
|
||||
if not skip_freeze:
|
||||
f_model = torch.jit.freeze(scripted)
|
||||
else:
|
||||
f_model = scripted
|
||||
else:
|
||||
f_model = pt_module
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user