From 87f6e34a563adf206f375ef2fa75a3c94ec953e7 Mon Sep 17 00:00:00 2001 From: Sebastian Golebiewski Date: Wed, 30 Aug 2023 11:52:36 +0200 Subject: [PATCH] [DOCS] Improving code snippets for quantization (#19479) * improve-snippets * Apply suggestions from code review Co-authored-by: Alexander Suslov * Update docs/optimization_guide/nncf/ptq/code/ptq_tensorflow.py Co-authored-by: Alexander Suslov * update-path * Update docs/optimization_guide/nncf/ptq/code/ptq_torch.py --------- Co-authored-by: Alexander Suslov --- docs/optimization_guide/nncf/ptq/code/ptq_onnx.py | 3 ++- docs/optimization_guide/nncf/ptq/code/ptq_openvino.py | 5 ++--- docs/optimization_guide/nncf/ptq/code/ptq_tensorflow.py | 3 ++- docs/optimization_guide/nncf/ptq/code/ptq_torch.py | 3 ++- 4 files changed, 8 insertions(+), 6 deletions(-) diff --git a/docs/optimization_guide/nncf/ptq/code/ptq_onnx.py b/docs/optimization_guide/nncf/ptq/code/ptq_onnx.py index fa9d1e35734..b8104d449c5 100644 --- a/docs/optimization_guide/nncf/ptq/code/ptq_onnx.py +++ b/docs/optimization_guide/nncf/ptq/code/ptq_onnx.py @@ -16,7 +16,8 @@ calibration_dataset = nncf.Dataset(calibration_loader, transform_fn) #! [dataset] #! [quantization] -model = ... # onnx.ModelProto object +import onnx +model = onnx.load("model_path") quantized_model = nncf.quantize(model, calibration_dataset) #! [quantization] diff --git a/docs/optimization_guide/nncf/ptq/code/ptq_openvino.py b/docs/optimization_guide/nncf/ptq/code/ptq_openvino.py index eb2f89edf36..81e19fd99f5 100644 --- a/docs/optimization_guide/nncf/ptq/code/ptq_openvino.py +++ b/docs/optimization_guide/nncf/ptq/code/ptq_openvino.py @@ -15,14 +15,13 @@ calibration_dataset = nncf.Dataset(calibration_loader, transform_fn) #! [dataset] #! [quantization] -model = ... # openvino.runtime.Model object +import openvino.runtime as ov +model = ov.Core().read_model("model_path") quantized_model = nncf.quantize(model, calibration_dataset) #! [quantization] #! [inference] -import openvino.runtime as ov - # compile the model to transform quantized operations to int8 model_int8 = ov.compile_model(quantized_model) diff --git a/docs/optimization_guide/nncf/ptq/code/ptq_tensorflow.py b/docs/optimization_guide/nncf/ptq/code/ptq_tensorflow.py index fcb29c8741e..e4fc817ddb3 100644 --- a/docs/optimization_guide/nncf/ptq/code/ptq_tensorflow.py +++ b/docs/optimization_guide/nncf/ptq/code/ptq_tensorflow.py @@ -15,7 +15,8 @@ calibration_dataset = nncf.Dataset(calibration_loader, transform_fn) #! [dataset] #! [quantization] -model = ... # tensorflow.Module object +import tensorflow as tf +model = tf.saved_model.load("model_path") quantized_model = nncf.quantize(model, calibration_dataset) #! [quantization] diff --git a/docs/optimization_guide/nncf/ptq/code/ptq_torch.py b/docs/optimization_guide/nncf/ptq/code/ptq_torch.py index a65f0998622..fbdf103b4f2 100644 --- a/docs/optimization_guide/nncf/ptq/code/ptq_torch.py +++ b/docs/optimization_guide/nncf/ptq/code/ptq_torch.py @@ -15,7 +15,8 @@ calibration_dataset = nncf.Dataset(calibration_loader, transform_fn) #! [dataset] #! [quantization] -model = ... # torch.nn.Module object +import torchvision +model = torchvision.models.resnet50(pretrained=True) quantized_model = nncf.quantize(model, calibration_dataset) #! [quantization]