diff --git a/docs/optimization_guide/nncf/ptq/basic_quantization_flow.md b/docs/optimization_guide/nncf/ptq/basic_quantization_flow.md index 2f315c04705..fb5ab52aa02 100644 --- a/docs/optimization_guide/nncf/ptq/basic_quantization_flow.md +++ b/docs/optimization_guide/nncf/ptq/basic_quantization_flow.md @@ -1,4 +1,4 @@ -# Basic Quantization Flow {#basic_qauntization_flow} +# Basic Quantization Flow {#basic_quantization_flow} @sphinxdirective diff --git a/docs/optimization_guide/nncf/ptq/ptq_introduction.md b/docs/optimization_guide/nncf/ptq/ptq_introduction.md index a87e5f9d293..2cd880b5060 100644 --- a/docs/optimization_guide/nncf/ptq/ptq_introduction.md +++ b/docs/optimization_guide/nncf/ptq/ptq_introduction.md @@ -6,17 +6,21 @@ :maxdepth: 1 :hidden: - basic_qauntization_flow + basic_quantization_flow quantization_w_accuracy_control -@endsphinxdirective -Neural Network Compression Framework (NNCF) provides a new post-training quantization API available in Python that is aimed at reusing the code for model training or validation that is usually available with the model in the source framework, for example, PyTorch* or TensroFlow*. The API is cross-framework and currently supports models representing in the following frameworks: PyTorch, TensorFlow 2.x, ONNX, and OpenVINO. +Neural Network Compression Framework (NNCF) provides a new post-training quantization API available in Python that is aimed at reusing the code for model training or validation that is usually available with the model in the source framework, for example, PyTorch or TensroFlow. The API is cross-framework and currently supports models representing in the following frameworks: PyTorch, TensorFlow 2.x, ONNX, and OpenVINO. + This API has two main capabilities to apply 8-bit post-training quantization: -* [Basic quantization](@ref basic_qauntization_flow) - the simplest quantization flow that allows to apply 8-bit integer quantization to the model. -* [Quantization with accuracy control](@ref quantization_w_accuracy_control) - the most advanced quantization flow that allows to apply 8-bit quantization to the model with accuracy control. -## See also +* :doc:`Basic quantization ` - the simplest quantization flow that allows to apply 8-bit integer quantization to the model. +* :doc:`Quantization with accuracy control ` - the most advanced quantization flow that allows to apply 8-bit quantization to the model with accuracy control. -* [NNCF GitHub](https://github.com/openvinotoolkit/nncf) -* [Optimizing Models at Training Time](@ref tmo_introduction) \ No newline at end of file +Additional Resources +#################### + +* `NNCF GitHub `__ +* :doc:`Optimizing Models at Training Time ` + +@endsphinxdirective diff --git a/docs/optimization_guide/nncf/ptq/quantization_w_accuracy_control.md b/docs/optimization_guide/nncf/ptq/quantization_w_accuracy_control.md index 65d5ede50e4..fec080c0b0a 100644 --- a/docs/optimization_guide/nncf/ptq/quantization_w_accuracy_control.md +++ b/docs/optimization_guide/nncf/ptq/quantization_w_accuracy_control.md @@ -5,12 +5,12 @@ Introduction #################### -This is the advanced quantization flow that allows to apply 8-bit quantization to the model with control of accuracy metric. This is achieved by keeping the most impactful operations within the model in the original precision. The flow is based on the :doc:`Basic 8-bit quantization ` and has the following differences: +This is the advanced quantization flow that allows to apply 8-bit quantization to the model with control of accuracy metric. This is achieved by keeping the most impactful operations within the model in the original precision. The flow is based on the :doc:`Basic 8-bit quantization ` and has the following differences: * Beside the calibration dataset, a **validation dataset** is required to compute accuracy metric. They can refer to the same data in the simplest case. * **Validation function**, used to compute accuracy metric is required. It can be a function that is already available in the source framework or a custom function. -* Since accuracy validation is run several times during the quantization process, quantization with accuracy control can take more time than the [Basic 8-bit quantization](@ref basic_qauntization_flow) flow. -* The resulted model can provide smaller performance improvement than the :doc:`Basic 8-bit quantization ` flow because some of the operations are kept in the original precision. +* Since accuracy validation is run several times during the quantization process, quantization with accuracy control can take more time than the [Basic 8-bit quantization](@ref basic_quantization_flow) flow. +* The resulted model can provide smaller performance improvement than the :doc:`Basic 8-bit quantization ` flow because some of the operations are kept in the original precision. .. note:: Currently, this flow is available only for models in OpenVINO representation. @@ -19,7 +19,7 @@ The steps for the quantization with accuracy control are described below. Prepare datasets #################### -This step is similar to the :doc:`Basic 8-bit quantization ` flow. The only difference is that two datasets, calibration and validation, are required. +This step is similar to the :doc:`Basic 8-bit quantization ` flow. The only difference is that two datasets, calibration and validation, are required. .. tab:: OpenVINO