Files
openvino/docs/optimization_guide/dldt_optimization_guide.md
Alexander Kozlov 1bbd92a8f8 Revised Tuning For Performance and Model optimization docs (#10276)
* Revised Tuning for performance and Model optimization docs

* Fixed links

* Fixed link

* Applied comments

* Fixed one more comment
2022-03-03 18:58:58 +03:00

2.0 KiB

Performance Optimization Guide

Before exploring optimization techniques, let us first define what performance is and how it is measured.

What Is Performance

Performance means how fast the model is in deployment. Two key metrics are used to measure performance: latency and throughput.

Latency measures inference time (ms) required to process a single input. When it comes to batch input need to measure throughput (images per second or frames per second, FPS). To calculate throughput, divide the number of frames that were processed by the processing time.

How to measure performance

To get performance numbers for OpenVINO, as well as tips how to measure it and compare with native framework, go to Getting performance numbers page.

How to Improve Performance

Note

: Make sure that your model can be successfully inferred with OpenVINO Inference Engine before reffering to the optimization topic.

Inside OpenVINO there are two ways how to get better performance numbers: optimize the model, which is called model optimization or tune parameters of execution, which is also deployment optimization. Note, that it is possible to combine both types of optimizations.

  • Model optimization includes model modification, such as quantization, pruning, optimization of preprocessing, etc. Fore more details, refer to this document.

  • Deployment optimization includes tuning inference parameters and optimizing model execution. To read more visit Deployment Optimization Guide.

Performance benchmarks

To estimate the performance and compare performance numbers, measured on various supported devices, a wide range of public models are available at Perforance benchmarks section.