2020-07-20 17:36:08 +03:00
# Benchmark C++ Tool {#openvino_inference_engine_samples_benchmark_app_README}
2018-11-23 16:19:43 +03:00
2021-09-13 15:40:36 +03:00
This topic demonstrates how to use the Benchmark C++ Tool to estimate deep learning inference performance on supported devices.
Performance can be measured for two inference modes: latency- and throughput-oriented.
2019-01-21 21:31:31 +03:00
2021-12-21 20:26:37 +03:00
> **NOTE**: This topic describes usage of C++ implementation of the Benchmark Tool. For the Python* implementation, refer to [Benchmark Python* Tool](../../../tools/benchmark_tool/README.md).
2019-01-21 21:31:31 +03:00
2018-11-23 16:19:43 +03:00
## How It Works
2021-09-13 15:40:36 +03:00
Upon start-up, the application reads command-line parameters and loads a network and inputs (images/binary files) to the specified device.
2018-11-23 16:19:43 +03:00
2022-03-24 22:27:29 +03:00
> **NOTE**: By default, OpenVINO™ Toolkit Samples, Tools and Demos expect input with BGR channels order. If you trained your model to work with RGB order, you need to manually rearrange the default channels order in the sample or demo application or reconvert your model using the Model Optimizer tool with `--reverse_input_channels` argument specified. For more information about the argument, refer to **When to Reverse Input Channels** section of [Embedding Preprocessing Computation](../../../docs/MO_DG/prepare_model/convert_model/Converting_Model.md).
2018-11-23 16:19:43 +03:00
2021-09-13 15:40:36 +03:00
Device-specific execution parameters (number of streams, threads, and so on) can be either explicitly specified through the command line
or left default. In the last case, the sample logic will select the values for the optimal throughput.
While experimenting with individual parameters allows to find the performance sweet spot, usually, the parameters are not very performance-portable,
so the values from one machine or device are not necessarily optimal for another.
From this perspective, the most portable way is experimenting only with the performance hints. To learn more, refer to the section on the command-line parameters below.
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
A number of execution steps is defined by one of the following parameters:
2019-04-12 18:25:53 +03:00
* Number of iterations specified with the `-niter` command-line argument
2019-08-09 19:02:42 +03:00
* Time duration specified with the `-t` command-line argument
* Both of them (execution will continue until both conditions are met)
* Predefined duration if `-niter` and `-t` are not specified. Predefined duration value depends on a device.
2018-11-23 16:19:43 +03:00
2021-09-13 15:40:36 +03:00
During the execution, the application calculates latency (if applicable) and overall throughput:
* By default, the median latency value is reported
* Throughput is calculated as overall_inference_time/number_of_processed_requests. Note that the throughput value also depends on batch size.
2019-04-12 18:25:53 +03:00
The application also collects per-layer Performance Measurement (PM) counters for each executed infer request if you
enable statistics dumping by setting the `-report_type` parameter to one of the possible values:
* `no_counters` report includes configuration options specified, resulting FPS and latency.
2019-08-09 19:02:42 +03:00
* `average_counters` report extends the `no_counters` report and additionally includes average PM counters values for each layer from the network.
* `detailed_counters` report extends the `average_counters` report and additionally includes per-layer PM counters and latency for each executed infer request.
2019-04-12 18:25:53 +03:00
2019-08-09 19:02:42 +03:00
Depending on the type, the report is stored to `benchmark_no_counters_report.csv` , `benchmark_average_counters_report.csv` ,
2019-04-12 18:25:53 +03:00
or `benchmark_detailed_counters_report.csv` file located in the path specified in `-report_folder` .
2020-07-20 17:36:08 +03:00
The application also saves executable graph information serialized to an XML file if you specify a path to it with the
2019-04-12 18:25:53 +03:00
`-exec_graph_path` parameter.
2018-11-23 16:19:43 +03:00
2019-10-04 19:26:43 +03:00
## Run the Tool
2021-02-02 11:29:12 +03:00
2020-07-20 17:36:08 +03:00
Note that the benchmark_app usually produces optimal performance for any device out of the box.
2019-08-09 19:02:42 +03:00
2019-10-04 19:26:43 +03:00
**So in most cases you don't need to play the app options explicitly and the plain device name is enough**, for example, for CPU:
```sh
./benchmark_app -m < model > -i < input > -d CPU
2019-08-09 19:02:42 +03:00
```
2019-10-04 19:26:43 +03:00
2022-01-27 19:39:49 +03:00
But it is still may be sub-optimal for some cases, especially for very small networks. More details can read in [Performance Optimization Guide ](../../../docs/optimization_guide/dldt_optimization_guide.md ).
2019-10-04 19:26:43 +03:00
2022-02-22 14:32:57 +03:00
As explained in the [Performance Optimization Guide ](../../../docs/optimization_guide/dldt_optimization_guide.md ) section, for all devices, including new [MULTI device ](../../../docs/OV_Runtime_UG/multi_device.md ) it is preferable to use the FP16 IR for the model.
2019-10-04 19:26:43 +03:00
Also if latency of the CPU inference on the multi-socket machines is of concern, please refer to the same
2022-01-27 19:39:49 +03:00
[Performance Optimization Guide ](../../../docs/optimization_guide/dldt_optimization_guide.md ).
2018-11-23 16:19:43 +03:00
Running the application with the `-h` option yields the following usage message:
2019-08-09 19:02:42 +03:00
```
2018-11-23 16:19:43 +03:00
./benchmark_app -h
benchmark_app [OPTION]
Options:
2022-03-21 19:31:17 +03:00
-h, --help Print a usage message
-m "< path > " Required. Path to an .xml/.onnx file with a trained model or to a .blob files with a trained compiled model.
-i "< path > " Optional. Path to a folder with images and/or binaries or to specific image or binary file.
In case of dynamic shapes networks with several inputs provide the same number of files for each input (except cases with single file for any input):"input1:1.jpg input2:1.bin", "input1:1.bin,2.bin input2:3.bin input3:4.bin,5.bin ". Also you can pass specific keys for inputs: "random" - for fillling input with random data, "image_info" - for filling input with image size.
You should specify either one files set to be used for all inputs (without providing input names) or separate files sets for every input of model (providing inputs names).
-d "< device > " Optional. Specify a target device to infer on (the list of available devices is shown below). Default value is CPU. Use "-d HETERO:< comma-separated_devices_list > " format to specify HETERO plugin. Use "-d MULTI:< comma-separated_devices_list > " format to specify MULTI plugin. The application looks for a suitable plugin for the specified device.
2022-04-04 16:08:38 +03:00
-extensions "< absolute_path > " Required for custom layers (extensions). Absolute path to a shared library with the kernels implementations.
2018-11-23 16:19:43 +03:00
Or
2022-03-21 19:31:17 +03:00
-c "< absolute_path > " Required for GPU custom kernels. Absolute path to an .xml file with the kernels description.
2022-05-06 12:42:59 +08:00
-hint "performance hint (latency or throughput or cumulative_throughput or none)" Optional. Performance hint allows the OpenVINO device to select the right network-specific settings.
2022-03-21 19:31:17 +03:00
'throughput' or 'tput': device performance mode will be set to THROUGHPUT.
2022-05-06 12:42:59 +08:00
'cumulative_throughput' or 'ctput': device performance mode will be set to CUMULATIVE_THROUGHPUT.
2022-03-21 19:31:17 +03:00
'latency': device performance mode will be set to LATENCY.
'none': no device performance mode will be set.
Using explicit 'nstreams' or other device-specific options, please set hint to 'none'
-api "< sync / async > " Optional (deprecated). Enable Sync/Async API. Default value is "async".
-niter "< integer > " Optional. Number of iterations. If not specified, the number of iterations is calculated depending on a device.
-nireq "< integer > " Optional. Number of infer requests. Default value is determined automatically for device.
-b "< integer > " Optional. Batch size value. If not specified, the batch size value is determined from Intermediate Representation.
-stream_output Optional. Print progress as a plain text. When specified, an interactive progress bar is replaced with a multiline output.
-t Optional. Time in seconds to execute topology.
-progress Optional. Show progress bar (can affect performance measurement). Default values is "false".
-shape Optional. Set shape for network input. For example, "input1[1,3,224,224],input2[1,4]" or "[1,3,224,224]" in case of one input size. This parameter affect model input shape and can be dynamic. For dynamic dimensions use symbol `?` or '-1'. Ex. [?,3,?,?]. For bounded dimensions specify range 'min..max'. Ex. [1..10,3,?,?].
-data_shape Required for networks with dynamic shapes. Set shape for input blobs. In case of one input size: "[1,3,224,224]" or "input1[1,3,224,224],input2[1,4]". In case of several input sizes provide the same number for each input (except cases with single shape for any input): "[1,3,128,128][3,3,128,128][1,3,320,320]", "input1[1,1,128,128][1,1,256,256],input2[80,1]" or "input1[1,192][1,384],input2[1,192][1,384],input3[1,192][1,384],input4[1,192][1,384]". If network shapes are all static specifying the option will cause an exception.
-layout Optional. Prompts how network layouts should be treated by application. For example, "input1[NCHW],input2[NC]" or "[NCHW]" in case of one input size.
-cache_dir "< path > " Optional. Enables caching of loaded models to specified directory. List of devices which support caching is shown at the end of this message.
-load_from_file Optional. Loads model from file directly without ReadNetwork. All CNNNetwork options (like re-shape) will be ignored
-latency_percentile Optional. Defines the percentile to be reported in latency metric. The valid range is [1, 100]. The default value is 50 (median).
Device-specific performance options:
-nstreams "< integer > " Optional. Number of streams to use for inference on the CPU, GPU or MYRIAD devices (for HETERO and MULTI device cases use format < dev1 > :< nstreams1 > ,< dev2 > :< nstreams2 > or just < nstreams > ). Default value is determined automatically for a device.Please note that although the automatic selection usually provides a reasonable performance, it still may be non - optimal for some cases, especially for very small networks. See sample's README for more details. Also, using nstreams>1 is inherently throughput-oriented option, while for the best-latency estimations the number of streams should be set to 1.
-nthreads "< integer > " Optional. Number of threads to use for inference on the CPU (including HETERO and MULTI cases).
-pin ("YES"|"CORE")/"HYBRID_AWARE"/("NO"|"NONE")/"NUMA" Optional. Explicit inference threads binding options (leave empty to let the OpenVINO to make a choice):
enabling threads->cores pinning("YES", which is already default for any conventional CPU),
letting the runtime to decide on the threads->different core types("HYBRID_AWARE", which is default on the hybrid CPUs)
threads->(NUMA)nodes("NUMA") or
completely disable("NO") CPU inference threads pinning
2020-04-15 19:01:57 +03:00
2019-04-12 18:25:53 +03:00
Statistics dumping options:
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
-report_type "< type > " Optional. Enable collecting statistics report. "no_counters" report contains configuration options specified, resulting FPS and latency.
"average_counters" report extends "no_counters" report and additionally includes average PM counters values for each layer from the network.
"detailed_counters" report extends "average_counters" report and additionally includes per-layer PM counters
and latency for each executed infer request.
2021-04-09 18:08:32 +03:00
-report_folder Optional. Path to a folder where statistics report is stored.
-exec_graph_path Optional. Path to a file where to store executable graph information serialized.
-pc Optional. Report performance counters.
2022-01-18 11:22:47 +03:00
-dump_config Optional. Path to JSON file to dump IE parameters, which were set by application.
-load_config Optional. Path to JSON file to load custom IE parameters. Please note, command line parameters have higher priority than parameters from configuration file.
2022-03-21 19:31:17 +03:00
Statistics dumping options:
-report_type "< type > " Optional. Enable collecting statistics report. "no_counters" report contains configuration options specified, resulting FPS and latency. "average_counters" report extends "no_counters" report and additionally includes average PM counters values for each layer from the network. "detailed_counters" report extends "average_counters" report and additionally includes per-layer PM counters and latency for each executed infer request.
-report_folder Optional. Path to a folder where statistics report is stored.
-json_stats Optional. Enables JSON-based statistics output (by default reporting system will use CSV format). Should be used together with -report_folder option. -exec_graph_path Optional. Path to a file where to store executable graph information serialized.
-pc Optional. Report performance counters.
-pcseq Optional. Report latencies for each shape in -data_shape sequence.
-dump_config Optional. Path to JSON file to dump IE parameters, which were set by application.
-load_config Optional. Path to JSON file to load custom IE parameters. Please note, command line parameters have higher priority then parameters from configuration file.
-infer_precision "< element type > "Optional. Inference precission
-ip < value > Optional. Specifies precision for all input layers of the network.
-op < value > Optional. Specifies precision for all output layers of the network.
-iop "< value > " Optional. Specifies precision for input and output layers by name.
Example: -iop "input:FP16, output:FP16".
Notice that quotes are required.
Overwrites precision from ip and op options for specified layers.
-iscale Optional. Scale values to be used for the input image per channel.
Values to be provided in the [R, G, B] format. Can be defined for desired input of the model.
Example: -iscale data[255,255,255],info[255,255,255]
-imean Optional. Mean values to be used for the input image per channel.
Values to be provided in the [R, G, B] format. Can be defined for desired input of the model,
Example: -imean data[255,255,255],info[255,255,255]
-inference_only Optional. Measure only inference stage. Default option for static models. Dynamic models are measured in full mode which includes inputs setup stage, inference only mode available for them with single input data shape only. To enable full mode for static models pass "false" value to this argument: ex. "-inference_only=false".
2018-11-23 16:19:43 +03:00
```
Running the application with the empty list of options yields the usage message given above and an error message.
2020-07-20 17:36:08 +03:00
Application supports topologies with one or more inputs. If a topology is not data-sensitive, you can skip the input parameter. In this case, inputs are filled with random values.
If a model has only image input(s), please provide a folder with images or a path to an image as input.
If a model has some specific input(s) (not images), please prepare a binary file(s) that is filled with data of appropriate precision and provide a path to them as input.
2019-08-09 19:02:42 +03:00
If a model has mixed input types, input folder should contain all required files. Image inputs are filled with image files one by one. Binary inputs are filled with binary inputs one by one.
Feature/merge 2021 3 to master (#5307)
* Feature/azaytsev/cldnn doc fixes (#4600)
* Legal fixes, removed the Generating docs section
* Removed info regarding generating docs
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Feature/azaytsev/gna model link fixes (#4599)
* Added info on DockerHub CI Framework
* Feature/azaytsev/change layout (#3295)
* Changes according to feedback comments
* Replaced @ref's with html links
* Fixed links, added a title page for installing from repos and images, fixed formatting issues
* Added links
* minor fix
* Added DL Streamer to the list of components installed by default
* Link fixes
* Link fixes
* ovms doc fix (#2988)
* added OpenVINO Model Server
* ovms doc fixes
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Updated openvino_docs.xml
* Link Fixes
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Fix for broken CC in CPU plugin (#4595)
* Azure CI: Add "ref: releases/2021/3"
* Fixed clone rt info (#4597)
* [.ci/azure] Enable CC build (#4619)
* Formula fix (#4624)
* Fixed transformation to pull constants into Loop body (cherry-pick of PR 4591) (#4607)
* Cherry-pick of PR 4591
* Fixed typo
* Moved a check into the parameter_unchanged_after_iteration function
* Fixed KW hits (#4638)
* [CPU] Supported ANY layout for inputs in inferRequest (#4621)
* [.ci/azure] Add windows_conditional_compilation.yml (#4648) (#4655)
* Fix for MKLDNN constant layers execution (#4642)
* Fix for MKLDNN constant layers execution
* Single mkldnn::engine for all MKLDNN graphs
* Add workaround for control edges to support TF 2.4 RNN (#4634)
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Corrected PyYAML dependency (#4598) (#4620)
5.4.2 is absent on PyPI
* [CPU] Statically analyzed issues. (#4637)
* Docs api (#4657)
* Updated API changes document
* Comment for CVS-49440
* Add documentation on how to convert QuartzNet model (#4664)
* Add documentation on how to convert QuartzNet model (#4422)
* Add documentation on how to convert QuartzNet model
* Apply review feedback
* Small fix
* Apply review feedback
* Apply suggestions from code review
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Add reference to file
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Fixed bug in assign elimination transformation. (#4644)
* [doc] Updated PyPI support OSes (#4643) (#4662)
* [doc] Updated PyPI support OSes (#4643)
* Updated PyPI support OSes
* Added python versions for win and mac
* Update pypi-openvino-dev.md
* Update pypi-openvino-dev.md
* Update pypi-openvino-rt.md
* Update pypi-openvino-dev.md
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* [IE][VPU]: Fix empty output of CTCGreedyDecoderSeqLen (#4653)
* Allow the second output of CTCGreedyDecoderSeqLen to be nullptr in cases when it is not used but calculated in the Myriad plugin. In this case, parse the second output as FakeData
* It is a cherry-pick of #4652
* Update the firmware to release version
* [VPU] WA for Segmentation fault on dlclose() issue (#4645)
* Document TensorFlow 2* Update: Layers Support and Remove Beta Status (#4474) (#4711)
* Document TensorFlow 2* Update: Layers Support and Remove Beta Status
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Update documentation based on latest test results and feedback
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Remove ConvLSTM2D from supported layers list
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Document Dot layer without limitation
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Address feedback upon DenseFeatures and RNN operations
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Do a grammar correction
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Do a grammar correction based on feedback
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Updated nGraph custom op documentation (#4604)
* Updated nGraph custom op documentation
* Fixed comments
* [IE CLDNN] Fix missing variable initializations and types (#4669)
* Fix NormalizeL2 creation in QueryNetwork (cherry pick from master PR 4310) (#4651)
* Updated documentation about the supported YOLOv3 model from ONNX (#4722) (#4726)
* Restored folded Operations for QueryNetwork (#4685)
* Restored folded Operations for QueryNetwork
* Fixed comment
* Add unfolded constant operations to supported layers map
* Add STN to list of supported models (#4728)
* Fix python API for Loop/TensorIterator/Assign/ReadValue operations
* Catch std::except in fuzz tests (#4695)
Fuzz tests must catch all expected exceptions from IE. IE is using C++ std
library which may raise standard exceptions which IE pass through.
* Docs update (#4626)
* Updated latency case desc to cover multi-socket machines
* updated opt guide a bit
* avoiding '#' which is interpreted as ref
* Update CPU.md
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Blocked dims hwc 2021/3 (#4729)
* Fix for BlockedDims
* Added test for HWC layout
* [GNA] Update documentation regarding splits and concatenations support (#4740)
* Added mo.py to wheel packages (#4731)
* Inserted a disclaimer (#4760)
* Fixed some klockwork issues in C API samples (#4767)
* Feature/vpu doc fixes 2021 3 (#4635)
* Documentation fixes and updates for VPU
* minor correction
* minor correction
* Fixed links
* updated supported layers list for vpu
* [DOCS] added iname/oname (#4735)
* [VPU] Limit dlclose() WA to be used for Ubuntu only (#4806)
* Fixed wrong link (#4817)
* MKLDNN weights cache key calculation algorithm changed (#4790)
* Updated PIP install instructions (#4821)
* Document YOLACT support (#4749)
* Document YOLACT support
* Add preprocessing section
* Apply suggestions from code review
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Add documentation on how to convert F3Net model (#4863)
* Add instruction for F3Net model pytorch->onnx conversion
* Fix style
* Fixed dead lock in telemetry (#4873)
* Fixed dead lock in telemetry
* Refactored TelemetrySender.send function
* Refactored send function implementation to avoid deadlocks
* Unit tests for telemetry sender function
* Added legal header
* avladimi/cvs-31369: Documented packages content to YUM/APT IGs (#4839)
* Documented runtime/dev packages content
* Minor formatting fixes
* Implemented review comments
* Update installing-openvino-apt.md
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* [DOC] Low-Precision 8-bit Integer Inference (#4834)
* [DOC] Low-Precision 8-bit Integer Inference
* [DOC] Low-Precision 8-bit Integer Inference: comment fixes
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* [DOC] LPT comments fix
* [DOC] LPT comments fix: absolute links are updated to relative
* Update Int8Inference.md
* Update Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* Avladimi/cherry pick from master (#4892)
* Fixed CVS-48061
* Reviewed and edited the Customization instructions
* Fixed broken links in the TOC
* Fixed links
* Fixed formatting in the IG for Raspberry
* Feature/benchmarks 2021 3 (#4910)
* added new topics, changed the intro text
* updated
* Updates
* Updates
* Updates
* Updates
* Updates
* Added yolo-v4-tf and unet-camvid-onnx graphs
* Date for pricing is updated to March 15th
* Feature/omz link changes (#4911)
* Changed labels for demos and model downloader
* Changed links to models and tools
* Changed links to models and tools
* Changed links to demos
* [cherry-pick] Extensibility docs review (#4915)
* Feature/ovsa docs 2021 3 (#4914)
* Updated to 2021-3, fixed formatting issues
* Fixed formatting issues
* Fixed formatting issues
* Fixed formatting issues
* Update ovsa_get_started.md
* Clarification of Low Latency Transformation and State API documentation (#4877)
* Assign/ReadValue, LowLatency and StateAPI clarifications
* Apply suggestions from code review: spelling mistakes
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* fixed wording
* cherry-pick missing commit to release branch: low latency documentation
* Resolve review remarks
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Svetlana Dolinina <svetlana.a.dolinina@intel.com>
* DevCloud call outs (#4904)
* [README.md] change latest release to 2021.3
* [49342] Update recommended CMake version on install guide in documentation (#4763)
* Inserted a disclaimer
* Another disclaimer
* Update installing-openvino-windows.md
* Update installing-openvino-windows.md
* Update installing-openvino-windows.md
* Feature/doc fixes 2021 3 (#4971)
* Made changes for CVS-50424
* Changes for CVS-49349
* Minor change for CVS-49349
* Changes for CVS-49343
* Cherry-pick #PR4254
* Replaced /opt/intel/openvino/ with /opt/intel/openvino_2021/ as the default target directory
* (CVS-50786) Added a new section Reference IMplementations to keep Speech Library and Speech Recognition Demos
* Doc fixes
* Replaced links to inference_engine_intro.md with Deep_Learning_Inference_Engine_DevGuide.md, fixed links
* Fixed link
* Fixes
* Fixes
* Reemoved Intel® Xeon® processor E family
* fixes for graphs (#5057)
* compression.configs.hardware config to package_data (#5066)
* update OpenCV version to 4.5.2 (#5069)
* update OpenCV version to 4.5.2
* Enable mo.front.common.extractors module (#5038)
* Enable mo.front.common.extractors module (#5018)
* Enable mo.front.common.extractors module
* Update package_BOM.txt
* Test MO wheel content
* fix doc iframe issue - 2021.3 (#5090)
* wrap with htmlonly
* wrap with htmlonly
* Add specification for ExperimentalDetectron* oprations (#5128)
* Feature/benchmarks 2021 3 ehl (#5191)
* Added EHL config
* Updated graphs
* improve table formatting
* Wrap <iframe> tag with \htmlonly \endhtmlonly to avoid build errors
* Updated graphs
* Fixed links to TDP and Price for 8380
* Add PyTorch section to the documentation (#4972)
* Add PyTorch section to the documentation
* Apply review feedback
* Remove section about loop
* Apply review feedback
* Apply review feedback
* Apply review feedback
* doc: add Red Hat docker registry (#5184) (#5253)
* Incorporate changes in master
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
Co-authored-by: Vladislav Volkov <vladislav.volkov@intel.com>
Co-authored-by: azhogov <alexander.zhogov@intel.com>
Co-authored-by: Ilya Churaev <ilya.churaev@intel.com>
Co-authored-by: Alina Kladieva <alina.kladieva@intel.com>
Co-authored-by: Evgeny Lazarev <evgeny.lazarev@intel.com>
Co-authored-by: Gorokhov Dmitriy <dmitry.gorokhov@intel.com>
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>
Co-authored-by: Mikhail Ryzhov <mikhail.ryzhov@intel.com>
Co-authored-by: Nikolay Shchegolev <nikolay.shchegolev@intel.com>
Co-authored-by: Ilya Lavrenov <ilya.lavrenov@intel.com>
Co-authored-by: Maxim Vafin <maxim.vafin@intel.com>
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Anastasia Popova <anastasia.popova@intel.com>
Co-authored-by: Maksim Doronin <maksim.doronin@intel.com>
Co-authored-by: Andrew Bakalin <andrew.bakalin@intel.com>
Co-authored-by: Mikhail Letavin <mikhail.letavin@intel.com>
Co-authored-by: Anton Chetverikov <Anton.Chetverikov@intel.com>
Co-authored-by: Ivan Tikhonov <ivan.tikhonov@intel.com>
Co-authored-by: Andrey Somsikov <andrey.somsikov@intel.com>
Co-authored-by: Maxim Shevtsov <maxim.y.shevtsov@intel.com>
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
Co-authored-by: Andrey Dmitriev <andrey.dmitriev@intel.com>
Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Edward Shogulin <edward.shogulin@intel.com>
Co-authored-by: Svetlana Dolinina <svetlana.a.dolinina@intel.com>
Co-authored-by: Alexey Suhov <alexey.suhov@intel.com>
Co-authored-by: Sergey Lyubimtsev <sergey.lyubimtsev@intel.com>
Co-authored-by: Dmitry Kurtaev <dmitry.kurtaev+github@gmail.com>
Co-authored-by: Nikolay Tyukaev <nikolay.tyukaev@intel.com>
Co-authored-by: Kate Generalova <kate.generalova@intel.com>
2021-04-19 20:19:17 +03:00
To run the tool, you can use [public ](@ref omz_models_group_public ) or [Intel's ](@ref omz_models_group_intel ) pre-trained models from the Open Model Zoo. The models can be downloaded using the [Model Downloader ](@ref omz_tools_downloader ).
2019-10-04 19:26:43 +03:00
2022-03-24 22:27:29 +03:00
> **NOTE**: Before running the tool with a trained model, make sure the model is converted to the OpenVINO IR (\*.xml + \*.bin) using the [Model Optimizer tool](../../../docs/MO_DG/Deep_Learning_Model_Optimizer_DevGuide.md).
2021-02-11 12:57:05 +03:00
>
2020-08-26 18:53:24 +03:00
> The sample accepts models in ONNX format (.onnx) that do not require preprocessing.
2019-10-04 19:26:43 +03:00
## Examples of Running the Tool
2022-03-21 19:31:17 +03:00
This section provides step-by-step instructions on how to run the Benchmark Tool with the `googlenet-v1` public model on CPU or GPU devices. The [dog.bmp ](https://storage.openvinotoolkit.org/data/test_data/images/224x224/dog.bmp ) file is used as an input.
2019-10-04 19:26:43 +03:00
2021-12-21 20:26:37 +03:00
> **NOTE**: The Internet access is required to execute the following steps successfully. If you have access to the Internet through the proxy server only, please make sure that it is configured in your OS environment.
2019-10-04 19:26:43 +03:00
2022-03-21 19:31:17 +03:00
1. Install OpenVINO Development Tools to work with Caffe* models:
``` sh
pip install openvino-dev[caffe]
2019-10-04 19:26:43 +03:00
```
2022-03-21 19:31:17 +03:00
2. Download the model. Go to the Model Downloader directory and run the `omz_downloader` script with specifying the model name and directory to download the model to:
2019-10-04 19:26:43 +03:00
```sh
2022-03-21 19:31:17 +03:00
omz_downloader --name googlenet-v1 -o < models_dir >
2019-10-04 19:26:43 +03:00
```
2022-03-21 19:31:17 +03:00
3. Convert the model to the OpenVINO IR format. Run the Model Optimizer using the `mo` command with the path to the model, model format and output directory to generate the IR files:
2019-10-04 19:26:43 +03:00
```sh
2021-12-21 20:26:37 +03:00
mo --input_model < models_dir > /public/googlenet-v1/googlenet-v1.caffemodel --data_type FP32 --output_dir < ir_dir >
2021-02-11 12:57:05 +03:00
```
2022-03-21 19:31:17 +03:00
4. Run the tool with specifying the `dog.bmp` file as an input image, the IR of the `googlenet-v1` model and a device to perform inference on. The following commands demonstrate running the Benchmark Tool in the asynchronous mode on CPU and GPU devices:
2021-02-11 12:57:05 +03:00
2019-10-04 19:26:43 +03:00
* On CPU:
```sh
2022-03-21 19:31:17 +03:00
./benchmark_app -m < ir_dir > /googlenet-v1.xml -i dog.bmp -d CPU -api async -progress
2019-10-04 19:26:43 +03:00
```
2021-09-13 14:01:49 +03:00
* On GPU:
2019-10-04 19:26:43 +03:00
```sh
2022-03-21 19:31:17 +03:00
./benchmark_app -m < ir_dir > /googlenet-v1.xml -i dog.bmp -d GPU -api async -progress
2019-10-04 19:26:43 +03:00
```
2018-11-23 16:19:43 +03:00
2020-07-20 17:36:08 +03:00
The application outputs the number of executed iterations, total duration of execution, latency, and throughput.
2019-10-04 19:26:43 +03:00
Additionally, if you set the `-report_type` parameter, the application outputs statistics report. If you set the `-pc` parameter, the application outputs performance counters. If you set `-exec_graph_path` , the application reports executable graph information serialized. All measurements including per-layer PM counters are reported in milliseconds.
2018-11-23 16:19:43 +03:00
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
Below are fragments of sample output static and dynamic networks:
2018-11-23 16:19:43 +03:00
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
* For static network:
2019-10-04 19:26:43 +03:00
```
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
[Step 10/11] Measuring performance (Start inference asynchronously, 4 inference requests using 4 streams for CPU, limits: 60000 ms duration)
[ INFO ] BENCHMARK IS IN INFERENCE ONLY MODE.
[ INFO ] Input blobs will be filled once before performance measurements.
[ INFO ] First inference took 26.26 ms
Progress: [................... ] 99% done
2018-11-23 16:19:43 +03:00
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
[Step 11/11] Dumping statistics report
[ INFO ] Count: 6640 iterations
[ INFO ] Duration: 60039.70 ms
[ INFO ] Latency:
[ INFO ] Median: 35.36 ms
[ INFO ] Avg: 36.12 ms
[ INFO ] Min: 18.55 ms
[ INFO ] Max: 88.96 ms
[ INFO ] Throughput: 110.59 FPS
2019-10-04 19:26:43 +03:00
```
2018-11-23 16:19:43 +03:00
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
* For dynamic network:
2019-10-04 19:26:43 +03:00
```
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
[Step 10/11] Measuring performance (Start inference asynchronously, 4 inference requests using 4 streams for CPU, limits: 60000 ms duration)
[ INFO ] BENCHMARK IS IN FULL MODE.
[ INFO ] Inputs setup stage will be included in performance measurements.
[ INFO ] First inference took 26.80 ms
Progress: [................... ] 99% done
2019-04-12 18:25:53 +03:00
2019-10-04 19:26:43 +03:00
[Step 11/11] Dumping statistics report
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
[ INFO ] Count: 5199 iterations
[ INFO ] Duration: 60043.34 ms
[ INFO ] Latency:
[ INFO ] Median: 41.58 ms
[ INFO ] Avg: 46.07 ms
[ INFO ] Min: 8.44 ms
[ INFO ] Max: 115.65 ms
[ INFO ] Latency for each data shape group:
[ INFO ] 1. data : [1, 3, 224, 224]
[ INFO ] Median: 38.37 ms
[ INFO ] Avg: 30.29 ms
[ INFO ] Min: 8.44 ms
[ INFO ] Max: 61.30 ms
[ INFO ] 2. data : [1, 3, 448, 448]
[ INFO ] Median: 68.21 ms
[ INFO ] Avg: 61.85 ms
[ INFO ] Min: 29.58 ms
[ INFO ] Max: 115.65 ms
[ INFO ] Throughput: 86.59 FPS
2019-10-04 19:26:43 +03:00
```
2019-04-12 18:25:53 +03:00
2018-11-23 16:19:43 +03:00
## See Also
2022-03-24 22:27:29 +03:00
* [Using OpenVINO Runtime Samples ](../../../docs/OV_Runtime_UG/Samples_Overview.md )
2020-07-20 17:36:08 +03:00
* [Model Optimizer ](../../../docs/MO_DG/Deep_Learning_Model_Optimizer_DevGuide.md )
Feature/merge 2021 3 to master (#5307)
* Feature/azaytsev/cldnn doc fixes (#4600)
* Legal fixes, removed the Generating docs section
* Removed info regarding generating docs
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Feature/azaytsev/gna model link fixes (#4599)
* Added info on DockerHub CI Framework
* Feature/azaytsev/change layout (#3295)
* Changes according to feedback comments
* Replaced @ref's with html links
* Fixed links, added a title page for installing from repos and images, fixed formatting issues
* Added links
* minor fix
* Added DL Streamer to the list of components installed by default
* Link fixes
* Link fixes
* ovms doc fix (#2988)
* added OpenVINO Model Server
* ovms doc fixes
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Updated openvino_docs.xml
* Link Fixes
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Fix for broken CC in CPU plugin (#4595)
* Azure CI: Add "ref: releases/2021/3"
* Fixed clone rt info (#4597)
* [.ci/azure] Enable CC build (#4619)
* Formula fix (#4624)
* Fixed transformation to pull constants into Loop body (cherry-pick of PR 4591) (#4607)
* Cherry-pick of PR 4591
* Fixed typo
* Moved a check into the parameter_unchanged_after_iteration function
* Fixed KW hits (#4638)
* [CPU] Supported ANY layout for inputs in inferRequest (#4621)
* [.ci/azure] Add windows_conditional_compilation.yml (#4648) (#4655)
* Fix for MKLDNN constant layers execution (#4642)
* Fix for MKLDNN constant layers execution
* Single mkldnn::engine for all MKLDNN graphs
* Add workaround for control edges to support TF 2.4 RNN (#4634)
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Corrected PyYAML dependency (#4598) (#4620)
5.4.2 is absent on PyPI
* [CPU] Statically analyzed issues. (#4637)
* Docs api (#4657)
* Updated API changes document
* Comment for CVS-49440
* Add documentation on how to convert QuartzNet model (#4664)
* Add documentation on how to convert QuartzNet model (#4422)
* Add documentation on how to convert QuartzNet model
* Apply review feedback
* Small fix
* Apply review feedback
* Apply suggestions from code review
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Add reference to file
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Fixed bug in assign elimination transformation. (#4644)
* [doc] Updated PyPI support OSes (#4643) (#4662)
* [doc] Updated PyPI support OSes (#4643)
* Updated PyPI support OSes
* Added python versions for win and mac
* Update pypi-openvino-dev.md
* Update pypi-openvino-dev.md
* Update pypi-openvino-rt.md
* Update pypi-openvino-dev.md
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* [IE][VPU]: Fix empty output of CTCGreedyDecoderSeqLen (#4653)
* Allow the second output of CTCGreedyDecoderSeqLen to be nullptr in cases when it is not used but calculated in the Myriad plugin. In this case, parse the second output as FakeData
* It is a cherry-pick of #4652
* Update the firmware to release version
* [VPU] WA for Segmentation fault on dlclose() issue (#4645)
* Document TensorFlow 2* Update: Layers Support and Remove Beta Status (#4474) (#4711)
* Document TensorFlow 2* Update: Layers Support and Remove Beta Status
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Update documentation based on latest test results and feedback
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Remove ConvLSTM2D from supported layers list
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Document Dot layer without limitation
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Address feedback upon DenseFeatures and RNN operations
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Do a grammar correction
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Do a grammar correction based on feedback
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Updated nGraph custom op documentation (#4604)
* Updated nGraph custom op documentation
* Fixed comments
* [IE CLDNN] Fix missing variable initializations and types (#4669)
* Fix NormalizeL2 creation in QueryNetwork (cherry pick from master PR 4310) (#4651)
* Updated documentation about the supported YOLOv3 model from ONNX (#4722) (#4726)
* Restored folded Operations for QueryNetwork (#4685)
* Restored folded Operations for QueryNetwork
* Fixed comment
* Add unfolded constant operations to supported layers map
* Add STN to list of supported models (#4728)
* Fix python API for Loop/TensorIterator/Assign/ReadValue operations
* Catch std::except in fuzz tests (#4695)
Fuzz tests must catch all expected exceptions from IE. IE is using C++ std
library which may raise standard exceptions which IE pass through.
* Docs update (#4626)
* Updated latency case desc to cover multi-socket machines
* updated opt guide a bit
* avoiding '#' which is interpreted as ref
* Update CPU.md
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Blocked dims hwc 2021/3 (#4729)
* Fix for BlockedDims
* Added test for HWC layout
* [GNA] Update documentation regarding splits and concatenations support (#4740)
* Added mo.py to wheel packages (#4731)
* Inserted a disclaimer (#4760)
* Fixed some klockwork issues in C API samples (#4767)
* Feature/vpu doc fixes 2021 3 (#4635)
* Documentation fixes and updates for VPU
* minor correction
* minor correction
* Fixed links
* updated supported layers list for vpu
* [DOCS] added iname/oname (#4735)
* [VPU] Limit dlclose() WA to be used for Ubuntu only (#4806)
* Fixed wrong link (#4817)
* MKLDNN weights cache key calculation algorithm changed (#4790)
* Updated PIP install instructions (#4821)
* Document YOLACT support (#4749)
* Document YOLACT support
* Add preprocessing section
* Apply suggestions from code review
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Add documentation on how to convert F3Net model (#4863)
* Add instruction for F3Net model pytorch->onnx conversion
* Fix style
* Fixed dead lock in telemetry (#4873)
* Fixed dead lock in telemetry
* Refactored TelemetrySender.send function
* Refactored send function implementation to avoid deadlocks
* Unit tests for telemetry sender function
* Added legal header
* avladimi/cvs-31369: Documented packages content to YUM/APT IGs (#4839)
* Documented runtime/dev packages content
* Minor formatting fixes
* Implemented review comments
* Update installing-openvino-apt.md
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* [DOC] Low-Precision 8-bit Integer Inference (#4834)
* [DOC] Low-Precision 8-bit Integer Inference
* [DOC] Low-Precision 8-bit Integer Inference: comment fixes
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* [DOC] LPT comments fix
* [DOC] LPT comments fix: absolute links are updated to relative
* Update Int8Inference.md
* Update Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* Avladimi/cherry pick from master (#4892)
* Fixed CVS-48061
* Reviewed and edited the Customization instructions
* Fixed broken links in the TOC
* Fixed links
* Fixed formatting in the IG for Raspberry
* Feature/benchmarks 2021 3 (#4910)
* added new topics, changed the intro text
* updated
* Updates
* Updates
* Updates
* Updates
* Updates
* Added yolo-v4-tf and unet-camvid-onnx graphs
* Date for pricing is updated to March 15th
* Feature/omz link changes (#4911)
* Changed labels for demos and model downloader
* Changed links to models and tools
* Changed links to models and tools
* Changed links to demos
* [cherry-pick] Extensibility docs review (#4915)
* Feature/ovsa docs 2021 3 (#4914)
* Updated to 2021-3, fixed formatting issues
* Fixed formatting issues
* Fixed formatting issues
* Fixed formatting issues
* Update ovsa_get_started.md
* Clarification of Low Latency Transformation and State API documentation (#4877)
* Assign/ReadValue, LowLatency and StateAPI clarifications
* Apply suggestions from code review: spelling mistakes
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* fixed wording
* cherry-pick missing commit to release branch: low latency documentation
* Resolve review remarks
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Svetlana Dolinina <svetlana.a.dolinina@intel.com>
* DevCloud call outs (#4904)
* [README.md] change latest release to 2021.3
* [49342] Update recommended CMake version on install guide in documentation (#4763)
* Inserted a disclaimer
* Another disclaimer
* Update installing-openvino-windows.md
* Update installing-openvino-windows.md
* Update installing-openvino-windows.md
* Feature/doc fixes 2021 3 (#4971)
* Made changes for CVS-50424
* Changes for CVS-49349
* Minor change for CVS-49349
* Changes for CVS-49343
* Cherry-pick #PR4254
* Replaced /opt/intel/openvino/ with /opt/intel/openvino_2021/ as the default target directory
* (CVS-50786) Added a new section Reference IMplementations to keep Speech Library and Speech Recognition Demos
* Doc fixes
* Replaced links to inference_engine_intro.md with Deep_Learning_Inference_Engine_DevGuide.md, fixed links
* Fixed link
* Fixes
* Fixes
* Reemoved Intel® Xeon® processor E family
* fixes for graphs (#5057)
* compression.configs.hardware config to package_data (#5066)
* update OpenCV version to 4.5.2 (#5069)
* update OpenCV version to 4.5.2
* Enable mo.front.common.extractors module (#5038)
* Enable mo.front.common.extractors module (#5018)
* Enable mo.front.common.extractors module
* Update package_BOM.txt
* Test MO wheel content
* fix doc iframe issue - 2021.3 (#5090)
* wrap with htmlonly
* wrap with htmlonly
* Add specification for ExperimentalDetectron* oprations (#5128)
* Feature/benchmarks 2021 3 ehl (#5191)
* Added EHL config
* Updated graphs
* improve table formatting
* Wrap <iframe> tag with \htmlonly \endhtmlonly to avoid build errors
* Updated graphs
* Fixed links to TDP and Price for 8380
* Add PyTorch section to the documentation (#4972)
* Add PyTorch section to the documentation
* Apply review feedback
* Remove section about loop
* Apply review feedback
* Apply review feedback
* Apply review feedback
* doc: add Red Hat docker registry (#5184) (#5253)
* Incorporate changes in master
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
Co-authored-by: Vladislav Volkov <vladislav.volkov@intel.com>
Co-authored-by: azhogov <alexander.zhogov@intel.com>
Co-authored-by: Ilya Churaev <ilya.churaev@intel.com>
Co-authored-by: Alina Kladieva <alina.kladieva@intel.com>
Co-authored-by: Evgeny Lazarev <evgeny.lazarev@intel.com>
Co-authored-by: Gorokhov Dmitriy <dmitry.gorokhov@intel.com>
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>
Co-authored-by: Mikhail Ryzhov <mikhail.ryzhov@intel.com>
Co-authored-by: Nikolay Shchegolev <nikolay.shchegolev@intel.com>
Co-authored-by: Ilya Lavrenov <ilya.lavrenov@intel.com>
Co-authored-by: Maxim Vafin <maxim.vafin@intel.com>
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Anastasia Popova <anastasia.popova@intel.com>
Co-authored-by: Maksim Doronin <maksim.doronin@intel.com>
Co-authored-by: Andrew Bakalin <andrew.bakalin@intel.com>
Co-authored-by: Mikhail Letavin <mikhail.letavin@intel.com>
Co-authored-by: Anton Chetverikov <Anton.Chetverikov@intel.com>
Co-authored-by: Ivan Tikhonov <ivan.tikhonov@intel.com>
Co-authored-by: Andrey Somsikov <andrey.somsikov@intel.com>
Co-authored-by: Maxim Shevtsov <maxim.y.shevtsov@intel.com>
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
Co-authored-by: Andrey Dmitriev <andrey.dmitriev@intel.com>
Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Edward Shogulin <edward.shogulin@intel.com>
Co-authored-by: Svetlana Dolinina <svetlana.a.dolinina@intel.com>
Co-authored-by: Alexey Suhov <alexey.suhov@intel.com>
Co-authored-by: Sergey Lyubimtsev <sergey.lyubimtsev@intel.com>
Co-authored-by: Dmitry Kurtaev <dmitry.kurtaev+github@gmail.com>
Co-authored-by: Nikolay Tyukaev <nikolay.tyukaev@intel.com>
Co-authored-by: Kate Generalova <kate.generalova@intel.com>
2021-04-19 20:19:17 +03:00
* [Model Downloader ](@ref omz_tools_downloader )