2020-07-20 17:36:08 +03:00
# Benchmark C++ Tool {#openvino_inference_engine_samples_benchmark_app_README}
2018-11-23 16:19:43 +03:00
2021-09-13 15:40:36 +03:00
This topic demonstrates how to use the Benchmark C++ Tool to estimate deep learning inference performance on supported devices.
Performance can be measured for two inference modes: latency- and throughput-oriented.
2019-01-21 21:31:31 +03:00
2021-12-21 20:26:37 +03:00
> **NOTE**: This topic describes usage of C++ implementation of the Benchmark Tool. For the Python* implementation, refer to [Benchmark Python* Tool](../../../tools/benchmark_tool/README.md).
2019-01-21 21:31:31 +03:00
2021-09-07 19:21:41 +03:00
> **TIP**: You can quick start with the Benchmark Tool inside the OpenVINO™ [Deep Learning Workbench](@ref openvino_docs_get_started_get_started_dl_workbench) (DL Workbench).
> [DL Workbench](@ref workbench_docs_Workbench_DG_Introduction) is the OpenVINO™ toolkit UI you to
2021-10-19 16:51:38 +03:00
> import a model, analyze its performance and accuracy, visualize the outputs, optimize and prepare the model for
2021-09-07 19:21:41 +03:00
> deployment on various Intel® platforms.
2018-11-23 16:19:43 +03:00
## How It Works
2021-09-13 15:40:36 +03:00
Upon start-up, the application reads command-line parameters and loads a network and inputs (images/binary files) to the specified device.
2018-11-23 16:19:43 +03:00
2021-09-13 15:40:36 +03:00
**NOTE** : By default, Inference Engine samples, tools and demos expect input with BGR channels order.
If you trained your model to work with RGB order, you need to manually rearrange the default channels order in the sample or demo application
or reconvert your model using the Model Optimizer tool with `--reverse_input_channels` argument specified.
For more information about the argument, refer to **When to Reverse Input Channels** section of
[Converting a Model Using General Conversion Parameters ](../../../docs/MO_DG/prepare_model/convert_model/Converting_Model_General.md ).
2018-11-23 16:19:43 +03:00
2021-09-13 15:40:36 +03:00
Device-specific execution parameters (number of streams, threads, and so on) can be either explicitly specified through the command line
or left default. In the last case, the sample logic will select the values for the optimal throughput.
While experimenting with individual parameters allows to find the performance sweet spot, usually, the parameters are not very performance-portable,
so the values from one machine or device are not necessarily optimal for another.
From this perspective, the most portable way is experimenting only with the performance hints. To learn more, refer to the section on the command-line parameters below.
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
A number of execution steps is defined by one of the following parameters:
2019-04-12 18:25:53 +03:00
* Number of iterations specified with the `-niter` command-line argument
2019-08-09 19:02:42 +03:00
* Time duration specified with the `-t` command-line argument
* Both of them (execution will continue until both conditions are met)
* Predefined duration if `-niter` and `-t` are not specified. Predefined duration value depends on a device.
2018-11-23 16:19:43 +03:00
2021-09-13 15:40:36 +03:00
During the execution, the application calculates latency (if applicable) and overall throughput:
* By default, the median latency value is reported
* Throughput is calculated as overall_inference_time/number_of_processed_requests. Note that the throughput value also depends on batch size.
2019-04-12 18:25:53 +03:00
The application also collects per-layer Performance Measurement (PM) counters for each executed infer request if you
enable statistics dumping by setting the `-report_type` parameter to one of the possible values:
* `no_counters` report includes configuration options specified, resulting FPS and latency.
2019-08-09 19:02:42 +03:00
* `average_counters` report extends the `no_counters` report and additionally includes average PM counters values for each layer from the network.
* `detailed_counters` report extends the `average_counters` report and additionally includes per-layer PM counters and latency for each executed infer request.
2019-04-12 18:25:53 +03:00
2019-08-09 19:02:42 +03:00
Depending on the type, the report is stored to `benchmark_no_counters_report.csv` , `benchmark_average_counters_report.csv` ,
2019-04-12 18:25:53 +03:00
or `benchmark_detailed_counters_report.csv` file located in the path specified in `-report_folder` .
2020-07-20 17:36:08 +03:00
The application also saves executable graph information serialized to an XML file if you specify a path to it with the
2019-04-12 18:25:53 +03:00
`-exec_graph_path` parameter.
2018-11-23 16:19:43 +03:00
2019-10-04 19:26:43 +03:00
## Run the Tool
2021-02-02 11:29:12 +03:00
2020-07-20 17:36:08 +03:00
Note that the benchmark_app usually produces optimal performance for any device out of the box.
2019-08-09 19:02:42 +03:00
2019-10-04 19:26:43 +03:00
**So in most cases you don't need to play the app options explicitly and the plain device name is enough**, for example, for CPU:
```sh
./benchmark_app -m < model > -i < input > -d CPU
2019-08-09 19:02:42 +03:00
```
2019-10-04 19:26:43 +03:00
2021-09-13 15:40:36 +03:00
But it is still may be sub-optimal for some cases, especially for very small networks. More details can read in [Introduction to Performance Topics ](../../../docs/IE_DG/Intro_to_Performance.md ).
2019-10-04 19:26:43 +03:00
2020-07-20 17:36:08 +03:00
As explained in the [Introduction to Performance Topics ](../../../docs/IE_DG/Intro_to_Performance.md ) section, for all devices, including new [MULTI device ](../../../docs/IE_DG/supported_plugins/MULTI.md ) it is preferable to use the FP16 IR for the model.
2019-10-04 19:26:43 +03:00
Also if latency of the CPU inference on the multi-socket machines is of concern, please refer to the same
2020-07-20 17:36:08 +03:00
[Introduction to Performance Topics ](../../../docs/IE_DG/Intro_to_Performance.md ) document.
2018-11-23 16:19:43 +03:00
Running the application with the `-h` option yields the following usage message:
2019-08-09 19:02:42 +03:00
```
2018-11-23 16:19:43 +03:00
./benchmark_app -h
InferenceEngine:
API version ............ < version >
Build .................. < number >
[ INFO ] Parsing input parameters
benchmark_app [OPTION]
Options:
2021-04-09 18:08:32 +03:00
-h, --help Print a usage message
-m "< path > " Required. Path to an .xml/.onnx/.prototxt file with a trained model or to a .blob files with a trained compiled model.
-i "< path > " Optional. Path to a folder with images and/or binaries or to specific image or binary file.
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
In case of dynamic shapes networks with several inputs provide the same number of files for each input (except cases with single file for any input):
"input1:1.jpg input2:1.bin", "input1:1.bin,2.bin input2:3.bin input3:4.bin,5.bin ".
Also you can pass specific keys for inputs: "random" - for fillling input with random data, "image_info" - for filling input with image size.
2021-04-09 18:08:32 +03:00
-d "< device > " Optional. Specify a target device to infer on (the list of available devices is shown below). Default value is CPU.
Use "-d HETERO:< comma-separated_devices_list > " format to specify HETERO plugin.
Use "-d MULTI:< comma-separated_devices_list > " format to specify MULTI plugin.
2021-10-19 16:51:38 +03:00
Use "-d GPU.X" format to specify device id for GPU devices.
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
The application looks for a suitable plugin for the specified device.
2021-04-09 18:08:32 +03:00
-l "< absolute_path > " Required for CPU custom layers. Absolute path to a shared library with the kernels implementations.
2018-11-23 16:19:43 +03:00
Or
2021-04-09 18:08:32 +03:00
-c "< absolute_path > " Required for GPU custom kernels. Absolute path to an .xml file with the kernels description.
2021-09-13 15:40:36 +03:00
-hint "< throughput ( or just ' tput ' ) / latency " >
Optional. Performance hint (optimize for latency or throughput).
The hint allows the OpenVINO device to select the right network-specific settings,
as opposite to just accepting specific values from the sample command line.
So you can specify only the hint without setting explicit 'nstreams' or other device-specific options.
-api "< sync / async > " Optional (deprecated). Enable Sync/Async API. Default value is "async".
2021-04-09 18:08:32 +03:00
-niter "< integer > " Optional. Number of iterations. If not specified, the number of iterations is calculated depending on a device.
-nireq "< integer > " Optional. Number of infer requests. Default value is determined automatically for a device.
-b "< integer > " Optional. Batch size value. If not specified, the batch size value is determined from Intermediate Representation.
-stream_output Optional. Print progress as a plain text. When specified, an interactive progress bar is replaced with a multiline output.
-t Optional. Time, in seconds, to execute topology.
-progress Optional. Show progress bar (can affect performance measurement). Default values is "false".
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
-shape Optional. Set shape for network input. For example, "input1[1,3,224,224],input2[1,4]" or "[1,3,224,224]" in case of one input size.
This parameter affect model input shape and can be dynamic. For dynamic dimensions use symbol `?` or '-1'. Ex. [?,3,?,?].
For bounded dimensions specify range 'min..max'. Ex. [1..10,3,?,?].
-data_shape Required for networks with dynamic shapes. Set shape for input blobs.
In case of one input size: "[1,3,224,224]" or "input1[1,3,224,224],input2[1,4]".
In case of several input sizes provide the same number for
each input (except cases with single shape for any input): "[1,3,128,128][3,3,128,128][1,3,320,320]",
"input1[1,1,128,128][1,1,256,256],input2[80,1]" or "input1[1,192][1,384],input2[1,192][1,384],input3[1,192][1,384],input4[1,192][1,384]".
If network shapes are all static specifying the option will cause an exception.
2021-04-09 18:08:32 +03:00
-layout Optional. Prompts how network layouts should be treated by application. For example, "input1[NCHW],input2[NC]" or "[NCHW]" in case of one input size.
2021-05-17 13:41:15 +03:00
-cache_dir "< path > " Optional. Enables caching of loaded models to specified directory.
-load_from_file Optional. Loads model from file directly without ReadNetwork.
2021-07-23 10:29:55 +03:00
-latency_percentile Optional. Defines the percentile to be reported in latency metric. The valid range is [1, 100]. The default value is 50 (median).
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
-inference_only Optional. Measure only inference stage. Default option for static models.
Dynamic models are measured in full mode which includes inputs setup stage,
inference only mode available for them with single input data shape only.
To enable full mode for static models pass \"false\" value to this argument: ex. -inference_only=false".
2019-04-12 18:25:53 +03:00
CPU-specific performance options:
2021-04-09 18:08:32 +03:00
-nstreams "< integer > " Optional. Number of streams to use for inference on the CPU, GPU or MYRIAD devices
(for HETERO and MULTI device cases use format < device1 > :< nstreams1 > ,< device2 > :< nstreams2 > or just < nstreams > ).
Default value is determined automatically for a device.
Please note that although the automatic selection usually provides a reasonable performance,
it still may be non-optimal for some cases, especially for very small networks.
Also, using nstreams>1 is inherently throughput-oriented option, while for the best-latency
estimations the number of streams should be set to 1.
-nthreads "< integer > " Optional. Number of threads to use for inference on the CPU (including HETERO and MULTI cases).
-enforcebf16="< true / false > " Optional. By default floating point operations execution in bfloat16 precision are enforced if supported by platform.
Openvino hybrid awareness (#5261)
* change the deprecated method to the recent
* first ver of the hybrid cores aware CPU streams (+debug info)
* more debug and fixed sum threads
* disabled NUMA pinning to experiment with affinity via OS
* further brushing of stream to core type logic
* hybrid CPU-aware getNumberOfCPUCores
* adding check on the efficiency
* experimental TBB package (that cmake should pull from the internal server)
* iterating over core types in the reversed order (so the big cores are populated first in case user specified less than all #threads)
* adding back the NUMA affinity code-path for the full validation (incl 2 sockets Windows Server)
* cpplint fix and tabbing the #if clauses for the readbility
* pre-production TBB from internal server
* wrapping over #cores/types
* wrapping over #cores/types, ver 2
* wrapping over #streams instead
* disabling warnings as errors for a while (to unlock testing)
* accomodating new TBB layout for dependencies.bat
* next tbb ver (with debug binaries that probably can unlock the commodity builds, without playing product_configs)
* minor brushing for experiments (so that pinning can be disabled)
* minor brushing from code review
* Updating the SHA hash which appeared when rebasing to the master
* WIP refactoring
* Completed refactoring of the "config" phase of the cpu stream executor and on-the-fly streams to core types mapping
* making the benchmark_app aware about new pinning mode
* Brushing a bit (in preparation for the "soft" affinity)
* map to vector to simplify the things
* updated executors comparison
* more fine-grained pinning scheme for the HYBRID (required to allow all cores on 2+8 1+4, and other LITTLE-skewed scenarios)
TODO: seprate little to big ratio for the fp322 and int8 (and pass the fp32Only flag to the MakeDefaultMultiTHreaded)
* separating fp32 and int8 intensive cases for hybrid execution, also leveraging the HT if the #big_cores is small, refactored. also switched to the 2021.2 oneTBB RC package
* code style
* stripped tbb archives from unused folders and files, also has to rename the LICENSE.txt to the LICENSE to match existing OV packaging tools
* assigning nodeId regradless of pinning mode
* tests OpenCV builds with same 2021.2 oneTBB, ubuntu 18/20
* cmake install paths for oneTBB, alos a ie_parallel.cmake warning on older ver of TBB
* Updated latency case desc to cover multi-socket machines
* adding centos8 OCV with oneTBB build
updating TBB drops with hwloc shared libs added.
* enabled internal OCV from THIRD_PARTY_SERVER to test thru CI..
Added Centos7 notbb OCV build (until g-api get ready for onetbb) to unlock the Centos7 CI build
* separate rpath log to respect one-tbb specific paths
* fixed SEQ code-path
* fixed doc misprint
* allowing all cores in 2+8 for int8 as well
* cleaned from debug printfs
* HYBRID_AWARE pinning option for the Python benchmark_app
* OpenVINO Hybrid CPUs support
* Remove custom::task_arena abstraction layout
* Get back to the custom::task_arena interface
* Add windows.h inclusion
* Fix typo in macro name
* Separate TBB and TBBbind packages
* Fix compile-time conditions
* Fix preprocessors conditions
* Fix typo
* Fix linking
* make linking private
* Fix typo
* Fix target_compile_definitions syntax
* Implement CMake install logic, update sha hash for the tbbbind_2_4 package
* Add tbbbind_2_4 required paths to setup_vars
* Update CI paths
* Include ie_parallel.hpp to ie_system_conf.cpp
* Try to update dependencies scripts
* Try to fix dependencies.bat
* Modify dependencies script
* Use static tbbbind_2_4 library
* Remove redundant paths from CI
* Revert "cleaned from debug printfs"
This reverts commit 82c9bd90c543a89c5cfeee7bf5b3ce0998ee49df.
# Conflicts:
# inference-engine/src/inference_engine/os/win/win_system_conf.cpp
# inference-engine/src/inference_engine/threading/ie_cpu_streams_executor.cpp
# inference-engine/src/mkldnn_plugin/config.cpp
* Update tbbbind package version
* fixed compilation
* removing the direct tbb::info calls from CPU plugin, to aggregate everything in the single module (that exposes the higher level APIs)
* Update tbbbind package version
(cherry picked from commit f66b8f6aa63e160eb1788cf3da18a6ac73cf4007)
* compilation fix
* brushing the headers a bit
* Make custom::task_arena inherited from tbb::task_arena
* change to the latest TBB API, and more debug printfs
* code-style
* ARM compilation
* aligned "failed system config" between OV and TBB (by using '-1')
* macos compilation fix
* default arena creation (to make sure all code-path have that fallback)
* Incapsulate all TBB versions related logic inside the custom namespace
* Move custom layer header to internal scope + minor improvements
* with all NUMA/Hybrid checks now consolidated in the custom_arena, cleaning the ugly ifdefs thta we had
* Introduce new ThreadBindingType + fix compilation
* fixing OMP compilation
* OpenVINO Hybrid CPUs support
* Remove custom::task_arena abstraction layout
* Get back to the custom::task_arena interface
* Add windows.h inclusion
* Fix typo in macro name
* Separate TBB and TBBbind packages
* Fix compile-time conditions
* Fix preprocessors conditions
* Fix typo
* Fix linking
* make linking private
* Fix typo
* Fix target_compile_definitions syntax
* Implement CMake install logic, update sha hash for the tbbbind_2_4 package
* Add tbbbind_2_4 required paths to setup_vars
* Update CI paths
* Include ie_parallel.hpp to ie_system_conf.cpp
* Try to update dependencies scripts
* Try to fix dependencies.bat
* Modify dependencies script
* Use static tbbbind_2_4 library
* Remove redundant paths from CI
* Update tbbbind package version
* Make custom::task_arena inherited from tbb::task_arena
* Incapsulate all TBB versions related logic inside the custom namespace
* Move custom layer header to internal scope + minor improvements
* Introduce new ThreadBindingType + fix compilation
* Fix compilation
* Use public tbbbind_2_4 package
* fixed macos build, corrected comments/desc
* reverted to the default binding selection logic ( to preserve the legacy beh)
* OpenVINO Hybrid CPUs support
* Remove custom::task_arena abstraction layout
* Get back to the custom::task_arena interface
* Add windows.h inclusion
* Fix typo in macro name
* Separate TBB and TBBbind packages
* Fix compile-time conditions
* Fix preprocessors conditions
* Fix typo
* Fix linking
* make linking private
* Fix typo
* Fix target_compile_definitions syntax
* Implement CMake install logic, update sha hash for the tbbbind_2_4 package
* Add tbbbind_2_4 required paths to setup_vars
* Update CI paths
* Include ie_parallel.hpp to ie_system_conf.cpp
* Try to update dependencies scripts
* Try to fix dependencies.bat
* Modify dependencies script
* Use static tbbbind_2_4 library
* Remove redundant paths from CI
* Update tbbbind package version
* Make custom::task_arena inherited from tbb::task_arena
* Incapsulate all TBB versions related logic inside the custom namespace
* Move custom layer header to internal scope + minor improvements
* Introduce new ThreadBindingType + fix compilation
* Fix compilation
* Use public tbbbind_2_4 package
* Apply review comments
* Fix compilation without tbbbind_2_4
* Fix compilation with different TBB versions
* code review remarks
* fix for the NONE pinning code-path under HYBRID_AWAR
* whitespace and cleaning the debug printfs (per review)
* code-review comments
* fixed code-style
Co-authored-by: Kochin, Ivan <ivan.kochin@intel.com>
Co-authored-by: Kochin Ivan <kochin.ivan@intel.com>
2021-04-28 17:42:58 +03:00
-pin "YES"/"HYBRID_AWARE"/"NUMA"/"NO"
Optional. Explicit inference threads binding options (leave empty to let the OpenVINO to make a choice):
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
enabling threads->cores pinning ("YES", which is already default for a conventional CPU),
letting the runtime to decide on the threads->different core types ("HYBRID_AWARE", which is default on the hybrid CPUs)
threads->(NUMA)nodes ("NUMA") or completely disable ("NO") CPU inference threads pinning.
2021-04-09 18:08:32 +03:00
-ip "U8"/"FP16"/"FP32" Optional. Specifies precision for all input layers of the network.
-op "U8"/"FP16"/"FP32" Optional. Specifies precision for all output layers of the network.
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
-iop Optional. Specifies precision for input and output layers by name. Example: -iop "input:FP16, output:FP16". Notice that quotes are required.
Overwrites precision from ip and op options for specified layers.
2020-04-15 19:01:57 +03:00
2019-04-12 18:25:53 +03:00
Statistics dumping options:
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
-report_type "< type > " Optional. Enable collecting statistics report. "no_counters" report contains configuration options specified, resulting FPS and latency.
"average_counters" report extends "no_counters" report and additionally includes average PM counters values for each layer from the network.
"detailed_counters" report extends "average_counters" report and additionally includes per-layer PM counters
and latency for each executed infer request.
2021-04-09 18:08:32 +03:00
-report_folder Optional. Path to a folder where statistics report is stored.
-exec_graph_path Optional. Path to a file where to store executable graph information serialized.
-pc Optional. Report performance counters.
-dump_config Optional. Path to XML/YAML/JSON file to dump IE parameters, which were set by application.
-load_config Optional. Path to XML/YAML/JSON file to load custom IE parameters. Please note, command line parameters have higher priority then parameters from configuration file.
2018-11-23 16:19:43 +03:00
```
Running the application with the empty list of options yields the usage message given above and an error message.
2020-07-20 17:36:08 +03:00
Application supports topologies with one or more inputs. If a topology is not data-sensitive, you can skip the input parameter. In this case, inputs are filled with random values.
If a model has only image input(s), please provide a folder with images or a path to an image as input.
If a model has some specific input(s) (not images), please prepare a binary file(s) that is filled with data of appropriate precision and provide a path to them as input.
2019-08-09 19:02:42 +03:00
If a model has mixed input types, input folder should contain all required files. Image inputs are filled with image files one by one. Binary inputs are filled with binary inputs one by one.
Feature/merge 2021 3 to master (#5307)
* Feature/azaytsev/cldnn doc fixes (#4600)
* Legal fixes, removed the Generating docs section
* Removed info regarding generating docs
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Feature/azaytsev/gna model link fixes (#4599)
* Added info on DockerHub CI Framework
* Feature/azaytsev/change layout (#3295)
* Changes according to feedback comments
* Replaced @ref's with html links
* Fixed links, added a title page for installing from repos and images, fixed formatting issues
* Added links
* minor fix
* Added DL Streamer to the list of components installed by default
* Link fixes
* Link fixes
* ovms doc fix (#2988)
* added OpenVINO Model Server
* ovms doc fixes
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Updated openvino_docs.xml
* Link Fixes
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Fix for broken CC in CPU plugin (#4595)
* Azure CI: Add "ref: releases/2021/3"
* Fixed clone rt info (#4597)
* [.ci/azure] Enable CC build (#4619)
* Formula fix (#4624)
* Fixed transformation to pull constants into Loop body (cherry-pick of PR 4591) (#4607)
* Cherry-pick of PR 4591
* Fixed typo
* Moved a check into the parameter_unchanged_after_iteration function
* Fixed KW hits (#4638)
* [CPU] Supported ANY layout for inputs in inferRequest (#4621)
* [.ci/azure] Add windows_conditional_compilation.yml (#4648) (#4655)
* Fix for MKLDNN constant layers execution (#4642)
* Fix for MKLDNN constant layers execution
* Single mkldnn::engine for all MKLDNN graphs
* Add workaround for control edges to support TF 2.4 RNN (#4634)
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Corrected PyYAML dependency (#4598) (#4620)
5.4.2 is absent on PyPI
* [CPU] Statically analyzed issues. (#4637)
* Docs api (#4657)
* Updated API changes document
* Comment for CVS-49440
* Add documentation on how to convert QuartzNet model (#4664)
* Add documentation on how to convert QuartzNet model (#4422)
* Add documentation on how to convert QuartzNet model
* Apply review feedback
* Small fix
* Apply review feedback
* Apply suggestions from code review
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Add reference to file
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Fixed bug in assign elimination transformation. (#4644)
* [doc] Updated PyPI support OSes (#4643) (#4662)
* [doc] Updated PyPI support OSes (#4643)
* Updated PyPI support OSes
* Added python versions for win and mac
* Update pypi-openvino-dev.md
* Update pypi-openvino-dev.md
* Update pypi-openvino-rt.md
* Update pypi-openvino-dev.md
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* [IE][VPU]: Fix empty output of CTCGreedyDecoderSeqLen (#4653)
* Allow the second output of CTCGreedyDecoderSeqLen to be nullptr in cases when it is not used but calculated in the Myriad plugin. In this case, parse the second output as FakeData
* It is a cherry-pick of #4652
* Update the firmware to release version
* [VPU] WA for Segmentation fault on dlclose() issue (#4645)
* Document TensorFlow 2* Update: Layers Support and Remove Beta Status (#4474) (#4711)
* Document TensorFlow 2* Update: Layers Support and Remove Beta Status
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Update documentation based on latest test results and feedback
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Remove ConvLSTM2D from supported layers list
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Document Dot layer without limitation
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Address feedback upon DenseFeatures and RNN operations
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Do a grammar correction
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Do a grammar correction based on feedback
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Updated nGraph custom op documentation (#4604)
* Updated nGraph custom op documentation
* Fixed comments
* [IE CLDNN] Fix missing variable initializations and types (#4669)
* Fix NormalizeL2 creation in QueryNetwork (cherry pick from master PR 4310) (#4651)
* Updated documentation about the supported YOLOv3 model from ONNX (#4722) (#4726)
* Restored folded Operations for QueryNetwork (#4685)
* Restored folded Operations for QueryNetwork
* Fixed comment
* Add unfolded constant operations to supported layers map
* Add STN to list of supported models (#4728)
* Fix python API for Loop/TensorIterator/Assign/ReadValue operations
* Catch std::except in fuzz tests (#4695)
Fuzz tests must catch all expected exceptions from IE. IE is using C++ std
library which may raise standard exceptions which IE pass through.
* Docs update (#4626)
* Updated latency case desc to cover multi-socket machines
* updated opt guide a bit
* avoiding '#' which is interpreted as ref
* Update CPU.md
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Blocked dims hwc 2021/3 (#4729)
* Fix for BlockedDims
* Added test for HWC layout
* [GNA] Update documentation regarding splits and concatenations support (#4740)
* Added mo.py to wheel packages (#4731)
* Inserted a disclaimer (#4760)
* Fixed some klockwork issues in C API samples (#4767)
* Feature/vpu doc fixes 2021 3 (#4635)
* Documentation fixes and updates for VPU
* minor correction
* minor correction
* Fixed links
* updated supported layers list for vpu
* [DOCS] added iname/oname (#4735)
* [VPU] Limit dlclose() WA to be used for Ubuntu only (#4806)
* Fixed wrong link (#4817)
* MKLDNN weights cache key calculation algorithm changed (#4790)
* Updated PIP install instructions (#4821)
* Document YOLACT support (#4749)
* Document YOLACT support
* Add preprocessing section
* Apply suggestions from code review
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Add documentation on how to convert F3Net model (#4863)
* Add instruction for F3Net model pytorch->onnx conversion
* Fix style
* Fixed dead lock in telemetry (#4873)
* Fixed dead lock in telemetry
* Refactored TelemetrySender.send function
* Refactored send function implementation to avoid deadlocks
* Unit tests for telemetry sender function
* Added legal header
* avladimi/cvs-31369: Documented packages content to YUM/APT IGs (#4839)
* Documented runtime/dev packages content
* Minor formatting fixes
* Implemented review comments
* Update installing-openvino-apt.md
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* [DOC] Low-Precision 8-bit Integer Inference (#4834)
* [DOC] Low-Precision 8-bit Integer Inference
* [DOC] Low-Precision 8-bit Integer Inference: comment fixes
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* [DOC] LPT comments fix
* [DOC] LPT comments fix: absolute links are updated to relative
* Update Int8Inference.md
* Update Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* Avladimi/cherry pick from master (#4892)
* Fixed CVS-48061
* Reviewed and edited the Customization instructions
* Fixed broken links in the TOC
* Fixed links
* Fixed formatting in the IG for Raspberry
* Feature/benchmarks 2021 3 (#4910)
* added new topics, changed the intro text
* updated
* Updates
* Updates
* Updates
* Updates
* Updates
* Added yolo-v4-tf and unet-camvid-onnx graphs
* Date for pricing is updated to March 15th
* Feature/omz link changes (#4911)
* Changed labels for demos and model downloader
* Changed links to models and tools
* Changed links to models and tools
* Changed links to demos
* [cherry-pick] Extensibility docs review (#4915)
* Feature/ovsa docs 2021 3 (#4914)
* Updated to 2021-3, fixed formatting issues
* Fixed formatting issues
* Fixed formatting issues
* Fixed formatting issues
* Update ovsa_get_started.md
* Clarification of Low Latency Transformation and State API documentation (#4877)
* Assign/ReadValue, LowLatency and StateAPI clarifications
* Apply suggestions from code review: spelling mistakes
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* fixed wording
* cherry-pick missing commit to release branch: low latency documentation
* Resolve review remarks
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Svetlana Dolinina <svetlana.a.dolinina@intel.com>
* DevCloud call outs (#4904)
* [README.md] change latest release to 2021.3
* [49342] Update recommended CMake version on install guide in documentation (#4763)
* Inserted a disclaimer
* Another disclaimer
* Update installing-openvino-windows.md
* Update installing-openvino-windows.md
* Update installing-openvino-windows.md
* Feature/doc fixes 2021 3 (#4971)
* Made changes for CVS-50424
* Changes for CVS-49349
* Minor change for CVS-49349
* Changes for CVS-49343
* Cherry-pick #PR4254
* Replaced /opt/intel/openvino/ with /opt/intel/openvino_2021/ as the default target directory
* (CVS-50786) Added a new section Reference IMplementations to keep Speech Library and Speech Recognition Demos
* Doc fixes
* Replaced links to inference_engine_intro.md with Deep_Learning_Inference_Engine_DevGuide.md, fixed links
* Fixed link
* Fixes
* Fixes
* Reemoved Intel® Xeon® processor E family
* fixes for graphs (#5057)
* compression.configs.hardware config to package_data (#5066)
* update OpenCV version to 4.5.2 (#5069)
* update OpenCV version to 4.5.2
* Enable mo.front.common.extractors module (#5038)
* Enable mo.front.common.extractors module (#5018)
* Enable mo.front.common.extractors module
* Update package_BOM.txt
* Test MO wheel content
* fix doc iframe issue - 2021.3 (#5090)
* wrap with htmlonly
* wrap with htmlonly
* Add specification for ExperimentalDetectron* oprations (#5128)
* Feature/benchmarks 2021 3 ehl (#5191)
* Added EHL config
* Updated graphs
* improve table formatting
* Wrap <iframe> tag with \htmlonly \endhtmlonly to avoid build errors
* Updated graphs
* Fixed links to TDP and Price for 8380
* Add PyTorch section to the documentation (#4972)
* Add PyTorch section to the documentation
* Apply review feedback
* Remove section about loop
* Apply review feedback
* Apply review feedback
* Apply review feedback
* doc: add Red Hat docker registry (#5184) (#5253)
* Incorporate changes in master
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
Co-authored-by: Vladislav Volkov <vladislav.volkov@intel.com>
Co-authored-by: azhogov <alexander.zhogov@intel.com>
Co-authored-by: Ilya Churaev <ilya.churaev@intel.com>
Co-authored-by: Alina Kladieva <alina.kladieva@intel.com>
Co-authored-by: Evgeny Lazarev <evgeny.lazarev@intel.com>
Co-authored-by: Gorokhov Dmitriy <dmitry.gorokhov@intel.com>
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>
Co-authored-by: Mikhail Ryzhov <mikhail.ryzhov@intel.com>
Co-authored-by: Nikolay Shchegolev <nikolay.shchegolev@intel.com>
Co-authored-by: Ilya Lavrenov <ilya.lavrenov@intel.com>
Co-authored-by: Maxim Vafin <maxim.vafin@intel.com>
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Anastasia Popova <anastasia.popova@intel.com>
Co-authored-by: Maksim Doronin <maksim.doronin@intel.com>
Co-authored-by: Andrew Bakalin <andrew.bakalin@intel.com>
Co-authored-by: Mikhail Letavin <mikhail.letavin@intel.com>
Co-authored-by: Anton Chetverikov <Anton.Chetverikov@intel.com>
Co-authored-by: Ivan Tikhonov <ivan.tikhonov@intel.com>
Co-authored-by: Andrey Somsikov <andrey.somsikov@intel.com>
Co-authored-by: Maxim Shevtsov <maxim.y.shevtsov@intel.com>
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
Co-authored-by: Andrey Dmitriev <andrey.dmitriev@intel.com>
Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Edward Shogulin <edward.shogulin@intel.com>
Co-authored-by: Svetlana Dolinina <svetlana.a.dolinina@intel.com>
Co-authored-by: Alexey Suhov <alexey.suhov@intel.com>
Co-authored-by: Sergey Lyubimtsev <sergey.lyubimtsev@intel.com>
Co-authored-by: Dmitry Kurtaev <dmitry.kurtaev+github@gmail.com>
Co-authored-by: Nikolay Tyukaev <nikolay.tyukaev@intel.com>
Co-authored-by: Kate Generalova <kate.generalova@intel.com>
2021-04-19 20:19:17 +03:00
To run the tool, you can use [public ](@ref omz_models_group_public ) or [Intel's ](@ref omz_models_group_intel ) pre-trained models from the Open Model Zoo. The models can be downloaded using the [Model Downloader ](@ref omz_tools_downloader ).
2019-10-04 19:26:43 +03:00
2020-07-20 17:36:08 +03:00
> **NOTE**: Before running the tool with a trained model, make sure the model is converted to the Inference Engine format (\*.xml + \*.bin) using the [Model Optimizer tool](../../../docs/MO_DG/Deep_Learning_Model_Optimizer_DevGuide.md).
2021-02-11 12:57:05 +03:00
>
2020-08-26 18:53:24 +03:00
> The sample accepts models in ONNX format (.onnx) that do not require preprocessing.
2019-10-04 19:26:43 +03:00
## Examples of Running the Tool
2021-09-15 16:49:11 +03:00
This section provides step-by-step instructions on how to run the Benchmark Tool with the `googlenet-v1` public model on CPU or GPU devices. As an input, the `car.png` file from the `<INSTALL_DIR>/samples/scripts/` directory is used.
2019-10-04 19:26:43 +03:00
2021-12-21 20:26:37 +03:00
> **NOTE**: The Internet access is required to execute the following steps successfully. If you have access to the Internet through the proxy server only, please make sure that it is configured in your OS environment.
2019-10-04 19:26:43 +03:00
1. Download the model. Go to the the Model Downloader directory and run the `downloader.py` script with specifying the model name and directory to download the model to:
```sh
2021-09-15 16:49:11 +03:00
cd < INSTAL_DIR > /extras/open_model_zoo/tools/downloader
2019-10-04 19:26:43 +03:00
```
```sh
python3 downloader.py --name googlenet-v1 -o < models_dir >
```
2021-12-21 20:26:37 +03:00
2. Convert the model to the Inference Engine IR format. Run the Model Optimizer using the `mo` command with the path to the model, model format (which must be FP32 for CPU and FPG) and output directory to generate the IR files:
2019-10-04 19:26:43 +03:00
```sh
2021-12-21 20:26:37 +03:00
mo --input_model < models_dir > /public/googlenet-v1/googlenet-v1.caffemodel --data_type FP32 --output_dir < ir_dir >
2021-02-11 12:57:05 +03:00
```
2021-09-15 16:49:11 +03:00
3. Run the tool with specifying the `<INSTALL_DIR>/samples/scripts/car.png` file as an input image, the IR of the `googlenet-v1` model and a device to perform inference on. The following commands demonstrate running the Benchmark Tool in the asynchronous mode on CPU and GPU devices:
2021-02-11 12:57:05 +03:00
2019-10-04 19:26:43 +03:00
* On CPU:
```sh
2021-09-15 16:49:11 +03:00
./benchmark_app -m < ir_dir > /googlenet-v1.xml -i < INSTALL_DIR > /samples/scripts/car.png -d CPU -api async --progress true
2019-10-04 19:26:43 +03:00
```
2021-09-13 14:01:49 +03:00
* On GPU:
2019-10-04 19:26:43 +03:00
```sh
2021-09-15 16:49:11 +03:00
./benchmark_app -m < ir_dir > /googlenet-v1.xml -i < INSTALL_DIR > /samples/scripts/car.png -d GPU -api async --progress true
2019-10-04 19:26:43 +03:00
```
2018-11-23 16:19:43 +03:00
2020-07-20 17:36:08 +03:00
The application outputs the number of executed iterations, total duration of execution, latency, and throughput.
2019-10-04 19:26:43 +03:00
Additionally, if you set the `-report_type` parameter, the application outputs statistics report. If you set the `-pc` parameter, the application outputs performance counters. If you set `-exec_graph_path` , the application reports executable graph information serialized. All measurements including per-layer PM counters are reported in milliseconds.
2018-11-23 16:19:43 +03:00
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
Below are fragments of sample output static and dynamic networks:
2018-11-23 16:19:43 +03:00
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
* For static network:
2019-10-04 19:26:43 +03:00
```
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
[Step 10/11] Measuring performance (Start inference asynchronously, 4 inference requests using 4 streams for CPU, limits: 60000 ms duration)
[ INFO ] BENCHMARK IS IN INFERENCE ONLY MODE.
[ INFO ] Input blobs will be filled once before performance measurements.
[ INFO ] First inference took 26.26 ms
Progress: [................... ] 99% done
2018-11-23 16:19:43 +03:00
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
[Step 11/11] Dumping statistics report
[ INFO ] Count: 6640 iterations
[ INFO ] Duration: 60039.70 ms
[ INFO ] Latency:
[ INFO ] Median: 35.36 ms
[ INFO ] Avg: 36.12 ms
[ INFO ] Min: 18.55 ms
[ INFO ] Max: 88.96 ms
[ INFO ] Throughput: 110.59 FPS
2019-10-04 19:26:43 +03:00
```
2018-11-23 16:19:43 +03:00
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
* For dynamic network:
2019-10-04 19:26:43 +03:00
```
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
[Step 10/11] Measuring performance (Start inference asynchronously, 4 inference requests using 4 streams for CPU, limits: 60000 ms duration)
[ INFO ] BENCHMARK IS IN FULL MODE.
[ INFO ] Inputs setup stage will be included in performance measurements.
[ INFO ] First inference took 26.80 ms
Progress: [................... ] 99% done
2019-04-12 18:25:53 +03:00
2019-10-04 19:26:43 +03:00
[Step 11/11] Dumping statistics report
Dynamic reshapes (#7788)
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
2021-12-17 12:20:43 +03:00
[ INFO ] Count: 5199 iterations
[ INFO ] Duration: 60043.34 ms
[ INFO ] Latency:
[ INFO ] Median: 41.58 ms
[ INFO ] Avg: 46.07 ms
[ INFO ] Min: 8.44 ms
[ INFO ] Max: 115.65 ms
[ INFO ] Latency for each data shape group:
[ INFO ] 1. data : [1, 3, 224, 224]
[ INFO ] Median: 38.37 ms
[ INFO ] Avg: 30.29 ms
[ INFO ] Min: 8.44 ms
[ INFO ] Max: 61.30 ms
[ INFO ] 2. data : [1, 3, 448, 448]
[ INFO ] Median: 68.21 ms
[ INFO ] Avg: 61.85 ms
[ INFO ] Min: 29.58 ms
[ INFO ] Max: 115.65 ms
[ INFO ] Throughput: 86.59 FPS
2019-10-04 19:26:43 +03:00
```
2019-04-12 18:25:53 +03:00
2018-11-23 16:19:43 +03:00
## See Also
2020-07-20 17:36:08 +03:00
* [Using Inference Engine Samples ](../../../docs/IE_DG/Samples_Overview.md )
* [Model Optimizer ](../../../docs/MO_DG/Deep_Learning_Model_Optimizer_DevGuide.md )
Feature/merge 2021 3 to master (#5307)
* Feature/azaytsev/cldnn doc fixes (#4600)
* Legal fixes, removed the Generating docs section
* Removed info regarding generating docs
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Feature/azaytsev/gna model link fixes (#4599)
* Added info on DockerHub CI Framework
* Feature/azaytsev/change layout (#3295)
* Changes according to feedback comments
* Replaced @ref's with html links
* Fixed links, added a title page for installing from repos and images, fixed formatting issues
* Added links
* minor fix
* Added DL Streamer to the list of components installed by default
* Link fixes
* Link fixes
* ovms doc fix (#2988)
* added OpenVINO Model Server
* ovms doc fixes
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Updated openvino_docs.xml
* Link Fixes
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
* Fix for broken CC in CPU plugin (#4595)
* Azure CI: Add "ref: releases/2021/3"
* Fixed clone rt info (#4597)
* [.ci/azure] Enable CC build (#4619)
* Formula fix (#4624)
* Fixed transformation to pull constants into Loop body (cherry-pick of PR 4591) (#4607)
* Cherry-pick of PR 4591
* Fixed typo
* Moved a check into the parameter_unchanged_after_iteration function
* Fixed KW hits (#4638)
* [CPU] Supported ANY layout for inputs in inferRequest (#4621)
* [.ci/azure] Add windows_conditional_compilation.yml (#4648) (#4655)
* Fix for MKLDNN constant layers execution (#4642)
* Fix for MKLDNN constant layers execution
* Single mkldnn::engine for all MKLDNN graphs
* Add workaround for control edges to support TF 2.4 RNN (#4634)
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Corrected PyYAML dependency (#4598) (#4620)
5.4.2 is absent on PyPI
* [CPU] Statically analyzed issues. (#4637)
* Docs api (#4657)
* Updated API changes document
* Comment for CVS-49440
* Add documentation on how to convert QuartzNet model (#4664)
* Add documentation on how to convert QuartzNet model (#4422)
* Add documentation on how to convert QuartzNet model
* Apply review feedback
* Small fix
* Apply review feedback
* Apply suggestions from code review
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Add reference to file
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Fixed bug in assign elimination transformation. (#4644)
* [doc] Updated PyPI support OSes (#4643) (#4662)
* [doc] Updated PyPI support OSes (#4643)
* Updated PyPI support OSes
* Added python versions for win and mac
* Update pypi-openvino-dev.md
* Update pypi-openvino-dev.md
* Update pypi-openvino-rt.md
* Update pypi-openvino-dev.md
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* [IE][VPU]: Fix empty output of CTCGreedyDecoderSeqLen (#4653)
* Allow the second output of CTCGreedyDecoderSeqLen to be nullptr in cases when it is not used but calculated in the Myriad plugin. In this case, parse the second output as FakeData
* It is a cherry-pick of #4652
* Update the firmware to release version
* [VPU] WA for Segmentation fault on dlclose() issue (#4645)
* Document TensorFlow 2* Update: Layers Support and Remove Beta Status (#4474) (#4711)
* Document TensorFlow 2* Update: Layers Support and Remove Beta Status
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Update documentation based on latest test results and feedback
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Remove ConvLSTM2D from supported layers list
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Document Dot layer without limitation
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Address feedback upon DenseFeatures and RNN operations
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Do a grammar correction
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Do a grammar correction based on feedback
Signed-off-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Updated nGraph custom op documentation (#4604)
* Updated nGraph custom op documentation
* Fixed comments
* [IE CLDNN] Fix missing variable initializations and types (#4669)
* Fix NormalizeL2 creation in QueryNetwork (cherry pick from master PR 4310) (#4651)
* Updated documentation about the supported YOLOv3 model from ONNX (#4722) (#4726)
* Restored folded Operations for QueryNetwork (#4685)
* Restored folded Operations for QueryNetwork
* Fixed comment
* Add unfolded constant operations to supported layers map
* Add STN to list of supported models (#4728)
* Fix python API for Loop/TensorIterator/Assign/ReadValue operations
* Catch std::except in fuzz tests (#4695)
Fuzz tests must catch all expected exceptions from IE. IE is using C++ std
library which may raise standard exceptions which IE pass through.
* Docs update (#4626)
* Updated latency case desc to cover multi-socket machines
* updated opt guide a bit
* avoiding '#' which is interpreted as ref
* Update CPU.md
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Update docs/optimization_guide/dldt_optimization_guide.md
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
* Blocked dims hwc 2021/3 (#4729)
* Fix for BlockedDims
* Added test for HWC layout
* [GNA] Update documentation regarding splits and concatenations support (#4740)
* Added mo.py to wheel packages (#4731)
* Inserted a disclaimer (#4760)
* Fixed some klockwork issues in C API samples (#4767)
* Feature/vpu doc fixes 2021 3 (#4635)
* Documentation fixes and updates for VPU
* minor correction
* minor correction
* Fixed links
* updated supported layers list for vpu
* [DOCS] added iname/oname (#4735)
* [VPU] Limit dlclose() WA to be used for Ubuntu only (#4806)
* Fixed wrong link (#4817)
* MKLDNN weights cache key calculation algorithm changed (#4790)
* Updated PIP install instructions (#4821)
* Document YOLACT support (#4749)
* Document YOLACT support
* Add preprocessing section
* Apply suggestions from code review
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Add documentation on how to convert F3Net model (#4863)
* Add instruction for F3Net model pytorch->onnx conversion
* Fix style
* Fixed dead lock in telemetry (#4873)
* Fixed dead lock in telemetry
* Refactored TelemetrySender.send function
* Refactored send function implementation to avoid deadlocks
* Unit tests for telemetry sender function
* Added legal header
* avladimi/cvs-31369: Documented packages content to YUM/APT IGs (#4839)
* Documented runtime/dev packages content
* Minor formatting fixes
* Implemented review comments
* Update installing-openvino-apt.md
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* [DOC] Low-Precision 8-bit Integer Inference (#4834)
* [DOC] Low-Precision 8-bit Integer Inference
* [DOC] Low-Precision 8-bit Integer Inference: comment fixes
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* Update docs/IE_DG/Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* [DOC] LPT comments fix
* [DOC] LPT comments fix: absolute links are updated to relative
* Update Int8Inference.md
* Update Int8Inference.md
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
* Avladimi/cherry pick from master (#4892)
* Fixed CVS-48061
* Reviewed and edited the Customization instructions
* Fixed broken links in the TOC
* Fixed links
* Fixed formatting in the IG for Raspberry
* Feature/benchmarks 2021 3 (#4910)
* added new topics, changed the intro text
* updated
* Updates
* Updates
* Updates
* Updates
* Updates
* Added yolo-v4-tf and unet-camvid-onnx graphs
* Date for pricing is updated to March 15th
* Feature/omz link changes (#4911)
* Changed labels for demos and model downloader
* Changed links to models and tools
* Changed links to models and tools
* Changed links to demos
* [cherry-pick] Extensibility docs review (#4915)
* Feature/ovsa docs 2021 3 (#4914)
* Updated to 2021-3, fixed formatting issues
* Fixed formatting issues
* Fixed formatting issues
* Fixed formatting issues
* Update ovsa_get_started.md
* Clarification of Low Latency Transformation and State API documentation (#4877)
* Assign/ReadValue, LowLatency and StateAPI clarifications
* Apply suggestions from code review: spelling mistakes
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
* fixed wording
* cherry-pick missing commit to release branch: low latency documentation
* Resolve review remarks
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Svetlana Dolinina <svetlana.a.dolinina@intel.com>
* DevCloud call outs (#4904)
* [README.md] change latest release to 2021.3
* [49342] Update recommended CMake version on install guide in documentation (#4763)
* Inserted a disclaimer
* Another disclaimer
* Update installing-openvino-windows.md
* Update installing-openvino-windows.md
* Update installing-openvino-windows.md
* Feature/doc fixes 2021 3 (#4971)
* Made changes for CVS-50424
* Changes for CVS-49349
* Minor change for CVS-49349
* Changes for CVS-49343
* Cherry-pick #PR4254
* Replaced /opt/intel/openvino/ with /opt/intel/openvino_2021/ as the default target directory
* (CVS-50786) Added a new section Reference IMplementations to keep Speech Library and Speech Recognition Demos
* Doc fixes
* Replaced links to inference_engine_intro.md with Deep_Learning_Inference_Engine_DevGuide.md, fixed links
* Fixed link
* Fixes
* Fixes
* Reemoved Intel® Xeon® processor E family
* fixes for graphs (#5057)
* compression.configs.hardware config to package_data (#5066)
* update OpenCV version to 4.5.2 (#5069)
* update OpenCV version to 4.5.2
* Enable mo.front.common.extractors module (#5038)
* Enable mo.front.common.extractors module (#5018)
* Enable mo.front.common.extractors module
* Update package_BOM.txt
* Test MO wheel content
* fix doc iframe issue - 2021.3 (#5090)
* wrap with htmlonly
* wrap with htmlonly
* Add specification for ExperimentalDetectron* oprations (#5128)
* Feature/benchmarks 2021 3 ehl (#5191)
* Added EHL config
* Updated graphs
* improve table formatting
* Wrap <iframe> tag with \htmlonly \endhtmlonly to avoid build errors
* Updated graphs
* Fixed links to TDP and Price for 8380
* Add PyTorch section to the documentation (#4972)
* Add PyTorch section to the documentation
* Apply review feedback
* Remove section about loop
* Apply review feedback
* Apply review feedback
* Apply review feedback
* doc: add Red Hat docker registry (#5184) (#5253)
* Incorporate changes in master
Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
Co-authored-by: Vladislav Volkov <vladislav.volkov@intel.com>
Co-authored-by: azhogov <alexander.zhogov@intel.com>
Co-authored-by: Ilya Churaev <ilya.churaev@intel.com>
Co-authored-by: Alina Kladieva <alina.kladieva@intel.com>
Co-authored-by: Evgeny Lazarev <evgeny.lazarev@intel.com>
Co-authored-by: Gorokhov Dmitriy <dmitry.gorokhov@intel.com>
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>
Co-authored-by: Mikhail Ryzhov <mikhail.ryzhov@intel.com>
Co-authored-by: Nikolay Shchegolev <nikolay.shchegolev@intel.com>
Co-authored-by: Ilya Lavrenov <ilya.lavrenov@intel.com>
Co-authored-by: Maxim Vafin <maxim.vafin@intel.com>
Co-authored-by: Anastasiya Ageeva <anastasiya.ageeva@intel.com>
Co-authored-by: Anastasia Popova <anastasia.popova@intel.com>
Co-authored-by: Maksim Doronin <maksim.doronin@intel.com>
Co-authored-by: Andrew Bakalin <andrew.bakalin@intel.com>
Co-authored-by: Mikhail Letavin <mikhail.letavin@intel.com>
Co-authored-by: Anton Chetverikov <Anton.Chetverikov@intel.com>
Co-authored-by: Ivan Tikhonov <ivan.tikhonov@intel.com>
Co-authored-by: Andrey Somsikov <andrey.somsikov@intel.com>
Co-authored-by: Maxim Shevtsov <maxim.y.shevtsov@intel.com>
Co-authored-by: Alina Alborova <alina.alborova@intel.com>
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
Co-authored-by: Andrey Dmitriev <andrey.dmitriev@intel.com>
Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Edward Shogulin <edward.shogulin@intel.com>
Co-authored-by: Svetlana Dolinina <svetlana.a.dolinina@intel.com>
Co-authored-by: Alexey Suhov <alexey.suhov@intel.com>
Co-authored-by: Sergey Lyubimtsev <sergey.lyubimtsev@intel.com>
Co-authored-by: Dmitry Kurtaev <dmitry.kurtaev+github@gmail.com>
Co-authored-by: Nikolay Tyukaev <nikolay.tyukaev@intel.com>
Co-authored-by: Kate Generalova <kate.generalova@intel.com>
2021-04-19 20:19:17 +03:00
* [Model Downloader ](@ref omz_tools_downloader )