* Extend throughput benchmark CLI parameters
* Added device name as the second CLI parameter with default CPU value
* Update samples/cpp/benchmark/throughput_benchmark/main.cpp
Co-authored-by: Zlobin Vladimir <vladimir.zlobin@intel.com>
* Fix comments to review
* Modified python version
* Modified documentation
* Fix comments to review
* Fixed the comment
* Modified python doc
* Fixed device name handling in python version
* Update main.cpp
* Update throughput_benchmark.py
---------
Co-authored-by: Zlobin Vladimir <vladimir.zlobin@intel.com>
* Improved warning message for BA + OpenCL case
* Update samples/cpp/benchmark_app/CMakeLists.txt
Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>
---------
Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>
The Python version uses `app_inputs_info` to represent different input configurations, but the C++ version extends that use case and uses `app_inputs_info` to represent different input images as well. That means that the assumption that if `app_input_info.size() > 1`, then input shape is dynamic, doesn’t always hold for C++
Ticket 117673
* Change `VPUX`/`VPU` occurrences to `NPU`
* Switch `HARDWARE_AWARE_IGNORED_PATTERNS` VPU to NPU
* Rename `MYRIAD plugin`
* Rename vpu_patterns to npu_patterns in tools/pot
* Rename vpu.json to npu.json in tools/pot
* Rename restrict_for_vpu to restrict_for_npu in tools/pot
* Change keembayOptimalBatchNum to npuOptimalBatchNum
---------
Co-authored-by: Dan <mircea-aurelian.dan@intel.com>
* [GNA] Fix for GeminiLake detection
* Added HWGeneration::GNA_1_0_E enumerator
Added DeviceVersion::GNAEmbedded1_0 enumerator, changed the meaning of DeviceVersion::GNA1_0.
Updated ConvLowPrecision test with all supported targets
* [GNA] Extended a few tests with GNA1.0
* Fix -api sync for single -data_shape
Tickets 111187 and 111185
I wasn’t able to find C++ equivalent of Python’s `info.original_shape.is_static`. Later I realized that it shouldn’t be considered because -shape cmd arg should have higher priority for shape inference than model’s shape. So I removed it from Python.
Replace
`if benchmark.inference_only and batch_size.is_dynamic:`
with
`if allow_inference_only_or_sync and batch_size.is_dynamic:`
to reset batch_size to static in case of dynamic shape with single -data_shape
* Check only app_input_info.size() == 1 because if it's gretaer than 1, input shape is dynamic and there are more that one static shapes. Apply TODO
* Deprecate ExecutableNetwork and InferRequest API
* Fixed some warnings
* Fixed some warnings
* Try to fix documentation
* Try to skip documentation warnings
* Build using conanfile.txt
* Update .ci/azure/linux_arm64.yml
* Several improvements
* Removed conanfile.py
* Try to use activate / deactivate
* Fixed clang-format code style
* Supported TBB version from Conan
* Added more NOMINMAX
* Fixed static build
* More improvements for static build
* Add usage of static snappy in case of static build
* More fixes
* Small fixes
* Final fixes
* Remove constructors for ov Exceptions
* Fixed linux build
* Fixed ONNX Frontend
* Fixed paddle
* Fixed exceptions in tests
* Deprecate constructors for ov::Exception
* Suppress some warnings
* Merge several exceptions
* Some small changes
* Suppress more warnings
* More warnings
* mode warnings
* Suppress more warnings
* More warnings