* Add test case for the loadNetwork with Auto Batching.
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Enable logic test case for GPU.
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Update.
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Enable property for config key 'AUTO_BATCH_DEVICE_CONFIG'.
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Omit {}.
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Add commont test for the property ALLOW_AUTO_BATCHING.
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Add commont test for AUTO Batching plugin.
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Moving PWL to ngraph
* improving the running time of php_search; refactoring the pwl operation
* fixed erros & refactored code
* moved PWL op to GNA
* Update src/plugins/intel_gna/ops/pwl.hpp
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
* Update src/plugins/intel_gna/ops/reference/pwl.hpp
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
* Update src/plugins/intel_gna/ops/pwl.cpp
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
* Update src/plugins/intel_gna/transformations/transpose_to_pwl.hpp
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
* Update src/plugins/intel_gna/transformations/transpose_to_pwl.cpp
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
* fixed compilation error
* Update inference-engine/tests/unit/gna/ngraph/transformations/gna_pwl.cpp
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
* added some tests; changed algorithm of checking accuracy of pwl; refactoring
* added first and last segments; added fq and fixed errors
* fixed after review & rewrote some tests on ngraph
* removed debug logs & fixed code style check error
* s/ngraph_helper/ngraph_util
* removed TRANSFORMATIONS_API in PWLApproximation class declaration
* removed OPENVINO_API in Pwl class declaration
* replaced the deprecated version of evaluate() with a new one
* fixed some problems after reviewing
* fixed a problem when a value of function of left point of segment is less than minimum of function
* corrected a value of the right point of last segments
* s/OPENVINO_RTTI/OPENVINO_OP
Co-authored-by: Elizaveta Lobanova <elizaveta.lobanova@intel.com>
+ Fix colorization-sig accuracy issue using oneDNN
Memory crash in case reuse_eltwise_sum_post in oneDNN and memory_pool
And print node in/out gpu_usm_mem addr at OV_GPU_Verbose >= 1
+ Check the size of z spatial axis for checking fulltensor.
+ Remove program_helpers's functions.
Co-authored-by: hyunback <hyunback.kim@intel.com>
Scenario:
- Node "Split" with multiple outputs (e.g. 3). All outputs are connected to "Result"s
- Add post-processing step (e.g. convert element type, can be also implicit)
Issue: after post-processing, 3 new results will be created, each will have "Split" friendly name - inconsistency with IRv10 rules
Fix:
- For nodes with multiple outputs, add '.<idx>' suffix to new output's friendly name
- If no post-processing is applied, return immediately, keeping original results as is
Tests:
- Split with 3 outputs where 2 outputs have post-processing.
- Split with 3 outputs, post-processing doesn't create any nodes
* [XLink] - tests to smoke scope
* [XLink] - small change in XLink related file to trigger ie-tests-windows-myriadx
* [XLink] - azure windows and linux
* [XLink] - azure windows and linux
* [XLink] - azure windows and linux - change dir?
* [XLink] - azure windows and linux - change dir?
* [XLink] - azure windows and linux - install?
* [XLink] - azure windows and linux - xlink cmake
* [XLink] - azure windows and linux - XLinkTests because another target with the same name already exists
* [XLink] - azure windows and linux - XLinkTests because another target with the same name already exists
* [XLink] - azure windows and linux - install TARGETS given target XLinkTests which does not exist
* [XLink] - azure windows and linux - remove smoke
Inserting padding into oneDNN primitive has issue with implicit concat behavior.
Deconv onedNN initialized output buffer to 0 including padding area. Padding area should be reserved.
Use oneDNN offset from program_node in/out lower_padding instead of oneDNN memory desc.
Signed-off-by: hyunback <hyunback.kim@intel.com>
* add 3D shape to test and rename crop4d to strided_slice
* remove ConvertStridedSliceToCropNegative2 since 3D is now supported
* add myriad functional tests to skip-list
* update Auto docs
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* update python snippets
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* remove vpu, fix a mistaken in python code
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* update MYRIAD device full name
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* update API name
old API use name Inference Engine API
NEW API usen name OpenVINO Runtime API 2.0
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* update tab name, and code format
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* fix AUTO4 format issue
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* update set_property code
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* auto draft
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* mv code into .cpp and .py
modify the devicelist part accoding to the review
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* remove priority list in code and document
modify the begning of the document
remove perfomance data
remove old API
use compile_model instead of set_property
add a image about cpu accelerate
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* fix mis print and code is not match document
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* try to fix doc build issue
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* fix snippets code compile issue
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* Added migration for deployment (#10800)
* Added migration for deployment
* Addressed comments
* more info after the What's new Sessions' questions (#10803)
* more info after the What's new Sessions' questions
* generalizing the optimal_batch_size vs explicit value message
* Update docs/OV_Runtime_UG/automatic_batching.md
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Update docs/OV_Runtime_UG/automatic_batching.md
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Update docs/OV_Runtime_UG/automatic_batching.md
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Update docs/OV_Runtime_UG/automatic_batching.md
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Update docs/OV_Runtime_UG/automatic_batching.md
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Update docs/OV_Runtime_UG/automatic_batching.md
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Perf Hints docs and General Opt Guide refactoring (#10815)
* Brushed the general optimization page
* Opt GUIDE, WIP
* perf hints doc placeholder
* WIP
* WIP2
* WIP 3
* added streams and few other details
* fixed titles, misprints etc
* Perf hints
* movin the runtime optimizations intro
* fixed link
* Apply suggestions from code review
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* some details on the FIL and other means when pure inference time is not the only factor
* shuffled according to general->use-case->device-specifics flow, minor brushing
* next iter
* section on optimizing for tput and latency
* couple of links to the features support matrix
* Links, brushing, dedicated subsections for Latency/FIL/Tput
* had to make the link less specific (otherwise docs compilations fails)
* removing the Temp/Should be moved to the Opt Guide
* shuffled the tput/latency/etc info into separated documents. also the following docs moved from the temp into specific feature, general product desc or corresponding plugins
- openvino_docs_IE_DG_Model_caching_overview
- openvino_docs_IE_DG_Int8Inference
- openvino_docs_IE_DG_Bfloat16Inference
- openvino_docs_OV_UG_NoDynamicShapes
* fixed toc for ov_dynamic_shapes.md
* referring the openvino_docs_IE_DG_Bfloat16Inference to avoid docs compilation errors
* fixed main product TOC, removed ref from the second-level items
* reviewers remarks
* reverted the openvino_docs_OV_UG_NoDynamicShapes
* reverting openvino_docs_IE_DG_Bfloat16Inference and openvino_docs_IE_DG_Int8Inference
* "No dynamic shapes" to the "Dynamic shapes" as TOC
* removed duplication
* minor brushing
* Caching to the next level in TOC
* brushing
* more on the perf counters ( for latency and dynamic cases)
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
* Updated common IE pipeline infer-request section (#10844)
* Updated common IE pipeline infer-reqest section
* Update ov_infer_request.md
* Apply suggestions from code review
Co-authored-by: Karol Blaszczak <karol.blaszczak@intel.com>
Co-authored-by: Maxim Shevtsov <maxim.y.shevtsov@intel.com>
Co-authored-by: Karol Blaszczak <karol.blaszczak@intel.com>
* DOCS: Removed useless 4 spaces in snippets (#10870)
* Updated snippets
* Added link to encryption
* [DOCS] ARM CPU plugin docs (#10885)
* initial commit
ARM_CPU.md added
ARM CPU is added to the list of supported devices
* Update the list of supported properties
* Update Device_Plugins.md
* Update CODEOWNERS
* Removed quotes in limitations section
* NVIDIA and Android are added to the list of supported devices
* Added See Also section and reg sign to arm
* Added Preprocessing acceleration section
* Update the list of supported layers
* updated list of supported layers
* fix typos
* Added support disclaimer
* update trade and reg symbols
* fixed typos
* fix typos
* reg fix
* add reg symbol back
Co-authored-by: Vitaly Tuzov <vitaly.tuzov@intel.com>
* Try to fix visualization (#10896)
* Try to fix visualization
* New try
* Update Install&Deployment for migration guide to 22/1 (#10933)
* updates
* update
* Getting started improvements (#10948)
* Onnx updates (#10962)
* onnx changes
* onnx updates
* onnx updates
* fix broken anchors api reference (#10976)
* add ote repo (#10979)
* DOCS: Increase content width (#10995)
* fixes
* fix
* Fixed compilation
Co-authored-by: Maxim Shevtsov <maxim.y.shevtsov@intel.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
Co-authored-by: Karol Blaszczak <karol.blaszczak@intel.com>
Co-authored-by: Aleksandr Voron <aleksandr.voron@intel.com>
Co-authored-by: Vitaly Tuzov <vitaly.tuzov@intel.com>
Co-authored-by: Ilya Churaev <ilya.churaev@intel.com>
Co-authored-by: Yuan Xu <yuan1.xu@intel.com>
Co-authored-by: Victoria Yashina <victoria.yashina@intel.com>
Co-authored-by: Nikolay Tyukaev <nikolay.tyukaev@intel.com>