Commit Graph

3454 Commits

Author SHA1 Message Date
Aleksandr Korolev
c87ac722b1 [VPU] Enable new InferRequest behavior tests with OV 2.0 (#9301) 2021-12-28 11:21:00 +03:00
Vladimir Paramuzov
f565e0f854 [GPU] Merge cldnn and plugin code (#8484) 2021-12-27 18:35:01 +03:00
Sergey Shlyapnikov
95d86eb2bf [GPU] Add parallel quantizes optimization (#9370) 2021-12-27 09:47:20 +03:00
Vladimir Paramuzov
0fa226a0c2 [GPU] Fixed uninit vairable and exceptions from nothrow methods (#9426) 2021-12-27 09:46:52 +03:00
Sergey Lyubimtsev
73143b8c03 Add batch plugin to openvino wheel (#9432) 2021-12-25 12:42:55 +03:00
Mikhail Ryzhov
43c45d3065 Moved gna library cmake to plugin dir (#9393) 2021-12-24 13:02:43 +03:00
Maxim Shevtsov
49b5e5728b Auto Batching impl (#7883)
* auto-batching POC squashed (all commits from auto-batch-2021.3 branch)

(cherry picked from commit d7742f2c747bc514a126cc9a4d5b99f0ff5cbbc7)

* applying/accomodating the API changes after rebase to the master

* replaying modified version of actual batch selection

* eearly experiments with model mem footprint

* changes from rebasing to the latest master

* experimenting with DG1 on the batch size selection, also collecting the mem footprint

* WIP:moving the auto-batching to the icore to let the MULT/AUTO support that, ALLOW_AUTO_BATCHING as a conventional config key. still fials hot device swap

* quick-n-dirty batch footpint vs device total mem

* code style

* testing which models perform badly due to kernels and NOT (batched) footprint

* stub  pipeline task to comunicate the readiness rather than promise/future

* quick-n-dirty timeout impl

* explicit _completionTasks,reverting BA to use the timeout

* inputs outputs copies, works with AUTO and demo now

* accomodate the config per device-id, after rebase to the latest master

* allowing the auto-batching only with tput hint to let more conventional tests pass

* fix the pre-mature timeout restaring via waiting for batch1 requests completion

* moved the bacthed request statring ( along with input copies) to the dedicated thread

* [IE CLDNN] Disable bs_fs_yx_bsv16_fsv16 format for int8 convolution

* code style

* increasing the timeout to test the ssd_* models perf (timeout?) issues

* reducing number of output stuff in BA to avoid bloating the logs in experiments

* more aggressive batching for experiments, not limited to 32 and also 4 as a min

* more accurate timeout debugging info

* getting the reqs limitation from the plugin SetConfig as well

* refactor the reshape logic a bit to accomodate CPU for bathcing, also added remeote context

* let the benchamrk_app to consume specific batch values for the auto-batching such as BATCH:GPU(4)

* auto-batching functional test (with results check vs ref) and GPU instance for that

* fixed arithemtic on blobs ptrs

* clang

* handling possible batched network failure

* BATCH as the constants device name in test

* ENABLE_BATCH

* func tests for CPU, also DetectionOutput hetero tests (CPU and GPU)

* DetectionOutput hetero test for the CPU

* reenabling the Auto-Batching in the AUTO

* auto-batching device enabled in the test

* fixed the DO test

* improve the loading loop logic

* brushed the config keys

* allow hetero code-path for explicit device name like BATCH:GPU(4), used in the hetero code-path tests

* fix the test after refactoring

* clang

* moving ThreadSafeQueue to the ie_parallel, as it is re-used in the AUTO/MULTI and BATCH now

* auto-batching hetero test (subgraph with DetectionOutput)

* fixed minor changes that were result of experiments with impl

* code-style

* brushing, disabling CPU's HETERO tests until planned activity for 22.2

* removing home-baked MAX_BATCH_SZIE and swicthing to the official impl by GPU team

* remote blobs tests for the auto-batching (old API)

* brushed names a bit

* CreateContext and LoadNEtwork with context for the Auto-Batching plus remote-blobs tests

* fixed the ieUnitTests with adding CreateContext stub to the MockICore

* clang

* improved remote-blobs tests

* revert the back BA from exeprimenents with AB + device_use_mem

* conformance tests for BATCH, alos batch size 1 is default for BATCH:DEVICE

* remote blobs 2.0 tests, issue with context having the orig device name

* debugging DG1 perf drop (presumably due to non-fitting the device-mem)

* disbaling WA with batch/=2 for excesive mem footptint, leaving only streams 2

* remote blobs 2.0 tests for different tensor sharing types

* converting assert to throw to accomodate legacy API where the lock() was possible to be called

* revert the timeout back to avoid mixing the studies, fixed the footprint calc

* reverting to estimating the max batch by extrapolating from bacth1 size

* more conservative footptint etimation (with bacth1), graceful bacth 1 handling without duplication

* even graceful batch 1 handling without duplication

* WA for MAX_BATCH_SIZE failure, removing batch4 as a min for the auto-batching

* AutoBatchPlugin -> ov_auto_batch_plugin

* WA for gcc 4.8

* clang

* fix misprint

* fixed errors resulted from recent OV's Variant to Any transition

* skip auto-batching for already-batched networks

* AUTO_BATCH_TIMEOUT and tests

* GPU-specific L3

* switched to pure config, also improved ALLOW_AUTO_BATCHING config key handling logic

* debugging device info

* enabling the config tests for the GPU and fixing the Auto-batching tests to pass

* making the default (when not recognized the driver) cache size more aggressive, to accomodate recent HW with old drivers

* skip auto-batching for RNNs and alikes (e.g. single CHW input)

* fixed fallback to the bacth1 and moved HETERO path under condition to avoid bloating

* brushing

* Auto plugin GetMetric support gpu auto-batch

Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>

* add test case

Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>

* add comments on test

Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>

* brushing the vars names, alos adding the excpetion handling

* disabling the auto-batching for the networks with non-batched outputs and faster-rcnn and alikes (CVS-74085) to minimize the of #failures

* add try catch

Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>

* brushing the code changed in the GPU plugin

* Auto-Batch requests tests

* brushed varibles a bit (ref)

* cleaned debug output from the ie_core

* cleaned cmake for the Auto-Batch

* removed batchN estimation from batch1

* cleaned from debug printf

* comments, cleanup

* WA the mock test errors introduced with merging the https://github.com/myshevts/openvino/pull/13

* Adding back  removed batchN estimation from batch1 to debug degradations on DG1 (resulted from too optimistic MAX_BATCH_SIZE?). This partially reverts commit e8f1738ac1.

* brushing ie_core.cpp

* fix 32bit compilation

* Code review: ENABLE_AUTO_BATCH

* consolidate the auot-batching logic in ie_core.cpp into single ApplyAutoBAtching

* renamed brushed the OPTIMAL_BATCH (now with_SIZE) and mimicks the MAX_BATCH_SZIE  wrt MODEL_PTR

* default value for the OPTIMAL_BATCH_SIZE

* clang

* accomodate new func tests location

* fix shuffle of headers after clang + copyrights

* fixed misprint made during code refactoring

* moving the common therad-safe containers (like ThreadSafeQueue) to the dedicated dev_api header

* switch from the device name to the OPTIMAL_BATCH_SIZE metric presence as a conditin to consider Auto-Batching

* switching from the unsafe size() and minimizing time under lock

* code style

* brushed the ApplyAutoBatching

* brushed the netric/config names and descriptions

* completed the core intergration tests for the auto-batching

* ExecGraphInfo and check for incorrect cfg

* removed explicit dependencies from cmake file of the plugin

* disabling Auto-Batching thru the tput hint (to preserve current product default), only excplicit like BATCH:GPU used in the tests

Co-authored-by: Roman Lyamin <roman.lyamin@intel.com>
Co-authored-by: Hu, Yuan2 <yuan2.hu@intel.com>
2021-12-24 12:55:22 +03:00
okhovan
31b6b034bc [GPU] MaxPool-8 (#9064) 2021-12-24 11:18:58 +03:00
serhii-pavlovskyi-altran
8315fe0e19 [GPU] Range v4 partial implementation (#8907) 2021-12-23 15:49:43 +03:00
Ilya Churaev
5e1d241c11 Renamed template plugin and tests (#9389)
* Renamed template plugin and plugin's tests

* Renamed template_extension
2021-12-23 14:59:24 +03:00
Andrey Sapozhnikov
da67ba135c [GNA] Remove GNA Library versioning (#9319) 2021-12-23 14:32:58 +03:00
Vladislav Volkov
60a11a6348 [CPU] Renamed CPU plugin to ov_intel_cpu_plugin (#9342) 2021-12-23 11:49:25 +03:00
Paul Youngsoo Ahn
bbceae3bc3 [GPU] Add INT32/UINT32 to available input data types when load type is aligned in GetJitLoad (#9300) (#9300)
- Modify fusibility checking to allow sub/div eltwise fusing for other primitives
- Modify dump checking code to use node name in exec graph
2021-12-23 17:48:34 +09:00
Mingyu Kim
16490959e6 [GPU] Use double blocked format if batch >= 16 (#9357) 2021-12-23 16:58:33 +09:00
Taylor Yeonbok Lee
3b03728807 [GPU] Fix get_estimated_device_mem_usage to handle mutable_data (#9297) 2021-12-23 16:20:04 +09:00
Sergey Shlyapnikov
507a498269 [GPU] Add OneDNN post ops description in graph dump mode (#9371) 2021-12-23 09:58:35 +03:00
Ilya Churaev
42350a705e Remove legacy targets (#9333)
* Remove some legacy targets

* Replace some targets

* Removed inference_engine_plugin_api dependency

* Minor comment for developer config

* Fixed include paths

* Small fixes for static build

* Try to fix build pyopenvino

* Fixed comments

* Try to fix build

* Include OpenVINODeveloperPackage inside InferenceEngineDeveloperPackageConfig

* Try to fix GAPI tests
2021-12-23 08:16:23 +03:00
Luwei Zhou
3d244a41ab [shape_infer]Implement shape inference of Roll, ROIAlign,Proposal (#8610)
* Implement the proposal and experimental_detecron_generate_proposals

* Implement the proposal shape infer

* Add ROI_Align OP shape infer implement.

* Fix building issue

* Fix bug.

* Update test cases.

* Add test cases for the OPs

* Apply the CI coding style check.

* Move the shape_infer API to the new folder.

* Update some fix.

* Applied review comments

* Move the shape infer tests into new folder.

* Apply review comments.

* Fix missing header when mering with master
2021-12-23 03:02:15 +00:00
Zhang Yi
529ab2b099 Yi3/shape infer 2nd batch (#8420)
* [ShapeInfer]shape infer 2nd batch

* [ShapeInfer]Impl Reverse Sequence

* [ShapeInfer]Fix typo

* [ShapeInfer]fix error

* [ShapeInfer]remove useless code

* [ShapeInfer]fix code style

* [ShapeInfer]enable shape_infer in mkldnn

* [ShapeInfer]use shape_inference in tests

* [ShapeInfer]add partialshape for tests

* [ShapeInfer]revise test cases

* [ShapeInfer]fix review comments

* [ShapeInfer]remove debug logs

* [ShapeInfer]fix ci build

* [ShapeInfer]Fix errors

* [ShapeInfer]fix build

* [ShapeInfer]fix bug

* [ShapeInfer]remove useless check

* [ShapeInfer]Fix vpu tests

* [ShapeInfer]Fix extract_image

* [ShapeInfer]apply reviews
2021-12-22 18:14:37 +03:00
Ilya Znamenskiy
eca9ff6981 [GPU] Fixed gemm_tiled_opt fusing error and code refactoring in fusing tests (#9303) 2021-12-22 14:38:24 +03:00
Min, Byungil
0e3c4cc103 Modify padding checking logic to enable new model (#9310)
+ Modified a way to add padding in prepare_padding
+ Changed condition of assertion for onednn padding

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2021-12-22 18:17:08 +09:00
Mang Guo
5fada94504 Mang/shape inference (#8412)
* Implement DFT and IDFT shape inference

* Implement CTCLoss shape inference

* Fix error message.

* Refactor test case.

* Apply review comments

* Apply review comments

* Fix clang format error

* Fix merge error

* Remove axes_vector.
2021-12-22 07:47:44 +03:00
Tingqian Li
b8e6b6368c [Shape inference] Pad_1/Topk_3/Split_1/VariadicSplit_1/ExperimentalDetectronROIFeatureExtractor_6/Bucketize_3/EmbeddingBagOffsetsSum_3/EmbeddingSegmentsSum_3/Range_4/RegionYolo_0/ReorgYolo_0 (#8413)
* [shape_infer] add shape_infer for ExperimentalDetectronROIFeatureExtractor op

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* add test

* Use compatible & merge for intersection checks

* Update

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* Add perf_test

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* Initial commit

* fix compile issue

* Add test

* fix clang format issue

* support for pads_begin/pads_end with different sizes

* fix bug in EDGE mode checking

* fix padding mode checks

* fix according to jane's review comment

* fix const reference

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* Initial commit

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* fix bugs

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* Switch to use single generic code with small helper template

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* Initial commit on Split

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* Convolution update

* Adds pragma once

* Reductions shape infer

* Shape nodes

* style

* Update

* add exp detectron roi feature

* Update

Signed-off-by: Li, Tingqian <tingqian.li@intel.com>

* Use get_data_as_int64 + constant_data

* Add test

* Add utils.hpp into cpuUnit shape inference test

* avoid using friend template function

* fix topk axis bug

* Add bucketize

* Add embeddingbag offsets sum

* Add embedding segments sum

* fix code style issue

* Add Range_4

* Update tests

* Add range

* Add region Yolo

* Add reorg

* fix according to Globev's comment

* call shape_infer in evaluate_variadic_split()

* fix CI issue

* fix CI issue

* fix CI issue, topk change revert

* fix flake8 E302

* fix myriad smoke test issue

* fix according to Vladislav's second round review

* fix format

* Add StridedSlice & Einsum

* fix pad_test.cpp build issue

* fix according to review comment

* insert directly into output shape

* revert infer_slice_shape() change since vpux compiler uses this function

* move tests

Co-authored-by: Stepyreva, Evgenya <evgenya.stepyreva@intel.com>
2021-12-22 06:54:12 +03:00
Jade Cho
7bcca1b82d [GPU] Fix some bugs of cldnn eltwise kernel at bs_fs_yx_bsv32_fsv16 format. (#9198)
* Fix some bugs of cldnn eltwise kernel at bs_fs_yx_bsv32_fsv16 format.

+ Add a condition so that the eltwise_simple_vload8 kernel is not selected when the tensor is not aligned as bsv32_fsv16 or bsv32_fsv32.
+ Optimize gws/lws of eltwise_ref kernel for bsv32_fsv16 format.

* Check feature align of b_fs_yx_fsv32 for eltwise vload8 kernel

+ Minor fix for OV_GPU_Help option
2021-12-22 11:43:04 +09:00
Andrey Zaytsev
4ae6258bed Feature/azaytsev/from 2021 4 (#9247)
* Added info on DockerHub CI Framework

* Feature/azaytsev/change layout (#3295)

* Changes according to feedback comments

* Replaced @ref's with html links

* Fixed links, added a title page for installing from repos and images, fixed formatting issues

* Added links

* minor fix

* Added DL Streamer to the list of components installed by default

* Link fixes

* Link fixes

* ovms doc fix (#2988)

* added OpenVINO Model Server

* ovms doc fixes

Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>

* Updated openvino_docs.xml

* Updated the link to software license agreements

* Revert "Updated the link to software license agreements"

This reverts commit 706dac500e.

* Docs to Sphinx (#8151)

* docs to sphinx

* Update GPU.md

* Update CPU.md

* Update AUTO.md

* Update performance_int8_vs_fp32.md

* update

* update md

* updates

* disable doc ci

* disable ci

* fix index.rst

Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
# Conflicts:
#	.gitignore
#	docs/CMakeLists.txt
#	docs/IE_DG/Deep_Learning_Inference_Engine_DevGuide.md
#	docs/IE_DG/Extensibility_DG/Custom_ONNX_Ops.md
#	docs/IE_DG/Extensibility_DG/VPU_Kernel.md
#	docs/IE_DG/InferenceEngine_QueryAPI.md
#	docs/IE_DG/Int8Inference.md
#	docs/IE_DG/Integrate_with_customer_application_new_API.md
#	docs/IE_DG/Model_caching_overview.md
#	docs/IE_DG/supported_plugins/GPU_RemoteBlob_API.md
#	docs/IE_DG/supported_plugins/HETERO.md
#	docs/IE_DG/supported_plugins/MULTI.md
#	docs/MO_DG/prepare_model/convert_model/Convert_Model_From_Caffe.md
#	docs/MO_DG/prepare_model/convert_model/Convert_Model_From_Kaldi.md
#	docs/MO_DG/prepare_model/convert_model/Convert_Model_From_MxNet.md
#	docs/MO_DG/prepare_model/convert_model/Convert_Model_From_ONNX.md
#	docs/MO_DG/prepare_model/convert_model/Converting_Model.md
#	docs/MO_DG/prepare_model/convert_model/Converting_Model_General.md
#	docs/MO_DG/prepare_model/convert_model/Cutting_Model.md
#	docs/MO_DG/prepare_model/convert_model/pytorch_specific/Convert_RNNT.md
#	docs/MO_DG/prepare_model/convert_model/tf_specific/Convert_EfficientDet_Models.md
#	docs/MO_DG/prepare_model/convert_model/tf_specific/Convert_WideAndDeep_Family_Models.md
#	docs/MO_DG/prepare_model/convert_model/tf_specific/Convert_YOLO_From_Tensorflow.md
#	docs/doxygen/Doxyfile.config
#	docs/doxygen/ie_docs.xml
#	docs/doxygen/ie_plugin_api.config
#	docs/doxygen/ngraph_cpp_api.config
#	docs/doxygen/openvino_docs.xml
#	docs/get_started/get_started_macos.md
#	docs/get_started/get_started_raspbian.md
#	docs/get_started/get_started_windows.md
#	docs/img/cpu_int8_flow.png
#	docs/index.md
#	docs/install_guides/VisionAcceleratorFPGA_Configure.md
#	docs/install_guides/VisionAcceleratorFPGA_Configure_Windows.md
#	docs/install_guides/deployment-manager-tool.md
#	docs/install_guides/installing-openvino-linux.md
#	docs/install_guides/installing-openvino-macos.md
#	docs/install_guides/installing-openvino-windows.md
#	docs/optimization_guide/dldt_optimization_guide.md
#	inference-engine/ie_bridges/c/include/c_api/ie_c_api.h
#	inference-engine/ie_bridges/python/docs/api_overview.md
#	inference-engine/ie_bridges/python/sample/ngraph_function_creation_sample/README.md
#	inference-engine/ie_bridges/python/sample/speech_sample/README.md
#	inference-engine/ie_bridges/python/src/openvino/inference_engine/ie_api.pyx
#	inference-engine/include/ie_api.h
#	inference-engine/include/ie_core.hpp
#	inference-engine/include/ie_version.hpp
#	inference-engine/samples/benchmark_app/README.md
#	inference-engine/samples/speech_sample/README.md
#	inference-engine/src/plugin_api/exec_graph_info.hpp
#	inference-engine/src/plugin_api/file_utils.h
#	inference-engine/src/transformations/include/transformations_visibility.hpp
#	inference-engine/tools/benchmark_tool/README.md
#	ngraph/core/include/ngraph/ngraph.hpp
#	ngraph/frontend/onnx_common/include/onnx_common/parser.hpp
#	ngraph/python/src/ngraph/utils/node_factory.py
#	openvino/itt/include/openvino/itt.hpp
#	thirdparty/ade
#	tools/benchmark/README.md

* Cherry-picked remove font-family (#8211)

* Cherry-picked: Update get_started_scripts.md (#8338)

* doc updates (#8268)

* Various doc changes

* theme changes

* remove font-family (#8211)

* fix  css

* Update uninstalling-openvino.md

* fix css

* fix

* Fixes for Installation Guides

Co-authored-by: Andrey Zaytsev <andrey.zaytsev@intel.com>
Co-authored-by: kblaszczak-intel <karol.blaszczak@intel.com>
# Conflicts:
#	docs/IE_DG/Bfloat16Inference.md
#	docs/IE_DG/InferenceEngine_QueryAPI.md
#	docs/IE_DG/OnnxImporterTutorial.md
#	docs/IE_DG/supported_plugins/AUTO.md
#	docs/IE_DG/supported_plugins/HETERO.md
#	docs/IE_DG/supported_plugins/MULTI.md
#	docs/MO_DG/prepare_model/convert_model/Convert_Model_From_Kaldi.md
#	docs/MO_DG/prepare_model/convert_model/tf_specific/Convert_YOLO_From_Tensorflow.md
#	docs/install_guides/installing-openvino-macos.md
#	docs/install_guides/installing-openvino-windows.md
#	docs/ops/opset.md
#	inference-engine/samples/benchmark_app/README.md
#	inference-engine/tools/benchmark_tool/README.md
#	thirdparty/ade

* Cherry-picked: doc script changes (#8568)

* fix openvino-sphinx-theme

* add linkcheck target

* fix

* change version

* add doxygen-xfail.txt

* fix

* AA

* fix

* fix

* fix

* fix

* fix
# Conflicts:
#	thirdparty/ade

* Cherry-pick: Feature/azaytsev/doc updates gna 2021 4 2 (#8567)

* Various doc changes

* Reformatted C++/Pythob sections. Updated with info from PR8490

* additional fix

* Gemini Lake replaced with Elkhart Lake

* Fixed links in IGs, Added 12th Gen
# Conflicts:
#	docs/IE_DG/supported_plugins/GNA.md
#	thirdparty/ade

* Cherry-pick: Feature/azaytsev/doc fixes (#8897)

* Various doc changes

* Removed the empty Learning path topic

* Restored the Gemini Lake CPIU list
# Conflicts:
#	docs/IE_DG/supported_plugins/GNA.md
#	thirdparty/ade

* Cherry-pick: sphinx copybutton doxyrest code blocks (#8992)

# Conflicts:
#	thirdparty/ade

* Cherry-pick: iframe video enable fullscreen (#9041)

# Conflicts:
#	thirdparty/ade

* Cherry-pick: fix untitled titles (#9213)

# Conflicts:
#	thirdparty/ade

* Cherry-pick: perf bench graph animation (#9045)

* animation

* fix
# Conflicts:
#	thirdparty/ade

* Cherry-pick: doc pytest (#8888)

* docs pytest

* fixes
# Conflicts:
#	docs/doxygen/doxygen-ignore.txt
#	docs/scripts/ie_docs.xml
#	thirdparty/ade

* Cherry-pick: restore deleted files (#9215)

* Added new operations to the doc structure (from removed ie_docs.xml)

* Additional fixes

* Update docs/IE_DG/InferenceEngine_QueryAPI.md

Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>

* Update docs/IE_DG/Int8Inference.md

Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>

* Update Custom_Layers_Guide.md

* Changes according to review  comments

* doc scripts fixes

* Update docs/IE_DG/Int8Inference.md

Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>

* Update Int8Inference.md

* update xfail

* clang format

* updated xfail

Co-authored-by: Trawinski, Dariusz <dariusz.trawinski@intel.com>
Co-authored-by: Nikolay Tyukaev <nikolay.tyukaev@intel.com>
Co-authored-by: kblaszczak-intel <karol.blaszczak@intel.com>
Co-authored-by: Yury Gorbachev <yury.gorbachev@intel.com>
Co-authored-by: Helena Kloosterman <helena.kloosterman@intel.com>
2021-12-21 20:26:37 +03:00
Andrew Kwangwoong Park
6f437fc1bd [GPU] Exclude peak memory usage from the result of memory statistics (#9332)
Signed-off-by: Andrew Kwangwoong Park <andrew.kwangwoong.park@intel.com>
2021-12-21 18:03:38 +03:00
Gorokhov Dmitriy
2d73eb6157 [CPU] Migration on oneDNN v2.4.4 (#9326)
Co-authored-by: Anton Voronov <anton.voronov@intel.com>
2021-12-21 14:45:44 +03:00
tgubanova-lohika
f9ece370f3 [GPU] Implement RandomUniform operation (#9043)
* [GPU] Implement Random Uniform kernel

* [GPU] Add random uniform kernel selector

* [GPU] Add random uniform primitive

* [GPU] Add random uniform instantiation helpers

* [GPU] Implement unit test for the random uniform operation

* [GPU] Add random uniform operation builder in cldnn_engine plugin

* [GPU] Add single layer tests for the random uniform operation

* [GPU] Add conversion for Random Uniform output type

* [GPU] Add RTTI macroses to RandomUniform operation
2021-12-21 13:24:07 +03:00
Aleksandr Korolev
4b0a2c9127 [VPU] Coverity fixes (#9214)
Some coverity fixes
2021-12-21 12:58:05 +03:00
Maxim Shevtsov
3032f20e03 brushed the #streams logic when #requests is limited (#9184) 2021-12-21 12:07:59 +03:00
Ivan Novoselov
fa9434494f [Snippets][CPU] Eltwise subgraphs code generation (#7936) 2021-12-20 22:45:01 +03:00
song, bell
1901f33bd3 auto dynamic shape (#8731)
* initial commit

Signed-off-by: fishbell <bell.song@intel.com>

* overwrite GetBlob/SetBlob to support dynamic shape when needed

Signed-off-by: fishbell <bell.song@intel.com>

* extend workers

Signed-off-by: fishbell <bell.song@intel.com>

* clean up code

Signed-off-by: fishbell <bell.song@intel.com>

* fix lock and add case for CPU,GPU

Signed-off-by: fishbell <bell.song@intel.com>

* fix mem leak

Signed-off-by: fishbell <bell.song@intel.com>

* clean up allocated worker

Signed-off-by: fishbell <bell.song@intel.com>

* fix test case failure

Signed-off-by: fishbell <bell.song@intel.com>

* enable case for chaining

Signed-off-by: fishbell <bell.song@intel.com>

* fix arm compile issue

Signed-off-by: fishbell <bell.song@intel.com>

* clang format, fix potential threading issue

Signed-off-by: fishbell <bell.song@intel.com>

* clean up idle request queue if infer request destroyed

Signed-off-by: fishbell <bell.song@intel.com>

* keep multi legacy behavior

Signed-off-by: fishbell <bell.song@intel.com>

* elegant way to handle recycling

Signed-off-by: fishbell <bell.song@intel.com>

* avoid the exception with set output blob in auto dynamic case

Signed-off-by: fishbell <bell.song@intel.com>

* remove increase workers logic

Signed-off-by: fishbell <bell.song@intel.com>

* formatting

Signed-off-by: fishbell <bell.song@intel.com>

* remove the irrelevant code to dynamic shape

Signed-off-by: fishbell <bell.song@intel.com>

* filter target device to CPU in the case of dynamic shape

Signed-off-by: fishbell <bell.song@intel.com>

* fix logic in selectdevice, add case for GPU

Signed-off-by: fishbell <bell.song@intel.com>

* add case for AUTO:GPU

Signed-off-by: fishbell <bell.song@intel.com>

* fix build error

Signed-off-by: fishbell <bell.song@intel.com>

* add comment for skipped cases

Signed-off-by: fishbell <bell.song@intel.com>

* fix the case failure caused by naming conflict

Signed-off-by: fishbell <bell.song@intel.com>

* move file to new folder

Signed-off-by: fishbell <bell.song@intel.com>

* change to 2.0 API in plugin

Signed-off-by: fishbell <bell.song@intel.com>

* refactor mkldnn code change

Signed-off-by: fishbell <bell.song@intel.com>

* formatting

Signed-off-by: fishbell <bell.song@intel.com>
2021-12-20 18:06:50 +03:00
Vladislav Golubev
ca7600edde [dynamism][Transformations] Sequences conversions: dynamic shapes support (#7766)
* [Transformations] Sequences conversion: refactoring & dynamic shapes support

* tests moved
2021-12-20 12:49:39 +03:00
Ilya Churaev
dbf8fcf556 Create ov_runtime library (#9199)
* Create ov_runtime library

* Added dev target

* Fixed SOURCE_DIR property for interface lib

* Fixed missed include path

* Fixed SOURCES for plugins and frontends

* Fixed Windows export

* Try to fix preproc build

* Fixed MKLDNNPlugin

* Fix C API

* Try to fix public CI

* Fixed comments

* Use TARGET_NAME variable

* Fixed static build

* Fixed target name in comfig

* Revert "Fixed target name in comfig"

This reverts commit f61c1e0c09.

* Try to remove WA with include path

* Try to fix TBB dependency for old CMake

* Revert install to lib for legacy tests

* Try to fix public tests

* Add WA for v7 reader

* Fixed dev export name

Co-authored-by: Ilya Lavrenov <ilya.lavrenov@intel.com>
2021-12-20 11:59:08 +03:00
Edward Shogulin
37ad512d98 [CPU] Optimize quantization scheme for SPR/ATS (#7549)
* [CPU] Optimize quantization scheme for SPR/ATS

* [CPU] [LPT] plugin tests

* [GPU] [LPT] plugin tests

* [CPU] limitation was removed

* [CPU] optimization FP32 old way support

* [LPT] comment fix

* [LPT] Multiply plugin test improvement

* [LPT] Multiply support

* [LPT] GPU tests fix

* [LPT] test quick fix

* [LPT] new ppi fix

* look like spent time for tests refactoring
2021-12-20 11:19:40 +03:00
Egor Duplensky
abee3ea4d4 [CPU] Refactoring. Avoid using align arg when appending post ops (#9225)
Always align legacy scale shift post ops
2021-12-20 10:23:32 +03:00
Jade Cho
a4518ae595 [GPU] Add DumpLayersLimitBatch config to GPU debug utils. (#9196)
* Add DumpLayersLimitBatch config to GPU debug utils.

+ Support OV_GPU_Help config
+ Only run first inference if OV_GPU_DumpLayersPath is set.
+ Fix dump graph bug.

* Apply some comments

* Remove unnecessary code.
2021-12-20 14:19:53 +09:00
Egor Duplensky
b394441c79 [CPU] Fix sse4 uni_ methods (Normalize and Input nodes) (#9268) 2021-12-17 15:13:16 +03:00
Sergey Shlyapnikov
71f3e2da10 [GPU] Fix fused ops reshaping for OneDNN FC (#9217) 2021-12-17 13:26:59 +03:00
Yaroslav Torzuk
3f5fca80db [GPU] Slice operation (#8267) 2021-12-17 12:41:05 +03:00
Aleksandr Korolev
4c4d006c5a [VPU] restore network P/R info in Import method (#8844)
In the new API, we need to recover information about function inputs/outputs.
2021-12-17 12:14:08 +03:00
Maxim Andronov
04dc16f861 [CPU] General fixes for dynamic shapes. Part 2 (#8871) 2021-12-17 11:12:07 +03:00
Min, Byungil
94c3731b50 [GPU] support multiple sum post-op of onednn kernel (#9221)
+ Applied WA for multiple sum post-ops of nednn kernel

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2021-12-17 10:43:21 +03:00
Sergey Lyubimtsev
b41cb57773 Fix build issue for openvino wheel package on Windows (#9265)
* Fix build issue for openvino wheel package on Windows

* fix env
2021-12-17 10:42:21 +03:00
Yury Gaydaychuk
58f0c75b91 [CPU] RoiAlign: case of integer tensor handled (#8069) 2021-12-16 22:02:09 +03:00
Vladislav Volkov
6ddc47a7ef [CPU] Convert precisions on inputs/outputs (#8805) 2021-12-16 16:58:19 +03:00
Sergey Shlyapnikov
2514c0ef38 [GPU] Add gemm_tiled_opt i8/u8 output support (#9202) 2021-12-16 15:20:28 +03:00
Alexander Zhogov
d10e8005c0 Revert "Fix build issue for openvino wheel package on Windows (#9231)" (#9260)
This reverts commit 460a6634fd.
2021-12-16 14:33:56 +03:00
Sergey Lyubimtsev
460a6634fd Fix build issue for openvino wheel package on Windows (#9231)
* fix build issue for openvino wheel package on Windows

* revert ngraph_libs
2021-12-16 13:01:03 +03:00
Roman Lyamin
2068d5838b [GPU] Add I420toRGB/I420toBGR operations (#9204) 2021-12-16 10:27:27 +03:00