Commit Graph

11265 Commits

Author SHA1 Message Date
Edward Shogulin
655c21adf1
[CPU] Quantized MHA extension for SmoothQuant (#17906) 2023-06-07 14:31:06 +00:00
Anton Voronov
2547301fa7
[CPU] gemm convolution: fixed bias offset (#17357) 2023-06-07 17:15:01 +04:00
Georgy Krivoruchko
ee659c1ce8
[TF FE] Workaround for Broadcast/Concat issue with empty tensors (#17864)
* Added transformation for Concat

* Added test

* CI fix

* Fixed behavior of the "empty tensor list" test
2023-06-07 14:17:20 +04:00
Tomasz Otoka
323dbec93c
Adds configuration file for cspell (#17355)
the file for future use with cspell
2023-06-07 12:16:28 +02:00
Tomasz Otoka
d8a78137a7
Adds Python wheel requirements info to docs (#17922) 2023-06-07 10:13:15 +00:00
Ilya Lavrenov
6df054b073
Added cmake_policy call to allow IN_LIST in if() (#17916) 2023-06-07 14:12:54 +04:00
Pawel Raasz
f023f5d672
Add interpolate from all opsets to cpu shape infer (#17875) 2023-06-07 11:28:45 +02:00
hyunback kim
13028397b7
Optimize permute gemm onednn (#17621)
* [GPU] Optimized out permute in permute-gemm(onednn) pattern.

Permute can be optimized out when permute's in and out are compatible and onednn gemm.

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-06-07 16:20:59 +09:00
Chen Xu
3a1326fb58
[CPU] Reduce node improve parallelism (#17615) 2023-06-07 05:11:18 +00:00
Wang Wangwang
84bd391369
[AUTO] Support setting log_level from config (#17569)
* [AUTO] Support setting log_level from config

* Add test cases

---------

Co-authored-by: Chen Peter <peter.chen@intel.com>
2023-06-07 10:31:27 +08:00
Wanglei Shen
4ca8d40e43
update scheduling core type value (#17771)
* update scheduling core type value

* update for comments

* update for comments

* fix code style issue
2023-06-06 22:14:40 +08:00
Alexandra Sidorova
2ec9fe915c
[Snippets] Added support of 3D MHA (#17817) 2023-06-06 16:55:01 +04:00
Nikita Malinin
29f06692d6
ENABLE_MMAP property pos (#17896) 2023-06-06 12:19:53 +00:00
Tatiana Savina
94597d8391
[DOCS] Port release doc fixes to master (#17901)
* [DOCS] Change downloads directory link  (#17846)

* installation link

* fix path

* change notebooks links (#17857)

* fix apt and yum links (#17877)

* [DOCS] Fix list and links to POT (#17887)

* change link to POT

* change header label

* fix typo
2023-06-06 12:48:18 +02:00
Ilya Lavrenov
13e9752012
Use tbb from vcpkg (#17891) 2023-06-06 09:12:16 +04:00
Ilya Lavrenov
ace4fb6ecd
Fixed compilation with clang (#17890) 2023-06-06 00:23:07 +04:00
Zlobin Vladimir
86e46e5e35
Work around deprecation error (#17888)
For some reason my MSVC gives the following error:
C:\Program Files\Microsoft Visual Studio\2022\Professional\VC\Tools\MSVC\14.35.32215\include\utility(176,5): error C4996: 'ngraph::SlicePlan::SlicePlan': T
he nGraph API is deprecated and will be removed in the 2024.0 release. For instructions on transitioning to the new API, please refer to https://docs.openv
ino.ai/latest/openvino_2_0_transition_guide.html [C:\Users\vzlobin\r\openvino\build\src\common\transformations\inference_engine_transformations_obj.vcxproj
]
C:\Program Files\Microsoft Visual Studio\2022\Professional\VC\Tools\MSVC\14.35.32215\include\xmemory(680,47): message : see reference to function 'std::pai
r<std::shared_ptr<ov::op::v1::StridedSlice>,ngraph::SlicePlan>::pair(std::pair<std::shared_ptr<ov::op::v1::StridedSlice>,ngraph::SlicePlan> &&)' [C:\Users\
vzlobin\r\openvino\build\src\common\transformations\inference_engine_transformations_obj.vcxproj]
C:\Program Files\Microsoft Visual Studio\2022\Professional\VC\Tools\MSVC\14.35.32215\include\utility(175,5): error C4996: 'ngraph::SlicePlan::SlicePlan': T
he nGraph API is deprecated and will be removed in the 2024.0 release. For instructions on transitioning to the new API, please refer to https://docs.openv
ino.ai/latest/openvino_2_0_transition_guide.html [C:\Users\vzlobin\r\openvino\build\src\common\transformations\inference_engine_transformations_obj.vcxproj
]
C:\Program Files\Microsoft Visual Studio\2022\Professional\VC\Tools\MSVC\14.35.32215\include\xmemory(680,47): message : see reference to function 'std::pai
r<std::shared_ptr<ov::op::v1::StridedSlice>,ngraph::SlicePlan>::pair(const std::pair<std::shared_ptr<ov::op::v1::StridedSlice>,ngraph::SlicePlan> &)' [C:\U
sers\vzlobin\r\openvino\build\src\common\transformations\inference_engine_transformations_obj.vcxproj]
2023-06-05 20:31:05 +04:00
Ilya Lavrenov
2799fe9855
Install all python build artifacts to a single folder (#17883) 2023-06-05 18:12:17 +02:00
Karol Blaszczak
6167619e28
DOCS post-release adjustments (#17874) 2023-06-05 16:39:19 +02:00
Chen Xu
1dafb405fd
[CPU] Improve performance in 5D scenario of Reduce node (#17828) 2023-06-05 14:52:18 +04:00
Yury Gaydaychuk
43bf90f90c
[CPU] Align jit and ref offsets handling in DeformableConvolution implementation (#10686) 2023-06-05 14:17:54 +04:00
Roman Zubarev
8e97010595
remove quantization config for vpu (#17507)
* remove quantization config for vpu

* remove VPU config testcases from test unify scales

---------

Co-authored-by: shokinal <alexander.shokin@intel.com>
Co-authored-by: Alexander Kozlov <alexander.kozlov@intel.com>
2023-06-05 13:14:48 +03:00
Piotr Krzemiński
0d9109acf3
[PT FE] Add aten::_shape_as_tensor (#17804)
* [PT FE] Add aten::_shape_as_tensor impl

* Update shape_as_tensor.cpp

* [PT FE] Fix headers, add explicit type, comment out shape detection

* [PT FE] Reverse example comments
2023-06-05 10:59:12 +02:00
Aleksandr Voron
0944295d61
[CPU] Enable SLT Eltwise tests on ARM (#17713) 2023-06-05 12:58:12 +04:00
Piotr Krzemiński
3d8a620ac3
[PT FE] Add aten::_native_multi_head_attention (#17550)
* [PT FE] Add implementation of MHA

* [PT FE] Add tests, add scaled dot product attention

* [PT FE] Fix missing transpose for Q,K,V & output Attention

* [PT FE] Formatting errors

* [PT FE] Fix testing class with nn.Linear

* [PT FE] Fix incorrect key franspose in dot product attention computation

* [PT FE] Fix incorrect matmul due to lack of transpose

* [PT FE] Enable support for all boolean masks

* [PT FE] Fix returned weights

* [PT FE] Remove debugging artifacts

* [PT FE] Remove unused nodes, optimize transpose nodes' usage, add comments to floating masks

* [PT FE] Further reduce node usage, return None instead of 0 for return_weights=false

* [PT FE] Allow for dynamic num_num_head, embed_dim

* [PT FE] Improve error comment, remove unnecessary Unsqueeze

* [PT FE] Clang format

* Update tests/layer_tests/pytorch_tests/test_native_multi_head_attention.py

Co-authored-by: Maxim Vafin <maxim.vafin@intel.com>

* [PT FE] Add masks comments, improve mask broadcasting

---------

Co-authored-by: Maxim Vafin <maxim.vafin@intel.com>
2023-06-05 10:55:03 +02:00
Yury Gaydaychuk
c0fb831c6e
[CPU][Commit slider] Rmtree fixed (#17858) 2023-06-05 12:52:58 +04:00
Ilya Churaev
36625404eb
[GPU] Fix GPU remote context name initialization (#17850) 2023-06-05 12:00:04 +04:00
Sergey Shlyapnikov
db8d23231a
[GPU] Change priority of CPU implementations (#17829) 2023-06-05 11:21:26 +04:00
Sebastian Golebiewski
a9ddc2b553
update-diagram (#17871) 2023-06-05 08:16:29 +02:00
Vladimir Paramuzov
1ce447674e
[GPU] Better device input memory reuse (#17853) 2023-06-05 09:30:22 +04:00
Asthestarsfalll
14da93c155
【PaddlePaddle Hackathon 4】add paddle set_value op (#15888)
* add set_value op

* Support for tensor input

* fix shape error

* refactor for dynamic shape

* update process of target_value_shape and add comments

* support arbitrary steps

* fix

* fix ends_node

* fix and add test cases

* fix error when slice operation return maximum number in int32

* remove redundant function call

* update for minus step

* add constraints for minus inputs

---------

Co-authored-by: mei, yang <yang.mei@intel.com>
2023-06-05 05:06:05 +02:00
Wanglei Shen
a0c8fdbd86
enable CPU map and multiple streams for ARM Linux (#17450)
* enable CPU map for ARM Linux based on freqency information

* fix code style issue

* fix code style issue

* remove 'streams = 1' WA for ARM linux

* update for typo and comments

* update for comments

* keep  WA

* keep WA of streams = 1 for ARM Linux

* update num_streams WA for ARM Linux in test case

* update for comments

* update for comments

* update for comments

* update for comments

* update for merge conflict

* update and add test case for MTL
2023-06-03 17:07:02 +08:00
Kelvin Choi
ec0daa5b10
[GPU] Apply m_pythondiv for fusing of eltwise div (#17590) 2023-06-02 17:29:02 -07:00
Alina Kladieva
acb4b1d37b
Revert "Stale PRs/Issues action limit extention (#17491)" (#17868)
This reverts commit c351335661.
2023-06-02 22:50:57 +00:00
Yaroslav Torziuk
eb588f0336
Add subgroup block reading in softmax_gpu_items_class_optimized.cl (#16223) 2023-06-02 12:59:55 -07:00
Taylor Yeonbok Lee
f670dc5a0d
[GPU] Enable runtime buffer fusing for dynamic shape (#17668)
* Initial impl for runtime buffer fusing
Passing unittest with static kernel

* pass unittest with dynamic impl

* Refactor allocate_output

* Separate header of buffer fusing

* Refactored buffer fusing :: matcher/optimize

* More cleanup

* Fix crash in dolly

* Reset can_be_optimized of primitive_inst when it is not

* Fix empty tensor : Primitive with empty data should be skipped

* Fix issue in dynamic padding : Static kernel should not contain dynamic padding dims
Fix missing reset of update_shape_done_by_other flag

* Not to add cache with emtpy kernel for optimized out inst

* Fix corner case error in buffer fusing
- Shapes of some preds may not be changed, but still needed to do update_impl because 1) paddings are changed 2) output memory should be updated
- optimizable impl should not be added to the cache

* Allowing reorder & permute_ref to be optimized concat predecessor

* Some more fixes :
runtime buffer fusing is available only when all preds/concat are dynamic
runtime buffer fusing is to be executed only if the node is dynamic

* Fix allocate_output parameter called by get_estimated_device_mem_usage according to the new change

* Fixed error in cascaded concatt

* Need to reinterprete even though the size is same
2023-06-02 12:39:28 -07:00
Marcus van Houdt
c3a54b0a6e
Add missing != operator, fixing compilation issue with Visual Studio 17.6 together with /std:c++latest (#17860) 2023-06-02 22:05:57 +04:00
Sofya Balandina
b430802d0c
[conformance] Add rel pass rate to highlight table (#17783) 2023-06-02 21:44:54 +04:00
Sergey Shlyapnikov
5afbd4cf92
[GPU] Remove clFinish call from USM memory lock function (#17830) 2023-06-02 16:17:05 +04:00
Edward Shogulin
43d67b0a32
[LPT] StridedSlice dequantization improvement (#17563)
* [LPT] StridedSlice dequantization improvement

* review comments: refactoring & simplification
2023-06-02 08:47:36 +01:00
Chen Peter
031f2cc7d1
Requirements for the HW plugin to integrate with AUTO (#17053)
* Requirements for the HW plugin to integrate with AUTO

Signed-off-by: Peter Chen <peter.chen@intel.com>

* Update property requirements and wording

1. Added purpose for each reqired property
2. Removed autobatching properties
3. Updated wording

Signed-off-by: Peter Chen <peter.chen@intel.com>

* Add one BA test and update purpose for model_name

Signed-off-by: Peter Chen <peter.chen@intel.com>

* Add request to ov::compilation_num_threads

Signed-off-by: Peter Chen <peter.chen@intel.com>

* Add link to intgration with AUTO

Signed-off-by: Peter Chen <peter.chen@intel.com>

* Wording with API 2.0

Co-authored-by: Ilya Churaev <ilyachur@gmail.com>

* Try to fix the link

* Remove ":doc:"

* Add postfix "__" for external link

* Apply suggestions from code review

Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>

* bash command and multiple devices description update

Signed-off-by: Peter Chen <peter.chen@intel.com>

---------

Signed-off-by: Peter Chen <peter.chen@intel.com>
Co-authored-by: Ilya Churaev <ilyachur@gmail.com>
Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
2023-06-02 15:04:33 +08:00
bstankix
e1ed6599f0
Add default platform type selection to ov graphs (#17844) 2023-06-01 15:28:37 +02:00
Karol Blaszczak
e0f241f382
DOCS homepage update (#17842) 2023-06-01 15:16:10 +02:00
dependabot[bot]
266764171f
Update networkx requirement from <=2.8.8 to <3.1.0 in /tools/mo (#15293) 2023-06-01 14:11:05 +02:00
Piotr Krzemiński
482c030408
[PT FE] Add aten::LogSoftmax (#17629)
* [PT FE] Add aten::LogSoftmax implementation & tests

* Update log_softmax.cpp

* Update src/frontends/pytorch/src/op/log_softmax.cpp

Co-authored-by: Maxim Vafin <maxim.vafin@intel.com>

* [PT FE] Add recommended comment, replace get_input_tensor with new implementation

* [PT FE] Align to f32 if no dtype provided

* [PT FE] Revert type align

---------

Co-authored-by: Maxim Vafin <maxim.vafin@intel.com>
2023-06-01 12:04:27 +00:00
Ilya Lavrenov
6d42a15e07
Use 2023.0.0 debian packages on CI (#17840) 2023-06-01 15:23:08 +04:00
Maciej Kwapulinski
4b1d0fbc37
[GNA] convertFunctionToICNNNetwork operation performance improovement (#17685)
* make CNNLayerCreator be persistent accross single convertFunctionToICNNNetwork operation

* [GNA] RR comments applied

* [GNA] RR comments applied
2023-06-01 13:21:54 +02:00
Katarzyna Mitrus
edf089bf22
[Pad-12] Introduce PadBase and Pad-12 operation (#17796)
* Introduce PadBase

* Update ov scope name

* Introduce Pad-12

* Common type_prop Pad tests

* Init Pad-12 ref tests

* Add Pad reference tests

* attr and op  check tests

* Move eval and clone inputs from PadBase

* Init opset12

* Headers clean up

* Update shape_inference map for CPU

* Update Pad evaluates to use ov::TensorVetor

* Update shape infer map with Pads

* Fix namespace

* Update op check test

* Add common Pad shape_inference tests

* Reuse PadBase shape_infer
2023-06-01 13:21:43 +02:00
Ilya Lavrenov
7697bacfca
Use OpenCL from CCI (#17839) 2023-06-01 15:04:32 +04:00
Yury Gaydaychuk
2467f655b7
[CPU][Tools] wa for tmp directory handling in linux (#17824) 2023-06-01 14:20:30 +04:00