Commit Graph

769 Commits

Author SHA1 Message Date
Xuejun Zhai
8509d0dd82 [Deprecated API] remove version (#16426)
* [Remove version] Remove version from py openvino

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Modify caused by remove version

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Fix clang format issue

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Revert "Fix clang format issue"

This reverts commit 132787286f.

* Fix CI format issue

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Fix CI format issue

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Fix merge conflict error

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

---------

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>
2023-03-22 16:09:14 +01:00
Sebastian Golebiewski
cbb25e9483 [DOCS] Proofreading developer documentation moved from wiki. (#15886)
Minor stylistic and grammar corrections. Fixing links

* Apply suggestions from code review

Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
2023-03-22 09:08:31 +01:00
hyunback kim
c14e6ef48e [GPU] Use 4dim directly for onednn in gemm (#16182)
* [GPU] Use 4-dim directly for onednn in gemm
   We were collapsing n-dim into 3d for onednn gemm, But it is not necessary, up to 4d.

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-22 17:08:10 +09:00
Andrew Kwangwoong Park
04a2c4ce61 [GPU] Add shape agnostic optimized FullyConnectedIMAD kernel (#16417)
* [GPU] Added shape agnostic kernel for fully_connected_gpu_imad

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add fully_connected_gpu_imad shape agnostic TCs for ov_gpu_unit_tests

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Apply comments

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-03-22 08:38:28 +01:00
Min, Byungil
a71c83d366 [GPU] Resolve eltwise kernel build failure (#16458)
Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-03-22 15:15:02 +09:00
Haiqi Pan
5e98696464 Fix Windows build warnings in template and core tests (#15967)
* fix C4305

* 1.0f

* Element

* fix c4244

* fix truncation from double to float in grn.cpp

* Revert "fix truncation from double to float in grn.cpp"

This reverts commit 5263b37cb2.

* fix grn.cpp

* add 4305

* fix low

* add TearDown

* revert softmax.cpp

* pragram

* fix conflicts

* fix conflicts

* size_t -> ov::label_t

* WIN32

---------

Co-authored-by: Ilya Lavrenov <ilya.lavrenov@intel.com>
2023-03-22 07:29:35 +04:00
hyunback kim
05e54e9f3d [GPU] Update the latest onedNN3.1 (#16381)
- Fix group conv regression issue

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-21 18:56:41 +09:00
Min, Byungil
5d6cd626bc Fix unit test on dGPU (#16295)
* Resolve failed cases and queue-type issue
+ Resolved out_of_order queue-type issue
+ Added get_test_default_config for setting default config of onednn
+ Cleared failed case

Signed-off-by: Min, Byungil <byungil.min@intel.com>
Co-authored-by: tuxedcat <tuxedcat@gmail.com>
2023-03-21 18:55:06 +09:00
Ilya Churaev
ec0a1e58d1 Fixed some leftovers for 2.0 dev api (#16421)
* Fixed some leftovers for 2.0 dev api

* Fixed build issue
2023-03-21 09:34:37 +00:00
Vladimir Paramuzov
98237b06b5 [GPU] Update memory_statistics property impl (#16399) 2023-03-21 08:52:52 +04:00
Pavel Durandin
4a4f06ba3b GPU documentation update (#16393)
* GPU documentation update

* GPU documentation update
2023-03-20 14:27:13 +04:00
Vladimir Paramuzov
a46fc47e6a [GPU] Enable tile with dynamic input (#16364) 2023-03-20 12:49:35 +04:00
Min, Byungil
bc15596c9e Remove redundant reorder (#15661)
+ Reorder 1d data
+ Reorder which only changes format

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-03-20 14:58:09 +09:00
hyunback kim
4f49d0e07e [GPU] enable dumpgraph in unit-test (#15388)
Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-20 11:51:51 +09:00
Vladimir Paramuzov
bb59672639 [GPU] Fixed shape agnostic scatter nd update kernel (#16319) 2023-03-17 09:57:25 +04:00
Wilson Seok
c5ccb3e954 add condition for activation ceil so it works when data type is fp32 or fp16 only (#16334) 2023-03-17 11:46:44 +09:00
hyunback kim
8d1139b61a Fix unet3d mlperf dump (#16253)
* Enable dump in unet3d_mlperf

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-17 10:51:40 +09:00
Andrew Kwangwoong Park
e4500c7d61 [GPU] Fixes for dynamic model in dGPU (#16298)
Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-03-16 11:09:55 -07:00
Tomasz Dołbniak
6762fe692d Interpolate-11 spec + core op (#16162) 2023-03-16 14:37:57 +01:00
Wilson Seok
461cc2aee8 change activation position in reorder_data_bfyx_to_blocked_format kernel (#16307) 2023-03-16 17:48:23 +09:00
Andrei Gorbachev
2f3ae4518e [GPU] Fix warnings (#16196)
* fix 1

* fix 2-10

* fixed code style

* fixed win plugin

* fixed linux plugin

* fixed a part of tests

* fixed test fot linux

* fixed pooling_gpu_test fot linux

* fixed pooling_gpu_test fot linux

* fix after review and enable wd4267 in makefile

* fix after review

* errors of unit test are fixed
2023-03-16 09:29:16 +04:00
Vladimir Paramuzov
28d3e1087e [GPU] Fix strided slice kernel with begin/end/stride as inputs (#16302) 2023-03-15 16:25:45 +04:00
Vladimir Paramuzov
e44fd03d2a [GPU] Shape agnostic concat kernel + refactoring (#16170) 2023-03-15 09:47:31 +04:00
Eddy Kim
e348481849 [GPU] Transformed IR serialization for dynamic models (#16169)
* IR serialization for dynamic models

* added ShapeOf1To3 transformation pass

* fixed input output type mismatch

* removed unnecessary codes

* moved ConvertShapeOf1To3 from common to GPU plugin

* updated copyright year

* fixed build errors
2023-03-14 11:03:02 -07:00
Mateusz Tabaka
8477bc8897 Reduce the number of validate and infer types in ConvertPrecision (#15277)
* Reduce the number of validate and infer types in ConvertPrecision

Currently, ConvertPrecision pass frequently runs validate and infer types.
This is due to the fact that it iterates over every precision pair, then over
the whole model followed by validate and infer types.
The proposed solution is to iterate over the model: for each node iterate
over precisions array, update the node if required followed by validate and
infer types.

Ticket: 81311

* use map

* clang format

* move enum hasher

* fix gpu

* revalidate

* reinvalidate if node has changed

* remove validate for input prec changes

* fix gpu

* review

* find

* fix pytorch case

* revalidate

---------

Co-authored-by: Michal Lukaszewski <michal.lukaszewski@intel.com>
2023-03-14 21:45:24 +04:00
hyunback kim
164db3def9 [GPU] Fix twin tranformer functional regression. (#16111)
* [GPU] Fix twin tranformer functional regression.

gemm/FC select_preferred_format select simple format depends on out rank size.

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-14 17:34:41 +09:00
Taylor Yeonbok Lee
3a96e06d4c Minor fixes (#16275)
- Fix are_data_types_suitable_for_onednn not to invalidate output layout
- Fix seg fault of printing resample node info
2023-03-14 08:09:54 +00:00
Taylor Yeonbok Lee
f0f1c47063 Fix concat to use ngraph shape infer (#16226)
Fix crop to return shape of original rank
2023-03-13 20:25:23 +00:00
Ilya Lavrenov
f080a0d9cf Added NCC style for frontends sources (#16200)
* Ability to provide several source dirs for ncc-style checks

* Fixed include headers; added NCC to TF common

* Fixed NCC for frontends

* Fixed NCC for frontends

* Extra fixes

* Fixest push --f

* Clang-format

* Apply comments

* Add an option to specify required clang-format version

* Update src/frontends/tensorflow/src/decoder_proto.cpp

* Update src/frontends/tensorflow/src/decoder_proto.cpp
2023-03-13 14:54:00 +00:00
Paul Youngsoo Ahn
e8b108ac6b [GPU] Change lws to avoid synchronization issue in nonzero_count (#16116)
* [GPU] Change lws to avoid synchronization issue in nonzero_count (#16116)

* [GPU] Add unit test (#16116)

* [GPU] update count_nonzero_ref kernel(#16116)
- Support the case total data size exceed max work group size
- Add dynamic shape test case

* [GPU] Change input indexing calculation and add random input generator in unit test (#16116)

* [GPU] update random generation input funciton in nonzero_count (#16116)

* [GPU] update unit test (#16116)

* [GPU] cldnn unit test: update random generation function for other test failure (fusings_gpu/conv_fp32_multi_eltwise_quantization.basic/0) (#16116)
2023-03-12 23:32:20 -07:00
Ilya Churaev
75314c2c53 Rename OPENVINO_UNREACHABLE to OPENVINO_THROW (#16201)
* Changed some exceptions to OPENVINO_THROW

* Changed samples throw exception

* Fixed some comments

* Remove OPENVINO_UNREACHABLE
2023-03-10 20:23:13 +04:00
Maciej Smyk
5e406a80d3 [DOCS] OpenVINO Wiki links update - master (#16219)
* wiki links
2023-03-10 16:16:14 +01:00
Roman Lyamin
b8e1dea345 [GPU] Fix binary_convolution non-constant weights (#15898)
* [GPU] Fix binary_convolution non-constant weights

* [GPU] Remove unused checks related to allowInputReordering
2023-03-10 14:36:12 +04:00
Ilya Churaev
45bdbf7486 Changed throw ov::Exception to macro (#16150)
* Changed throw ov::Exception to macro

* Fixed code style

* Revert myriad headers

* CPPlint fixes

* Fixed typo
2023-03-10 11:14:50 +04:00
Mykhailo Hnap
d5e98cbdce [GPU] IsFinite, IsInf, IsNaN operations (#15979)
* [GPU] Enabled ComparisonLayerTest in single layer tests.

It seems that before, these tests were disabled cause of some failures. Now I cannot see any errors, so I just enabled all of them.

* [GPU] Run clang format for comparison single layer tests.

* [GPU] Added handling of f16 type to IsInfLayerTest.

* [GPU] Added single-layer tests for IsFinite and IsNaN operations.

* [GPU] Added single-layer test for IsInf operation.

* [GPU] Implemented IsFinite, IsInf, and IsNaN operations as activation functions.

But notice that currently, the activation kernel support only the same output data type as the input data type. So an additional reorder would be needed to convert to the correct output data type for these ops. Also worth noting is that activation functions are fused in reorder kernel. But for now, it's not working for these ops because in reorder activation call, there is a hard conversion of input data to output data type before activation. I don't know why it's added there, but it breaks fusion. So need to fix this activation fusion or disable this fusion for these ops.

* Revert "[GPU] Implemented IsFinite, IsInf, and IsNaN operations as activation functions."

This reverts commit 3f9ffe617ecddce6dbbcdeab9584a7ddeb6d1845.

* [GPU] Implemented IsFinite, IsInf, and IsNaN operations as eltwise op.

* [GPU] Changed CLDNN_ERROR_MESSAGE to OPENVINO_ASSERT in check_inputs_count method.
2023-03-09 16:10:48 -08:00
Andrew Kwangwoong Park
3ec386a741 [GPU] Minor fixes for dynamic BERT models (#16158)
* [GPU] Minor fix for dynamic bert-base-uncased-qqp

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix to check full tensor only for static shape during creating onednn gemm

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-03-09 14:48:08 -08:00
Taylor Yeonbok Lee
dff7f2451b Revert PR15386's change (#16172)
- Previously, PR15386 changed allocation of memory of primitives which are to be used as shape infer dep to host memory, for better shape infer perf.
- However this causes cache coherence issue in dGPU.
- Reverting this change so that the memory will be allocated to devicet
2023-03-09 22:44:32 +00:00
Sungeun Kim
0365ebf5ad disable test case: fusings_gpu/lrn_fp16_eltwise_activation.basic/7 (#16149) 2023-03-09 08:38:33 +00:00
Jade Cho
aaeace9740 [GPU] Fix stable diffusion failure (#16052)
* [dGPU] Enable stable diffusion

+ Prevent to fuse swish into oneDNN reorder.
+ Makes concat explicitly if batch size is greater than 1 and the siblings are oneDNN impl.
2023-03-09 14:35:31 +09:00
Andrew Kwangwoong Park
b7ff3a1d64 [GPU] Added shape agnostic Pad kernel implementation (#16160)
Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-03-08 15:36:43 -08:00
Vladimir Paramuzov
75b48f2153 [GPU] Changed impls cache key type to avoid hash collisions (#16130) 2023-03-08 16:09:55 +04:00
hyunback kim
a9cbccd829 Broadcast for post ops enable enable onednngemm (#16074)
* [GPU] Add data broadcasting for OneDNN binary ops for Gemm primitive
* Based on https://github.com/openvinotoolkit/openvino/pull/15790 and enable onednn gemm from support multiple users and non constant input.

--------

Signed-off-by: hyunback <hyunback.kim@intel.com>
Co-authored-by: Sergey Shlyapnikov <sergey.shlyapnikov@intel.com>
2023-03-08 13:55:51 +09:00
Roman Lyamin
681faadce3 [GPU] Added shape agnostic kernels for GatherElements and Tile (#15798)
* [GPU] Added shape agnostic kernel for GatherElements

* [GPU] Added shape agnostic kernel for Tile
2023-03-08 08:34:24 +04:00
Vladimir Paramuzov
a1eb76ad06 [GPU] Move is_local_block_io_supported WA to kernel selector (#15235) 2023-03-07 15:12:08 +04:00
Min, Byungil
87b18a21c1 [GPU] Optimize eltwise kernel for blocked format (#15717)
* [GPU] Optimize eltwise kernel for blocked format

+ Optimize etlwise_blocked_opt
+ Replace deprecated kernels with eltwise_blocked_opt
+ Remove eltwise_b_fs_yx_fsv16, b_fs_yx_fsv4 kernels
+ Add test-cases in eltwise_gpu_test

Signed-off-by: byungilm <byungil.min@intel.com>
2023-03-07 14:21:09 +09:00
Vladimir Paramuzov
eff0bce7e3 [GPU] Move some op parameters from node to primitive class (#16070)
* [GPU] Move parameters of conv and quantize primitive from node to primitive

---------

Co-authored-by: Eddy Kim <eddy.kim@intel.com>
2023-03-07 08:56:00 +04:00
Andrew Kwangwoong Park
7123e8879e [GPU] Added shape agnostic optimized SoftMax kernel (#15834)
* [GPU] Added shape agnostic optimized SoftMax kernel

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Update SoftmaxKernelBaseBF::Validate policy for shape agnostic kernel

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add softmax_gpu_bf shape agnostic TC for ov_gpu_unit_tests

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix failed TCs for ie-tests-linux-ubuntu20-gpu

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Update to use stack array instead of global buffer

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Remove global buffer usage completely

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add #undef directive

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-03-06 09:10:29 -08:00
Andrew Kwangwoong Park
4ce35fd851 [GPU] Minor fixes for dynamic model (#16075)
Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-03-06 15:50:38 +04:00
Xiping Yan
8b66b35bf7 [CPU]Remove C4250 warning suppress, and fix the corresponding warning. (#15966) 2023-03-06 12:43:53 +04:00
Xuejun Zhai
9b97235902 Xuejun/remove api in ov any (#15667)
* [Remove APIs] remove ov::any api  &

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>

* [Remove APIs] remove ov::any api

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>

* [Remove APIs] remove interfaces in ov::any  Base* operator->() & const Base* operator->()

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>

* [Remove APIs] remove ov::any interfaces Base* get() & const Base* get()

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>

* [Remove APIs] remove ov::any interfaces call(const Any& any) & dynamic_pointer_cast(const ::ov::Any& any) & static_pointer_cast(const ::ov::Any& any)

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>

* [Remove APIs] fix code format issues in ov::any

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>

* [Remove APIs] fix review issue

Signed-off-by: xuejun <xuejun.zhai@intel.com>

* [Remove APIs] clear code

Signed-off-by: xuejun <xuejun.zhai@intel.com>

* [Remove APIs] fix review issue

Signed-off-by: xuejun <xuejun.zhai@intel.com>

* [Remove APIs] fix compiler issue

Signed-off-by: xuejun <xuejun.zhai@intel.com>

* [Remove APIs] fix compiler issue

Signed-off-by: xuejun <xuejun.zhai@intel.com>

* [Remove APIs] fix compiler issue

Signed-off-by: xuejun <xuejun.zhai@intel.com>

* Fix variant error

Signed-off-by: xuejun <xuejun.zhai@intel.com>

---------

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>
Signed-off-by: xuejun <xuejun.zhai@intel.com>
2023-03-06 10:24:08 +04:00