Commit Graph

805 Commits

Author SHA1 Message Date
Vladimir Paramuzov
b2a64e8c3a [GPU] Enable dynamic shapes support for gather elements (#16727) 2023-04-06 11:14:36 +04:00
Roman Lyamin
38c8a3d15b [GPU] Added custom canonicalize_shapes for Gather (#16733) 2023-04-06 10:50:57 +04:00
Oleg Pipikin
1c564226f3 Deprecate util functions in public api (#16716)
* Deprecate util functions in public api

* Add deprecation suppression for usage inside openvino

* Fix clang-format

* Fix1
2023-04-06 06:32:04 +04:00
Sungeun Kim
fef04e468a [GPU] add WA to avoid hang issue. (#16724) 2023-04-05 16:32:42 +04:00
Andrew Kwangwoong Park
44cfbea9ab [GPU] Fix synchronization issue from wrong stream in multi-stream mode on dGPU (#16671)
Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-04-05 16:29:47 +04:00
Vladimir Paramuzov
f5e199c494 [GPU] Don't reorder weights when can reinterpret (#16714)
* [GPU] Don't reorder weights when can reinterpret

* [GPU] Test fixes
2023-04-05 16:20:51 +04:00
Pavel Esir
f9bd2d2c1e [ie transformations] improve SoftMax fusion for better mixed precision inference (#16574)
* improve SoftMax fusion

* style and unit-test fix

* more precise SoftMax unit-tests

* rewritten SoftMaxFusion with single matcher

* fixes for align_mixed_fp32_fp16_types_test.cpp and mark_subgraph_to_keep_in_mixed_precision_test.cpp

* add include for pass/pattern/op/or.hpp

* get rank only when necessary

* style-fix

* add comment why SoftmaxFusion is called manually

* fix copy_runtime_info
2023-04-05 11:28:48 +04:00
Eddy Kim
90615cf26a [GPU] Fix OneDNN primitive attr serialization logic (#16654)
* fix onednn primitive attr serialization logic

* added an onednn fc fusing serialization test

* added gemm fusing serialization tests
2023-04-03 18:24:40 -07:00
Vladimir Paramuzov
f7d15e12c8 [GPU] Refactor dimensions jitter and max rank related code (#16603)
* [GPU] Refactor dimensions jitter and max rank related code
2023-04-03 13:34:06 +02:00
Yaroslav Torzuk
8491f15ba7 [GPU] Softmax for stable diffusion (#15863) 2023-04-03 10:21:02 +04:00
Sergey Shlyapnikov
bb93bfd90f [GPU] Add clDNN shape agnostic kernels usage as an initial impls for dGPU (#16018)
* [GPU] Add clDNN shape agnostic kernels usage as an initial impls for dGPU

* [GPU] Use layout as a key of weights cache, implement logic for weights cache capacity calculation based on available memory
2023-03-31 13:05:59 +04:00
Ilya Lavrenov
ee0bb79ed6 Fixed LTO build (#16629)
* Partially fixed LTO

* Fixed issues with cnpy LTO

* CPU

* Disabled failing GPU test
2023-03-31 11:34:42 +04:00
Ilya Lavrenov
e94f7b25c0 Fixed cmake dev warnings (#16655) 2023-03-30 21:01:41 +04:00
Ilya Lavrenov
ccf9c19f61 Deprecated UNDEFINED values for execution / performance hints (#16563)
* Deprecated UNDEFINED values for execution / performance hints

* Update src/tests/functional/plugin/gpu/shared_tests_instances/behavior/ov_plugin/properties_tests.cpp

* Fixes

* Fixes
2023-03-30 13:48:19 +04:00
Ilya Lavrenov
0250f62d11 Revert inference precision to be a hint (#16634) 2023-03-29 18:59:33 +04:00
Ilya Churaev
f3dcf93f96 Remove suppression Wno-delete-non-abstract-non-virtual-dtor (#16560)
* Remove suppression Wno-delete-non-abstract-non-virtual-dtor

* Fixed Allocator warning

* Suppress warning for GPU plugin

* Skip warning for GNA

* Fixed preprocessing

* Added virtual constructor for base plugin class

* Some fix for CPU

* Suppress for CPU

* Fixed any

* Fixed meta

* Disable warning for paddle

* Fixed Allocator tests

* Move suppress to paddle

* Fixed benchmark_app
2023-03-29 14:19:30 +04:00
Min, Byungil
7a95830d24 [GPU] Disable failed onednn tests (#16614)
* Resolved failed unit-tests for fully connected

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-03-29 15:53:04 +09:00
Min, Byungil
ea6e3481cd [GPU] Fix failed onednn tests (#16410)
* Fix failed unit-tests on dGPU

+ modified fully_connected_random_test_i8_3d not to have ambiguous
+ oneDNN does NOT support i64 type for reorder. Added exception.
+ bugfix in prepare_primitive_fusing about exception of activation function
+ Add exception logic for dynamic to select ocl type in is_node_for_onednn

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-03-29 15:50:09 +09:00
Taylor Yeonbok Lee
daf562832f [GPU] Fix malfunction in crop static kernel in dynamic shape scenario (#16586)
* Fix malfunction in crop static kernel in dynamic shape execution

* Add unittest

* Fix lint errort
2023-03-29 04:19:24 +00:00
Sergey Shlyapnikov
6c766a81b5 [GPU] Treat warnings C4267 as errors for Windows (#16345) 2023-03-28 22:56:47 +00:00
Wilson Seok
79b267033c [GPU] Fix program::replace() to copy duplicated connection from single constant (#16529)
* fix program::replace() to copy duplicated connection from single constant

* add unit test

* modified with review feedback
2023-03-28 19:25:22 +00:00
Pawel Raasz
796bd98913 Review convolution classes for shape inference aspects (#16375)
* Review adaptive max pool shape inference

* Review AvgPool and MaxPool

* Review convolution operator

* Review GroupConvolution shape inference

* Review ConvolutionBackpropData operator

* Review GroupConvolutionBackpropData op

* Review BinaryConvolution operator
- add common bases for convolution ops
- refactor convolution ops

* Review DeformableConvolution operator

* Use new convolution shape_infer in GPU

* Fix build and test issues

* Correct set output spatial shape
in default constructed back prop convolutions

* The convolution shape_infer use pads as parameters
the external padding can be operators or other class padding properties shape_infer should not modify operators padding when
called from plugin

* Apply code formatting

* Fix padding validation and update

* Use shape inference with padding instead fallback
for DeformableConvolution from opset1

* Update convertPadding function to be template
2023-03-28 19:10:08 +00:00
Eddy Kim
e169c7cd38 fix a bug in permute_bfzyx_to_bfyxz (#16599) 2023-03-28 18:19:35 +00:00
Paul Youngsoo Ahn
253e4eb366 [GPU] Remove duplicated OpenCL kernel compilation on static model (#16262)
* * update kernel_ids using hash value
* Change set to unordered_map for kernels_code
* replace unique_id to hash value
* Remove hash_val params
* remove redundant codes (#16262)
** Remove unique_id in program_node
** Remove gen_kernel_id
** Remove set_kernels_source
** Remove remove_kernels
** Remove kernel_idx in kernels_cache

* * Use kernel_impl_params instead of kernel_id
* Divide batch when entry_point are duplicated
* rollback removing unique_id

* * Fix get_kernel failure issue (#102467)
 - Modify has function of custom_gpu_primitive and generic_layer
 - Add ==operation of generic_layer for _kernels map in kernels_cache
 - Fix invalid kernel_impl_params related to unique_ptr life cycle issue

* Improve kernels_cache (#102467)
* Move add_kernels_source step to build_implementations
* Change replace kernels_code key to kernel_impl_params
* Return kernel vector in get_kernels

* Modify function name to get_kernels (#102467)

* Fix functions related graph serialization (#102467)

* Fix failure to run dynamic model (#102467)

* Add unit test

* Code review follow-up
- Add const to input params
- Add missing code to check kernel duplication in kernels_cache

* Add const to input params (#102467)

* [GPU] update hash and ==operator for generic_layer and custom_gpu_primitive (#102467)

* [GPU] override get_kernels_source in generic_layer and custom_gpu_primitive (#102467)

* [GPU] Fix onednn build error (#102467)

* [GPU] Fix Lin build error (#102467)

* [GPU] kernels_cache::get_kernels return vector of clone of cldnn::kernel (#102467)

* Updated serialization logics for improved kernel caches (#16262)

* primitive key kernel cache for serialization
* kernel serialization with binaries hash
* fix kernel cache init function for deserialization
* removed unnecessary codes

* [GPU] Update commnet and fix test failure (#16262)

* [GPU] Fix custom_gpu_primitive unit test failures (#16262)

* [GPU] Improved kernels cache serialization (#16262)
* removed hash in serialization logic
* update not to create a new kernels_cache for serialization
* code refactoring in serialization logic

* [GPU] Follow-up code review (#16262)

* [GPU] modify lock(#16262)

* [GPU] Fix custom_gpu_primitive unit test failure (#16262)

---------

Co-authored-by: Eddy Kim <eddy.kim@intel.com>
2023-03-28 18:48:19 +02:00
Vladimir Paramuzov
906939a1f1 [GPU] Fixed invalid is_dynamic flag value for scalar inputs (#16565) 2023-03-28 10:03:51 +04:00
hyunback kim
d06a22f4e4 [GPU] Support FC+eltwise fusion in fp16 for OneDNN (#16303)
* [GPU] Support FC+eltwise fusion in fp16

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-28 14:49:49 +09:00
Roman Lyamin
5113a5538c [GPU] Added shape canonicalization mechanism (#16166) 2023-03-27 15:02:06 +04:00
Sergey Shlyapnikov
ce67ac09d3 [GPU] Disable OneDNN primitive cache (#16525) 2023-03-26 23:29:47 +04:00
Andrew Kwangwoong Park
2956717118 [GPU] Added shape agnostic TopK kernel (#16161)
* [GPU] Added shape agnostic TopK kernel implementation

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Update kernel to use internal buffers for shape agnostic kernel

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add WA to compile_graph for shape agnostic arg_max_min_axis with non-const k input

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix is_dynamic pameter for FillCLKernelData with the case where the output is static shape

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix corner case where inbuf size becomes 0 when ops_size is 1

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-03-25 22:32:17 -07:00
Ilya Lavrenov
580b99c99b Align plugins in caching properties (#16528)
* Align plugins in caching properties

* Fixed caching mock tests

* Added new TestNoCachingProperties test

* Fixed test

* Added ov::caching_properties to API 1.0 metrics as well
2023-03-25 00:26:18 +04:00
Taylor Yeonbok Lee
6a25143045 [GPU] Prevent memory reset at runtime allocation for dynamic shape, fix wrong padding handling (#16351)
* Prevent memory reset at runtime allocation for dynamic shape

* Set default alloc to reset mem

* Additional fixes :
- If there is any convolution/deconvolution users which requires padded input, enqueue reset buffer when reuse buffer.
- Removed cl finish from gpu_buffer::fill. (Hopefully it should be waited only when needed. Otherwise sync is to be done by event)
- Removed buffer reset from on_execute of nonzero count, which is not needed any more.

* Remove unused API

* Fix tensor offset to project the padding

* Added unittest

* Applied review comment
2023-03-24 13:10:33 -07:00
Andrei Gorbachev
cabb917b1f [GPU] Fix warnings (#16516)
* fix a few warnings

* cast size_t to uint32_t
2023-03-24 13:26:24 +04:00
Kelvin Choi
8518a3a8e8 [GPU] Disable converting gather8 to 7 pass because GPU plugin supports gather8:nagative-index feature (#15868)
* Add GatherV7 and gatherV8 for convert_gather_0d pattern

* Add updating output_shape using reorder/reshape for scalar indice instead of using ConvertGather0D pass

* Add WA for NMS-gather8 pattern
2023-03-23 23:12:12 -07:00
Sergey Shlyapnikov
e434c320f5 [GPU] Update tuning params of shape agnostic version of fully_connected_bf_tiled kernel for dGPUs (#16482) 2023-03-24 09:08:09 +04:00
Paul Youngsoo Ahn
74870f9b0b [GPU] Fix gpu dynamic model multistream test issue (#16510) (#16510) 2023-03-23 10:51:57 -07:00
Kelvin Choi
6bf2fe11ae [GPU] Need to exclude fused mem_dep from shape_infer_dep (#16300) 2023-03-22 13:00:29 -07:00
Xuejun Zhai
8509d0dd82 [Deprecated API] remove version (#16426)
* [Remove version] Remove version from py openvino

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Modify caused by remove version

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Fix clang format issue

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Revert "Fix clang format issue"

This reverts commit 132787286f.

* Fix CI format issue

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Fix CI format issue

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

* Fix merge conflict error

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>

---------

Signed-off-by: Zhai, Xuejun <xuejun.zhai@intel.com>
2023-03-22 16:09:14 +01:00
Sebastian Golebiewski
cbb25e9483 [DOCS] Proofreading developer documentation moved from wiki. (#15886)
Minor stylistic and grammar corrections. Fixing links

* Apply suggestions from code review

Co-authored-by: Tatiana Savina <tatiana.savina@intel.com>
2023-03-22 09:08:31 +01:00
hyunback kim
c14e6ef48e [GPU] Use 4dim directly for onednn in gemm (#16182)
* [GPU] Use 4-dim directly for onednn in gemm
   We were collapsing n-dim into 3d for onednn gemm, But it is not necessary, up to 4d.

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-22 17:08:10 +09:00
Andrew Kwangwoong Park
04a2c4ce61 [GPU] Add shape agnostic optimized FullyConnectedIMAD kernel (#16417)
* [GPU] Added shape agnostic kernel for fully_connected_gpu_imad

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add fully_connected_gpu_imad shape agnostic TCs for ov_gpu_unit_tests

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Apply comments

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-03-22 08:38:28 +01:00
Min, Byungil
a71c83d366 [GPU] Resolve eltwise kernel build failure (#16458)
Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-03-22 15:15:02 +09:00
Haiqi Pan
5e98696464 Fix Windows build warnings in template and core tests (#15967)
* fix C4305

* 1.0f

* Element

* fix c4244

* fix truncation from double to float in grn.cpp

* Revert "fix truncation from double to float in grn.cpp"

This reverts commit 5263b37cb2.

* fix grn.cpp

* add 4305

* fix low

* add TearDown

* revert softmax.cpp

* pragram

* fix conflicts

* fix conflicts

* size_t -> ov::label_t

* WIN32

---------

Co-authored-by: Ilya Lavrenov <ilya.lavrenov@intel.com>
2023-03-22 07:29:35 +04:00
hyunback kim
05e54e9f3d [GPU] Update the latest onedNN3.1 (#16381)
- Fix group conv regression issue

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-21 18:56:41 +09:00
Min, Byungil
5d6cd626bc Fix unit test on dGPU (#16295)
* Resolve failed cases and queue-type issue
+ Resolved out_of_order queue-type issue
+ Added get_test_default_config for setting default config of onednn
+ Cleared failed case

Signed-off-by: Min, Byungil <byungil.min@intel.com>
Co-authored-by: tuxedcat <tuxedcat@gmail.com>
2023-03-21 18:55:06 +09:00
Ilya Churaev
ec0a1e58d1 Fixed some leftovers for 2.0 dev api (#16421)
* Fixed some leftovers for 2.0 dev api

* Fixed build issue
2023-03-21 09:34:37 +00:00
Vladimir Paramuzov
98237b06b5 [GPU] Update memory_statistics property impl (#16399) 2023-03-21 08:52:52 +04:00
Pavel Durandin
4a4f06ba3b GPU documentation update (#16393)
* GPU documentation update

* GPU documentation update
2023-03-20 14:27:13 +04:00
Vladimir Paramuzov
a46fc47e6a [GPU] Enable tile with dynamic input (#16364) 2023-03-20 12:49:35 +04:00
Min, Byungil
bc15596c9e Remove redundant reorder (#15661)
+ Reorder 1d data
+ Reorder which only changes format

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-03-20 14:58:09 +09:00
hyunback kim
4f49d0e07e [GPU] enable dumpgraph in unit-test (#15388)
Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-03-20 11:51:51 +09:00