Commit Graph

959 Commits

Author SHA1 Message Date
yanlan song
05e8bd375e Bell/auto api 2.0 (#17805)
* 2.0 innitial

Signed-off-by: fishbell <bell.song@intel.com>

* enable all tests

Signed-off-by: fishbell <bell.song@intel.com>

* remove unecessary files

Signed-off-by: fishbell <bell.song@intel.com>

* move container header to auto foler, remove uncessary macro define

Signed-off-by: fishbell <bell.song@intel.com>

* enable caching

Signed-off-by: fishbell <bell.song@intel.com>

* enable query_model

Signed-off-by: fishbell <bell.song@intel.com>

* support loaded_from_cache property

Signed-off-by: fishbell <bell.song@intel.com>

* fix some build warning

Signed-off-by: fishbell <bell.song@intel.com>

fake inputs/outputs if needed

Signed-off-by: fishbell <bell.song@intel.com>

* resolve conflict

Signed-off-by: fishbell <bell.song@intel.com>

* skip unsupported test

Signed-off-by: fishbell <bell.song@intel.com>

* use mock icore from common foler

Signed-off-by: fishbell <bell.song@intel.com>

* fix failure for remote tensors

Signed-off-by: fishbell <bell.song@intel.com>

* apply ppp related fix in auto

Signed-off-by: fishbell <bell.song@intel.com>

* fix build warning on windows

Signed-off-by: fishbell <bell.song@intel.com>

* fix ppp output layout issue

Signed-off-by: fishbell <bell.song@intel.com>

* fix ppp output layout issue

Signed-off-by: fishbell <bell.song@intel.com>

* clean up headers

Signed-off-by: fishbell <bell.song@intel.com>

* log formatting

Signed-off-by: fishbell <bell.song@intel.com>

* enable fps logging for binder mode

Signed-off-by: fishbell <bell.song@intel.com>

* apply review comments

apply review comments

Signed-off-by: fishbell <bell.song@intel.com>

* remove all legacy namings, exenetwork/network/metric/IE etc

Signed-off-by: fishbell <bell.song@intel.com>

* update readme

Signed-off-by: fishbell <bell.song@intel.com>

* fix build lto issue

Signed-off-by: fishbell <bell.song@intel.com>

* minor wording

Signed-off-by: fishbell <bell.song@intel.com>

* case fix

Signed-off-by: fishbell <bell.song@intel.com>

---------

Signed-off-by: fishbell <bell.song@intel.com>
Co-authored-by: Chen Peter <peter.chen@intel.com>
2023-06-21 00:10:59 +08:00
Wilson Seok
3519050ef0 skip all user format check when dynamic shape in get_preferred_format() to avoid endless recursive call (#18096) 2023-06-19 18:52:58 -07:00
Patman11
b9575d9586 [GPU] Disable threaded kernel compilation when running in Windows Store app (#18062) 2023-06-19 17:55:47 +04:00
Min, Byungil
9943ffc259 [GPU] Fix unit-tests for dGPU (#18125)
+ Resolved unit-tests failure on dGPU
+ Applied get_test_default_config for testing config

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-06-19 11:41:47 +04:00
Min, Byungil
555c083336 [GPU] Optimize out Gather by converting to implicit crop (#17743)
+ Changed Gather if it divides input tensor along batch axis
+ Converted Gather to cldnn Crop in CreateGatherOpBase
+ Added implicit Crop condition for batch axis

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-06-19 05:05:22 +00:00
Vladimir Paramuzov
3d79bd1ac5 [GPU] Minor layout optimizer refactoring (#17553) 2023-06-16 10:33:53 +04:00
Pavel Esir
aa32ff1df3 keep Const + DecompressionConvert for CPU (#15930)
* keep Const+DecompressionConvert pattern for CPU

* temporary disabled failing unit-tests

* disable CF by modifying bounds evaluate as well; minor corrections

* added TODOs with ticket numbers

* join const+decompression markings

* minimized convert_precision.cpp changes

* minor corrections

* refactor fp16 transformations: moved into separate fp16_compression folder

* style-fix

* minor fixes

* do not disable evaluate and CF in shape path

* safer disabling of Const conversion

* style-fix and minor corrections

* restore original placement of ConvertPrecision
2023-06-15 13:07:22 +04:00
Andrei Gorbachev
52834659c4 [GPU] additional checks fixed for fully_connected (#18068) 2023-06-15 09:11:38 +04:00
Mykhailo Hnap
bae926de22 [GPU] Unique-10 operation implementation. (#16412)
* [GPU] Unique-10 operation implementation.

* Handled flattened case.

* Created results for all outputs in single layer test.

* Save total unique count as fifth output.

* Handled axis case.

* Added unique reshape kernel.

* Moved data types to unique primitive constructor.

* Added shape agnostic Unique ref kernel.

* Added blocked layout support to Unique-10.

* Use int in bubble sort.

* Added unit tests.

* Added support for blocked layouts to flattened mode.

* Fixed usage of shape_info in kernel.

* Use correct total data size for dynamic shapes.

* Commented some functional tests.

For some reasons big shapes cause std::bad_alloc.

* Initialize out_counts with zeros.

* Implemented new approach for reducing memory footprint.

Changed first kernel to only count unique values and changed second kernel to fill all outputs.

* Revert "Commented some functional tests."

This reverts commit a7f9763c575e71e14b85ee37adf1e98f10785c15.

* Fixed calc output layouts for flattened case when rank in greater than 4.

* Added temporary fix for axis case when rank is greater than 4.

* Revert "Added temporary fix for axis case when rank is greater than 4."

This reverts commit 236640d2f0e9d5b1f8dcbbf9482763badd7fde66.

* Renamed "unique" to "unique_count" and "unique_reshape" to "unique_gather" primitives.

* Quick fix for add_intermediate_node to consider dep_idx of multiple output

* Fix bug for multiple output:
1) get_reorder was getting reorder from cache regardless of the dep_idx.
2) remove_redundant_reorder was not considering original dep_idx

* Fixed conflicts.

* Fixed win build issue.

* Fixed build issue.

* Revert "Fix bug for multiple output:"

This reverts commit d4a2c4f32eabe9108df31d4837fed8995c93bd1c.

* Revert "Quick fix for add_intermediate_node to consider dep_idx of multiple output"

This reverts commit 2dfd2aaefdf32067a7469505b35f7096632ac5f2.

* Added some tests to skip config.

---------

Co-authored-by: Taylor Yeonbok Lee <taylor.lee@intel.com>
2023-06-14 10:41:51 -07:00
Andrei Gorbachev
1761427ab1 fixed fp16 x fp16 overflow in NonMaxSuppression (#18038) 2023-06-14 15:58:49 +04:00
Roman Lyamin
63a5ec5762 [GPU] Several fixes for format traits (#18018) 2023-06-14 14:33:58 +04:00
Sergey Shlyapnikov
e631f65a9b [GPU] Fix in-order queue synchronization issue related to OCL/OneDNN impls interaction with CPU impls (#17976) 2023-06-14 10:15:04 +09:00
Ilya Churaev
0743e9bfb5 Removed legacy methods SetBatch and SetBlob (#17984)
* Removed legacy methods SetBatch and SetBlob

* Fixed GPU plugin build

* Remove DYN_BATCH_LIMIT from tests

* Revert some changes in GPU plugin
2023-06-12 18:54:23 +00:00
Ilya Churaev
df44f92a97 Remove NV12 and I420 blobs and deprecate some legacy API (#17919)
* Remove NV12 and I420 blobs and deprecate some legacy API

* Fixed some errors

* Remove NV12 blobs

* Remote NV12 conversion

* Fixed other warnings

* Suppress version

* Fix some warnings

* Fixed version

* Try to fix some warnings

* Suppress warnings in C header

* Suppress warnings in C

* Fixed Windows exceptions

* Try to fix warnings

* Try to fix C bindings build

* Suppress InferRequest

* Fixed some build issues

* Fixed some errors
2023-06-12 21:15:02 +04:00
Sergey Shlyapnikov
70e0caca4f [GPU] Fix dynamic padding processing of static dimension (#17978) 2023-06-12 08:39:42 +04:00
Wilson Seok
cff083f83d [GPU] gather nd shape agnostic kernel implementation (#17940)
* gather nd shape agnostic kernel implementation

* add func test

* fix minor bugs

* minor bug fixes

* fix win build error
2023-06-10 00:28:00 -07:00
Andrew Kwangwoong Park
c413825845 [GPU] Fuse type conversion only reorders to the prev nodes (#17881)
* Fuse convert reorder to prev MVN/Concat node

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add dynamic TCs for ov_gpu_unit_test

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add descriptions for changes

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix kernel selection failure

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add is_type_conversion_only function for reorder_node

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-06-09 16:07:01 -07:00
Ilya Lavrenov
a0119fe33c Android debug build (#17955) 2023-06-09 08:03:10 +04:00
Sergey Shlyapnikov
58d79aa3a6 [GPU] Add shape_of subgraphs markup and initial cpu implementations (#17762)
* [GPU] Add shape of subgraphs markup and initial cpu implementations for some of primitives

* Apply review comments

* Exclude eltwise with boolean mode types from shape of subgraphs and fix leftovers
2023-06-08 13:46:21 +04:00
Taylor Yeonbok Lee
f246015dd7 [GPU] Fix issue in runtime buffer fusing (#17909)
* There were two issues in runtime buffer fusing
1) Missing condition in matcher for dyanmic tensor
2) If the node is marked as can_be_optimized = true at build time and then turned out to false at runtime, the kernel compilation has been skipped becuaes it was checking node->can_be_optimized
=> To resolve this issue, added can_be_optimzied to impl_param and let the impl create check can_be_optimized in impl_param instead of that in node.

* Fixed primtiive::can_be_optimize to be set through function
2023-06-07 19:39:26 -07:00
hyunback kim
13028397b7 Optimize permute gemm onednn (#17621)
* [GPU] Optimized out permute in permute-gemm(onednn) pattern.

Permute can be optimized out when permute's in and out are compatible and onednn gemm.

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-06-07 16:20:59 +09:00
Ilya Churaev
36625404eb [GPU] Fix GPU remote context name initialization (#17850) 2023-06-05 12:00:04 +04:00
Sergey Shlyapnikov
db8d23231a [GPU] Change priority of CPU implementations (#17829) 2023-06-05 11:21:26 +04:00
Vladimir Paramuzov
1ce447674e [GPU] Better device input memory reuse (#17853) 2023-06-05 09:30:22 +04:00
Kelvin Choi
ec0daa5b10 [GPU] Apply m_pythondiv for fusing of eltwise div (#17590) 2023-06-02 17:29:02 -07:00
Yaroslav Torziuk
eb588f0336 Add subgroup block reading in softmax_gpu_items_class_optimized.cl (#16223) 2023-06-02 12:59:55 -07:00
Taylor Yeonbok Lee
f670dc5a0d [GPU] Enable runtime buffer fusing for dynamic shape (#17668)
* Initial impl for runtime buffer fusing
Passing unittest with static kernel

* pass unittest with dynamic impl

* Refactor allocate_output

* Separate header of buffer fusing

* Refactored buffer fusing :: matcher/optimize

* More cleanup

* Fix crash in dolly

* Reset can_be_optimized of primitive_inst when it is not

* Fix empty tensor : Primitive with empty data should be skipped

* Fix issue in dynamic padding : Static kernel should not contain dynamic padding dims
Fix missing reset of update_shape_done_by_other flag

* Not to add cache with emtpy kernel for optimized out inst

* Fix corner case error in buffer fusing
- Shapes of some preds may not be changed, but still needed to do update_impl because 1) paddings are changed 2) output memory should be updated
- optimizable impl should not be added to the cache

* Allowing reorder & permute_ref to be optimized concat predecessor

* Some more fixes :
runtime buffer fusing is available only when all preds/concat are dynamic
runtime buffer fusing is to be executed only if the node is dynamic

* Fix allocate_output parameter called by get_estimated_device_mem_usage according to the new change

* Fixed error in cascaded concatt

* Need to reinterprete even though the size is same
2023-06-02 12:39:28 -07:00
Sergey Shlyapnikov
5afbd4cf92 [GPU] Remove clFinish call from USM memory lock function (#17830) 2023-06-02 16:17:05 +04:00
Andrei Gorbachev
97113b317f [GPU] fix incorrect deformable_group_idx calculation (#17759) 2023-06-01 10:51:48 +04:00
Vladimir Paramuzov
ac26216869 [GPU] Functional fixes for nvidia (#17735) 2023-06-01 09:45:30 +04:00
Maciej Smyk
dc36ec11b5 [DOCS] Link adjustment for dev docs + fix to build.md CPU link for master (#17744)
* link-update-1

* link update

* Update build.md

* dl workbench

* Update README.md
2023-05-31 13:27:20 +04:00
Pawel Raasz
5299c3378b Review interpolate for shape inference aspects (#17667)
* Review interpolate shapes and label propagation

* Review shape_infer template implementation

* Update shape infer of interpolate in GPU plugin
- Add new tensor accessor for ov::Tensor map

* Correct casting in dim::scale function

* Remove validation of size of input 1 in v0

* Relax inputs check for interpolate v4

* Correct GPU shape inference

* Use ov::Tensors in interpolate's evaluate
- Remove some duplicated code
- Apply comments from review

* Set shape in interpolate's eval for output tensor
2023-05-30 14:49:54 +04:00
Pavel Durandin
dfb6c8ae38 [GPU] Mvn skipconfig update and typos fix (#17660) 2023-05-29 09:09:36 +04:00
Eddy Kim
ef041565a8 [GPU] primitive serialization (#17670)
* primitive serialization

* updated primitive::desc() to use impl_param instead of program_node

* added hash caching unit tests

* added missed calls to save and load of parent

* updated copyright year
2023-05-25 18:31:32 -07:00
Andrew Kwangwoong Park
eeb552cc93 [GPU] Added shape agnostic optimized Permute_tile_8x8_4x4 kernel (#17652)
* [GPU] Added shape agnostic optimized Permute_tile_8x8_4x4 kernel

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add permute_gpu_tile_8x8_4x4 shape agnostic TCs for ov_gpu_unit_tests

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix calculation for required local mem size

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Update not to condisder x and feature dimension for tile size on shape agnostic kernel case

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-05-25 15:07:07 -07:00
Andrei Gorbachev
71dcdf8a28 [GPU] Remove reorder fix (#17691) 2023-05-25 14:32:32 +04:00
guozhong wang
b93b863bac 1.Remove MULTI test cases 2.Reduce iteration from 50 to 10 (#17551)
Co-authored-by: Chen Peter <peter.chen@intel.com>
2023-05-25 11:10:02 +08:00
Egor Duplenskii
2e8548ca36 [IE_TESTS] Correct random data generation (#17244) 2023-05-24 15:50:30 +04:00
Sofya Balandina
be96f5438c [apiConformance] Fix issues in core properties tests (#17608) 2023-05-24 15:29:31 +04:00
Maciej Smyk
13c966f293 [DOCS] Link adjustment (Snippets) for master (#17659)
* link fix

* diagram_workflow

Removal of not used diagram_workflow.svg images

---------

Co-authored-by: Karol Blaszczak <karol.blaszczak@intel.com>
2023-05-24 14:03:54 +04:00
Min, Byungil
0d3b636d1c [GPU] BugFix reduce_b_fs_yx_fsv16 kernel (#17477)
+ Invalid calculation in reducing un-aligned feature axis for b_fs_yx_fsv16
+ Some reduce modes are not invariant by using 0 value out of range
+ Added jit ZERO_INVARIANT_REDUCTION
+ Enable blocked unit-tests on dGPU by PR#15873

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-05-24 13:56:55 +09:00
Bo Liu
703e5421ca extend PaddlePaddle elementwise broadcast type support (#17102)
* enable PaddlePaddle elementwise broadcast

* fix CI fail issue

* Apply suggestions from code review

* fix CI fail issue

* only B to A broadcast is supported for PDPD

* fix GPU plugin testcase fail issue

* keep PDPD broadcast_merge cpu plugin implement align with ov core

* add type prop test case for pdpd broadcast dst shape smaller than src shape
2023-05-23 14:25:56 +04:00
Taylor Yeonbok Lee
de2302a711 Prevented gather fusion test for dgpu dynamic shape (#17616) 2023-05-23 01:57:56 +02:00
Ilya Lavrenov
4c92ffa563 Build wheel arm64 packages using cross-compilation (#17635) 2023-05-23 00:31:35 +04:00
Andrew Kwangwoong Park
4ccb6794a4 [GPU] Minor fix for shape inference of dynmaic reshape (#17565)
Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-05-22 10:44:23 -07:00
Ilya Lavrenov
84db7d0ee6 Build using conanfile.txt (#17580)
* Build using conanfile.txt

* Update .ci/azure/linux_arm64.yml

* Several improvements

* Removed conanfile.py

* Try to use activate / deactivate

* Fixed clang-format code style

* Supported TBB version from Conan

* Added more NOMINMAX

* Fixed static build

* More improvements for static build

* Add usage of static snappy in case of static build

* More fixes

* Small fixes

* Final fixes
2023-05-19 14:01:39 +04:00
Pavel Durandin
54bbc9e603 [GPU] Fix out of range check for pooling (#17612)
* [GPU] Fix out of range check for pooling

* [GPU] Fix out of range check for pooling
2023-05-19 14:01:27 +04:00
Vladimir Paramuzov
b95aa84b45 [GPU] Removed some redundant internal passes (#17552) 2023-05-19 13:34:42 +04:00
Eddy Kim
7c84a586f9 [GPU] Fixed deserialization logic for dynamic batch (#17486)
* deserialization of dynamic batch

* updated multi stream tests

* added unit tests

* updated cache dir name

* resolved type conversion warning

* removed teardown()

* added const
2023-05-18 15:40:04 -07:00
Kelvin Choi
fac6668ed1 [GPU] Shape agnostic ref kernels implementation for convolution (#16593) 2023-05-18 15:21:30 -07:00