Commit Graph

898 Commits

Author SHA1 Message Date
Vladimir Paramuzov
8251399222 [GPU] convolution primitive updates (#17351) 2023-05-12 17:19:29 +04:00
Roman Lyamin
ad03c631b9 [GPU] Added os_is_yx_osv16_isv2 format (#17476) 2023-05-12 15:11:09 +04:00
Taylor Yeonbok Lee
078c5cb7e9 [GPU] Dynamic padding support (#16668) 2023-05-12 09:22:58 +04:00
Pavel Durandin
a6a9a44a0a fix eltwise with one input (#17480) 2023-05-11 18:26:39 +04:00
Pawel Raasz
c13423e2ca Introduce data accessor function for infer in IStaticShapeInfer (#15574)
* Tensor accessor for shape inference
- as functor for getting data from tensor vector or map.
- as lambda in GPU plugin on tile op

* Make tensor data adapter pure virtual
- function accessor to data returns pointer to interface

* Refactor tensor data accessor and adapter

* Extract memory adapter make it GPU graph internal
- can't be part of GPU runtime memory core dev API not visible there

* Expand IStaticShapeInfer by port map
- update factory map for new infer interface with port map information
- add bit util to generate bit mask use it in PortMask

* Pass tensor accessor as reference not fun object
- Add cldnn data adapter and accessor
- Reduce dynamic allocations in data accessors

* Fix compilation issues

* Use ov::Tensor for data accessor
- remove data adapters are they not required

* Update comments

* Fix build issues

* Fix tile shape infer test

* Add empty null tensor accessor as specialization

* Apply style formatting

* Move data accessor from dev API to shape inference

* Fix linking issues
2023-05-11 13:30:30 +04:00
hyunback kim
30395c3e96 Revert "[GPU] Fix proposal sort condition (#16981)" (#17399)
Daily Perf regression occured becuase of changing sort condition.
mask_rcnn_resnet101_atrous_coco,int8
mask_rcnn_resnet50_atrous_coco,int8
faster_rcnn_inception_v2_coco,int8

This reverts commit bd8ca523b9.
2023-05-11 13:32:50 +09:00
Kelvin Choi
370b39e475 [GPU] rnnt accuracy degradation (#17247) 2023-05-11 02:59:57 +02:00
Eddy Kim
e7d94ba020 [GPU] fix adaptive pooling kernel (#17429)
* fixed not to use ceil

* added an unit test
2023-05-09 23:56:43 -07:00
Andrew Kwangwoong Park
00eacd2a96 [GPU] Fix accuracy issue for assign and fully_connected_mmad (#17430)
* Add reorder with usr's output data type for assign

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix incorrect input index for handling leftovers

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add TCs for ov_gpu_unit_tests

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-05-09 20:31:18 -07:00
hyunback kim
3b90165581 [GPU] Improve dump naming rule in debug feature. (#17275)
* [GPU] Improve dump naming rule in debug feature.

Now, we can support below dump naming rule.
- Exec_graph name
- Wildcard letter for target names ('*', '?')
- Case-insensitive name searching
- Apply to show loop body primitives.

New introduces OV_GPU_xxx
- OV_GPU_ListLayers = 1 (Show layer names and exit)
- OV_GPU_VerboseColor = 1 (Show verbose with color)

Add file, line, function in log prefix.


Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-05-09 21:30:50 +09:00
Jade Cho
78cdf22040 [GPU] Add a debug feature to dump by iteration (#17035)
+ 108686
2023-05-08 06:24:27 +00:00
Eddy Kim
634c58903d [GPU] fix gen9_common_conv_fwd_data_f32 kernel to support op fusion when #input_channel == 3 (#17318)
* added op fusion code

* fixed typo

* added unit tests

* size reduction
2023-05-07 20:54:03 -07:00
Edward Shogulin
9c3186b243 [LPT] Concat: different branch precisions support (#17330)
* [LPT] Concat: different branch precisions support
2023-05-07 11:38:32 +01:00
Taylor Yeonbok Lee
9062b81edb [GPU] Fix bug in reorder_redundant_reorder (#17329)
* Fix bug
1) reshape w/ fused primiitive should not be optimized out
2) Wrong usage of slice mem / concat mem in loop
3) LWS not set in lstm_elt

* Added unittest
2023-05-03 16:43:34 -07:00
Paul Youngsoo Ahn
0afe835811 [GPU] update padding info for optimized out node (#17286)
* [GPU] update padding info for optimized out node

* [GPU] Fix padding propagation issue for crop's user nodes(#17286)
- Fix unit test failure
2023-05-03 11:01:23 -07:00
Sofya Balandina
bda41b7321 [ie tests] Move CheckOperationInProfilingInfo to common part (#17282) 2023-05-03 14:42:03 +04:00
Sofya Balandina
96c5cd16d5 [conformance] Rebalance props tests (#17124) 2023-05-03 14:22:43 +04:00
Vladimir Paramuzov
e282a25d94 [GPU] Minor binary size reduction (#17308) 2023-05-03 09:07:52 +04:00
Sofya Balandina
56290fee01 [apiConformance] Fix rebalance missing changes (#17270) 2023-05-02 15:42:19 +00:00
Sergey Shlyapnikov
3b8bb1fb17 [GPU] Use BFS processing order for out_of_order queue (#17250) 2023-05-02 15:14:38 +04:00
Vladimir Paramuzov
b6761beb55 [GPU] Don't throw exception if no devices are found (#17288) 2023-05-01 13:57:06 +00:00
Taylor Yeonbok Lee
52bf9abb8c Added batch_hash and entry_hash of corresponding opencl kernel to be used in debugging (#17277) 2023-04-29 01:11:24 +00:00
Vladimir Paramuzov
c6ec6d457f [GPU] Move functional tests to intel_gpu/tests (#17248) 2023-04-28 22:10:18 +04:00
Roman Lyamin
fbc14a2efb [GPU] Coverity fixes (#17241) 2023-04-27 22:14:27 +04:00
Vladimir Paramuzov
b019868653 [GPU] Added missing second input to splitted reshape in handle reshape pass (#17254) 2023-04-27 17:21:17 +01:00
Wilson Seok
03a428f50c [GPU] Fix remove redundant reorder to skip reorder fusing when sibling node doesn't support fused padding (#17041)
* initial fix

* add corresponding unit test

* skip reorder fusing when sibling node does not support fused padding

* fix data type of axis for win build

* Revert "fix data type of axis for win build"

This reverts commit 719ea75d7826aafc7bb94c1971586c33a9842f10.

* add static casting for win build
2023-04-26 16:53:23 +00:00
Vladimir Paramuzov
6ed85178d5 [GPU] Fix layout propagation logic (#17199) 2023-04-26 14:20:48 +01:00
Edward Shogulin
e593cf8545 [LPT] Precision restriction customization extending (#17147)
* [LPT] Precision restriction customization extending

* comments fix: refactoring
2023-04-26 13:29:09 +01:00
Vladimir Paramuzov
ce5f65af14 [GPU] Use hash of test name for random generator initialization (#17213) 2023-04-26 12:52:38 +04:00
Eddy Kim
09265083ed [GPU] fixed a missing data type (#17200)
* fixed missing data type

* updated the resolution for better accuracy check
2023-04-26 08:28:18 +00:00
Sungeun Kim
3c485feea8 removed case to choose onednn impl for deconv (#17108)
- in_dt(f16) wei_dt(f16) out_dt(f32)
2023-04-26 13:20:11 +09:00
Sergey Shlyapnikov
0a5975bdfa [GPU] Add real kernels' execution timings collection for DumpProfilingData debug option (#15797) 2023-04-25 14:33:08 +04:00
Sungeun Kim
8c09a128ac [GPU] update weights_layout for GroupConv 1d spatial (#17109)
* update weights_layout for GroupConv 1d spatial
2023-04-25 18:54:54 +09:00
Andrew Kwangwoong Park
72906ca242 [GPU] Fix i8/u8 representation error for clamp due to overflow (#17183)
* [GPU] Fix i8 representation error for clamp due to overflow

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix to not include in ocl code

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-04-25 09:41:01 +00:00
Vladimir Paramuzov
f736c71feb [GPU] Fix reshape split for dynamic models + accuracy fix for SAM (#16911) 2023-04-25 09:21:31 +00:00
hyunback kim
19f8f5a3a7 [GPU] Disable oneDNN post-op Prelu in FC,gemm (#17084)
* [GPU] Disable oneDNN post-op relu

Only disable Prelu fusion in Fc, gemm
 - check additional data input

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-04-25 18:06:22 +09:00
Vladimir Paramuzov
ca1102b855 [GPU] Support MVN cases with axis=-1 w/o decomposition (#17020) 2023-04-25 12:59:03 +04:00
Vladimir Paramuzov
d1f1fa2b39 [GPU] Enable broadcast transition pass (#17172) 2023-04-25 09:04:37 +04:00
Vladimir Paramuzov
3bb0fb61f6 [GPU] Support 8d tensors in activation and quantize primitives (#16947) 2023-04-25 09:02:54 +04:00
Taylor Yeonbok Lee
ce23ce00f1 [GPU] Fixed fused_primitive_desc to have -1 value for dep_start_idx (#17099)
* Fixed fused_primitive_desc to have -1 value for dep_start_idxt b

* Fixed dgpu i8 errors
2023-04-24 22:21:58 +00:00
Eddy Kim
d972a71b4c [GPU] Fixed the prepare_quantization pass to support grouped_weights_shape (#17093)
* fixed to support grouped_weights_shape

* added grouped_weights unit tests
2023-04-24 14:21:50 -07:00
Jade Cho
5c21dcec4d [GPU] Fix detection output kernel build error on dGPU (#17150)
+ Check local memory size used in the kernel and choose proper kernel.
+ 	Select DO_STAGE_0_CAFFE instead of DO_STAGE_0_CAFFE_OPT
2023-04-25 01:00:26 +04:00
Ilya Lavrenov
a34ef680f2 Made plugins.hpp generation to be CONFIG dependent (#17139) 2023-04-24 14:48:45 +04:00
Vladimir Paramuzov
e8ae1e41ea [GPU] Skip FC fake alignment for some vector by matrix multiplications (#17051) 2023-04-24 14:34:50 +04:00
hyunback kim
63f5c2f0e7 [GPU] Fix levit-128s accuracy issue (#17136)
* [GPU] Fix levit-128s accuracy issue

Wrong batch dims for fused eltwise of gemm.
-> The issue is getting incorrect batch size of fused eltwise used by gemm.
     Its rank is different from src tensor. Eltwise tensor rank was reduced by mistake.
     It is only reproduce in batch 1 and full tensor. 
     The batch size in here means all of non spatial dims, but previous implementation was default batch dim role.

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-04-24 18:16:00 +09:00
Vladimir Paramuzov
219a0eebdc [GPU] Fix 1d onednn convolutions (#17038) 2023-04-24 09:24:56 +04:00
Min, Byungil
bb0be3c177 [GPU] Resolve failed onednn tests (#16990)
* [GPU] Resolve failed unit-tests on dGPU

+ Modified unit-tests of asymetric conv with per channel(WA for oneDNN issue)
+ Modified conv unit-tests with padded input or output
+ For testing oneDNN conv, it needs to query oneDNN about format. Applied this to conv tests.
+ Modified accuracy checking logic in unit-tests which have different format on dGPU.
+ reorder from fsv16 to bfyx should not be optimized out if not aligned by 16

Signed-off-by: Min, Byungil <byungil.min@intel.com>
2023-04-24 14:11:35 +09:00
Jade Cho
8fbd78fb07 [GPU] Fix a bug of fusing eltwise sum post-op. (#17078)
+ When input of eltwise is full-tensor constant layer, use binary add
instead of sum as post-op on oneDNN.
2023-04-21 20:17:35 +09:00
Mingyu Kim
ba23e2290e [GPU] Choose onednn impl for reorder (#17077)
* [GPU] Choose onednn impl for reorder
* [GPU] Add unit test
2023-04-21 13:56:58 +09:00
hyunback kim
344db564fc [GPU] Fix dump graph failure issue in levit-128s model. (#17055)
* [GPU] Fix dump_graph failure issue in levit-128s model.

1. to_string() in strided_slice always access begin/end/stride param id from dependencies
    regardless of max dependencies.
2. Add an exception in dump_full_node(). It helps below.
   - Avoid a dump failure. Usually, graph dump are used during debugging,
      which reduces unnecessary debugging time due to graph dump failure.
   - You can immediately see which node has failed, making it easy to find it.

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-04-21 09:14:47 +09:00