Commit Graph

694 Commits

Author SHA1 Message Date
Ilya Lavrenov
87bcbc1747 Supported OpenSUSE 15.3 (#15897) 2023-02-23 11:25:33 +04:00
Dohyun Kim (Felix)
1028c7b5d5 [GPU] Fix weight reorder bug (#15672) 2023-02-23 14:48:46 +09:00
Jade Cho
c749163f72 [GPU] Update unit tests for swap XY (#15833) 2023-02-23 14:38:10 +09:00
Dohyun Kim (Felix)
1f196bacd3 [GPU][DG2] Fix some testcases (#15774)
* C++ exception with description write lock_type thrown in the test body. 
   Use get_output_values_to_float()
   * fusings_gpu/gemm_2in_act_scale_quantize_eltwise_i8.basic/2
   * fusings_gpu/gemm_2in_act_scale_eltwise.basic/2
* Remove WA test code of [GPU][DG2] Fix fusings_gpu/gemm_2in_scale.basic/7 #15353
   * Now non full-tensor post-ops are broadcasted
2023-02-23 14:23:40 +09:00
Dohyun Kim (Felix)
ed65583957 [GPU] Fix OV_GPU_DumpGraphs option (#15800) 2023-02-23 14:10:21 +09:00
Taylor Yeonbok Lee
4fd38844a2 [GPU] Fix remote blob creation to use original shape (#15864)
* Fix remote blob creation to use original shape

* Revert "Fix remote blob creation to use original shape"

This reverts commit 35c674aa97.

* Fix cldnn tensor adjusted blob to be reinterpreted with actual input layout
2023-02-21 22:22:51 -08:00
Eddy Kim
a6ff809ad7 [GPU] Model caching unit tests (#15413)
* gpu model caching unit tests

* added serialization unit tests

* added save and load for quantize primitive_inst

* reduced the range of inputs for Gemm tests

* updated the copyright year
2023-02-22 05:53:43 +00:00
Konstantin Beluchenko
7f3f576151 [GPU] Permute 5d optimization (#14170) 2023-02-21 14:39:53 +09:00
Dohyun Kim (Felix)
b7bcef6864 [GPU] Improve OV_GPU_DumpLayers debug configuration (#15719)
Co-authored-by: Kim,SungEun <sungeun.kim@intel.com>
2023-02-19 14:57:19 +00:00
Ilya Lavrenov
1d5839fb92 Fixed compilation with clang (#15801) 2023-02-19 16:22:18 +04:00
Ilya Lavrenov
ed5fa69b41 Fixed compilation on CI (#15787) 2023-02-17 22:28:48 +04:00
Roman Lyamin
efb51b058c [GPU] Added operator== for cldnn primitives (#15736) 2023-02-17 19:09:12 +04:00
Xuejun Zhai
91df0a8aa9 [API remove] remove variantImpl & variantwrapper related class/interfaces (#15580)
* [API remove] remove variantImpl & variantwrapper related class/interfaces

Signed-off-by: xuejun <xuejun.zhai@intel.com>

* [Remove APIs] fix code format issue

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>

* [Remove api] fix python compiler issue caused by deprecated varient

Signed-off-by: xuejun <Xuejun.Zhai@intel.com>

* [Remove APIs] fix code format issue

Signed-off-by: xuejun <xuejun.zhai@intel.com>

---------

Signed-off-by: xuejun <xuejun.zhai@intel.com>
Signed-off-by: xuejun <Xuejun.Zhai@intel.com>
2023-02-17 16:31:26 +04:00
Jade Cho
71cff0ae62 [GPU] Fix a bug of permute optimization (#15701)
* [GPU] Fix a bug of permute optimization

For int8 models, if there is FakeQuantize between permute and convolution, an operation like data type casting could be fused to permute. In this case, do not optimize permute.
2023-02-16 11:32:23 +00:00
Maciej Smyk
70cb829992 [DOCS] Move of developer documentation from wiki to md documents - master (#15372)
* CPU Plugin README creation

* debug capabilities

* Update debug_capabilities.md

* performance_analysis_ITT_counters

* cpu-emulation

* runtime_parameters_cache

* Update README.md

* internal_cpu_plugin_optimization

* See Also update for CPU Plugin

* See Also update for CPU Plugin 2

* intel_gpu

* Update README.md

* source code structure & See Also update for CPU plugin

* Update README.md

* See also update

* basic_data_structure

* memory_allocation_gpu_plugin

* Update memory_allocation_gpu_plugin.md

* simplified workflow

* graph optimization passes

* execution_of_inference

* GPU Plugin

* GPU Plugin fix

* Snippets

* Update README.md

* Update README.md

* fixes

* Snippets fix

* Update README.md

* component description

* Key Contacts

* Apply suggestions from code review

Co-authored-by: Ilya Churaev <ilyachur@gmail.com>

* Update src/plugins/intel_gpu/README.md

* Update src/plugins/intel_cpu/docs/internal_cpu_plugin_optimization.md

* Update src/plugins/intel_cpu/docs/internal_cpu_plugin_optimization.md

* Update src/plugins/intel_cpu/docs/internal_cpu_plugin_optimization.md

* Apply suggestions from code review

* Apply suggestions from code review

* Apply suggestions from code review

Text graphs to mermaid

* Update src/plugins/intel_gpu/docs/simplified_workflow.md

* Delete ov_intel_gpu_plugin_diagram.png

Removal of ov_intel_gpu_plugin_diagram.png file as the mermaid version is replacing it.

* Apply suggestions from code review

* Update src/common/snippets/README.md

---------

Co-authored-by: Sebastian Golebiewski <sebastianx.golebiewski@intel.com>
Co-authored-by: Ilya Churaev <ilyachur@gmail.com>
2023-02-16 11:03:11 +04:00
Taylor Yeonbok Lee
523b516835 [GPU] Support empty tensor (#15631)
* Support empty tensor in gpu plugin

* Common kernel setup for skipping

* Refactor

* Cleanup

* Fix for shape agnostic kernel

* Fix error due to memory allocation conflict for an empty input blob with other input blob

* Fix output blob parsing error

* Fixed quantize unittest error

* Fixed wrong TC

* Rename set_skip_kernels to update_kernels_list_to_skip

* Refactor output blob processing

* Applied review comments : more cleanup
2023-02-15 21:53:22 -08:00
Sungeun Kim
0a224a8804 [GPU] Geti: Add jit value for fused_ops in strided_slice_ref (#15634)
* Add jit value for fused_ops in strided_slice_ref
* add test-cases for strided_slice_fusion
* add checking supported fusing ops for strided_slice
2023-02-16 12:12:54 +09:00
Andrew Kwangwoong Park
4028d83990 [GPU] Added shape agnostic optimized MVN kernel (#15603)
* [GPU] Added shape agnostic optimized MVN kernel

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Apply code review

Signed-off-by: Andrew Park <andrew.park@intel.com>

* replace built-in log2 with function macro to calculate power from integer

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Move compile-time JIT constants to cl code

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-02-16 02:51:01 +00:00
Taylor Yeonbok Lee
acced012d9 Minor fix for dynamic shape (#15745)
Should not reshape to 4D for < 4D tensor when new shape infer is used.
2023-02-16 01:02:53 +00:00
Wilson Seok
c186a56f35 fix mem_lock_type of get_sliced_mem() in loop (#15705) 2023-02-15 10:17:07 -08:00
Kelvin Choi
bd4d74d3dc [GPU] Update TopK to support non const top_k input (#15429) 2023-02-14 13:51:24 -08:00
Wilson Seok
60de5a3454 [GPU] fix getConstTensor() for 1d from [1,N,1,1] to [N,1,1,1] (#15569)
* fix getConstTensor() for 1d from [1,N,1,1} to [N,1,1,1]

* add eltwise comparison op in 1d constant check of eltwise arithmetic/logical
2023-02-14 13:02:22 -08:00
Dohyun Kim (Felix)
5853c78b7f [GPU][DG2] Fix fusings_gpu/gemm_2in_act_scale_eltwise.basic/4 (#14888)
* Handle hsigmoid by hard_sigmoid
* add onednn_post_op_type::eltwise_hardsigmoid
2023-02-14 17:28:33 +09:00
Dohyun Kim (Felix)
dd0bf817b7 update (#15568) 2023-02-14 17:20:46 +09:00
Dohyun Kim (Felix)
148c7cc617 [GPU][DG2] Adjust expected fuse counts (#15592) 2023-02-14 17:20:07 +09:00
Zlobin Vladimir
ee2e9d497c benchmark_app: factor out advanced options (#15605)
* benchmark_app: factor out advanced options

* cpp/benchmark_app: add usage word

* move api to advanced, group b/shape/data_shape/layout into Input shapes, factor out Statistics dumping options

* Factor out Device-specific performance options

* Factor out Preprocessing options

* Minor regroup
2023-02-13 17:45:17 +04:00
Roman Lyamin
803a927e70 [GPU] Added shape agnostic kernel support for Range + fix dynamic tests (#15640) 2023-02-13 17:14:29 +04:00
Vladimir Paramuzov
9b840a789c [GPU] Minor fixes for dynamic models (#15543) 2023-02-13 10:34:41 +04:00
Eddy Kim
c0888b45da [GPU] Serialization of the primitive class (#15335)
* serialization of proposal

* serialization of anchors in proposal

* added unit tests for gpu proposal

* updated the proposal primitive to be partially serialized

* serialization of primitive class

* removed unnecessary codes

* removed white spaces

* serialization of loop primitive

* serialization of nms

* fixed implicit concat logic in serialization

* added RUN_ALL_MODEL_CACHING_TESTS directive

* fixed an error related with specialization of boolean vector
2023-02-13 05:34:35 +01:00
Roman Lyamin
75735ca597 [GPU] Added ScatterUpdate shape agnostic kernel (#15579) 2023-02-10 07:06:30 +01:00
Nadezhda Ageeva
ccebc629e2 [GPU] Update caching properties for GPU (#15463) 2023-02-09 10:57:38 +01:00
Andrew Kwangwoong Park
e77c2ab6d7 [GPU] Added shape agnostic NonZero ref kernel (#15534)
* [GPU] Added shape agnostic NonZero kernels

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Add dynamic TC for ov_gpu_unit_tests

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Apply code review

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-02-08 17:53:18 -08:00
Wilson Seok
0d06e525db [GPU] add different type support in range kernel (#15422)
* add different type support in range kernel

* add functional test case for mixed input data type
2023-02-08 16:36:25 -08:00
Taylor Yeonbok Lee
b9107ac7ca Optimized permute kernel cannot be fused with both eltwise + reorder. (#15542) 2023-02-08 10:32:13 -08:00
Roman Lyamin
23b863ffe8 [GPU] Added ScatterNDUpdate shape agnostic kernel (#15567) 2023-02-08 16:27:34 +01:00
Ilya Lavrenov
1f3e469c5e Added -Wall for Clang and GCC (#15513)
* Added -Wall for Clang and GCC

* Fixes

* Don't use /J

* Fixed warnings

* Fixed warnings

* More fixes

* Fixed for MSVC

* Fixed more warnings on Windows

* Suppressed some warnings in template plugin

* Update src/tests/functional/plugin/shared/include/behavior/plugin/caching_tests.hpp

* Added suppression for PT FE

* Suppressed warnings in TF FE

* Suppressed warnings on Core unit tests

* Suppress warnings in python

* Suppressed Windows warning for 3rd party modules

* Suppresed one more warning
2023-02-08 15:01:00 +04:00
Sergey Shlyapnikov
7b649c4150 [GPU] Fix reset_execution() method with wait option enabled for in_order queue type (#15562) 2023-02-08 09:18:46 +04:00
Ilya Churaev
26108b1b67 Deprecate clone_model API, use model->clone() instead of (#15482)
* Deprecate clone_model API, use model->clone() instead of

* Renamed clone_nodes function

* Fixed exception
2023-02-07 23:53:17 +04:00
Dohyun Kim (Felix)
7659551d71 [GPU][DG2] Fix fusings_gpu/gemm_2in_scale.basic/7 (#15353)
* Onednn only supports 2D/3D gemm but openvino GPU plugin policy enforces 4D~6D. 
  This API mismatch causes problems in the post-op axis and requires massive code changes. 
  Therefore we decided to insert throw code for now and fix this issue later 
   if some models require non-(per tensor/full tensor) post-ops.
* Specifically, per-channel(=f) axis in this testcase becomes y-axis 
   because onednn gemm merges b,f axes into one batch axis.
2023-02-07 16:37:26 +09:00
Sungeun Kim
00d9ed0da4 [GPU] fix bug on resample_opt (#15434)
* fix bug: wrong feature slice num
2023-02-07 16:29:18 +09:00
Mingyu Kim
6fa31fbed2 [GPU] Show num_ccs for RANGE_FOR_STREAMS (#15525) 2023-02-07 15:22:45 +09:00
Eddy Kim
8e84531b58 [GPU] Serialization of read_value and assign (#15007)
* serialization of read_value and assign primitives

* lines should be <= 160 characters long

* added unit tests for read_value and assign

* updated to store is_output_evnet in primitive_inst

* removing _is_output_event in typed_primitive_impl_ocl

* added comments for mem_allocated and is_output_null
2023-02-06 11:10:59 -08:00
Roman Lyamin
014a35c3ce [GPU] Added strided_slice shape agnostic kernel (#15477) 2023-02-06 13:03:00 +04:00
Sergey Shlyapnikov
e003bf3af7 [GPU] Shape agnostic FC opt tiled kernel (#15396) 2023-02-06 12:17:55 +04:00
Sergey Shlyapnikov
cd48d76009 [GPU] Limit legacy fusions usage in Convolution kernels (#15465) 2023-02-06 12:08:22 +04:00
Kelvin Choi
8ed71a22fa [GPU] Update ScatterNDUpdate Op to use ngraph shape infer (#15176) 2023-02-05 21:31:33 -08:00
Ilya Lavrenov
de1631d67d Generalized OpenCL handling (#15253)
* Squashed commit of the following:

commit 62c992f6a0bc3a2f559faac6912be9c5632a359f
Author: Ilya Lavrenov <ilya.lavrenov@intel.com>
Date:   Sun Jan 22 11:38:18 2023 +0400

    Generalized OpenCL handling

* Updates

* Fixes

* Update thirdparty/CMakeLists.txt

test

* Fixed build with CL/cl2.hpp

* Fixes

* Fixes

* Fixed compilation flags

* Fixed build with target OpenCL 120

* Don't use cache
2023-02-03 15:36:47 +00:00
Pavel Esir
4103a931c2 [FP16] call marking for mixed precision inside ConvertPrecision (#14965)
* call marking for mixed precision inside ConvertPrecision

* fix typo in precisions list; moved conversion from f64->f32 to the very beginning

* remove obsolete convert_compressed_to_mixed_precision_test.cpp

* typo fix after merge

* corrected namespace prefix

* fixed align_mixed_fp32_fp16_types_test.cpp by removing redundant ConvertPrecision

* updated ConvertPrecison tests for mixed precision

* style fix
2023-02-03 13:47:57 +04:00
Andrew Kwangwoong Park
ab509ce164 [GPU] Added shape agnostic optimized GEMM kernel (#15317)
* [GPU] Shape agnositc optimized gemm kernel

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix CI failure

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Apply code review

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Fix dynamic shape accuracy drop on SQuAD v1.1

- F1: 91.81%, EM: 85.25% @bert-small-uncased-whole-word-masking-squad-0001

Signed-off-by: Andrew Park <andrew.park@intel.com>

* Apply code review

Signed-off-by: Andrew Park <andrew.park@intel.com>

---------

Signed-off-by: Andrew Park <andrew.park@intel.com>
2023-02-03 09:26:35 +04:00
hyunback kim
9d8532e998 [GPU] Use onednn fc/gemm in dGPU. (#15143)
* [GPU] Fix the functional issue using fc:onednn in bert model.

* The issue had happened when input dims are 3 with post-po eltwise.
* oneDNN FC out supports 2-dims only, so OV need to update output and post-op too.
* Fix ACC issue in b16 onednn FC. cldnn updates yxfb format in b16 for opt kernel, but no need in onednn.
* Remove W.A code for running fc cldnn.
* Support gemm primtiive and multi types ForceImplTypes
* Change env name OV_GPU_ForceImplTypes
* Do not change elstwise post-op shape from original node: it caused the ACC issue when multiple users.

Signed-off-by: hyunback <hyunback.kim@intel.com>
2023-02-03 09:58:00 +09:00