* [GPU] Minor fix for dynamic bert-base-uncased-qqp
Signed-off-by: Andrew Park <andrew.park@intel.com>
* Fix to check full tensor only for static shape during creating onednn gemm
Signed-off-by: Andrew Park <andrew.park@intel.com>
---------
Signed-off-by: Andrew Park <andrew.park@intel.com>
- Previously, PR15386 changed allocation of memory of primitives which are to be used as shape infer dep to host memory, for better shape infer perf.
- However this causes cache coherence issue in dGPU.
- Reverting this change so that the memory will be allocated to devicet
* [TF FE] Support EmptyTensorList and TensorListPushBack operations
Signed-off-by: Kazantsev, Roman <roman.kazantsev@intel.com>
* Rename a script to generate the test model
* Correct test model generating script
---------
Signed-off-by: Kazantsev, Roman <roman.kazantsev@intel.com>
* flush fp32 subnormals to zero in IR
* style fix in test_offline_api.py
* simplified call of FlushFP32SubnormalsToZero: is called form offline_transformations.cpp
* reverted offline_transformations.py
* use fpclassify
* style-fix
* Update src/common/transformations/tests/common_optimizations/flush_fp32_subnormals_to_zero_test.cpp
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>
---------
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>
* initial version of implementation
* styles applied
* fixed and registration
* add more unit tests
* fixed and in legacy opset
* review remarks
* refactor of version name range
* [dGPU] Enable stable diffusion
+ Prevent to fuse swish into oneDNN reorder.
+ Makes concat explicitly if batch size is greater than 1 and the siblings are oneDNN impl.
* Small CoreImpl refactoring
* Removed cache_dirhandling from CPU plugin
* clang-format
* Fixed python tests
* Fix
* Fixed bugs in HETERO case
* Fixed clang-format and warnings in auto plugin
* Added import_export as capability for TEMPLATE plugin
* Commented throw exception from loaded_from_cache
* Fixed clang-formatof ro template plugin
This is a corner case because body graph nodes have named output ports.
This allows to support custom RetinaNet model.
Signed-off-by: Kazantsev, Roman <roman.kazantsev@intel.com>
* remove ov::device::thermal
ov::device::thermal was only supported on myriad
* additional cleanup
* remove myriad from AUTO and MULTI
auto n multi n hetero
+ remove mentions of listing myriad devices
* two final fixes
* Update ov_auto.py
---------
Co-authored-by: Ilya Churaev <ilya.churaev@intel.com>
* Remove warning suppression: wd4018, wd4309
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Remove linux warning suppression no-sign-compare
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* ov::intel_cpu::VectorDims base value type is size_t;
dnnl::memory::dims base value type is int64_t;
All compare data up to int64_t can fix warning and there is potential issue.
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* channelAxis maybe == -1; means: no exist any more.
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Fix recursive macro: "one_of", "everyone_is" sign-compare warning.
Must pass same value type.
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Fix Windows sign unsign compare warning
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* There are 2 instances:
using ov::Dimension::value_type = int64_t
using ov::intel_cpu::StaticDimension::value_type = size_t
All up to int64.
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* linux have too many sign-compare issue.
Complete windows sign-compare firstly.
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Fix clang issues.
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Fix warning.
Because instantiate T1=unsigned int, T2=int
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Fix warning for tests unit reorder_node_test.cpp
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Fix warning : ASSERT_GE(step, 1u);
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Fix tests: warning C4018
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* Remove auto, using int64_t is more reasonable.
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
---------
Signed-off-by: Yan, Xiping <xiping.yan@intel.com>
* [GPU] Add data broadcasting for OneDNN binary ops for Gemm primitive
* Based on https://github.com/openvinotoolkit/openvino/pull/15790 and enable onednn gemm from support multiple users and non constant input.
--------
Signed-off-by: hyunback <hyunback.kim@intel.com>
Co-authored-by: Sergey Shlyapnikov <sergey.shlyapnikov@intel.com>
* Support setting and getting element type, shape and value in PyTorch FE InputModel
* Fix code style
* Fix code style
* Fix rsub layer test
* Fix py style
* Apply review feedback
* Fix code style
* Fix initial values of input and output flags in Place
* Enable AUTO to support execution mode hint.
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Add test case.
* Set default value "PERFORMANCE" for ov::hint::execution_mode.
* Update.
* Update.
* Correct default ov::hint::execution_mode value for the default value checking test case.
* Update.
* Delete obsolete config.hpp file.
---------
Signed-off-by: Wang, Yang <yang4.wang@intel.com>
* Init auto_batch plugin unit test
* Add more mock test
* Add to ci yml file
* Fix clang issue
* Resolve compilation issue
* Fix symbol multiple definition in static build
* Add test cases for AutoBatchInferRequest
* Add test cases for AutoBatchAsyncInferRequest
* qFixed build error after PR-15229
* Resolve blocked issue when call StartAsync test cases
* add more test for auto batch async inference
---------
Co-authored-by: Chen Peter <peter.chen@intel.com>