fixes the FP32 and FP16 swap of results.
* SPR-6448Y updated results
Updated all files with data for SPR 6448Y. New results in benchmark-data.csv, benchmark-performance-data.xls, benchmark-info-detailed.xls and platform-list_22.3.pdf
* Use ov::Tensor in are_equal
* Use ov::Tensor in bounds evaluation
- remove from node evaluates bounds for HostTensor
- update validation utils
- update operators
* Fix code style
* Fix windows compilation errors
* Fix windows compilation errors in scatter update
* Improve reshape output shape calculation
during shape infer
* Add common bound evaluator
- rename evaluate_bound_t -> evaluate_bound
* Fix code formatting
* Move bound evaluation util functions to dev API
* Fix compilation issues
* Remove from dev API not required bound eval
- extract bound eval functions to separate source
* Add missing doxygen comment
* Use ref evaluate in bounds calc for shape_of
because input data can have dynamic type and
create tensor is not possible and data type has
no meaning for shape calculation.
* Add shape_infer function for GatherND
* GatherND shape infer improvements
* Align test to trigger correct error message
* Add new and improve GatherND type_prop tests
* Update tests to use ov namespace
* Add GatherND common shape_infer tests
* Init shape infer tests for not common cases
* Tests refactor
* Add default ctor tests
* Add more test cases
* Register shape_infer for GatherND V5 and V8
* Enable more tests and print params
* Move GatherNDTestParams
* Review ctc loss operator for
- partial shape and label propagation
- template implementation of shape_infer
- update/extend tests
* Use namespace ov in ctc loss operator
* [GPU] Optimize permute for acdb format
Target subgraphs to be optimized-out
- input(bfyx) - permute(byxf) - conv
- conv(byxf) - permute(bfyx) - output
+ Fix test_device_mem_usage_estimation unit test failed.
* Add test to verify add_extension with relative path
* Fix code style
* Use std::string::find instead of std::regex
* Remove unnecessary include
* Add comments about generating relative path
* Don't add empty tokens when splitting path
* [TF FE] Refactor CropAndResize support
Make it more reshape-oriented. It allows to convert Mask R-CNN model without config file.
Signed-off-by: Kazantsev, Roman <roman.kazantsev@intel.com>
* Update src/frontends/tensorflow_common/src/op/crop_and_resize.cpp
* Use Gather for coordinates swapping
* Update src/frontends/tensorflow_common/src/op/crop_and_resize.cpp
* Update src/frontends/tensorflow_common/src/op/crop_and_resize.cpp
---------
Signed-off-by: Kazantsev, Roman <roman.kazantsev@intel.com>
added 3-axis interpolation for linear-onnx mode
fixed resample_opt for onnx mode, it didn't work in case of padding
added tests for the new implementation and fix
@OlehKravchyshyn
* [GPU] improved impl cache key (#14797)
- Add hash function for primitive and program_node
- Filter task before entering async compilation queue
* [GPU] improved impl cache key (#14797)
- Multiply magic prime number at input value of hash_combine to avoid hash collision
* [GPU] Update codes to follow up review comments (#14797)
- Change func name from pop_front_task to erase_front_task
- Change func name from get_layout_key to get_impl_key
- Remove average_unpooling.hpp because it was alread removed
- Replace std::list to std::deque in compilation_context
- Modify layout::hash() to get hash of shape from partial shape
- Remove calculation code to get hash from static layout in program_node => layout hash is calculated outside of program_node
* [GPU] Update gpu functional test for improved impl key (#14797)
* [GPU] update compilation queue (#14797)
* [GPU] Move type_string hash to primitive (#14797)
- Add hash for num_outputs in program_node
* [GPU] update hash functions for program_node (#14797)
- add hash for number of inputs in program_node
- program node::hash() had separated into void program node::caclulate_hash() and size_t program_node::get_hash()
* [GPU] Fix gpu unit test failures (#14797)
- move the location to calculate all nodes from compile_graph to program ctor
* [GPU] Fix build issue after rebase (#14797)
* [GPU] Update impl if optimized kernel is in impl_cache even if the shape does not change. (#14797)
- Apply improved hash key to mem kernels cache in update_weight
- Add missing hash value for broadcast
- Add simple unit test to check hash value for program_node, primitive and program_inst
* [GPU] The draft for integration oneDNN3.0
Initial PR.
1. Support oneDNN3.0 API
2. Use binary_mul post_opt instead of oscale channel-wise mask(2)
3. Disable some post-opt fusing because of no eltwise scale API
eltw(non_linear)+eltw(linear), eltw+sum+eltw(linear)
Signed-off-by: hyunback <hyunback.kim@intel.com>
* Fix hardwish issue in 3.0
hard coded hardswish parameter(2.7) is changed alpha and beta from user's required input.
Signed-off-by: hyunback <hyunback.kim@intel.com>
* clean up code
Signed-off-by: hyunback <hyunback.kim@intel.com>
* Apply code review comment and fix ci issue
Signed-off-by: hyunback <hyunback.kim@intel.com>
* Remove setting dst scale
- ACC issue
- No perf gain compared binary_mul
Signed-off-by: hyunback <hyunback.kim@intel.com>
* gpu serialization for onednn 3.0
* missed changes
* add onednn engine creator when loading model from cache
* fixed to use mem_dep index
* updated to save zero_point_mask for serialization
* fixed onednn fc serialization logic
* updated the logic to check if onednn is enabled
---------
Signed-off-by: hyunback <hyunback.kim@intel.com>
Co-authored-by: hyunback <hyunback.kim@intel.com>
* Adds base class and first test for tflite_layer tests
* adds layer tests for unary ops
* adds functionality to get tensors from ops
* 1. adds functionality to use custom funcs for input generation
2. removed UNIQUE op from testing ops
* adds functionality to use custom dtypes
* Cast operation support
* Enhanced tfl layer tests
* Trigger tfl layer tests in .ci
* Apply suggestions from code review
---------
Co-authored-by: Evgenya Stepyreva <evgenya.stepyreva@intel.com>
Co-authored-by: Evgenya Stepyreva <eva.my.link@gmail.com>
Co-authored-by: missjane <estepyreva@gmail.com>