* Add and implement the logic of property SCHEDULE_POLICY for MULTI plugin.
* Updated.
* Enable test case for schedule policy test.
* enable test case for property ov::intel_auto::schedule_policy.
* Update.
* Updated.
* Updated.
* Update.
* Update the lock logic here by considering the runtime fallback case.
* Update.
* Update.
* Update.
* Update default value of schedule policy to DEVICE_PRIORITY
* Enable the function test case for schedule policy.
* Add description for inference requests schedule policy within AUTO plugin cumulative mode.
* Updated.
* Python bindings for enum SchedulePolicy and property ov::intel_auto::schedule_policy.
* Update.
* Update.
* Update.
* Updated.
---------
Co-authored-by: Chen Peter <peter.chen@intel.com>
Co-authored-by: Wanglei Shen <wanglei.shen@intel.com>
* FakeConvert spec init
* Add spec files to the opset docs
* Remove apply_scale attribute and update inputs description
* Experimental op notice
* Update short description
* Update detailed description.
* Add BF16 type to supported inputs
* Update note about "Fake"
* Add more details
* Add formula of the operation
* migrate threading related interface from API 1.0 to 2.0
* fix code style
* fix @ref issue in doc
* change <> to quotation marks
* restore threading related interface API 1.0
* restore the changes of legacy code
* introduced ov.save_model(...) to the ptq code examples
* replied to comments
* fixed rendering
---------
Co-authored-by: Karol Blaszczak <karol.blaszczak@intel.com>
* Added experimental ScaledDotProductAttention operation in opset12. Supported in PT FE for aten::scaled_dot_product_attention translation. Decomposed in the common optimizations as functional reference.
* Better ScaledDotProductAttention
- Moved decomposition to the decomposing transformation
- Implemented more ctors for the op
- Renamed is_causal to causal
- Shape/type inference native code instead of using decomposition
- Moved the op from opset12 to opset13
- Added Python wrapper for ScaledDotProductAttention
* Fix test that counts ops in the opsets
* Update src/core/src/op/scaled_dot_product_attention.cpp
Co-authored-by: Katarzyna Mitrus <katarzyna.mitrus@intel.com>
* Update src/core/src/op/scaled_dot_product_attention.cpp
Co-authored-by: Katarzyna Mitrus <katarzyna.mitrus@intel.com>
* Move ScaledDotProductAttentionDecomposition from fusions to decompositions.
* Remove not used legacy shape inference in ScaledDotProductAttention
* Better namespace usage
* Register all nodes in ScaledDotProductDecomposition for correct tracking of nodes and running next mather passes on all new nodes.
* Don't use register_new_node_
* ScaledDotProductAttention specification (with an extra scale argument)
* Code style fix
* Scale input implementation for ScaledDotProductAttention
* Handle attention_mask=0 case in the op spec
* Better description of scale input
* N->M in scale description
* Code style fix, remove debug print.
* Apply suggestions from code review
Co-authored-by: Katarzyna Mitrus <katarzyna.mitrus@intel.com>
Co-authored-by: Mateusz Mikolajczyk <mateusz.mikolajczyk@intel.com>
* Fix for case when is_causal is not passed
* Extended description of ScaledDotProduct op
* Better description in py op wrapper
* Basic shape propagation tests for ScaledDotProductAttention
* Added ScaledDotProductAttention to toc.
* Add op impl check
---------
Co-authored-by: Katarzyna Mitrus <katarzyna.mitrus@intel.com>
Co-authored-by: Mateusz Mikolajczyk <mateusz.mikolajczyk@intel.com>
* Extend throughput benchmark CLI parameters
* Added device name as the second CLI parameter with default CPU value
* Update samples/cpp/benchmark/throughput_benchmark/main.cpp
Co-authored-by: Zlobin Vladimir <vladimir.zlobin@intel.com>
* Fix comments to review
* Modified python version
* Modified documentation
* Fix comments to review
* Fixed the comment
* Modified python doc
* Fixed device name handling in python version
* Update main.cpp
* Update throughput_benchmark.py
---------
Co-authored-by: Zlobin Vladimir <vladimir.zlobin@intel.com>
* Update list of TF formats.
* Minor correction.
* Added comment.
* Update docs/articles_en/openvino_workflow/model_preparation/Convert_Model_From_TensorFlow.md
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>
* Model changed.
* Update docs/articles_en/openvino_workflow/model_preparation/Convert_Model_From_TensorFlow.md
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>
---------
Co-authored-by: Roman Kazantsev <roman.kazantsev@intel.com>