* keep Const+DecompressionConvert pattern for CPU
* temporary disabled failing unit-tests
* disable CF by modifying bounds evaluate as well; minor corrections
* added TODOs with ticket numbers
* join const+decompression markings
* minimized convert_precision.cpp changes
* minor corrections
* refactor fp16 transformations: moved into separate fp16_compression folder
* style-fix
* minor fixes
* do not disable evaluate and CF in shape path
* safer disabling of Const conversion
* style-fix and minor corrections
* restore original placement of ConvertPrecision
* Mark as deprecated nGraph API
* Fixed code style
* Added IN_OV_LIBRARY define
* Suppress warnings for log
* Suppress warning
* Updated nGraph headers
* Fixed build for macOS
* Fixed lpt and snippets
* Fixed build all on macOS
* Suppress some warnings
* Fixed some new warnings
* Fixed new warnings
* Try to fix some warnings
* More warnings
* Soome change
* Suppress more warnings
* Suppress warnings for transformations
* Suppress warnings for LPT
* One more fix
* Suppress more warnings
* Try to fix opset error
* Remove opset constructor
* Cannot fix opset warning
* Suppress warnings for offline transfromations
* Fixed some warnings for Windows
* Fixed code style
* Suppress some warnings for onnx FE
* Revert "Suppress some warnings for onnx FE"
This reverts commit 75d23b64fc.
* Revert "Fixed code style"
This reverts commit c6eba63116.
* Revert "Fixed some warnings for Windows"
This reverts commit 23d7ed88b6.
* Revert "Suppress warnings for offline transfromations"
This reverts commit 0b9f6317bf.
* Revert "Cannot fix opset warning"
This reverts commit 19ea658639.
* Revert "Remove opset constructor"
This reverts commit 06afb1bc20.
* Revert "Suppress warnings for LPT"
This reverts commit 58b1c0f5a0.
* Revert "Suppress warnings for transformations"
This reverts commit f8bb9814a1.
* Revert "Suppress more warnings"
This reverts commit f9f0da9acb.
* Revert "Soome change"
This reverts commit e545d4984e.
* Remove deprecation for ngraph::OpSet and FactoryRegistry
* [LPT] reused reference FQ implementation in fold_fake_quantize
* [LPT] Removed legacy parameters
* Added plugin tests with per-channel FQ for GrConv wo reshape
* Apply folding only in the case when FQ data input is constant
* EliminateFQ fix
* [LPT][TESTS] GrConv: added test cases with per channel dq on weights and without reshape
* FoldFQ: don't transform FQ with quantization by several dimensions
* ConvolutionTransformation: supported GrConv with per channel dq on weights and without reshape
* fold_reshape: refactoring
* fix special FQ with zero range in quantized models
* fix format & comments
* Add test case
* remove dot interval test case from smoke_LPT/FakeQuantizeTransformation.CompareFunctions
* Remove dot interval gpu test case because Pooling is also folded
* handle review comment
* fix code style
* update docs
* remove fold_zero_multiply
* Review interval shape and labels propagation
* Review shape infer template implementation
- add test for default ctor
* Add evaluate upper, lower and label
- add new default label evaluator which propagates labels
from inputs list
* default_label_evaluator for 0 input only is wrapper for
generic evaluator implementation
* Use default_label_evaluator in ScatterUpdate
* Fix build issues
* Added -Wall for Clang and GCC
* Fixes
* Don't use /J
* Fixed warnings
* Fixed warnings
* More fixes
* Fixed for MSVC
* Fixed more warnings on Windows
* Suppressed some warnings in template plugin
* Update src/tests/functional/plugin/shared/include/behavior/plugin/caching_tests.hpp
* Added suppression for PT FE
* Suppressed warnings in TF FE
* Suppressed warnings on Core unit tests
* Suppress warnings in python
* Suppressed Windows warning for 3rd party modules
* Suppresed one more warning
* [LPT] FQ Decomposition modified to create FQ without constants with several consumers
* [LPT] Added subgraph test
* CPUTestsBase: Added the ability to check only fusing results
* [CPU] Added subgraph test
* LPT review comments applied
* CPUTestsUtils: added special string 'anytype' for selectedType check ignore
ConvolutionBackpropData weights have following shape {input channels, output_channels, ...},
so unlike forward convolution, the output channels are on the second dimension.
* Move transformation and lp transformations tests from IeFuncTests and IeUnitTests
* Enable tests in Azure CI
Co-authored-by: Ilya Churaev <ilya.churaev@intel.com>
* Move GeluFusion into ov namespace
It required to move utils as well.
* Move AddFakeQuantizeFusion into ov namespace
* Move AlignEltwiseInputRanks into ov namespace
* Move BatchToSpaceFusion into ov namespace
* Move BinarizeWeights into ov namespace
* Move BroadcastElementwiseFusion into ov namespace
* Move ClampFusion into ov namespace
* Move Concats and Eliminations into ov namespace
* Move many passes and such into ov namespace
* Move many passes and such into ov namespace
ngraph_ops
op_conversions
* Move control_flow into ov namespace
* Apply style
* Move smart_reshape into ov namespace
* Move opset_conversions into ov namespace
* Move low_precision into ov namespace
* Apply style
* Move fused_names into ov namespace
* Add missing aliases
* Move few passes into ov namespace
ConvertPrecision
FixRtInfo
InitNodeInfo
* Move CommonOptimizations into ov namespace
* Fix convert_precision conflicting namespaces
* Apply code style
* ov-nize Type Relaxed definitions
* Remove useless aliases
bulkly added by script
* Revert opsets
* Address review comments
* Remove NGRAPH_RTTI
* Clean up remains of yet another implicit inclusion
Co-authored-by: Andrei Kochin <andrei.kochin@intel.com>
Co-authored-by: Ivan Tikhonov <ivan.tikhonov@intel.com>