* [dGPU] Enable user scratchpad mode.
* Reuse intermediate buffer.
* Add own id to the memory dependencies at the c-tor of program_node
+ Allocate intermediate memory with memory_pool::get_memory() function.
+ Assign scratchpad memory desc in load() function for onednn primitive
serialization
* Allocate device mem for onednn scratchpad mem
* [Bug Template] New format bug template
This new bug template provides an easy interface for customers to fill in the information before submitting a new issue.
* [Update][Bug Template] New format bug template
- use 2021.4 LTS as an example for the OpenVINO version (this is where the customer will fill up the version)
- Adding macOS Systems for Intel CPU, macOS Systems for Apple Silicon, Android Systems, Raspbian Stretch OS, and WebAssembly to the OS list
- Adding GNA, AUTO, HETERO, and BATCH to the devices list
- Remove "Engine" in "Inference Engine"
* Update bug.yml v3
removed #endhere
* Apply suggestions from code review
add some suggested changes by @p-wysicki
Co-authored-by: Przemyslaw Wysocki <przemyslaw.wysocki@intel.com>
* Update bug.yml
---------
Co-authored-by: Przemyslaw Wysocki <przemyslaw.wysocki@intel.com>
* ResolveNamesCollisions transformation refactoring; enable it in MOC
* fix the description
* call ResolveNamesCollisions transformation in the frontends; resolve review comments
* Resolve review comments
* fix EliminateUnsqueezeGather and AlignMixedTypes transformations
* [Snippets][CI] Added Snippets Func Tests to Azure Windows
* [Snippets][CI] Added Snippets Func Tests to Azure Linux
* [Snippets][CI] Added Snippets Func Tests to GitHib workflow Linux
* [Snippets][CI] Added Snippets Func Tests to GitHib workflow Windows
* [Snippets][CI] Added Snippets Func Tests to Azure Linux Debian
* Remove inits, update main one
* Fix stacklevel
* Testing wrong solution
* Testing test test
* Fix test test test test
* mo modules mo problems
* Xfail test that check stdout/err?
* not so correct solution to circular imports
* Fix or not to fix
* CMake magic, co-authors: my best team
* Fix package imports
* Fix tools inits
* Fix ovc tf
* Fix Q000
* Fix F401
* Fix linters
* Add save_model
* Remove LayoutMap
* Move test_utils to 'internal modules'
* First testing
* Missing Type
* Expand main namespace
* Change some more tests
* Add OVAny to namespace
* Add Constant and Parameter to namespace
* More tests changes
* Fix inits
* Add layout_helpers to main namespace
* Revert CMake and linux.yml with ovc
* Update main inits
* Remove MO from tools inits
* changes to init files
* Fix tests
---------
Co-authored-by: Sergey Lyalin <sergey.lyalin@intel.com>
* [GNA] Fix for GeminiLake detection
* Added HWGeneration::GNA_1_0_E enumerator
Added DeviceVersion::GNAEmbedded1_0 enumerator, changed the meaning of DeviceVersion::GNA1_0.
Updated ConvLowPrecision test with all supported targets
* [GNA] Extended a few tests with GNA1.0
* Add debug messages
* Add more debug messages
* Add more messages
* Disable proxy plugin
* Enable proxy and suppress NVIDIA
* Remove disabling NVIDIA
* Update hidden device only if proxy settings were changed
* Use static unavailable device to load unavailable device only one time
for the process
* Removed debug messages and add global mutexes for unavailable plugins
* Modify the condition making batch interpretation true/false
- When the user is Convert for Constant node, and tensor is 1d,
- Set needBatchInterpretation to true
* Narrow down the range of the condition
* Merge the condition
* Add additional condition not to check self node
* Fix incomplete condition
* Check if all inputs to binary eltwise is 1d
* Change code style
* [MO] compress_to_fp16=True by default (2dn attempt)
* fix unit-tests
* second round of fixin unit-tests
* set compress_to_fp16 default to True in ovc/cli_parser.py
* use save_model in mo_python_api_tests
* enforce compress_to_fp16=False in test_zero_copy
* selectively compress depending on the path user has chosen to generate IR
* corrected doc
* allow compress_to_fp16=False/True for ovc
* doc and unit-tests failing fix
* user save_model in ovc cli tool
* revert back serialize and compress_model but into main instead of moc_emit_ir
* cover more argument combinations for cli tool and convert_model
* [GPU] Improvement for buffer dump
+ added OV_GPU_DumpLayersInput to support dump input layers
+ added OV_GPU_DumpLayersRawBinary to make binary dump
+ added OV_GPU_LoadDumpRawBinary to use binary dump as input
+ binary dump naming rule layername_datatype_tensor_format.bin
Signed-off-by: Min, Byungil <byungil.min@intel.com>
* Previously reorder / permute was not allocating its memory at build time thought the shape has upper bound
* Update src/plugins/intel_gpu/src/graph/permute.cpp
Co-authored-by: Sergey Shlyapnikov <Sergeishlyapnikov@gmail.com>
* Fix as review comment
---------
Co-authored-by: Sergey Shlyapnikov <Sergeishlyapnikov@gmail.com>