* cpu also need to check precision
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* fix test case failed issue
Signed-off-by: Hu, Yuan2 <yuan2.hu@intel.com>
* Create ov_runtime library
* Added dev target
* Fixed SOURCE_DIR property for interface lib
* Fixed missed include path
* Fixed SOURCES for plugins and frontends
* Fixed Windows export
* Try to fix preproc build
* Fixed MKLDNNPlugin
* Fix C API
* Try to fix public CI
* Fixed comments
* Use TARGET_NAME variable
* Fixed static build
* Fixed target name in comfig
* Revert "Fixed target name in comfig"
This reverts commit f61c1e0c09.
* Try to remove WA with include path
* Try to fix TBB dependency for old CMake
* Revert install to lib for legacy tests
* Try to fix public tests
* Add WA for v7 reader
* Fixed dev export name
Co-authored-by: Ilya Lavrenov <ilya.lavrenov@intel.com>
* Move OMZ to thirdparty/open_model_zoo
* Improve diagnostic of issues
* Update path to OMZ in the Python wheel setup.py
* Update POT tests for OMZ location in openvino/thirdparty
* fix mo path
Co-authored-by: azhogov <alexander.zhogov@intel.com>
* Add DumpLayersLimitBatch config to GPU debug utils.
+ Support OV_GPU_Help config
+ Only run first inference if OV_GPU_DumpLayersPath is set.
+ Fix dump graph bug.
* Apply some comments
* Remove unnecessary code.
* Moved Speech sample to OpenVINO 2.0
* improved samples's score
* changed code style
* added GNA configs
* renamed function due to new API
* added dynamic batch
* reordered includes
* speech_sample has more than 1 input
* added oname processing
* added multi input
* fixed notes
* removed getFullDeviceName with old api
* getFullDeviceName for benchmark
* move conv_ie deconv_ie to legacy with related passes. Change headers to legacy
* update convert_nms_gather_path_to_unsigned.cpp
* add legacy dependecy and revert changes in tests
* remove INFERENCE_ENGINE_API_CLASS
* add support for f16 and bf16 for outputs_to_dict
* create test for f16
* add test for infer new request return type
* fix infer return type test
* update Type import
* update asserts in infer_new_request_return_type
* fix import in test_infer_request
* clean test file
* style fix
* fix ci checkout in the face of changed submodules
* update cldnn plugin name in lohika ci script
Co-authored-by: Serhii Pavlovskyi <spavlovskyi@lohika.com>
* add performance hint to time infer
* Update time tests with options
* fix mistake in help
* update test conf with GPU models
* move ie configs from load plugin timer
* Merged and compiling
* Fix for dynamic shape type
* review fixes
* renamed blob shape to tensor shape, small improvements
* fix code style
* added parsing of multiple shapes
* store latency per group, add isIdleRequestAvailable() to Infer Queue
* added cached random inputs
* redesign pipeline, added new metrics(avg, max, min), added metrics per groups
* fixed code style
* small improvements
* modified tensor parameters parsing
* modified -i parameter parsing: added possibility to specify input names
* implemented image cashing
* added cashed blobs creating
* added -pcseq flag, modified batch filling, changes fps formula
* improvements
* code formatting
* code formatting2
* apply suggestions from review
* replaced Buffer class with InferenceEngine Blobs
* use batch size in blobs filling
* added shared blob allocator to handle blob's data
* fixed warnings & code style
* allocate blobs
* fix for networks with image info input
* added comments & fixed codestyle
* clear data in free() in SharedBlobAllocator
* remove unnecessary check
* Delimeter is changed to ::
* stylefix
* added layout from string function, small improvements
* modified parsing to enable : in input parameters
* small fixes
* small fixes
* added missed blob allocation, fixes
* [TEST]added support for remote blobs
* fix remote blobs
* new inputs/files output format
* removed vectors resize which caused bugs
* made cl::Buffer type under ifdef, fix inputs filling
* changed batch() function to not throwing exceptions
* removed unused var
* fix code style
* replace empty name in input files with name from net input
* restored old behaviour for static models
* fix code style
* fix warning - made const iterator
* fix warning - remove reference in loop variable
* added random and image_info input types to -i, fix problem with layout
* replaced batch() with getBatchSize() in main
* fix layout, shape, tensor shape parameters parsing
* upd help messages for input, tensor shape and pcseq command
* added buffer for cl output blobs, small fixes
Signed-off-by: ivikhrev <ivan.vikhrev@intel.com>
* added legacy mode
* restore setBlob
* code style formatting
* move collecting latency for groups under flag
* removed not applicable layouts
* added hint to error message when wrong input name in -tensor_shape was specified
* added new metrics to statistics report
* Apply suggestions from code review
* fix binary blobs filling when layout is CN
* apply suggestions
* moved file in the right place after rebase
* improved -pcseq output
* updated args and readme
* removed TEMPLATE plugin registration
* fix -shape arg decsription
* enable providing several -i args as input
* renamed legacy_mode to inference_only and made it default for static models, renamed tensor_shape to data_shape
* upd readme
* use getBlob() in inference only mode
* fix old input type for static case
* fix typo
* upd readme
* move log about benchmark mode to the measuring perfomance step
* added class for latency metrics
* upd readme, fix typos, renamed funcs
* fix warning and upd parsing to avoid error with : in file paths
* fix error on centos : error: use of deleted function ‘std::basic_stringstream<char>::basic_stringstream(const std::basic_stringstream<char>&)
* added check for key in inputs
* renamed input to inputs
* adjust batch size for binary blobs
* replaced warning with exception in bench mode defining
* align measurement cycle with master
Co-authored-by: ivikhrev <ivan.vikhrev@intel.com>
* Refactor ONNX,IR,Common FrontEnds
* Rename frontends dir to frontend
* Rename frontend_manager, frontend_defs -> manager, defs; move Place, InputModel to src for Paddle FE
* remove exports for ir/input_model
* fix unit tests
* Fix unit tests
* revert ONNX_API/IR_API defines
* move manager.hpp to frontend dir
* delete common folder; rename defs to visibility; fix includes
* revert local changes
* codestyle
* fix build: set correct includes
* fix includes
* fix build; resolve review comments
* fix install paths
* correct installation folder for onnx_import
* use IE_THROW in ir frontend