OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
Go to file
Bartosz Sochacki 8b87e1a477
[GNA] Fix for concat layer with >2 inputs (#1475)
* Fix for concat layer with more than 2 inputs

Signed-off-by: Bartosz Sochacki <bartosz.sochacki@intel.com>

* Fixed check if affine is used for crop layer

Signed-off-by: Bartosz Sochacki <bartosz.sochacki@intel.com>

* code cleanup for fix affine layer check

Signed-off-by: Bartosz Sochacki <bartosz.sochacki@intel.com>

* added test for concat layer with multiple inputs

* simplified test to use less number of layers

* fixed code style

* fixed coding style

* addressed review comments and one more issue that appeared during testing

* fixed code style errors

* scale factor propagation for concat layer with multiple inputs

* fix for a case when all inputs to concat are activation layers

* fix for linux compilation - C++14 is not enabled and fails on lambda with auto parameters

* corrected current year in headers in concat multi input tests

* fixes for code review issues raised by Denis Orlov

* enabled integer mode computation in GNA concat multi input test

* removed 1 space per review comment

* a fix to fail when not all scale factors are equal

* added GNA_DEVICE_MODE config to concat multi input test

* corrected searching for a next input to concat layer

* changed selection of 2nd candidate for source quant value

* code style fix - else and brackets should be in the same line

* small code improvement

* fix for mixing line endings

* addressed with endless requantization loop and fixed failing tests
2020-09-09 14:55:07 +03:00
.ci/openvino-onnx Removed NGRAPH_IE_ENABLE flag because it is always ON if unit tests are enabled (#2003) 2020-09-01 06:03:59 +03:00
.github Update issue templates 2020-08-12 13:17:34 +03:00
cmake Simplified plugin interfaces (#1745) 2020-08-14 12:11:54 +03:00
docs GRU/RNN/LSTM sequence ops, reference implementations, single layer tests (#1594) 2020-09-08 10:31:44 +03:00
inference-engine [GNA] Fix for concat layer with >2 inputs (#1475) 2020-09-09 14:55:07 +03:00
model-optimizer Support python 3.8 by the Model Optimizer tool in default configuration (#2078) 2020-09-09 08:34:43 +03:00
ngraph Fixed AddressSanitizer issue (#2122) 2020-09-09 14:31:12 +03:00
openvino Add static library with nGraph reference implementations (#1810) 2020-08-17 19:43:11 +03:00
scripts Refactoring: install_openvino_dependencies.sh script (#2112) 2020-09-07 19:54:56 +03:00
tests Sort memcheck timeline report for failures (#2088) 2020-09-09 10:10:22 +03:00
tools Adds first inference time measurements in benchmark_app (#1487) 2020-07-27 16:45:07 +03:00
.gitattributes Doc Migration (master) (#1377) 2020-07-20 17:36:08 +03:00
.gitignore publish master branch snapshot, revision 8d31237e2c3f673cbb0f0ba110fc10f5cce1d2bb 2020-05-22 02:23:12 +03:00
.gitmodules add submodules for mkl-dnn, gflags and gtest 2020-05-21 23:00:55 +03:00
azure-pipelines.yml Azure CI: Enable all MklDnnFunctionalTests (#1881) 2020-08-20 20:53:01 +03:00
build-instruction.md Add python executable for RPI compilation Docker (#1530) 2020-08-10 23:10:46 +03:00
CMakeLists.txt Removed NGRAPH_IE_ENABLE flag because it is always ON if unit tests are enabled (#2003) 2020-09-01 06:03:59 +03:00
CODEOWNERS Added code owners for scripts folder (#2130) 2020-09-08 17:23:27 +03:00
CONTRIBUTING_DOCS.md docs contribution guides (#1535) 2020-08-07 15:33:11 +03:00
CONTRIBUTING.md Create CONTRIBUTING.md 2020-05-19 19:04:27 +03:00
get-started-linux.md Separate MO configuration for TensorFlow 2 model conversion (#1685) 2020-08-11 18:02:05 +03:00
install_dependencies.sh [Docs] Fixes in readme files: (#750) 2020-06-03 20:14:35 +03:00
Jenkinsfile [Jenkinsfile] Add failFast parameter (#721) 2020-06-02 20:22:25 +03:00
LICENSE Publishing R3 2018-10-16 13:45:03 +03:00
README.md docs contribution guides (#1535) 2020-08-07 15:33:11 +03:00

OpenVINO™ Toolkit - Deep Learning Deployment Toolkit repository

Stable release Apache License Version 2.0

This toolkit allows developers to deploy pre-trained deep learning models through a high-level C++ Inference Engine API integrated with application logic.

This open source version includes two components: namely Model Optimizer and Inference Engine, as well as CPU, GPU and heterogeneous plugins to accelerate deep learning inferencing on Intel® CPUs and Intel® Processor Graphics. It supports pre-trained models from the Open Model Zoo, along with 100+ open source and public models in popular formats such as Caffe*, TensorFlow*, MXNet* and ONNX*.

Repository components:

License

Deep Learning Deployment Toolkit is licensed under Apache License Version 2.0. By contributing to the project, you agree to the license and copyright terms therein and release your contribution under these terms.

Documentation

How to Contribute

See CONTRIBUTING for contribution to the code. See CONTRIBUTING_DOCS for contribution to the documentation. Thank you!

Support

Please report questions, issues and suggestions using:


* Other names and brands may be claimed as the property of others.