* Doc Migration from Gitlab (#1289) * doc migration * fix * Update FakeQuantize_1.md * Update performance_benchmarks.md * Updates graphs for FPGA * Update performance_benchmarks.md * Change DL Workbench structure (#1) * Changed DL Workbench structure * Fixed tags * fixes * Update ie_docs.xml * Update performance_benchmarks_faq.md * Fixes in DL Workbench layout * Fixes for CVS-31290 * [DL Workbench] Minor correction * Fix for CVS-30955 * Added nGraph deprecation notice as requested by Zoe * fix broken links in api doxy layouts * CVS-31131 fixes * Additional fixes * Fixed POT TOC * Update PAC_Configure.md PAC DCP 1.2.1 install guide. * Update inference_engine_intro.md * fix broken link * Update opset.md * fix * added opset4 to layout * added new opsets to layout, set labels for them * Update VisionAcceleratorFPGA_Configure.md Updated from 2020.3 to 2020.4 Co-authored-by: domi2000 <domi2000@users.noreply.github.com>
5.5 KiB
Inference Engine development configuration document
To create MakeFiles use following process or run build-after-clone.sh script located in the root folder if you use Ubuntu 16.04. To create Visual Studio project run create_vs_proj_x64.cmd from scripts folder.
Setting up the environment for development
- Update/init submodules bu running
git submodule init
git submodule update --recursive
-
Install Git LFS extension. It's required to download models from the repo Below is step by step guide to install Git LFS.
2.1 Linux
wget https://github.com/git-lfs/git-lfs/releases/download/v2.3.4/git-lfs-linux-amd64-2.3.4.tar.gz tar xf git-lfs-linux-amd64-2.3.4.tar.gz cd git-lfs-2.3.4 sudo PREFIX=/usr/ ./install.sh git config --global http.sslverify false2.1 Windows 2.1.1 Download Git LFS and install it. 2.1.2 Run console command
bash git config --global http.sslverify falseNote
: HTTPS protocol is used to download files by Git LFS. You either have to disable HTTPS proxy for local resources like GitLab server gitlab-icv.inn.intel.com by setting
no_proxy=localhost,gitlab-icv.inn.intel.comor switch tohttp://proxy-chain.intel.com:911proxy server, because it disables proxy for local servers automatically. -
Use Cmake to fetch project dependencies and create Unix makefiles
mkdir build cd build
There are number of options which turn on some components during builds and initiate downloading of the models
-DENABLE_TESTS=ON - to build functional and behavior tests
this will copy necessary dependencies to ./temp folder, or to ENV.DL_SDK_TEMP folder if environment variable set
-DENABLE_FUNCTIONAL_TESTS=ON - to build functional tests
-DCMAKE_BUILD_TYPE=Debug/Release - to point debug or release configuration. Missing this option will generate something between
Release and Debug and you might be surprised by certain aspects of the compiled binaries
-DENABLE_PRIVATE_MODELS=ON - copy private models from https://gitlab-icv.inn.intel.com/inference-engine-models/private-ir with restricted access
The full command line enough for development is following:
cmake -DENABLE_TESTS=ON -DENABLE_FUNCTIONAL_TESTS=ON -DCMAKE_BUILD_TYPE=Debug ..
The full command line enough for validation before push to the server
cmake -DENABLE_TESTS=ON -DENABLE_FUNCTIONAL_TESTS=ON -DCMAKE_BUILD_TYPE=Release ..
- Build project and tests:
make -j16
-
To build documentation: a. Install doxygen and graphviz:
apt-get install doxygen && apt-get install graphviz && apt-get install texliveb. Go to the documentation build directory:
cd to scripts/build_documentationc. Run the
build_docs.shscript: * To build the documentation set that includes documentation from the current branch of theinference-enginerepo and specific branches of theopenvino-documentation,modelsandmodel-optimizer-tensorflowrepos, specify three branches as parameters:sh ./build_docs.sh ovinodoc:<OPENVINO_BRANCH> models:<MODELS_BRANCH> mo:<MO_BRANCH>* To build the documentation set that includes only documentation from the current branch of theinference-enginerepo, run the script with no parameters:sh ./build_docs.shNote
: You should run the script either with specifying all three parameters or without any parameters.
d. Find the generated documentation in the
root_directory/docdirectoryNote
: If you make any changes in the documentation source files, it is recommended to cleanup the documentation build directory and continue with step 3:
cd scripts/build_documentation ./clean.shNote
: The scripts for building documentation use SSH for cloning repositories. Please, make sure that you have added your SSH key to git-lab. For more information about it, please visit the instructions page
Compilers supported and verified
All others may be compatible but Inference Engine does not guarantee that.
- Linux : gcc(5.4)*, clang(3.9)
- MacOS : gcc(5.4), clang(3.9)*
- Windows: MSVC(14), ICC(17.0)* * - is target compiler for platform and used for public external drops
TeamCity CI
TeamCity CI server is available here
To get access to the server, go to AGS and search "DevTools -- INDE xOS - Project Developer".
Troubleshooting steps
- Issue: Build of the "mkldnn" project failed on Windows with "Error MSB6003 The specified task executable "cmd.exe" could not be run. The working directory "\mkl\tools" does not exist". Solution: open InferenceEngine.sln -> goto "mkldnn" project Properties -> Configuration Properties -> Intel Performance Libraries -> Use Intel MKL -> choose "No"