Files
openvino/docs/install_guides/installing-openvino-macos.md
Yuan Xu e71a05d94d Minor updates to Linux/macOS/Windows installation guide (#10955)
* Add Overview page

* Revert "Add Overview page"

* CVS-71745

* fix links & errors

* fix formatting

* update bullet points
2022-03-15 12:06:30 +03:00

9.0 KiB
Raw Blame History

Install and Configure Intel® Distribution of OpenVINO™ toolkit for macOS

Note

: Since the OpenVINO™ 2022.1 release, the following development tools: Model Optimizer, Post-Training Optimization Tool, Model Downloader and other Open Model Zoo tools, Accuracy Checker, and Annotation Converter are not part of the installer. These tools are now only available on pypi.org.

Note

: The Intel® Distribution of OpenVINO™ toolkit is supported on macOS version 10.15 with Intel® processor-based machines.

System Requirements

@sphinxdirective .. tab:: Operating Systems

macOS 10.15

.. tab:: Hardware

Optimized for these processors:

  • 6th to 12th generation Intel® Core™ processors and Intel® Xeon® processors
  • 3rd generation Intel® Xeon® Scalable processor (formerly code named Cooper Lake)
  • Intel® Xeon® Scalable processor (formerly Skylake and Cascade Lake)
  • Intel® Neural Compute Stick 2

.. note:: The current version of the Intel® Distribution of OpenVINO™ toolkit for macOS supports inference on Intel CPUs and Intel® Neural Compute Stick 2 devices only.

.. tab:: Software Requirements

  • CMake 3.13 or higher <https://cmake.org/download/>_ (choose "macOS 10.13 or later"). Add /Applications/CMake.app/Contents/bin to path (for default install).
  • Python 3.6 - 3.9 <https://www.python.org/downloads/mac-osx/>_ (choose 3.6 - 3.9). Install and add to path.
  • Apple Xcode Command Line Tools. In the terminal, run xcode-select --install from any directory
  • (Optional) Apple Xcode IDE (not required for OpenVINO™, but useful for development)

@endsphinxdirective

Overview

This guide provides step-by-step instructions on how to install the Intel® Distribution of OpenVINO™ toolkit for macOS. The following steps will be covered:

  1. Install the Intel® Distribution of OpenVINO™ Toolkit
  2. Configure the Environment
  3. Download additional components (Optional)
  4. Configure the Intel® Neural Compute Stick 2 (Optional)
  5. Whats next?

Step 1: Install the Intel® Distribution of OpenVINO™ Toolkit Core Components

  1. Download the Intel® Distribution of OpenVINO™ toolkit package file from Intel® Distribution of OpenVINO™ toolkit for macOS. Select the Intel® Distribution of OpenVINO™ toolkit for macOS package from the dropdown menu.

  2. Go to the directory where you downloaded the Intel® Distribution of OpenVINO™ toolkit. This document assumes this is your Downloads directory. By default, the disk image file is saved as m_openvino_toolkit_p_<version>.dmg.

  3. Double-click the m_openvino_toolkit_p_<version>.dmg file to mount. The disk image is mounted to /Volumes/m_openvino_toolkit_p_<version> and automatically opens in a separate window.

  4. Run the installation wizard application bootstrapper.app. You should see the following dialog box open up:

    @sphinxdirective

    .. image:: _static/images/openvino-install.png :width: 400px :align: center

    @endsphinxdirective

  5. Follow the instructions on your screen. During the installation you will be asked to accept the license agreement. Your acceptance is required to continue. Click on the image to see the details.

    By default, the Intel® Distribution of OpenVINO™ is installed in the following directory, referred to as <INSTALL_DIR> elsewhere in the documentation:

    /opt/intel/openvino_<version>/

    For simplicity, a symbolic link to the latest installation is also created: /opt/intel/openvino_2022/.

To check Release Notes please visit: Release Notes.

The core components are now installed. Continue to the next section to configure environment.

Step 2: Configure the Environment

You must update several environment variables before you can compile and run OpenVINO™ applications. Set environment variables as follows:

source <INSTALL_DIR>/setupvars.sh

If you have more than one OpenVINO™ version on your machine, you can easily switch its version by sourcing setupvars.sh of your choice.

Note

: You can also run this script every time when you start new terminal session. Open ~/.bashrc in your favorite editor, and add source <INSTALL_DIR>/setupvars.sh. Next time when you open a terminal, you will see [setupvars.sh] OpenVINO™ environment initialized. Changing .bashrc is not recommended when you have many OpenVINO™ versions on your machine and want to switch among them, as each may require different setup.

The environment variables are set. Continue to the next section if you want to download any additional components.

Step 3 (Optional): Download Additional Components

Note

: Since the OpenVINO™ 2022.1 release, the following development tools: Model Optimizer, Post-Training Optimization Tool, Model Downloader and other Open Model Zoo tools, Accuracy Checker, and Annotation Converter are not part of the installer. The OpenVINO™ Development Tools can only be installed via PyPI now. See Install OpenVINO™ Development Tools for detailed steps.

@sphinxdirective

.. dropdown:: OpenCV

OpenCV is necessary to run demos from Open Model Zoo (OMZ). Some OpenVINO samples and demos also use OpenCV as a dependency. The Intel® Distribution of OpenVINO™ provides a script to install OpenCV: <INSTALL_DIR>/extras/scripts/download_opencv.sh.

.. note:: Make sure you have 2 prerequisites installed: curl and tar.

Depending on how you have installed the Intel® Distribution of OpenVINO™, the script should be run either as root or regular user. After the execution of the script, you will find OpenCV extracted to <INSTALL_DIR>/extras/opencv.

@endsphinxdirective

Step 4 (Optional): Configure the Intel® Neural Compute Stick 2

@sphinxdirective

If you want to run inference on Intel® Neural Compute Stick 2 use the following instructions to setup the device: :ref:NCS2 Setup Guide <ncs guide macos>.

@endsphinxdirective

Step 5: What's next?

Now you are ready to try out the toolkit. You can use the following tutorials to write your applications using Python and C++.

Developing in Python:

Developing in C++:

  • [Image Classification Async C++ Sample](@ref openvino_inference_engine_samples_classification_sample_async_README)
  • [Hello Classification C++ Sample](@ref openvino_inference_engine_samples_hello_classification_README)
  • [Hello Reshape SSD C++ Sample](@ref openvino_inference_engine_samples_hello_reshape_ssd_README)

Uninstall the Intel® Distribution of OpenVINO™ Toolkit

To uninstall the toolkit, follow the steps on the Uninstalling page.

@sphinxdirective .. raw:: html

@endsphinxdirective

@sphinxdirective

.. dropdown:: Additional Resources

  • Convert models for use with OpenVINO™: :ref:Model Optimizer Developer Guide <deep learning model optimizer>
  • Write your own OpenVINO™ applications: :ref:OpenVINO™ Runtime User Guide <deep learning inference engine>
  • Information on sample applications: :ref:OpenVINO™ Toolkit Samples Overview <code samples>
  • Information on a supplied set of models: :ref:Overview of OpenVINO™ Toolkit Pre-Trained Models <model zoo>
  • IoT libraries and code samples in the GitHUB repository: Intel® IoT Developer Kit_

To learn more about converting models from specific frameworks, go to:

  • :ref:Convert Your Caffe Model <openvino_docs_MO_DG_prepare_model_convert_model_Convert_Model_From_Caffe>
  • :ref:Convert Your TensorFlow Model <openvino_docs_MO_DG_prepare_model_convert_model_Convert_Model_From_TensorFlow>
  • :ref:Convert Your MXNet Model <openvino_docs_MO_DG_prepare_model_convert_model_Convert_Model_From_MxNet>
  • :ref:Convert Your Kaldi Model <openvino_docs_MO_DG_prepare_model_convert_model_Convert_Model_From_Kaldi>
  • :ref:Convert Your ONNX Model <openvino_docs_MO_DG_prepare_model_convert_model_Convert_Model_From_ONNX>

.. _Intel® IoT Developer Kit: https://github.com/intel-iot-devkit

@endsphinxdirective