30 KiB
Inference Engine API Changes History
The sections below contain detailed list of changes made to the Inference Engine API in recent releases.
Deprecation Notice
| Deprecation Begins | June 1, 2020 |
| Removal Date | December 1, 2020 |
Starting with the OpenVINO™ toolkit 2020.2 release, all of the features previously available through nGraph have been merged into the OpenVINO™ toolkit. As a result, all the features previously available through ONNX RT Execution Provider for nGraph have been merged with ONNX RT Execution Provider for OpenVINO™ toolkit.
Therefore, ONNX RT Execution Provider for nGraph will be deprecated starting June 1, 2020 and will be completely removed on December 1, 2020. Users are recommended to migrate to the ONNX RT Execution Provider for OpenVINO™ toolkit as the unified solution for all AI inferencing on Intel® hardware.
2021.1
Deprecated API
Utility functions to convert Unicode paths
- InferenceEngine::stringToFileName - use OS-specific native conversion functions
- InferenceEngine::fileNameToString - use OS-specific native conversion functions
Removed API
Plugin API:
- InferenceEngine::InferencePlugin C++ plugin wrapper class
- InferenceEngine::IInferencePlugin plugin interface
- InferenceEngine::PluginDispatcher class
- InferenceEngine::InferenceEnginePluginPtr typedef
- InferenceEngine::ICNNNetReader reader interface
- InferenceEngine::CNNNetReader class
Extensibility API:
- InferenceEngine::ILayerImplFactory class
- InferenceEngine::IShapeInferImpl class
- InferenceEngine::IShapeInferExtension class
- InferenceEngine::IExtension::getFactoryFor(ILayerImplFactory*& factory, const CNNLayer* cnnLayer, ResponseDesc* resp) noexcept method
- InferenceEngine::IExtension::getPrimitiveTypes(char**& types, unsigned int& size, ResponseDesc* resp) noexcept method
- InferenceEngine::ShapeInferImpl class
- InferenceEngine::Extension::getFactoryFor(ILayerImplFactory*& factory, const CNNLayer* cnnLayer, ResponseDesc* resp) noexcept method
- InferenceEngine::Extension::getPrimitiveTypes(char**& types, unsigned int& size, ResponseDesc* resp) noexcept method
Network API:
- InferenceEngine::details::CNNNetworkIterator class
- InferenceEngine::CNNNetwork::getPrecision() const method
- InferenceEngine::CNNNetwork::getLayerByName(const char* layerName) const method
- InferenceEngine::CNNNetwork::size() const method
- InferenceEngine::CNNNetwork::begin() const method
- InferenceEngine::CNNNetwork::end() const method
- InferenceEngine::CNNNetwork::AddExtension(const IShapeInferExtensionPtr& extension) method
- InferenceEngine::ICNNNetwork::getPrecision() const noexcept method
- InferenceEngine::ICNNNetwork::getName(char* pName, size_t len) const noexcept method
- InferenceEngine::ICNNNetwork::getData(const char* dname) noexcept method
- InferenceEngine::ICNNNetwork::addLayer(const CNNLayerPtr& layer) noexcept method
- InferenceEngine::ICNNNetwork::getLayerByName(const char* layerName, CNNLayerPtr& out, ResponseDesc* resp) const noexcept method
- InferenceEngine::ICNNNetwork::AddExtension(const IShapeInferExtensionPtr& extension, ResponseDesc* resp) noexcept method
- InferenceEngine::ICNNNetwork::getStats(ICNNNetworkStats** stats, ResponseDesc* resp) const noexcept method
- InferenceEngine::ICNNNetworkStats class
- InferenceEngine::NetworkNodeStats class
- InferenceEngine::Data::getCreatorLayer() method
- InferenceEngine::Data::getInputTo() method
- InferenceEngine::LayerParams class
Layer API:
- InferenceEngine::CNNLayer class
- InferenceEngine::WeightableLayer class
- InferenceEngine::BatchNormalizationLayer class
- InferenceEngine::BatchToSpaceLayer class
- InferenceEngine::BinaryConvolutionLayer class
- InferenceEngine::BroadcastLayer class
- InferenceEngine::BucketizeLayer class
- InferenceEngine::ClampLayer class
- InferenceEngine::ConcatLayer class
- InferenceEngine::ConvolutionLayer class
- InferenceEngine::CropLayer class
- InferenceEngine::DeconvolutionLayer class
- InferenceEngine::DeformableConvolutionLayer class
- InferenceEngine::DepthToSpaceLayer class
- InferenceEngine::EltwiseLayer class
- InferenceEngine::ExperimentalDetectronPriorGridGenerator class
- InferenceEngine::ExperimentalDetectronPriorGridGeneratorLayer class
- InferenceEngine::ExperimentalSparseWeightedReduceLayer class
- InferenceEngine::FillLayer class
- InferenceEngine::FullyConnectedLayer class
- InferenceEngine::GRNLayer class
- InferenceEngine::GRUCell class
- InferenceEngine::GatherLayer class
- InferenceEngine::GemmLayer class
- InferenceEngine::LSTMCell class
- InferenceEngine::MVNLayer class
- InferenceEngine::MathLayer class
- InferenceEngine::NonMaxSuppression class
- InferenceEngine::NormLayer class
- InferenceEngine::OneHotLayer class
- InferenceEngine::PReLULayer class
- InferenceEngine::PadLayer class
- InferenceEngine::PoolingLayer class
- InferenceEngine::PowerLayer class
- InferenceEngine::QuantizeLayer class
- InferenceEngine::RNNCell class
- InferenceEngine::RNNCellBase class
- InferenceEngine::RNNSequenceLayer class
- InferenceEngine::RangeLayer class
- InferenceEngine::ReLU6Layer class
- InferenceEngine::ReLULayer class
- InferenceEngine::ReduceLayer class
- InferenceEngine::ReshapeLayer class
- InferenceEngine::ReverseSequenceLayer class
- InferenceEngine::ScaleShiftLayer class
- InferenceEngine::ScatterLayer class
- InferenceEngine::SelectLayer class
- InferenceEngine::ShuffleChannelsLayer class
- InferenceEngine::SoftMaxLayer class
- InferenceEngine::SpaceToBatchLayer class
- InferenceEngine::SpaceToDepthLayer class
- InferenceEngine::SparseFillEmptyRowsLayer class
- InferenceEngine::SparseSegmentReduceLayer class
- InferenceEngine::SparseToDenseLayer class
- InferenceEngine::SplitLayer class
- InferenceEngine::StridedSliceLayer class
- InferenceEngine::TensorIterator class
- InferenceEngine::TileLayer class
- InferenceEngine::TopKLayer class
- InferenceEngine::UniqueLayer class
2020.4
New API
CPU Plugin API:
- InferenceEngine::PluginConfigParams::KEY_ENFORCE_BF16 config key
Metrics and values for Query API:
- METRIC_KEY(OPTIMIZATION_CAPABILITIES)
- METRIC_VALUE(BF16)
Deprecated API
Myriad Plugin API:
- VPU_CONFIG_KEY(IGNORE_IR_STATISTIC)
Removed API
Inference Engine NN Builder API:
- InferenceEngine::Builder::EltwiseLayer
- InferenceEngine::Builder::MemoryLayer
- InferenceEngine::Builder::ROIPoolingLayer
- InferenceEngine::Builder::DeconvolutionLayer
- InferenceEngine::Builder::ReLULayer
- InferenceEngine::Builder::TanHLayer
- InferenceEngine::Builder::InputLayer
- InferenceEngine::Builder::PoolingLayer
- InferenceEngine::Builder::CropLayer
- InferenceEngine::Builder::GRUSequenceLayer
- InferenceEngine::Builder::NormLayer
- InferenceEngine::Builder::LSTMSequenceLayer
- InferenceEngine::Builder::ClampLayer
- InferenceEngine::Builder::PSROIPoolingLayer
- InferenceEngine::Builder::Layer
- InferenceEngine::Builder::RNNSequenceLayer
- InferenceEngine::Builder::ReorgYoloLayer
- InferenceEngine::Builder::NormalizeLayer
- InferenceEngine::Builder::PriorBoxClusteredLayer
- InferenceEngine::Builder::MVNLayer
- InferenceEngine::Builder::PermuteLayer
- InferenceEngine::Builder::SimplerNMSLayer
- InferenceEngine::Builder::ConstLayer
- InferenceEngine::Builder::DeformableConvolutionLayer
- InferenceEngine::Builder::FullyConnectedLayer
- InferenceEngine::Builder::PriorBoxLayer
- InferenceEngine::Builder::SoftMaxLayer
- InferenceEngine::Builder::OutputLayer
- InferenceEngine::Builder::TileLayer
- InferenceEngine::Builder::SplitLayer
- InferenceEngine::Builder::PReLULayer
- InferenceEngine::Builder::RegionYoloLayer
- InferenceEngine::Builder::ReshapeLayer
- InferenceEngine::Builder::ConvolutionLayer
- InferenceEngine::Builder::DetectionOutputLayer
- InferenceEngine::Builder::ConcatLayer
- InferenceEngine::Builder::ELULayer
- InferenceEngine::Builder::GRNLayer
- InferenceEngine::Builder::LRNLayer
- InferenceEngine::Builder::ArgMaxLayer
- InferenceEngine::Builder::ReLU6Layer
- InferenceEngine::Builder::ScaleShiftLayer
- InferenceEngine::Builder::ProposalLayer
- InferenceEngine::Builder::SigmoidLayer
- InferenceEngine::Builder::ResampleLayer
- InferenceEngine::Builder::CTCGreedyDecoderLayer
- InferenceEngine::Builder::BatchNormalizationLayer
- InferenceEngine::Builder::LayerDecorator
- InferenceEngine::Builder::PowerLayer
- InferenceEngine::Builder::Network
- InferenceEngine::Builder::PortInfo
- InferenceEngine::Builder::Connection
- InferenceEngine::Builder::PortData
- InferenceEngine::Builder::Port
- InferenceEngine::Builder::ILayer
- InferenceEngine::Builder::INetworkIterator
- InferenceEngine::Builder::INetwork
- InferenceEngine::Builder::ILayer
2020.2
New API
Extensibility API:
- InferenceEngine::IExtension::getImplTypes(const std::shared_ptrngraph::Node& node) method
- InferenceEngine::IExtension::getImplementation(const std::shared_ptrngraph::Node& node, const std::string& implType) method
Deprecated API
Extensibility API:
- InferenceEngine::ILayerImplFactory class
- InferenceEngine::IShapeInferImpl class
- InferenceEngine::IShapeInferImpl class
- InferenceEngine::IShapeInferExtension class
- InferenceEngine::IExtension::getFactoryFor(ILayerImplFactory*& factory, const CNNLayer* cnnLayer, ResponseDesc* resp) noexcept method
- InferenceEngine::IExtension::getPrimitiveTypes(char**& types, unsigned int& size, ResponseDesc* resp) noexcept method
- InferenceEngine::ShapeInferImpl class
- InferenceEngine::Extension::getFactoryFor(ILayerImplFactory*& factory, const CNNLayer* cnnLayer, ResponseDesc* resp) noexcept method
- InferenceEngine::Extension::getPrimitiveTypes(char**& types, unsigned int& size, ResponseDesc* resp) noexcept method
Network API:
- InferenceEngine::details::CNNNetworkIterator class
- InferenceEngine::CNNNetwork::getPrecision() const method
- InferenceEngine::CNNNetwork::getLayerByName(const char* layerName) const method
- InferenceEngine::CNNNetwork::size() const method
- InferenceEngine::CNNNetwork::begin() const method
- InferenceEngine::CNNNetwork::end() const method
- InferenceEngine::CNNNetwork::AddExtension(const IShapeInferExtensionPtr& extension) method
- InferenceEngine::ICNNNetwork::getPrecision() const noexcept method
- InferenceEngine::ICNNNetwork::getName(char* pName, size_t len) const noexcept method
- InferenceEngine::ICNNNetwork::getData(const char* dname) noexcept method
- InferenceEngine::ICNNNetwork::addLayer(const CNNLayerPtr& layer) noexcept method
- InferenceEngine::ICNNNetwork::getLayerByName(const char* layerName, CNNLayerPtr& out, ResponseDesc* resp) const noexcept method
- InferenceEngine::ICNNNetwork::AddExtension(const IShapeInferExtensionPtr& extension, ResponseDesc* resp) noexcept method
- InferenceEngine::ICNNNetwork::getStats(ICNNNetworkStats** stats, ResponseDesc* resp) const noexcept method
- InferenceEngine::ICNNNetworkStats class
- InferenceEngine::NetworkNodeStats class
- InferenceEngine::Data::getCreatorLayer() method
- InferenceEngine::Data::getInputTo() method
- InferenceEngine::LayerParams class
Layer API:
- InferenceEngine::CNNLayer class
- InferenceEngine::WeightableLayer class
- InferenceEngine::BatchNormalizationLayer class
- InferenceEngine::BatchToSpaceLayer class
- InferenceEngine::BinaryConvolutionLayer class
- InferenceEngine::BroadcastLayer class
- InferenceEngine::BucketizeLayer class
- InferenceEngine::ClampLayer class
- InferenceEngine::ConcatLayer class
- InferenceEngine::ConvolutionLayer class
- InferenceEngine::CropLayer class
- InferenceEngine::DeconvolutionLayer class
- InferenceEngine::DeformableConvolutionLayer class
- InferenceEngine::DepthToSpaceLayer class
- InferenceEngine::EltwiseLayer class
- InferenceEngine::ExperimentalDetectronPriorGridGenerator class
- InferenceEngine::ExperimentalDetectronPriorGridGeneratorLayer class
- InferenceEngine::ExperimentalSparseWeightedReduceLayer class
- InferenceEngine::FillLayer class
- InferenceEngine::FullyConnectedLayer class
- InferenceEngine::GRNLayer class
- InferenceEngine::GRUCell class
- InferenceEngine::GatherLayer class
- InferenceEngine::GemmLayer class
- InferenceEngine::LSTMCell class
- InferenceEngine::MVNLayer class
- InferenceEngine::MathLayer class
- InferenceEngine::NonMaxSuppression class
- InferenceEngine::NormLayer class
- InferenceEngine::OneHotLayer class
- InferenceEngine::PReLULayer class
- InferenceEngine::PadLayer class
- InferenceEngine::PoolingLayer class
- InferenceEngine::PowerLayer class
- InferenceEngine::QuantizeLayer class
- InferenceEngine::RNNCell class
- InferenceEngine::RNNCellBase class
- InferenceEngine::RNNSequenceLayer class
- InferenceEngine::RangeLayer class
- InferenceEngine::ReLU6Layer class
- InferenceEngine::ReLULayer class
- InferenceEngine::ReduceLayer class
- InferenceEngine::ReshapeLayer class
- InferenceEngine::ReverseSequenceLayer class
- InferenceEngine::ScaleShiftLayer class
- InferenceEngine::ScatterLayer class
- InferenceEngine::SelectLayer class
- InferenceEngine::ShuffleChannelsLayer class
- InferenceEngine::SoftMaxLayer class
- InferenceEngine::SpaceToBatchLayer class
- InferenceEngine::SpaceToDepthLayer class
- InferenceEngine::SparseFillEmptyRowsLayer class
- InferenceEngine::SparseSegmentReduceLayer class
- InferenceEngine::SparseToDenseLayer class
- InferenceEngine::SplitLayer class
- InferenceEngine::StridedSliceLayer class
- InferenceEngine::TensorIterator class
- InferenceEngine::TileLayer class
- InferenceEngine::TopKLayer class
- InferenceEngine::UniqueLayer class
2020.1
New API
Integration with ngraph API:
- InferenceEngine::CNNNetwork(const std::shared_ptrngraph::Function& network) ctor from ngraph::Function
- InferenceEngine::CNNNetwork::getFunction() const noexcept method
- InferenceEngine::ICNNNetwork::getFunction() const noexcept method
- InferenceEngine::Parameter(const std::shared_ptrngraph::Variant& var) ctor
- InferenceEngine::Parameter::asVariant() const method
- InferenceEngine::Parameter::operator std::shared_ptrngraph::Variant() const operator
- InferenceEngine::Core::ReadNetwork(const std::wstring& modelPath, const std::wstring& binPath) method
- InferenceEngine::Core::ReadNetwork(const std::string& modelPath, const std::string& binPath = "") method
- InferenceEngine::Core::ReadNetwork(const std::string& model, const Blob::CPtr& weights) method
- InferenceEngine::Code::AddExtension(const IExtensionPtr& extension) method
- InferenceEngine::IExtension::getOpSets() method
Offline compilation: import / export to std::stream:
- InferenceEngine::ExecutableNetwork::Export(std::ostream& networkModel) method
- InferenceEngine::Core::ImportNetwork(std::istream& networkModel, const std::string& deviceName = {}, const std::map<std::string, std::string>& config = {}) method
- InferenceEngine::IExecutableNetwork::Export(std::ostream& networkModel, ResponseDesc *resp) noexcept method
RemoteBlob accelerator memory sharing API:
- InferenceEngine::RemoteContext class
- InferenceEngine::RemoteBlob class
- InferenceEngine::Core::CreateContext(const std::string& deviceName, const ParamMap& params) method
- InferenceEngine::Core::GetDefaultContext(const std::string& deviceName) method
- InferenceEngine::Core::LoadNetwork(CNNNetwork network, RemoteContext::Ptr context, const std::map<std::string, std::string>& config = std::map<std::string, std::string>()) method
GNA firmware model image generation:
- GNA_CONFIG_KEY(FIRMWARE_MODEL_IMAGE_GENERATION) config key
- GNA_CONFIG_VALUE(GEN) value
- GNA_CONFIG_VALUE(GEN_EXACT) value
- GNA_CONFIG_VALUE(SSE) value
- GNA_CONFIG_VALUE(SSE_EXACT) value
- GNA_CONFIG_VALUE(AVX1) value
- GNA_CONFIG_VALUE(AVX1_EXACT) value
- GNA_CONFIG_VALUE(AVX2) value
- GNA_CONFIG_VALUE(AVX2_EXACT) value
MemoryBlob mapping of memory to the user space:
- InferenceEngine::MemoryBlob::rwmap() noexcept method
- InferenceEngine::MemoryBlob::rmap() noexcept method
- InferenceEngine::MemoryBlob::wmap() noexcept method
Memory interoperability on acceleration devices. General classes and GPU helper functions
- InferenceEngine::RemoteBlob class
- InferenceEngine::RemoteContext class
- InferenceEngine::Core::CreateContext(const std::string& deviceName, const ParamMap& params) method
- InferenceEngine::Core::GetDefaultContext(const std::string& deviceName) method
- InferenceEngine::make_shared_blob(const TensorDesc& desc, RemoteContext::Ptr ctx) function
- InferenceEngine::gpu::make_shared_blob_nv12(size_t height, size_t width, RemoteContext::Ptr ctx, VASurfaceID nv12_surf) function
- InferenceEngine::gpu::make_shared_context(Core& core, std::string deviceName, VADisplay device) function
- InferenceEngine::gpu::make_shared_blob(const TensorDesc& desc, RemoteContext::Ptr ctx, VASurfaceID surface, uint32_t plane = 0) function
- InferenceEngine::gpu::make_shared_blob_nv12(RemoteContext::Ptr ctx, cl::Image2D& nv12_image_plane_y, cl::Image2D& nv12_image_plane_uv) function
- InferenceEngine::gpu::make_shared_context(Core& core, std::string deviceName, cl_context ctx) function
- InferenceEngine::gpu::make_shared_blob(const TensorDesc& desc, ClContext::Ptr ctx) function
- InferenceEngine::gpu::make_shared_blob(const TensorDesc& desc, RemoteContext::Ptr ctx, cl::Buffer& buffer) function
- InferenceEngine::gpu::make_shared_blob(const TensorDesc& desc, RemoteContext::Ptr ctx, cl_mem buffer) function
- InferenceEngine::gpu::make_shared_blob(const TensorDesc& desc, RemoteContext::Ptr ctx, cl::Image2D& image) function
Deprecated API
Inference Engine NN Builder API:
- InferenceEngine::Builder::EltwiseLayer
- InferenceEngine::Builder::MemoryLayer
- InferenceEngine::Builder::ROIPoolingLayer
- InferenceEngine::Builder::DeconvolutionLayer
- InferenceEngine::Builder::ReLULayer
- InferenceEngine::Builder::TanHLayer
- InferenceEngine::Builder::InputLayer
- InferenceEngine::Builder::PoolingLayer
- InferenceEngine::Builder::CropLayer
- InferenceEngine::Builder::GRUSequenceLayer
- InferenceEngine::Builder::NormLayer
- InferenceEngine::Builder::LSTMSequenceLayer
- InferenceEngine::Builder::ClampLayer
- InferenceEngine::Builder::PSROIPoolingLayer
- InferenceEngine::Builder::Layer
- InferenceEngine::Builder::RNNSequenceLayer
- InferenceEngine::Builder::ReorgYoloLayer
- InferenceEngine::Builder::NormalizeLayer
- InferenceEngine::Builder::PriorBoxClusteredLayer
- InferenceEngine::Builder::MVNLayer
- InferenceEngine::Builder::PermuteLayer
- InferenceEngine::Builder::SimplerNMSLayer
- InferenceEngine::Builder::ConstLayer
- InferenceEngine::Builder::DeformableConvolutionLayer
- InferenceEngine::Builder::FullyConnectedLayer
- InferenceEngine::Builder::PriorBoxLayer
- InferenceEngine::Builder::SoftMaxLayer
- InferenceEngine::Builder::OutputLayer
- InferenceEngine::Builder::TileLayer
- InferenceEngine::Builder::SplitLayer
- InferenceEngine::Builder::PReLULayer
- InferenceEngine::Builder::RegionYoloLayer
- InferenceEngine::Builder::ReshapeLayer
- InferenceEngine::Builder::ConvolutionLayer
- InferenceEngine::Builder::DetectionOutputLayer
- InferenceEngine::Builder::ConcatLayer
- InferenceEngine::Builder::ELULayer
- InferenceEngine::Builder::GRNLayer
- InferenceEngine::Builder::LRNLayer
- InferenceEngine::Builder::ArgMaxLayer
- InferenceEngine::Builder::ReLU6Layer
- InferenceEngine::Builder::ScaleShiftLayer
- InferenceEngine::Builder::ProposalLayer
- InferenceEngine::Builder::SigmoidLayer
- InferenceEngine::Builder::ResampleLayer
- InferenceEngine::Builder::CTCGreedyDecoderLayer
- InferenceEngine::Builder::BatchNormalizationLayer
- InferenceEngine::Builder::LayerDecorator
- InferenceEngine::Builder::PowerLayer
- InferenceEngine::Builder::Network
- InferenceEngine::Builder::PortInfo
- InferenceEngine::Builder::Connection
- InferenceEngine::Builder::PortData
- InferenceEngine::Builder::Port
- InferenceEngine::Builder::ILayer
- InferenceEngine::Builder::INetworkIterator
- InferenceEngine::Builder::INetwork
- InferenceEngine::Builder::ILayer
Plugin API:
- InferenceEngine::InferencePlugin C++ plugin wrapper class
- InferenceEngine::IInferencePlugin plugin interface
- InferenceEngine::PluginDispatcher class
- InferenceEngine::InferenceEnginePluginPtr typedef
- InferenceEngine::ICNNNetReader reader interface
- InferenceEngine::CNNNetReader class
Blob API:
- Blob::element_size() const noexcept method
- Blob::buffer() noexcept method
- Blob::cbuffer() noexcept method
- MemoryBlob::buffer() noexcept method
- MemoryBlob::cbuffer() noexcept method
Removed API
Removed all Inference Engine API which deprecated in 2019'R2
2019 R3
New API
New supported layers:
- InferenceEngine::SparseFillEmptyRowsLayer new class
- InferenceEngine::UniqueLayer new class
- InferenceEngine::NonMaxSuppressionLayer new class
- InferenceEngine::ScatterLayer new class
FPGA plugin streaming support:
- DLIA_METRIC_VALUE(INPUT_STREAMING) value to METRIC_KEY(OPTIMIZATION_CAPABILITIES)
- DLIA_CONFIG_KEY(ENABLE_STREAMING) config key
Removed API
- InferenceEngine::EltwiseLayer::Select from InferenceEngine::EltwiseLayer::eOperation enumeration
2019 R2
New API
Inference Engine Core API:
- Introduced InferenceEngine::Core high level class to manage devices
Query API extensions to InferenceEngine::ExecutableNetwork and InferenceEngine::IExecutableNetwork:
- InferenceEngine::ExecutableNetwork::SetConfig method
- InferenceEngine::ExecutableNetwork::GetConfig method
- InferenceEngine::ExecutableNetwork::GetMetric method
- InferenceEngine::IExecutableNetwork::SetConfig method
- InferenceEngine::IExecutableNetwork::GetConfig method
- InferenceEngine::IExecutableNetwork::GetMetric method
Metrics and values for Query API:
- METRIC_KEY(AVAILABLE_DEVICES)
- METRIC_KEY(SUPPORTED_METRICS)
- METRIC_KEY(SUPPORTED_CONFIG_KEYS)
- METRIC_KEY(FULL_DEVICE_NAME)
- METRIC_KEY(OPTIMIZATION_CAPABILITIES)
- METRIC_VALUE(FP32)
- METRIC_VALUE(FP16)
- METRIC_VALUE(INT8)
- METRIC_VALUE(BIN)
- METRIC_VALUE(WINOGRAD)
- DLIA_METRIC_VALUE(FP11)
- METRIC_KEY(RANGE_FOR_STREAMS)
- METRIC_KEY(NUMBER_OF_WAITING_INFER_REQUESTS)
- METRIC_KEY(NUMBER_OF_EXEC_INFER_REQUESTS)
- METRIC_KEY(DEVICE_THERMAL)
- METRIC_KEY(RANGE_FOR_ASYNC_INFER_REQUESTS)
- EXEC_NETWORK_METRIC_KEY(NETWORK_NAME)
- EXEC_NETWORK_METRIC_KEY(OPTIMAL_NUMBER_OF_INFER_REQUESTS)
Common API:
- CLDNN_CONFIG_KEY(INT8_ENABLED) config key
- CONFIG_KEY(GPU_THROUGHPUT_AUTO)
- CONFIG_KEY(GPU_THROUGHPUT_STREAMS)
- DLIA_CONFIG_KEY(IO_TRANSFORMATIONS_NATIVE) config key
- DLIA_CONFIG_KEY(DUMP_SUPPORTED_LAYERS_INFORMATION) config key
- GNA_CONFIG_VALUE(SW_FP32) config value for GNA_CONFIG_KEY(DEVICE_MODE) key
- MULTI_CONFIG_KEY(DEVICE_PRIORITIES) config key for
MULTIdevice - InferenceEngine::CNNNetReader::ReadNetwork(const std::wstring &filepath) new method
- InferenceEngine::CNNNetReader::ReadWeights(const std::wstring &filepath) new method
- InferenceEngine::ExecutableNetwork::ExecutableNetwork(IExecutableNetwork::Ptr actual, InferenceEnginePluginPtr plg) constructor with additional
plgparameter - InferenceEngine::InferRequest::InferRequest(IInferRequest::Ptr request, InferenceEnginePluginPtr plg) constructor with additional
plgparameter - InferenceEngine::Data::setName method
- InferenceEngine::QueryNetworkResult::supportedLayersMap
- InferenceEngine::Precision::I64 extension to InferenceEngine::Precision::ePrecision enumeration
New supported primitives:
- InferenceEngine::Builder::DeformableConvolutionLayer new class
- InferenceEngine::DeformableConvolutionLayer new class
- InferenceEngine::EltwiseLayer::Logical_NOT, InferenceEngine::EltwiseLayer::Mean, InferenceEngine::EltwiseLayer::Select extensions to InferenceEngine::EltwiseLayer::eOperation enumeration
- InferenceEngine::OneHotLayer new class
- InferenceEngine::SelectLayer new class
- InferenceEngine::BroadcastLayer new class
- InferenceEngine::MathLayer new class
- InferenceEngine::ReduceLayer new class
- InferenceEngine::TopKLayer new class
Extensions to Blob creation API:
- InferenceEngine::Blob::is method
- InferenceEngine::Blob::is const method
- InferenceEngine::Blob::as method
- InferenceEngine::Blob::as const method
- InferenceEngine::Blob::getAllocator abstract method
- InferenceEngine::Blob::getHandle abstract method
- InferenceEngine::MemoryBlob class
- InferenceEngine::ColorFormat enumeration
- InferenceEngine::PreProcessInfo::setColorFormat method
- InferenceEngine::PreProcessInfo::getColorFormat method
- InferenceEngine::CompoundBlob class to work with blobs consisting of several planes
- InferenceEngine::NV12Blob class representing NV12 blob with two planes
Deprecated API
The methods listed below are deprecated and will be removed in 2019 R4 release:
Common API:
- InferenceEngine::InputInfo::getInputPrecision method
- InferenceEngine::InputInfo::setInputPrecision method
- InferenceEngine::InputInfo::getDims method
- InferenceEngine::CNNLayer::GetParamsAsBool method
- InferenceEngine::CNNNetwork::CNNNetwork(ICNNNetwork* actual) constructor
- InferenceEngine::CNNNetwork::setTargetDevice method
- HETERO_CONFIG_KEY(DUMP_DLA_MESSAGES) config key
- InferenceEngine::ILayerImplFactory::getShapes method
- InferenceEngine::IShapeInferImpl::inferShapes(const std::vector&, const std::map<std::string, std::string>& , const std::map<std::string, Blob::Ptr>&, std::vector&, ResponseDesc*) method
- InferenceEngine::Data::setBatchSize method
- InferenceEngine::QueryNetworkResult::supportedLayers field
- InferenceEngine::ICNNNetwork::setBatchSize(const size_t size) method
- InferenceEngine::Blob::Resize method
- InferenceEngine::Blob::Reshape method
- InferenceEngine::TBlob::set method
InferenceEngine::IInferencePlugin and InferenceEngine:InferencePlugin obsolete methods:
- InferenceEngine::InferencePlugin::LoadNetwork(ICNNNetwork &network) method
- InferenceEngine::InferencePlugin::Infer method
- InferenceEngine::InferencePlugin::GetPerformanceCounts method
- InferenceEngine::InferencePlugin::QueryNetwork(const ICNNNetwork &network, QueryNetworkResult &res) const method
- InferenceEngine::IInferencePlugin::LoadNetwork(ICNNNetwork &network, ResponseDesc *resp) method
- InferenceEngine::IInferencePlugin::Infer(const Blob &input, Blob &result, ResponseDesc *resp) method
- InferenceEngine::IInferencePlugin::Infer(const BlobMap &input, BlobMap &result, ResponseDesc *resp) method
- InferenceEngine::IInferencePlugin::GetPerformanceCounts method
- InferenceEngine::IInferencePlugin::QueryNetwork(const ICNNNetwork& network, QueryNetworkResult& res) const method
Fields in InferenceEngine::Data class are replaced with appropriate methods:
- InferenceEngine::Data::precision field
- InferenceEngine::Data::layout field
- InferenceEngine::Data::dims field
- InferenceEngine::Data::creatorLayer field
- InferenceEngine::Data::name field
- InferenceEngine::Data::inputTo field
- InferenceEngine::Data::userObject field
Heterogeneous plugin:
- InferenceEngine::IHeteroDeviceLoader class
- InferenceEngine::IHeteroInferencePlugin class
- InferenceEngine::HeteroPluginPtr class
- operator InferenceEngine::InferencePlugin::HeteroPluginPtr operator
Blob creation API with dimensions in reverse order:
- InferenceEngine::Blob::Blob(Precision p) constructor
- InferenceEngine::Blob::Blob(Precision p, Layout l) constructor
- InferenceEngine::Blob::Blob(Precision p, const SizeVector &dims) constructor
- InferenceEngine::Blob::Blob(Precision p, Layout l, const SizeVector &dims) constructor
- InferenceEngine::TBlob::TBlob(Precision p, Layout l) constructor
- InferenceEngine::TBlob::TBlob(Precision p, Layout l, const SizeVector& dims) constructor
- InferenceEngine::TBlob::TBlob(Precision p, Layout l, const SizeVector& dims, T* ptr, size_t data_size) constructor
- InferenceEngine::TBlob::TBlob(Precision p, Layout l, const SizeVector &dims, std::shared_ptr alloc) constructor
- InferenceEngine::Blob::type() method
- InferenceEngine::Blob::precision() method
- InferenceEngine::Blob::layout() method
- InferenceEngine::Blob::dims() method
- InferenceEngine::make_shared_blob(Precision p, Layout l, const SizeVector &dims) function
- InferenceEngine::make_shared_blob(Precision p, const SizeVector &dims) function
- InferenceEngine::make_shared_blob(Precision p, Layout l, const TArg &arg) function
- InferenceEngine::make_shared_blob(Precision p, const TArg &arg) function
- InferenceEngine::make_shared_blob(TBlob &&arg) function
- InferenceEngine::make_shared_blob(Precision p, Layout l) function
- InferenceEngine::make_shared_blob(Precision p, Layout l, SizeVector dims, const std::vector &arg) function
- InferenceEngine::make_shared_blob(Precision p, Layout l, const std::vector &arg) function
- InferenceEngine::make_shared_blob(Precision p, const std::vector &arg) function
- InferenceEngine::make_shared_blob(Precision p, Layout l, const SizeVector &dims, TypeTo * ptr, size_t size) function
- InferenceEngine::make_shared_blob(Precision p, const SizeVector &dims, TypeTo * ptr, size_t size) function
- InferenceEngine::I_N variable
- InferenceEngine::I_C variable
- InferenceEngine::I_H variable
- InferenceEngine::I_W variable
- InferenceEngine::LayoutOffsetCounter class
- InferenceEngine::ConvertLayout function
API working with device enumeration:
- InferenceEngine::TargetDevice enumeration
- InferenceEngine::TargetDeviceInfo class
- InferenceEngine::getDeviceName function
- InferenceEngine::FindPluginRequest class
- InferenceEngine::FindPluginResponse class
- InferenceEngine::findPlugin(const FindPluginRequest &req, FindPluginResponse &result, ResponseDesc *resp) function
- InferenceEngine::ICNNNetwork::setTargetDevice method
- InferenceEngine::ICNNNetwork::getTargetDevice method
- InferenceEngine::PluginDispatcher::getPluginByDevice method
- InferenceEngine::PluginDispatcher::getSuitablePlugin method