From 5128b1f23d3800081c10d954344deb12cdda5235 Mon Sep 17 00:00:00 2001 From: Ilya Lavrenov Date: Thu, 14 Oct 2021 17:48:26 +0300 Subject: [PATCH] Added more checks in remote_tensors tests (#8014) --- .../inference_engine/include/openvino/runtime/core.hpp | 10 +++++----- .../include/openvino/runtime/executable_network.hpp | 2 +- .../include/openvino/runtime/gpu/dx.hpp | 8 ++++---- .../include/openvino/runtime/gpu/ocl.hpp | 8 ++++---- .../include/openvino/runtime/gpu/va.hpp | 8 ++++---- 5 files changed, 18 insertions(+), 18 deletions(-) diff --git a/inference-engine/src/inference_engine/include/openvino/runtime/core.hpp b/inference-engine/src/inference_engine/include/openvino/runtime/core.hpp index e739a989622..140d757bc8c 100644 --- a/inference-engine/src/inference_engine/include/openvino/runtime/core.hpp +++ b/inference-engine/src/inference_engine/include/openvino/runtime/core.hpp @@ -49,7 +49,7 @@ public: * See register_plugins for more details. * * @param xmlConfigFile A path to .xml file with plugins to load from. If XML configuration file is not specified, - * then default Inference Engine plugins are loaded from the default plugin.xml file. + * then default OpenVINO Runtime plugins are loaded from the default plugin.xml file. */ explicit Core(const std::string& xmlConfigFile = {}); @@ -231,7 +231,7 @@ public: std::vector get_available_devices() const; /** - * @brief Register new device and plugin which implement this device inside Inference Engine. + * @brief Register new device and plugin which implement this device inside OpenVINO Runtime. * * @param pluginName A name of plugin. Depending on platform pluginName is wrapped with shared library suffix and * prefix to identify library full name @@ -242,15 +242,15 @@ public: void register_plugin(const std::string& pluginName, const std::string& deviceName); /** - * @brief Unloads previously loaded plugin with a specified name from Inference Engine + * @brief Unloads previously loaded plugin with a specified name from OpenVINO Runtime * The method is needed to remove plugin instance and free its resources. If plugin for a * specified device has not been created before, the method throws an exception. * - * @param deviceName Device name identifying plugin to remove from Inference Engine + * @param deviceName Device name identifying plugin to remove from OpenVINO Runtime */ void unload_plugin(const std::string& deviceName); - /** @brief Registers plugin to Inference Engine Core instance using XML configuration file with + /** @brief Registers plugin to OpenVINO Runtime Core instance using XML configuration file with * plugins description. * * XML file has the following structure: diff --git a/inference-engine/src/inference_engine/include/openvino/runtime/executable_network.hpp b/inference-engine/src/inference_engine/include/openvino/runtime/executable_network.hpp index 419f94057d4..0587e9afcc7 100644 --- a/inference-engine/src/inference_engine/include/openvino/runtime/executable_network.hpp +++ b/inference-engine/src/inference_engine/include/openvino/runtime/executable_network.hpp @@ -34,7 +34,7 @@ class Core; */ class OPENVINO_RUNTIME_API ExecutableNetwork { std::shared_ptr _so; - std::shared_ptr _impl; + std::shared_ptr _impl; /** * @brief Constructs ExecutableNetwork from the initialized std::shared_ptr diff --git a/inference-engine/src/inference_engine/include/openvino/runtime/gpu/dx.hpp b/inference-engine/src/inference_engine/include/openvino/runtime/gpu/dx.hpp index 93ec80eb54e..817afdd5234 100644 --- a/inference-engine/src/inference_engine/include/openvino/runtime/gpu/dx.hpp +++ b/inference-engine/src/inference_engine/include/openvino/runtime/gpu/dx.hpp @@ -25,7 +25,7 @@ namespace gpu { /** * @brief This class represents an abstraction for GPU plugin remote tensor * which is shared with Direct3D 11 buffer. - * The plugin object derived from this class can be obtained with create_tensor() call. + * The plugin object derived from this class can be obtained with D3DContext::create_tensor() call. * @note User can also obtain OpenCL buffer handle from this class. */ class D3DBufferTensor : public ClBufferTensor { @@ -52,7 +52,7 @@ public: /** * @brief This class represents an abstraction for GPU plugin remote tensor * which is shared with Direct3D 11 2D texture. - * The plugin object derived from this class can be obtained with create_tensor() call. + * The plugin object derived from this class can be obtained with D3DContext::create_tensor() call. * @note User can also obtain OpenCL 2D image handle from this class. */ class D3DSurface2DTensor : public ClImage2DTensor { @@ -89,7 +89,7 @@ public: * @brief This class represents an abstraction for GPU plugin remote context * which is shared with Direct3D 11 device. * The plugin object derived from this class can be obtained either with - * GetContext() method of Executable network or using CreateContext() Core call. + * ExecutableNetwork::get_context() or Core::create_context() calls. * @note User can also obtain OpenCL context handle from this class. */ class D3DContext : public ClContext { @@ -116,7 +116,7 @@ public: /** * @brief Constructs D3DContext remote context object from ID3D11Device - * @param core Inference Engine Core object instance + * @param core OpenVINO Runtime Core object instance * @param deviceName A name of to create a remote context for * @param device A pointer to ID3D11Device to be used to create a remote context */ diff --git a/inference-engine/src/inference_engine/include/openvino/runtime/gpu/ocl.hpp b/inference-engine/src/inference_engine/include/openvino/runtime/gpu/ocl.hpp index 3f2c02a4b1d..6ca2f74bc6a 100644 --- a/inference-engine/src/inference_engine/include/openvino/runtime/gpu/ocl.hpp +++ b/inference-engine/src/inference_engine/include/openvino/runtime/gpu/ocl.hpp @@ -31,7 +31,7 @@ using gpu_handle_param = void*; /** * @brief This class represents an abstraction for GPU plugin remote tensor * which can be shared with user-supplied OpenCL buffer. - * The plugin object derived from this class can be obtained with create_tensor() call. + * The plugin object derived from this class can be obtained with ClContext::create_tensor() call. * @note User can obtain OpenCL buffer handle from this class. */ class ClBufferTensor : public RemoteTensor { @@ -75,7 +75,7 @@ public: /** * @brief This class represents an abstraction for GPU plugin remote tensor * which can be shared with user-supplied OpenCL 2D Image. - * The plugin object derived from this class can be obtained with create_tensor() call. + * The plugin object derived from this class can be obtained with ClContext::create_tensor() call. * @note User can obtain OpenCL image handle from this class. */ class ClImage2DTensor : public RemoteTensor { @@ -120,7 +120,7 @@ public: * @brief This class represents an abstraction for GPU plugin remote context * which is shared with OpenCL context object. * The plugin object derived from this class can be obtained either with - * GetContext() method of Executable network or using CreateContext() Core call. + * ExecutableNetwork::get_context() or Core::create_context() calls. */ class ClContext : public RemoteContext { using RemoteContext::create_tensor; @@ -138,7 +138,7 @@ public: /** * @brief Constructs context object from user-supplied OpenCL context handle - * @param core A reference to Inference Engine Core object + * @param core A reference to OpenVINO Runtime Core object * @param deviceName A name of device to create a remote context for * @param ctx A OpenCL context to be used to create shared remote context */ diff --git a/inference-engine/src/inference_engine/include/openvino/runtime/gpu/va.hpp b/inference-engine/src/inference_engine/include/openvino/runtime/gpu/va.hpp index 2f01d7b4689..83542558138 100644 --- a/inference-engine/src/inference_engine/include/openvino/runtime/gpu/va.hpp +++ b/inference-engine/src/inference_engine/include/openvino/runtime/gpu/va.hpp @@ -27,7 +27,7 @@ namespace gpu { /** * @brief This class represents an abstraction for GPU plugin remote tensor * which is shared with VA output surface. - * The plugin object derived from this class can be obtained with create_tensor() call. + * The plugin object derived from this class can be obtained with VAContext::create_tensor() call. * @note User can also obtain OpenCL 2D image handle from this class. */ class VASurfaceTensor : public ClImage2DTensor { @@ -43,7 +43,7 @@ public: {GPU_PARAM_KEY(SHARED_MEM_TYPE), {GPU_PARAM_VALUE(VA_SURFACE)}}}); } /** - * @brief VASurfaceID conversion operator for the VASurfaceBlob object. + * @brief VASurfaceID conversion operator for the VASurfaceTensor object. * @return `VASurfaceID` handle */ operator VASurfaceID() { @@ -63,7 +63,7 @@ public: * @brief This class represents an abstraction for GPU plugin remote context * which is shared with VA display object. * The plugin object derived from this class can be obtained either with - * GetContext() method of Executable network or using CreateContext() Core call. + * ExecutableNetwork::get_context() or Core::create_context() calls. * @note User can also obtain OpenCL context handle from this class. */ class VAContext : public ClContext { @@ -90,7 +90,7 @@ public: /** * @brief Constructs remote context object from VA display handle - * @param core Inference Engine Core object + * @param core OpenVINO Runtime Core object * @param deviceName A device name to create a remote context for * @param device A `VADisplay` to create remote context from */