From e5f2903c832edc066deb32d74384b0fac63d9557 Mon Sep 17 00:00:00 2001 From: Ilya Churaev Date: Tue, 28 Feb 2023 02:27:12 +0400 Subject: [PATCH] Changed template plugin namespace (#15962) * Changed template plugin namespace * Fixed documentation --- docs/IE_PLUGIN_DG/InferRequest.md | 16 +++---- .../template/src/async_infer_request.cpp | 8 ++-- .../template/src/async_infer_request.hpp | 8 ++-- src/plugins/template/src/compiled_model.cpp | 42 +++++++++---------- src/plugins/template/src/compiled_model.hpp | 6 ++- src/plugins/template/src/plugin.cpp | 42 +++++++++---------- src/plugins/template/src/plugin.hpp | 8 ++-- ...fer_request.cpp => sync_infer_request.cpp} | 29 ++++++------- ...fer_request.hpp => sync_infer_request.hpp} | 8 ++-- src/plugins/template/src/template_config.cpp | 2 +- src/plugins/template/src/template_config.hpp | 6 ++- src/plugins/template/src/template_itt.hpp | 6 ++- 12 files changed, 96 insertions(+), 85 deletions(-) rename src/plugins/template/src/{infer_request.cpp => sync_infer_request.cpp} (89%) rename src/plugins/template/src/{infer_request.hpp => sync_infer_request.hpp} (89%) diff --git a/docs/IE_PLUGIN_DG/InferRequest.md b/docs/IE_PLUGIN_DG/InferRequest.md index 2c6aacefc3e..db03bb3b06d 100644 --- a/docs/IE_PLUGIN_DG/InferRequest.md +++ b/docs/IE_PLUGIN_DG/InferRequest.md @@ -12,7 +12,7 @@ Inference Engine Plugin API provides the helper InferenceEngine::IInferRequestIn to use as a base class for a synchronous inference request implementation. Based of that, a declaration of a synchronous request class can look as follows: -@snippet src/infer_request.hpp infer_request:header +@snippet src/sync_infer_request.hpp infer_request:header #### Class Fields @@ -34,7 +34,7 @@ The example class has several fields: The constructor initializes helper fields and calls methods which allocate blobs: -@snippet src/infer_request.cpp infer_request:ctor +@snippet src/sync_infer_request.cpp infer_request:ctor > **NOTE**: Call InferenceEngine::CNNNetwork::getInputsInfo and InferenceEngine::CNNNetwork::getOutputsInfo to specify both layout and precision of blobs, which you can set with InferenceEngine::InferRequest::SetBlob and get with InferenceEngine::InferRequest::GetBlob. A plugin uses these hints to determine its internal layouts and precisions for input and output blobs if needed. @@ -42,7 +42,7 @@ The constructor initializes helper fields and calls methods which allocate blobs Decrements a number of created inference requests: -@snippet src/infer_request.cpp infer_request:dtor +@snippet src/sync_infer_request.cpp infer_request:dtor ### `InferImpl()` @@ -50,13 +50,13 @@ Decrements a number of created inference requests: - Checks blobs set by users - Calls the `InferImpl` method defined in a derived class to call actual pipeline stages synchronously -@snippet src/infer_request.cpp infer_request:infer_impl +@snippet src/sync_infer_request.cpp infer_request:infer_impl #### 1. `inferPreprocess` Below is the code of the `inferPreprocess` method to demonstrate Inference Engine common preprocessing step handling: -@snippet src/infer_request.cpp infer_request:infer_preprocess +@snippet src/sync_infer_request.cpp infer_request:infer_preprocess **Details:** * `InferImpl` must call the InferenceEngine::IInferRequestInternal::execDataPreprocessing function, which executes common Inference Engine preprocessing step (for example, applies resize or color conversion operations) if it is set by the user. The output dimensions, layout and precision matches the input information set via InferenceEngine::CNNNetwork::getInputsInfo. @@ -66,18 +66,18 @@ Below is the code of the `inferPreprocess` method to demonstrate Inference Engin Executes a pipeline synchronously using `_executable` object: -@snippet src/infer_request.cpp infer_request:start_pipeline +@snippet src/sync_infer_request.cpp infer_request:start_pipeline #### 3. `inferPostprocess` Converts output blobs if precisions of backend output blobs and blobs passed by user are different: -@snippet src/infer_request.cpp infer_request:infer_postprocess +@snippet src/sync_infer_request.cpp infer_request:infer_postprocess ### `GetPerformanceCounts()` The method sets performance counters which were measured during pipeline stages execution: -@snippet src/infer_request.cpp infer_request:get_performance_counts +@snippet src/sync_infer_request.cpp infer_request:get_performance_counts The next step in the plugin library implementation is the [Asynchronous Inference Request](@ref openvino_docs_ie_plugin_dg_async_infer_request) class. diff --git a/src/plugins/template/src/async_infer_request.cpp b/src/plugins/template/src/async_infer_request.cpp index f92f259ad7b..29532650b9d 100644 --- a/src/plugins/template/src/async_infer_request.cpp +++ b/src/plugins/template/src/async_infer_request.cpp @@ -4,13 +4,13 @@ #include "async_infer_request.hpp" -#include "infer_request.hpp" #include "openvino/runtime/iinfer_request.hpp" +#include "sync_infer_request.hpp" #include "template_itt.hpp" // ! [async_infer_request:ctor] -TemplatePlugin::AsyncInferRequest::AsyncInferRequest( - const std::shared_ptr& request, +ov::template_plugin::AsyncInferRequest::AsyncInferRequest( + const std::shared_ptr& request, const std::shared_ptr& task_executor, const std::shared_ptr& wait_executor, const std::shared_ptr& callback_executor) @@ -47,7 +47,7 @@ TemplatePlugin::AsyncInferRequest::AsyncInferRequest( // ! [async_infer_request:ctor] // ! [async_infer_request:dtor] -TemplatePlugin::AsyncInferRequest::~AsyncInferRequest() { +ov::template_plugin::AsyncInferRequest::~AsyncInferRequest() { ov::IAsyncInferRequest::stop_and_wait(); } // ! [async_infer_request:dtor] diff --git a/src/plugins/template/src/async_infer_request.hpp b/src/plugins/template/src/async_infer_request.hpp index a15d7ee388a..092d8e342ff 100644 --- a/src/plugins/template/src/async_infer_request.hpp +++ b/src/plugins/template/src/async_infer_request.hpp @@ -6,11 +6,12 @@ #include -#include "infer_request.hpp" #include "openvino/runtime/iasync_infer_request.hpp" #include "openvino/runtime/iinfer_request.hpp" +#include "sync_infer_request.hpp" -namespace TemplatePlugin { +namespace ov { +namespace template_plugin { // ! [async_infer_request:header] class AsyncInferRequest : public ov::IAsyncInferRequest { @@ -27,4 +28,5 @@ private: }; // ! [async_infer_request:header] -} // namespace TemplatePlugin +} // namespace template_plugin +} // namespace ov diff --git a/src/plugins/template/src/compiled_model.cpp b/src/plugins/template/src/compiled_model.cpp index c4cfc59dfa3..4d420d09304 100644 --- a/src/plugins/template/src/compiled_model.cpp +++ b/src/plugins/template/src/compiled_model.cpp @@ -15,13 +15,11 @@ #include "template_itt.hpp" #include "transformations/utils/utils.hpp" -using namespace TemplatePlugin; - // ! [executable_network:ctor_cnnnetwork] -TemplatePlugin::CompiledModel::CompiledModel(const std::shared_ptr& model, - const std::shared_ptr& plugin, - const std::shared_ptr& task_executor, - const Configuration& cfg) +ov::template_plugin::CompiledModel::CompiledModel(const std::shared_ptr& model, + const std::shared_ptr& plugin, + const std::shared_ptr& task_executor, + const Configuration& cfg) : ov::ICompiledModel(model, plugin, task_executor), // Disable default threads creation _cfg(cfg), m_model(model) { @@ -45,7 +43,7 @@ TemplatePlugin::CompiledModel::CompiledModel(const std::shared_ptr& m // forward declaration void transform_model(const std::shared_ptr& model); -void TemplatePlugin::CompiledModel::compile_model(const std::shared_ptr& model) { +void ov::template_plugin::CompiledModel::compile_model(const std::shared_ptr& model) { // apply plugins transformations transform_model(model); // Perform any other steps like allocation and filling backend specific memory handles and so on @@ -53,44 +51,44 @@ void TemplatePlugin::CompiledModel::compile_model(const std::shared_ptr TemplatePlugin::CompiledModel::create_infer_request() const { +std::shared_ptr ov::template_plugin::CompiledModel::create_infer_request() const { auto internal_request = create_sync_infer_request(); - auto async_infer_request = - std::make_shared(std::static_pointer_cast(internal_request), - get_task_executor(), - get_template_plugin()->_waitExecutor, - get_callback_executor()); + auto async_infer_request = std::make_shared( + std::static_pointer_cast(internal_request), + get_task_executor(), + get_template_plugin()->_waitExecutor, + get_callback_executor()); return async_infer_request; } -std::shared_ptr TemplatePlugin::CompiledModel::create_sync_infer_request() const { +std::shared_ptr ov::template_plugin::CompiledModel::create_sync_infer_request() const { return std::make_shared( - std::static_pointer_cast(shared_from_this())); + std::static_pointer_cast(shared_from_this())); } // ! [executable_network:create_infer_request] -void TemplatePlugin::CompiledModel::set_property(const ov::AnyMap& properties) { +void ov::template_plugin::CompiledModel::set_property(const ov::AnyMap& properties) { OPENVINO_NOT_IMPLEMENTED; } -ov::RemoteContext TemplatePlugin::CompiledModel::get_context() const { +ov::RemoteContext ov::template_plugin::CompiledModel::get_context() const { OPENVINO_NOT_IMPLEMENTED; } -std::shared_ptr TemplatePlugin::CompiledModel::get_runtime_model() const { +std::shared_ptr ov::template_plugin::CompiledModel::get_runtime_model() const { return m_model; } -std::shared_ptr TemplatePlugin::CompiledModel::get_template_plugin() const { +std::shared_ptr ov::template_plugin::CompiledModel::get_template_plugin() const { auto plugin = get_plugin(); OPENVINO_ASSERT(plugin); - auto template_plugin = std::static_pointer_cast(plugin); + auto template_plugin = std::static_pointer_cast(plugin); OPENVINO_ASSERT(template_plugin); return template_plugin; } // ! [executable_network:get_config] -ov::Any TemplatePlugin::CompiledModel::get_property(const std::string& name) const { +ov::Any ov::template_plugin::CompiledModel::get_property(const std::string& name) const { const auto& add_ro_properties = [](const std::string& name, std::vector& properties) { properties.emplace_back(ov::PropertyName{name, ov::PropertyMutability::RO}); }; @@ -152,7 +150,7 @@ ov::Any TemplatePlugin::CompiledModel::get_property(const std::string& name) con // ! [executable_network:get_config] // ! [executable_network:export] -void TemplatePlugin::CompiledModel::export_model(std::ostream& modelStream) const { +void ov::template_plugin::CompiledModel::export_model(std::ostream& modelStream) const { OV_ITT_SCOPED_TASK(itt::domains::TemplatePlugin, "ExecutableNetwork::Export"); std::stringstream xmlFile, binFile; diff --git a/src/plugins/template/src/compiled_model.hpp b/src/plugins/template/src/compiled_model.hpp index a138f3f6b68..ffc632708e4 100644 --- a/src/plugins/template/src/compiled_model.hpp +++ b/src/plugins/template/src/compiled_model.hpp @@ -10,7 +10,8 @@ #include "openvino/runtime/tensor.hpp" #include "template_config.hpp" -namespace TemplatePlugin { +namespace ov { +namespace template_plugin { class Plugin; class InferRequest; @@ -55,4 +56,5 @@ private: }; // ! [executable_network:header] -} // namespace TemplatePlugin +} // namespace template_plugin +} // namespace ov diff --git a/src/plugins/template/src/plugin.cpp b/src/plugins/template/src/plugin.cpp index e5343f53fc6..3b2765172ed 100644 --- a/src/plugins/template/src/plugin.cpp +++ b/src/plugins/template/src/plugin.cpp @@ -19,15 +19,13 @@ #include "transformations/op_conversions/convert_reduce_to_pooling.hpp" #include "transformations/template_pattern_transformation.hpp" -using namespace TemplatePlugin; - namespace { static constexpr const char* wait_executor_name = "TemplateWaitExecutor"; static constexpr const char* stream_executor_name = "TemplateStreamsExecutor"; } // namespace // ! [plugin:ctor] -Plugin::Plugin() { +ov::template_plugin::Plugin::Plugin() { // TODO: fill with actual device name, backend engine set_device_name("TEMPLATE"); @@ -40,7 +38,7 @@ Plugin::Plugin() { // ! [plugin:ctor] // ! [plugin:dtor] -Plugin::~Plugin() { +ov::template_plugin::Plugin::~Plugin() { // Plugin should remove executors from executor cache to avoid threads number growth in the whole application get_executor_manager()->clear(stream_executor_name); get_executor_manager()->clear(wait_executor_name); @@ -49,11 +47,11 @@ Plugin::~Plugin() { } // ! [plugin:dtor] -ov::RemoteContext TemplatePlugin::Plugin::create_context(const ov::AnyMap& remote_properties) const { +ov::RemoteContext ov::template_plugin::Plugin::create_context(const ov::AnyMap& remote_properties) const { OPENVINO_NOT_IMPLEMENTED; } -ov::RemoteContext TemplatePlugin::Plugin::get_default_context(const ov::AnyMap& remote_properties) const { +ov::RemoteContext ov::template_plugin::Plugin::get_default_context(const ov::AnyMap& remote_properties) const { OPENVINO_NOT_IMPLEMENTED; } @@ -85,8 +83,9 @@ void transform_model(const std::shared_ptr& model) { // ! [plugin:transform_network] // ! [plugin:load_exe_network_impl] -std::shared_ptr TemplatePlugin::Plugin::compile_model(const std::shared_ptr& model, - const ov::AnyMap& properties) const { +std::shared_ptr ov::template_plugin::Plugin::compile_model( + const std::shared_ptr& model, + const ov::AnyMap& properties) const { OV_ITT_SCOPED_TASK(itt::domains::TemplatePlugin, "Plugin::compile_model"); auto fullConfig = Configuration{properties, _cfg}; @@ -101,16 +100,17 @@ std::shared_ptr TemplatePlugin::Plugin::compile_model(const return compiled_model; } -std::shared_ptr TemplatePlugin::Plugin::compile_model(const std::shared_ptr& model, - const ov::AnyMap& properties, - const ov::RemoteContext& context) const { +std::shared_ptr ov::template_plugin::Plugin::compile_model( + const std::shared_ptr& model, + const ov::AnyMap& properties, + const ov::RemoteContext& context) const { OPENVINO_NOT_IMPLEMENTED; } // ! [plugin:load_exe_network_impl] // ! [plugin:import_network] -std::shared_ptr TemplatePlugin::Plugin::import_model(std::istream& model, - const ov::AnyMap& properties) const { +std::shared_ptr ov::template_plugin::Plugin::import_model(std::istream& model, + const ov::AnyMap& properties) const { OV_ITT_SCOPED_TASK(itt::domains::TemplatePlugin, "Plugin::import_model"); auto fullConfig = Configuration{properties, _cfg}; @@ -141,16 +141,16 @@ std::shared_ptr TemplatePlugin::Plugin::import_model(std::is return compiled_model; } -std::shared_ptr TemplatePlugin::Plugin::import_model(std::istream& model, - const ov::RemoteContext& context, - const ov::AnyMap& properties) const { +std::shared_ptr ov::template_plugin::Plugin::import_model(std::istream& model, + const ov::RemoteContext& context, + const ov::AnyMap& properties) const { OPENVINO_NOT_IMPLEMENTED; } // ! [plugin:import_network] // ! [plugin:query_network] -ov::SupportedOpsMap TemplatePlugin::Plugin::query_model(const std::shared_ptr& model, - const ov::AnyMap& properties) const { +ov::SupportedOpsMap ov::template_plugin::Plugin::query_model(const std::shared_ptr& model, + const ov::AnyMap& properties) const { OV_ITT_SCOPED_TASK(TemplatePlugin::itt::domains::TemplatePlugin, "Plugin::query_model"); Configuration fullConfig{properties, _cfg, false}; @@ -194,13 +194,13 @@ ov::SupportedOpsMap TemplatePlugin::Plugin::query_model(const std::shared_ptr& properties) { properties.emplace_back(ov::PropertyName{name, ov::PropertyMutability::RO}); }; @@ -283,5 +283,5 @@ ov::Any TemplatePlugin::Plugin::get_property(const std::string& name, const ov:: // ! [plugin:create_plugin_engine] static const ov::Version version = {CI_BUILD_NUMBER, "openvino_template_plugin"}; -OV_DEFINE_PLUGIN_CREATE_FUNCTION(Plugin, version) +OV_DEFINE_PLUGIN_CREATE_FUNCTION(ov::template_plugin::Plugin, version) // ! [plugin:create_plugin_engine] diff --git a/src/plugins/template/src/plugin.hpp b/src/plugins/template/src/plugin.hpp index aa5b9077312..0c04798e3a1 100644 --- a/src/plugins/template/src/plugin.hpp +++ b/src/plugins/template/src/plugin.hpp @@ -12,7 +12,8 @@ #include "template_config.hpp" //! [plugin:header] -namespace TemplatePlugin { +namespace ov { +namespace template_plugin { class Plugin : public ov::IPlugin { public: @@ -46,7 +47,7 @@ public: const ov::AnyMap& properties) const override; private: - friend class TemplatePlugin::CompiledModel; + friend class CompiledModel; friend class InferRequest; std::shared_ptr _backend; @@ -54,5 +55,6 @@ private: std::shared_ptr _waitExecutor; }; -} // namespace TemplatePlugin +} // namespace template_plugin +} // namespace ov //! [plugin:header] diff --git a/src/plugins/template/src/infer_request.cpp b/src/plugins/template/src/sync_infer_request.cpp similarity index 89% rename from src/plugins/template/src/infer_request.cpp rename to src/plugins/template/src/sync_infer_request.cpp index 8d2863c224c..b6f2a6edf2c 100644 --- a/src/plugins/template/src/infer_request.cpp +++ b/src/plugins/template/src/sync_infer_request.cpp @@ -2,7 +2,7 @@ // SPDX-License-Identifier: Apache-2.0 // -#include "infer_request.hpp" +#include "sync_infer_request.hpp" #include #include @@ -31,7 +31,7 @@ void allocate_tensor_impl(ov::Tensor& tensor, const ov::element::Type& element_t } // namespace // ! [infer_request:ctor] -TemplatePlugin::InferRequest::InferRequest(const std::shared_ptr& model) +ov::template_plugin::InferRequest::InferRequest(const std::shared_ptr& model) : ov::ISyncInferRequest(model) { // TODO: allocate infer request device and host buffers if needed, fill actual list of profiling tasks @@ -75,23 +75,24 @@ TemplatePlugin::InferRequest::InferRequest(const std::shared_ptr> TemplatePlugin::InferRequest::query_state() const { +std::vector> ov::template_plugin::InferRequest::query_state() const { OPENVINO_NOT_IMPLEMENTED; } -std::shared_ptr TemplatePlugin::InferRequest::get_template_model() const { +std::shared_ptr ov::template_plugin::InferRequest::get_template_model() + const { auto& compiled_model = get_compiled_model(); - auto template_model = std::dynamic_pointer_cast(compiled_model); + auto template_model = std::dynamic_pointer_cast(compiled_model); OPENVINO_ASSERT(template_model); return template_model; } // ! [infer_request:dtor] -TemplatePlugin::InferRequest::~InferRequest() = default; +ov::template_plugin::InferRequest::~InferRequest() = default; // ! [infer_request:dtor] // ! [infer_request:infer_impl] -void TemplatePlugin::InferRequest::infer() { +void ov::template_plugin::InferRequest::infer() { // TODO: fill with actual list of pipeline stages, which are executed synchronously for sync infer requests infer_preprocess(); start_pipeline(); @@ -101,7 +102,7 @@ void TemplatePlugin::InferRequest::infer() { // ! [infer_request:infer_impl] // ! [infer_request:infer_preprocess] -void TemplatePlugin::InferRequest::infer_preprocess() { +void ov::template_plugin::InferRequest::infer_preprocess() { OV_ITT_SCOPED_TASK(itt::domains::TemplatePlugin, m_profiling_task[Preprocess]); auto start = Time::now(); convert_batched_tensors(); @@ -168,7 +169,7 @@ void TemplatePlugin::InferRequest::infer_preprocess() { // ! [infer_request:infer_preprocess] // ! [infer_request:start_pipeline] -void TemplatePlugin::InferRequest::start_pipeline() { +void ov::template_plugin::InferRequest::start_pipeline() { OV_ITT_SCOPED_TASK(itt::domains::TemplatePlugin, m_profiling_task[StartPipeline]) auto start = Time::now(); m_executable->call(m_backend_output_tensors, m_backend_input_tensors); @@ -176,7 +177,7 @@ void TemplatePlugin::InferRequest::start_pipeline() { } // ! [infer_request:start_pipeline] -void TemplatePlugin::InferRequest::wait_pipeline() { +void ov::template_plugin::InferRequest::wait_pipeline() { OV_ITT_SCOPED_TASK(itt::domains::TemplatePlugin, m_profiling_task[WaitPipeline]) auto start = Time::now(); // TODO: Wait pipeline using driver API or other synchronizations methods @@ -185,7 +186,7 @@ void TemplatePlugin::InferRequest::wait_pipeline() { } // ! [infer_request:infer_postprocess] -void TemplatePlugin::InferRequest::infer_postprocess() { +void ov::template_plugin::InferRequest::infer_postprocess() { OV_ITT_SCOPED_TASK(itt::domains::TemplatePlugin, m_profiling_task[Postprocess]); auto start = Time::now(); OPENVINO_ASSERT(get_outputs().size() == m_backend_output_tensors.size()); @@ -206,8 +207,8 @@ void TemplatePlugin::InferRequest::infer_postprocess() { // ! [infer_request:infer_postprocess] // ! [infer_request:set_blobs_impl] -void TemplatePlugin::InferRequest::set_tensors_impl(const ov::Output port, - const std::vector& tensors) { +void ov::template_plugin::InferRequest::set_tensors_impl(const ov::Output port, + const std::vector& tensors) { for (const auto& input : get_inputs()) { if (input == port) { m_batched_tensors[input.get_tensor_ptr()] = tensors; @@ -219,7 +220,7 @@ void TemplatePlugin::InferRequest::set_tensors_impl(const ov::Output TemplatePlugin::InferRequest::get_profiling_info() const { +std::vector ov::template_plugin::InferRequest::get_profiling_info() const { std::vector info; const auto fill_profiling_info = [](const std::string& name, const std::chrono::duration& time) -> ov::ProfilingInfo { diff --git a/src/plugins/template/src/infer_request.hpp b/src/plugins/template/src/sync_infer_request.hpp similarity index 89% rename from src/plugins/template/src/infer_request.hpp rename to src/plugins/template/src/sync_infer_request.hpp index 35b9c6f75b5..49b9bdaa790 100644 --- a/src/plugins/template/src/infer_request.hpp +++ b/src/plugins/template/src/sync_infer_request.hpp @@ -16,7 +16,8 @@ #include "openvino/itt.hpp" #include "openvino/runtime/isync_infer_request.hpp" -namespace TemplatePlugin { +namespace ov { +namespace template_plugin { // forward declaration class CompiledModel; @@ -24,7 +25,7 @@ class CompiledModel; // ! [infer_request:header] class InferRequest : public ov::ISyncInferRequest { public: - explicit InferRequest(const std::shared_ptr& compiled_model); + explicit InferRequest(const std::shared_ptr& compiled_model); ~InferRequest(); void infer() override; @@ -54,4 +55,5 @@ private: }; // ! [infer_request:header] -} // namespace TemplatePlugin +} // namespace template_plugin +} // namespace ov diff --git a/src/plugins/template/src/template_config.cpp b/src/plugins/template/src/template_config.cpp index 840d1ae8541..d24afae2885 100644 --- a/src/plugins/template/src/template_config.cpp +++ b/src/plugins/template/src/template_config.cpp @@ -9,7 +9,7 @@ #include "template/config.hpp" -using namespace TemplatePlugin; +using namespace ov::template_plugin; Configuration::Configuration() {} diff --git a/src/plugins/template/src/template_config.hpp b/src/plugins/template/src/template_config.hpp index 20363312397..a0a2bf8b41d 100644 --- a/src/plugins/template/src/template_config.hpp +++ b/src/plugins/template/src/template_config.hpp @@ -10,7 +10,8 @@ #include "openvino/runtime/properties.hpp" #include "openvino/runtime/threading/istreams_executor.hpp" -namespace TemplatePlugin { +namespace ov { +namespace template_plugin { // ! [configuration:header] using ConfigMap = std::map; @@ -37,4 +38,5 @@ struct Configuration { }; // ! [configuration:header] -} // namespace TemplatePlugin +} // namespace template_plugin +} // namespace ov diff --git a/src/plugins/template/src/template_itt.hpp b/src/plugins/template/src/template_itt.hpp index 50a410bb07d..b376682439b 100644 --- a/src/plugins/template/src/template_itt.hpp +++ b/src/plugins/template/src/template_itt.hpp @@ -11,10 +11,12 @@ #include -namespace TemplatePlugin { +namespace ov { +namespace template_plugin { namespace itt { namespace domains { OV_ITT_DOMAIN(TemplatePlugin); } } // namespace itt -} // namespace TemplatePlugin +} // namespace template_plugin +} // namespace ov