diff --git a/inference-engine/ie_bridges/c/include/c_api/ie_c_api.h b/inference-engine/ie_bridges/c/include/c_api/ie_c_api.h index a7caa653804..4801d6e2ea5 100644 --- a/inference-engine/ie_bridges/c/include/c_api/ie_c_api.h +++ b/inference-engine/ie_bridges/c/include/c_api/ie_c_api.h @@ -352,7 +352,7 @@ INFERENCE_ENGINE_C_API(void) ie_core_free(ie_core_t **core); * @brief Gets version information of the device specified. Use the ie_core_versions_free() method to free memory. * @ingroup Core * @param core A pointer to ie_core_t instance. - * @param device_name Name to indentify device. + * @param device_name Name to identify device. * @param versions A pointer to versions corresponding to device_name. * @return Status code of the operation: OK(0) for success. */ @@ -409,7 +409,7 @@ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_core_load_network(ie_core_t * @ingroup Core * @param core A pointer to ie_core_t instance. * @param ie_core_config Device configuration. - * @param device_name An optinal name of a device. If device name is not specified, + * @param device_name An optional name of a device. If device name is not specified, * the config is set for all the registered devices. * @return Status code of the operation: OK(0) for success. */ @@ -537,7 +537,7 @@ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_exec_network_get_metric(con /** * @brief Sets configuration for current executable network. Currently, the method can be used - * when the network run on the Multi device and the configuration paramter is only can be "MULTI_DEVICE_PRIORITIES" + * when the network run on the Multi device and the configuration parameter is only can be "MULTI_DEVICE_PRIORITIES" * @ingroup ExecutableNetwork * @param ie_exec_network A pointer to ie_executable_network_t instance. * @param param_config A pointer to device configuration.. @@ -551,7 +551,7 @@ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_exec_network_set_config(ie_ * @ingroup ExecutableNetwork * @param ie_exec_network A pointer to ie_executable_network_t instance. * @param metric_config A configuration parameter name to request. - * @param param_result A configuration value corresponding to a configuration paramter name. + * @param param_result A configuration value corresponding to a configuration parameter name. * @return Status code of the operation: OK(0) for success. */ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_exec_network_get_config(const ie_executable_network_t *ie_exec_network, \ @@ -653,7 +653,7 @@ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_infer_request_set_batch(ie_ */ /** - * @brief When netowrk is loaded into the Infernece Engine, it is not required anymore and should be released + * @brief When network is loaded into the Infernece Engine, it is not required anymore and should be released * @ingroup Network * @param network The pointer to the instance of the ie_network_t to free. */ @@ -765,7 +765,7 @@ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_network_set_input_resize_al * @param network A pointer to ie_network_t instance. * @param input_name Name of input data. * @param colformat_result The pointer to the color format used for input blob creation. - * @reutrn Status code of the operation: OK(0) for success. + * @return Status code of the operation: OK(0) for success. */ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_network_get_color_format(const ie_network_t *network, const char *input_name, colorformat_e *colformat_result); @@ -775,7 +775,7 @@ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_network_get_color_format(co * @param network A pointer to ie_network_t instance. * @param input_name Name of input data. * @param color_format Color format of the input data. - * @reutrn Status code of the operation: OK(0) for success. + * @return Status code of the operation: OK(0) for success. */ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_network_set_color_format(ie_network_t *network, const char *input_name, const colorformat_e color_format); @@ -801,7 +801,7 @@ INFERENCE_ENGINE_C_API(IE_NODISCARD IEStatusCode) ie_network_reshape(ie_network_ /** * @brief Gets number of output for the network. * @ingroup Network - * @param network A pointer to the instance of the ie_network_t to get number of ouput information. + * @param network A pointer to the instance of the ie_network_t to get number of output information. * @param size_result A number of the network's output information. * @return Status code of the operation: OK(0) for success. */ diff --git a/inference-engine/ie_bridges/c/src/ie_c_api.cpp b/inference-engine/ie_bridges/c/src/ie_c_api.cpp index ba87c7ca714..d4eb6eb09c5 100644 --- a/inference-engine/ie_bridges/c/src/ie_c_api.cpp +++ b/inference-engine/ie_bridges/c/src/ie_c_api.cpp @@ -145,7 +145,7 @@ std::map config2ParamMap(const ie_config_t *config) } /** - *@brief convert the paramter. + *@brief convert the parameter. */ void parameter2IEparam(const IE::Parameter param, ie_param_t *ie_param) { if (param.is()) { diff --git a/inference-engine/ie_bridges/java/cpp/infer_request.cpp b/inference-engine/ie_bridges/java/cpp/infer_request.cpp index e675f2ba90d..3ceace6085a 100644 --- a/inference-engine/ie_bridges/java/cpp/infer_request.cpp +++ b/inference-engine/ie_bridges/java/cpp/infer_request.cpp @@ -172,8 +172,8 @@ JNIEXPORT jobject JNICALL Java_org_intel_openvino_InferRequest_GetPerformanceCou try { InferRequest *infer_request = (InferRequest *)addr; - std::map perfomance; - perfomance = infer_request->GetPerformanceCounts(); + std::map performance; + performance = infer_request->GetPerformanceCounts(); jclass hashMap_class = env->FindClass("java/util/LinkedHashMap"); jmethodID init_method_id = env->GetMethodID(hashMap_class, "", "()V"); @@ -188,7 +188,7 @@ JNIEXPORT jobject JNICALL Java_org_intel_openvino_InferRequest_GetPerformanceCou jclass layerStatus_enum = env->FindClass("org/intel/openvino/InferenceEngineProfileInfo$LayerStatus"); jmethodID valueOf_method_id = env->GetStaticMethodID(layerStatus_enum,"valueOf","(I)Lorg/intel/openvino/InferenceEngineProfileInfo$LayerStatus;"); - for (const auto& itr : perfomance) { + for (const auto& itr : performance) { InferenceEngine::InferenceEngineProfileInfo ie_prof_info = itr.second; auto it = layer_status_map.find(ie_prof_info.status); diff --git a/inference-engine/ie_bridges/python/src/openvino/inference_engine/ie_api_impl.cpp b/inference-engine/ie_bridges/python/src/openvino/inference_engine/ie_api_impl.cpp index f6f086014c9..a40f598c557 100644 --- a/inference-engine/ie_bridges/python/src/openvino/inference_engine/ie_api_impl.cpp +++ b/inference-engine/ie_bridges/python/src/openvino/inference_engine/ie_api_impl.cpp @@ -74,7 +74,7 @@ PyObject *parse_parameter(const InferenceEngine::Parameter ¶m) { auto val = param.as(); return PyLong_FromLong((long)val); } - // Check for unsinged int + // Check for unsigned int else if (param.is()) { auto val = param.as(); return PyLong_FromLong((unsigned long)val); diff --git a/inference-engine/include/cpp/ie_executable_network.hpp b/inference-engine/include/cpp/ie_executable_network.hpp index 5b8cfa7d899..31d246c0779 100644 --- a/inference-engine/include/cpp/ie_executable_network.hpp +++ b/inference-engine/include/cpp/ie_executable_network.hpp @@ -209,7 +209,7 @@ public: * * Wraps IExecutableNetwork::GetConfig * @param name - config key, can be found in ie_plugin_config.hpp - * @return Configuration paramater value + * @return Configuration parameter value */ Parameter GetConfig(const std::string& name) const { Parameter configValue; @@ -222,7 +222,7 @@ public: * * Wraps IExecutableNetwork::GetMetric * @param name - metric name to request - * @return Metric paramater value + * @return Metric parameter value */ Parameter GetMetric(const std::string& name) const { Parameter metricValue; diff --git a/inference-engine/include/gna/gna_config.hpp b/inference-engine/include/gna/gna_config.hpp index a6290f0e606..195b1ac2378 100644 --- a/inference-engine/include/gna/gna_config.hpp +++ b/inference-engine/include/gna/gna_config.hpp @@ -55,7 +55,7 @@ DECLARE_GNA_CONFIG_KEY(PRECISION); DECLARE_GNA_CONFIG_KEY(FIRMWARE_MODEL_IMAGE); /** -* @brief inforamtion on GNA generation chosen for firmware model dump, can be overriden by GNA3 +* @brief information on GNA generation chosen for firmware model dump, can be overridden by GNA3 */ DECLARE_GNA_CONFIG_KEY(FIRMWARE_MODEL_IMAGE_GENERATION); diff --git a/inference-engine/include/ie_blob.h b/inference-engine/include/ie_blob.h index a586a7b58a7..6a6514e80c9 100644 --- a/inference-engine/include/ie_blob.h +++ b/inference-engine/include/ie_blob.h @@ -404,7 +404,7 @@ public: * * The memory been addressed in the MemoryBlob in general case can be allocated on remote device. * This function copies remote memory to the memory in the virtual process space and after - * destruction of the LockedMemory it will not upload host memory back, bacause it is expected that + * destruction of the LockedMemory it will not upload host memory back, because it is expected that * content is not changed. * * To have an ability change content, you can use rwmap() and wmap() functions. @@ -428,10 +428,10 @@ public: * * The memory been addressed in the MemoryBlob in general case can be allocated on remote device. * This function does not copy of the content from the device to the memory in the virtual process - * space, the content of the memory just after calling of this functin is not specified. After + * space, the content of the memory just after calling of this function is not specified. After * destruction of the LockedMemory, content will be upload host memory. * In the same time there is no abilities to restrict reading from the memory, you need to care of - * reading from memory got by wmap(), it might have sence in some cases like filling of content and + * reading from memory got by wmap(), it might have sense in some cases like filling of content and * before uploading to device * * To access data stored in the blob, you can use rwmap() and rmap() functions. diff --git a/inference-engine/include/ie_common.h b/inference-engine/include/ie_common.h index 4149e96d08b..79f9a88b790 100644 --- a/inference-engine/include/ie_common.h +++ b/inference-engine/include/ie_common.h @@ -79,7 +79,7 @@ enum Layout : uint8_t { SCALAR = 95, //!< A scalar layout // bias layouts - C = 96, //!< A bias layout for opearation + C = 96, //!< A bias layout for operation // Single image layouts CHW = 128, //!< A single image layout (e.g. for mean image) @@ -166,7 +166,7 @@ struct InferenceEngineProfileInfo { * @brief Defines the general status of the layer */ enum LayerStatus { - NOT_RUN, //!< A layer is not exectued + NOT_RUN, //!< A layer is not executed OPTIMIZED_OUT, //!< A layer is optimized out during graph optimization phase EXECUTED //!< A layer is executed }; @@ -237,7 +237,7 @@ struct ResponseDesc { /** - * @brief Responce structure encapsulating information about supported layer + * @brief Response structure encapsulating information about supported layer */ struct QueryNetworkResult { /** diff --git a/inference-engine/include/ie_core.hpp b/inference-engine/include/ie_core.hpp index 7f2b9efdc5c..c2e5445c828 100644 --- a/inference-engine/include/ie_core.hpp +++ b/inference-engine/include/ie_core.hpp @@ -44,7 +44,7 @@ public: /** * @brief Returns plugins version information * - * @param deviceName Device name to indentify plugin + * @param deviceName Device name to identify plugin * @return A vector of versions */ std::map GetVersions(const std::string& deviceName) const; @@ -184,7 +184,7 @@ public: /** * @brief Sets configuration for device, acceptable keys can be found in ie_plugin_config.hpp * - * @param deviceName An optinal name of a device. If device name is not specified, the config is set for all the + * @param deviceName An optional name of a device. If device name is not specified, the config is set for all the * registered devices. * * @param config Map of pairs: (config parameter name, config parameter value) diff --git a/inference-engine/include/ie_iexecutable_network.hpp b/inference-engine/include/ie_iexecutable_network.hpp index 487df0f908e..491c24e0736 100644 --- a/inference-engine/include/ie_iexecutable_network.hpp +++ b/inference-engine/include/ie_iexecutable_network.hpp @@ -160,7 +160,7 @@ public: /** * @brief Gets shared context used to create an executable network. * - * @param pContext Refernce to a pointer that will receive resulting shared context object ptr + * @param pContext Reference to a pointer that will receive resulting shared context object ptr * @param resp Pointer to the response message that holds a description of an error if any occurred * @return code of the operation. InferenceEngine::OK if succeeded */ diff --git a/inference-engine/include/ie_remote_context.hpp b/inference-engine/include/ie_remote_context.hpp index 5140555d76e..bd8f00a6328 100644 --- a/inference-engine/include/ie_remote_context.hpp +++ b/inference-engine/include/ie_remote_context.hpp @@ -194,7 +194,7 @@ public: * @brief A wrapper of CreateBlob method of RemoteContext to keep consistency with * plugin-specific wrappers. * @param desc Defines the layout and dims of the blob - * @param ctx Poniter to the plugin object derived from RemoteContext. + * @param ctx Pointer to the plugin object derived from RemoteContext. * @return A pointer to plugin object that implements RemoteBlob interface. */ inline RemoteBlob::Ptr make_shared_blob(const TensorDesc& desc, RemoteContext::Ptr ctx) {