diff --git a/samples/cpp/benchmark_app/README.md b/samples/cpp/benchmark_app/README.md index 995ea6c8a28..45e587d062f 100644 --- a/samples/cpp/benchmark_app/README.md +++ b/samples/cpp/benchmark_app/README.md @@ -164,7 +164,7 @@ Options: -pcseq Optional. Report latencies for each shape in -data_shape sequence. -dump_config Optional. Path to JSON file to dump IE parameters, which were set by application. -load_config Optional. Path to JSON file to load custom IE parameters. Please note, command line parameters have higher priority then parameters from configuration file. - -infer_precision ""Optional. Inference precission + -infer_precision ""Optional. Inference precision -ip Optional. Specifies precision for all input layers of the network. -op Optional. Specifies precision for all output layers of the network. -iop "" Optional. Specifies precision for input and output layers by name. @@ -271,4 +271,4 @@ Below are fragments of sample output static and dynamic networks: ## See Also * [Using OpenVINO Runtime Samples](../../../docs/OV_Runtime_UG/Samples_Overview.md) * [Model Optimizer](../../../docs/MO_DG/Deep_Learning_Model_Optimizer_DevGuide.md) -* [Model Downloader](@ref omz_tools_downloader) \ No newline at end of file +* [Model Downloader](@ref omz_tools_downloader) diff --git a/samples/cpp/benchmark_app/benchmark_app.hpp b/samples/cpp/benchmark_app/benchmark_app.hpp index 19209bccf49..0d3d41f8c9b 100644 --- a/samples/cpp/benchmark_app/benchmark_app.hpp +++ b/samples/cpp/benchmark_app/benchmark_app.hpp @@ -199,7 +199,7 @@ static const char load_from_file_message[] = "Optional. Loads model from file di " All CNNNetwork options (like re-shape) will be ignored"; // @brief message for inference_precision -static const char inference_precision_message[] = "Optional. Inference precission"; +static const char inference_precision_message[] = "Optional. Inference precision"; static constexpr char inputs_precision_message[] = "Optional. Specifies precision for all input layers of the network.";