Files
openvino/inference-engine/samples/benchmark_app/main.cpp

681 lines
32 KiB
C++
Raw Normal View History

// Copyright (C) 2018-2021 Intel Corporation
2018-11-23 16:19:43 +03:00
// SPDX-License-Identifier: Apache-2.0
//
#include <algorithm>
#include <chrono>
#include <memory>
#include <map>
#include <string>
#include <vector>
#include <utility>
#include <inference_engine.hpp>
2019-04-12 18:25:53 +03:00
#include <vpu/vpu_plugin_config.hpp>
2019-08-09 19:02:42 +03:00
#include <cldnn/cldnn_config.hpp>
#include <gna/gna_config.hpp>
2018-11-23 16:19:43 +03:00
#include <samples/common.hpp>
#include <samples/slog.hpp>
#include <samples/args_helper.hpp>
2019-04-12 18:25:53 +03:00
#include "benchmark_app.hpp"
#include "infer_request_wrap.hpp"
#include "progress_bar.hpp"
#include "statistics_report.hpp"
2019-08-09 19:02:42 +03:00
#include "inputs_filling.hpp"
#include "utils.hpp"
2018-11-23 16:19:43 +03:00
using namespace InferenceEngine;
2019-08-09 19:02:42 +03:00
static const size_t progressBarDefaultTotalCount = 1000;
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
uint64_t getDurationInMilliseconds(uint32_t duration) {
return duration * 1000LL;
}
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
uint64_t getDurationInNanoseconds(uint32_t duration) {
return duration * 1000000000LL;
}
2018-11-23 16:19:43 +03:00
2019-04-12 18:25:53 +03:00
bool ParseAndCheckCommandLine(int argc, char *argv[]) {
2019-08-09 19:02:42 +03:00
// ---------------------------Parsing and validating input arguments--------------------------------------
2019-04-12 18:25:53 +03:00
slog::info << "Parsing input parameters" << slog::endl;
gflags::ParseCommandLineNonHelpFlags(&argc, &argv, true);
2019-08-09 19:02:42 +03:00
if (FLAGS_help || FLAGS_h) {
2019-04-12 18:25:53 +03:00
showUsage();
2019-08-09 19:02:42 +03:00
showAvailableDevices();
2019-04-12 18:25:53 +03:00
return false;
}
2018-11-23 16:19:43 +03:00
2019-04-12 18:25:53 +03:00
if (FLAGS_m.empty()) {
showUsage();
2019-08-09 19:02:42 +03:00
throw std::logic_error("Model is required but not set. Please set -m option.");
2019-04-12 18:25:53 +03:00
}
2018-11-23 16:19:43 +03:00
2019-04-12 18:25:53 +03:00
if (FLAGS_api != "async" && FLAGS_api != "sync") {
2019-08-09 19:02:42 +03:00
throw std::logic_error("Incorrect API. Please set -api option to `sync` or `async` value.");
2019-04-12 18:25:53 +03:00
}
2018-11-23 16:19:43 +03:00
2019-04-12 18:25:53 +03:00
if (!FLAGS_report_type.empty() &&
FLAGS_report_type != noCntReport && FLAGS_report_type != averageCntReport && FLAGS_report_type != detailedCntReport) {
2019-08-09 19:02:42 +03:00
std::string err = "only " + std::string(noCntReport) + "/" + std::string(averageCntReport) + "/" + std::string(detailedCntReport) +
" report types are supported (invalid -report_type option value)";
2019-04-12 18:25:53 +03:00
throw std::logic_error(err);
}
2019-10-04 19:26:43 +03:00
if ((FLAGS_report_type == averageCntReport) && ((FLAGS_d.find("MULTI") != std::string::npos))) {
throw std::logic_error("only " + std::string(detailedCntReport) + " report type is supported for MULTI device");
}
2019-04-12 18:25:53 +03:00
return true;
}
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
static void next_step(const std::string additional_info = "") {
static size_t step_id = 0;
static const std::map<size_t, std::string> step_names = {
{ 1, "Parsing and validating input arguments" },
{ 2, "Loading Inference Engine" },
{ 3, "Setting device configuration" },
{ 4, "Reading network files" },
{ 5, "Resizing network to match image sizes and given batch" },
{ 6, "Configuring input of the model" },
{ 7, "Loading the model to the device" },
{ 8, "Setting optimal runtime parameters" },
{ 9, "Creating infer requests and filling input blobs with images" },
{ 10, "Measuring performance" },
{ 11, "Dumping statistics report" }
2019-08-09 19:02:42 +03:00
};
step_id++;
if (step_names.count(step_id) == 0)
THROW_IE_EXCEPTION << "Step ID " << step_id << " is out of total steps number " << step_names.size();
std::cout << "[Step " << step_id << "/" << step_names.size() << "] " << step_names.at(step_id)
<< (additional_info.empty() ? "" : " (" + additional_info + ")") << std::endl;
}
2019-10-04 19:26:43 +03:00
template <typename T>
T getMedianValue(const std::vector<T> &vec) {
std::vector<T> sortedVec(vec);
std::sort(sortedVec.begin(), sortedVec.end());
return (sortedVec.size() % 2 != 0) ?
sortedVec[sortedVec.size() / 2ULL] :
(sortedVec[sortedVec.size() / 2ULL] + sortedVec[sortedVec.size() / 2ULL - 1ULL]) / static_cast<T>(2.0);
}
2019-04-12 18:25:53 +03:00
/**
2019-08-09 19:02:42 +03:00
* @brief The entry point of the benchmark application
2019-04-12 18:25:53 +03:00
*/
int main(int argc, char *argv[]) {
2019-10-04 19:26:43 +03:00
std::shared_ptr<StatisticsReport> statistics;
2019-04-12 18:25:53 +03:00
try {
2020-02-11 22:48:49 +03:00
ExecutableNetwork exeNetwork;
2019-08-09 19:02:42 +03:00
// ----------------- 1. Parsing and validating input arguments -------------------------------------------------
next_step();
2019-04-12 18:25:53 +03:00
if (!ParseAndCheckCommandLine(argc, argv)) {
return 0;
2018-11-23 16:19:43 +03:00
}
2020-02-11 22:48:49 +03:00
bool isNetworkCompiled = fileExt(FLAGS_m) == "blob";
if (isNetworkCompiled) {
slog::info << "Network is compiled" << slog::endl;
}
std::vector<gflags::CommandLineFlagInfo> flags;
StatisticsReport::Parameters command_line_arguments;
gflags::GetAllFlags(&flags);
for (auto &flag : flags) {
if (!flag.is_default) {
command_line_arguments.push_back({ flag.name, flag.current_value });
2019-10-04 19:26:43 +03:00
}
}
if (!FLAGS_report_type.empty()) {
2019-10-04 19:26:43 +03:00
statistics = std::make_shared<StatisticsReport>(StatisticsReport::Config{FLAGS_report_type, FLAGS_report_folder});
statistics->addParameters(StatisticsReport::Category::COMMAND_LINE_PARAMETERS, command_line_arguments);
}
auto isFlagSetInCommandLine = [&command_line_arguments] (const std::string& name) {
return (std::find_if(command_line_arguments.begin(), command_line_arguments.end(),
[ name ] (const std::pair<std::string, std::string>& p) { return p.first == name;}) != command_line_arguments.end());
};
std::string device_name = FLAGS_d;
// Parse devices
auto devices = parseDevices(device_name);
2019-10-04 19:26:43 +03:00
// Parse nstreams per device
std::map<std::string, std::string> device_nstreams = parseNStreamsValuePerDevice(devices, FLAGS_nstreams);
// Load device config file if specified
std::map<std::string, std::map<std::string, std::string>> config;
#ifdef USE_OPENCV
if (!FLAGS_load_config.empty()) {
load_config(FLAGS_load_config, config);
}
#endif
2019-04-12 18:25:53 +03:00
/** This vector stores paths to the processed images **/
2019-08-09 19:02:42 +03:00
std::vector<std::string> inputFiles;
parseInputFilesArguments(inputFiles);
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
// ----------------- 2. Loading the Inference Engine -----------------------------------------------------------
next_step();
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
Core ie;
2020-02-11 22:48:49 +03:00
if (FLAGS_d.find("CPU") != std::string::npos && !FLAGS_l.empty()) {
// CPU (MKLDNN) extensions is loaded as a shared library and passed as a pointer to base extension
2021-03-05 12:08:01 +03:00
const auto extension_ptr = std::make_shared<InferenceEngine::Extension>(FLAGS_l);
ie.AddExtension(extension_ptr);
2020-02-11 22:48:49 +03:00
slog::info << "CPU (MKLDNN) extensions is loaded " << FLAGS_l << slog::endl;
}
// Load clDNN Extensions
if ((FLAGS_d.find("GPU") != std::string::npos) && !FLAGS_c.empty()) {
// Override config if command line parameter is specified
if (!config.count("GPU"))
config["GPU"] = {};
config["GPU"][CONFIG_KEY(CONFIG_FILE)] = FLAGS_c;
}
if (config.count("GPU") && config.at("GPU").count(CONFIG_KEY(CONFIG_FILE))) {
auto ext = config.at("GPU").at(CONFIG_KEY(CONFIG_FILE));
ie.SetConfig({{ CONFIG_KEY(CONFIG_FILE), ext }}, "GPU");
slog::info << "GPU extensions is loaded " << ext << slog::endl;
2018-11-23 16:19:43 +03:00
}
2019-08-09 19:02:42 +03:00
slog::info << "InferenceEngine: " << GetInferenceEngineVersion() << slog::endl;
slog::info << "Device info: " << slog::endl;
std::cout << ie.GetVersions(device_name) << std::endl;
2018-11-23 16:19:43 +03:00
2020-02-11 22:48:49 +03:00
// ----------------- 3. Setting device configuration -----------------------------------------------------------
2019-08-09 19:02:42 +03:00
next_step();
bool perf_counts = false;
// Update config per device according to command line parameters
for (auto& device : devices) {
if (!config.count(device)) config[device] = {};
std::map<std::string, std::string>& device_config = config.at(device);
// Set performance counter
if (isFlagSetInCommandLine("pc")) {
// set to user defined value
device_config[CONFIG_KEY(PERF_COUNT)] = FLAGS_pc ? CONFIG_VALUE(YES) : CONFIG_VALUE(NO);
} else if (device_config.count(CONFIG_KEY(PERF_COUNT)) &&
(device_config.at(CONFIG_KEY(PERF_COUNT)) == "YES")) {
slog::warn << "Performance counters for " << device <<
" device is turned on. To print results use -pc option." << slog::endl;
} else if (FLAGS_report_type == detailedCntReport || FLAGS_report_type == averageCntReport) {
slog::warn << "Turn on performance counters for " << device <<
" device since report type is " << FLAGS_report_type << "." << slog::endl;
device_config[CONFIG_KEY(PERF_COUNT)] = CONFIG_VALUE(YES);
} else if (!FLAGS_exec_graph_path.empty()) {
slog::warn << "Turn on performance counters for " << device <<
" device due to execution graph dumping." << slog::endl;
device_config[CONFIG_KEY(PERF_COUNT)] = CONFIG_VALUE(YES);
} else {
// set to default value
device_config[CONFIG_KEY(PERF_COUNT)] = FLAGS_pc ? CONFIG_VALUE(YES) : CONFIG_VALUE(NO);
2020-04-13 21:17:23 +03:00
}
perf_counts = (device_config.at(CONFIG_KEY(PERF_COUNT)) == CONFIG_VALUE(YES)) ? true : perf_counts;
auto setThroughputStreams = [&] () {
const std::string key = device + "_THROUGHPUT_STREAMS";
if (device_nstreams.count(device)) {
// set to user defined value
std::vector<std::string> supported_config_keys = ie.GetMetric(device, METRIC_KEY(SUPPORTED_CONFIG_KEYS));
if (std::find(supported_config_keys.begin(), supported_config_keys.end(), key) == supported_config_keys.end()) {
throw std::logic_error("Device " + device + " doesn't support config key '" + key + "'! " +
"Please specify -nstreams for correct devices in format <dev1>:<nstreams1>,<dev2>:<nstreams2>" +
" or via configuration file.");
}
device_config[key] = device_nstreams.at(device);
} else if (!device_config.count(key) && (FLAGS_api == "async")) {
slog::warn << "-nstreams default value is determined automatically for " << device << " device. "
2021-02-16 13:08:54 +09:00
"Although the automatic selection usually provides a reasonable performance, "
"but it still may be non-optimal for some cases, for more information look at README." << slog::endl;
if (std::string::npos == device.find("MYRIAD")) // MYRIAD sets the default number of streams implicitly (without _AUTO)
device_config[key] = std::string(device + "_THROUGHPUT_AUTO");
}
if (device_config.count(key))
device_nstreams[device] = device_config.at(key);
};
2020-04-13 21:17:23 +03:00
2019-08-09 19:02:42 +03:00
if (device == "CPU") { // CPU supports few special performance-oriented keys
// limit threading for CPU portion of inference
if (isFlagSetInCommandLine("nthreads"))
device_config[CONFIG_KEY(CPU_THREADS_NUM)] = std::to_string(FLAGS_nthreads);
if (isFlagSetInCommandLine("enforcebf16"))
device_config[CONFIG_KEY(ENFORCE_BF16)] = FLAGS_enforcebf16 ? CONFIG_VALUE(YES) : CONFIG_VALUE(NO);
if (isFlagSetInCommandLine("pin")) {
// set to user defined value
device_config[CONFIG_KEY(CPU_BIND_THREAD)] = FLAGS_pin;
} else if (!device_config.count(CONFIG_KEY(CPU_BIND_THREAD))) {
if ((device_name.find("MULTI") != std::string::npos) &&
(device_name.find("GPU") != std::string::npos)) {
slog::warn << "Turn off threads pinning for " << device <<
" device since multi-scenario with GPU device is used." << slog::endl;
device_config[CONFIG_KEY(CPU_BIND_THREAD)] = CONFIG_VALUE(NO);
} else {
// set to default value
device_config[CONFIG_KEY(CPU_BIND_THREAD)] = FLAGS_pin;
}
2019-10-04 19:26:43 +03:00
}
2019-08-09 19:02:42 +03:00
// for CPU execution, more throughput-oriented execution via streams
setThroughputStreams();
2019-08-09 19:02:42 +03:00
} else if (device == ("GPU")) {
// for GPU execution, more throughput-oriented execution via streams
setThroughputStreams();
2019-10-04 19:26:43 +03:00
if ((device_name.find("MULTI") != std::string::npos) &&
(device_name.find("CPU") != std::string::npos)) {
slog::warn << "Turn on GPU trottling. Multi-device execution with the CPU + GPU performs best with GPU trottling hint," <<
"which releases another CPU thread (that is otherwise used by the GPU driver for active polling)"<< slog::endl;
device_config[CLDNN_CONFIG_KEY(PLUGIN_THROTTLE)] = "1";
2019-10-04 19:26:43 +03:00
}
2019-08-09 19:02:42 +03:00
} else if (device == "MYRIAD") {
device_config[CONFIG_KEY(LOG_LEVEL)] = CONFIG_VALUE(LOG_WARNING);
setThroughputStreams();
} else if (device == "GNA") {
if (FLAGS_qb == 8)
device_config[GNA_CONFIG_KEY(PRECISION)] = "I8";
else
device_config[GNA_CONFIG_KEY(PRECISION)] = "I16";
if (isFlagSetInCommandLine("nthreads"))
device_config[GNA_CONFIG_KEY(LIB_N_THREADS)] = std::to_string(FLAGS_nthreads);
} else {
std::vector<std::string> supported_config_keys = ie.GetMetric(device, METRIC_KEY(SUPPORTED_CONFIG_KEYS));
auto supported = [&] (const std::string& key) {
return std::find(std::begin(supported_config_keys), std::end(supported_config_keys), key)
!= std::end(supported_config_keys);
};
if (supported(CONFIG_KEY(CPU_THREADS_NUM)) && isFlagSetInCommandLine("nthreads")) {
device_config[CONFIG_KEY(CPU_THREADS_NUM)] = std::to_string(FLAGS_nthreads);
}
if (supported(CONFIG_KEY(CPU_THROUGHPUT_STREAMS)) && isFlagSetInCommandLine("nstreams")) {
device_config[CONFIG_KEY(CPU_THROUGHPUT_STREAMS)] = FLAGS_nstreams;
}
if (supported(CONFIG_KEY(CPU_BIND_THREAD)) && isFlagSetInCommandLine("pin")) {
device_config[CONFIG_KEY(CPU_BIND_THREAD)] = FLAGS_pin;
}
2019-08-09 19:02:42 +03:00
}
2018-11-23 16:19:43 +03:00
}
for (auto&& item : config) {
ie.SetConfig(item.second, item.first);
}
2020-02-11 22:48:49 +03:00
auto double_to_string = [] (const double number) {
std::stringstream ss;
ss << std::fixed << std::setprecision(2) << number;
return ss.str();
};
2020-02-11 22:48:49 +03:00
auto get_total_ms_time = [] (Time::time_point& startTime) {
return std::chrono::duration_cast<ns>(Time::now() - startTime).count() * 0.000001;
};
2019-08-09 19:02:42 +03:00
2020-02-11 22:48:49 +03:00
size_t batchSize = FLAGS_b;
Precision precision = Precision::UNSPECIFIED;
std::string topology_name = "";
benchmark_app::InputsInfo app_inputs_info;
std::string output_name;
2020-02-11 22:48:49 +03:00
if (!isNetworkCompiled) {
// ----------------- 4. Reading the Intermediate Representation network ----------------------------------------
next_step();
slog::info << "Loading network files" << slog::endl;
auto startTime = Time::now();
CNNNetwork cnnNetwork = ie.ReadNetwork(FLAGS_m);
auto duration_ms = double_to_string(get_total_ms_time(startTime));
slog::info << "Read network took " << duration_ms << " ms" << slog::endl;
if (statistics)
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"read network time (ms)", duration_ms}
2020-02-11 22:48:49 +03:00
});
const InputsDataMap inputInfo(cnnNetwork.getInputsInfo());
if (inputInfo.empty()) {
throw std::logic_error("no inputs info is provided");
}
// ----------------- 5. Resizing network to match image sizes and given batch ----------------------------------
next_step();
2020-04-13 21:17:23 +03:00
batchSize = cnnNetwork.getBatchSize();
// Parse input shapes if specified
bool reshape = false;
app_inputs_info = getInputsInfo<InputInfo::Ptr>(FLAGS_shape, FLAGS_layout, FLAGS_b, inputInfo, reshape);
if (reshape) {
InferenceEngine::ICNNNetwork::InputShapes shapes = {};
for (auto& item : app_inputs_info)
shapes[item.first] = item.second.shape;
slog::info << "Reshaping network: " << getShapesString(shapes) << slog::endl;
startTime = Time::now();
cnnNetwork.reshape(shapes);
auto duration_ms = double_to_string(get_total_ms_time(startTime));
slog::info << "Reshape network took " << duration_ms << " ms" << slog::endl;
if (statistics)
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"reshape network time (ms)", duration_ms}
});
2020-02-11 22:48:49 +03:00
}
// use batch size according to provided layout and shapes
batchSize = (!FLAGS_layout.empty()) ? getBatchSize(app_inputs_info) : cnnNetwork.getBatchSize();
2020-02-11 22:48:49 +03:00
topology_name = cnnNetwork.getName();
2020-04-13 21:17:23 +03:00
slog::info << (FLAGS_b != 0 ? "Network batch size was changed to: " : "Network batch size: ") << batchSize << slog::endl;
2020-02-11 22:48:49 +03:00
// ----------------- 6. Configuring input ----------------------------------------------------------------------
next_step();
for (auto& item : inputInfo) {
if (app_inputs_info.at(item.first).isImage()) {
2020-02-11 22:48:49 +03:00
/** Set the precision of input data provided by the user, should be called before load of the network to the device **/
app_inputs_info.at(item.first).precision = Precision::U8;
item.second->setPrecision(app_inputs_info.at(item.first).precision);
2020-02-11 22:48:49 +03:00
}
}
// ----------------- 7. Loading the model to the device --------------------------------------------------------
next_step();
startTime = Time::now();
exeNetwork = ie.LoadNetwork(cnnNetwork, device_name);
2020-02-11 22:48:49 +03:00
duration_ms = double_to_string(get_total_ms_time(startTime));
slog::info << "Load network took " << duration_ms << " ms" << slog::endl;
if (statistics)
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"load network time (ms)", duration_ms}
2020-02-11 22:48:49 +03:00
});
} else {
next_step();
slog::info << "Skipping the step for compiled network" << slog::endl;
next_step();
slog::info << "Skipping the step for compiled network" << slog::endl;
next_step();
slog::info << "Skipping the step for compiled network" << slog::endl;
// ----------------- 7. Loading the model to the device --------------------------------------------------------
next_step();
auto startTime = Time::now();
exeNetwork = ie.ImportNetwork(FLAGS_m, device_name, {});
auto duration_ms = double_to_string(get_total_ms_time(startTime));
slog::info << "Import network took " << duration_ms << " ms" << slog::endl;
if (statistics)
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"import network time (ms)", duration_ms}
2020-02-11 22:48:49 +03:00
});
app_inputs_info = getInputsInfo<InputInfo::CPtr>(FLAGS_shape, FLAGS_layout, FLAGS_b, exeNetwork.GetInputsInfo());
2020-02-11 22:48:49 +03:00
if (batchSize == 0) {
batchSize = 1;
}
}
2019-08-09 19:02:42 +03:00
// ----------------- 8. Setting optimal runtime parameters -----------------------------------------------------
next_step();
// Update number of streams
for (auto&& ds : device_nstreams) {
const std::string key = ds.first + "_THROUGHPUT_STREAMS";
device_nstreams[ds.first] = ie.GetConfig(ds.first, key).as<std::string>();
}
2019-08-09 19:02:42 +03:00
// Number of requests
uint32_t nireq = FLAGS_nireq;
if (nireq == 0) {
2020-04-13 21:17:23 +03:00
if (FLAGS_api == "sync") {
nireq = 1;
} else {
std::string key = METRIC_KEY(OPTIMAL_NUMBER_OF_INFER_REQUESTS);
try {
nireq = exeNetwork.GetMetric(key).as<unsigned int>();
} catch (const std::exception& ex) {
2020-04-13 21:17:23 +03:00
THROW_IE_EXCEPTION
<< "Every device used with the benchmark_app should "
<< "support OPTIMAL_NUMBER_OF_INFER_REQUESTS ExecutableNetwork metric. "
<< "Failed to query the metric for the " << device_name << " with error:" << ex.what();
}
2018-11-23 16:19:43 +03:00
}
}
2019-08-09 19:02:42 +03:00
// Iteration limit
uint32_t niter = FLAGS_niter;
if ((niter > 0) && (FLAGS_api == "async")) {
niter = ((niter + nireq - 1)/nireq)*nireq;
if (FLAGS_niter != niter) {
slog::warn << "Number of iterations was aligned by request number from "
<< FLAGS_niter << " to " << niter << " using number of requests " << nireq << slog::endl;
}
}
2019-04-12 18:25:53 +03:00
2019-08-09 19:02:42 +03:00
// Time limit
uint32_t duration_seconds = 0;
if (FLAGS_t != 0) {
// time limit
duration_seconds = FLAGS_t;
} else if (FLAGS_niter == 0) {
// default time limit
duration_seconds = deviceDefaultDeviceDurationInSeconds(device_name);
2019-04-12 18:25:53 +03:00
}
2019-08-09 19:02:42 +03:00
uint64_t duration_nanoseconds = getDurationInNanoseconds(duration_seconds);
2019-04-12 18:25:53 +03:00
2019-10-04 19:26:43 +03:00
if (statistics) {
statistics->addParameters(StatisticsReport::Category::RUNTIME_CONFIG,
{
{"topology", topology_name},
{"target device", device_name},
{"API", FLAGS_api},
{"precision", std::string(precision.name())},
{"batch size", std::to_string(batchSize)},
{"number of iterations", std::to_string(niter)},
{"number of parallel infer requests", std::to_string(nireq)},
{"duration (ms)", std::to_string(getDurationInMilliseconds(duration_seconds))},
2019-10-04 19:26:43 +03:00
});
for (auto& nstreams : device_nstreams) {
std::stringstream ss;
ss << "number of " << nstreams.first << " streams";
statistics->addParameters(StatisticsReport::Category::RUNTIME_CONFIG,
{
{ss.str(), nstreams.second},
2019-10-04 19:26:43 +03:00
});
}
}
2019-08-09 19:02:42 +03:00
// ----------------- 9. Creating infer requests and filling input blobs ----------------------------------------
next_step();
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
InferRequestsQueue inferRequestsQueue(exeNetwork, nireq);
fillBlobs(inputFiles, batchSize, app_inputs_info, inferRequestsQueue.requests);
2019-04-12 18:25:53 +03:00
2019-08-09 19:02:42 +03:00
// ----------------- 10. Measuring performance ------------------------------------------------------------------
size_t progressCnt = 0;
size_t progressBarTotalCount = progressBarDefaultTotalCount;
size_t iteration = 0;
2019-04-12 18:25:53 +03:00
2019-08-09 19:02:42 +03:00
std::stringstream ss;
ss << "Start inference " << FLAGS_api << "hronously";
2019-08-09 19:02:42 +03:00
if (FLAGS_api == "async") {
if (!ss.str().empty()) {
ss << ", ";
}
ss << nireq << " inference requests";
std::stringstream device_ss;
for (auto& nstreams : device_nstreams) {
if (!device_ss.str().empty()) {
device_ss << ", ";
}
device_ss << nstreams.second << " streams for " << nstreams.first;
}
if (!device_ss.str().empty()) {
ss << " using " << device_ss.str();
2019-04-12 18:25:53 +03:00
}
}
2019-08-09 19:02:42 +03:00
ss << ", limits: ";
if (duration_seconds > 0) {
ss << getDurationInMilliseconds(duration_seconds) << " ms duration";
2018-11-23 16:19:43 +03:00
}
2019-08-09 19:02:42 +03:00
if (niter != 0) {
if (duration_seconds == 0) {
progressBarTotalCount = niter;
}
if (duration_seconds > 0) {
ss << ", ";
}
ss << niter << " iterations";
}
next_step(ss.str());
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
// warming up - out of scope
auto inferRequest = inferRequestsQueue.getIdleRequest();
if (!inferRequest) {
THROW_IE_EXCEPTION << "No idle Infer Requests!";
}
2019-04-12 18:25:53 +03:00
if (FLAGS_api == "sync") {
inferRequest->infer();
2019-08-09 19:02:42 +03:00
} else {
inferRequest->startAsync();
}
inferRequestsQueue.waitAll();
auto duration_ms = double_to_string(inferRequestsQueue.getLatencies()[0]);
slog::info << "First inference took " << duration_ms << " ms" << slog::endl;
if (statistics)
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"first inference time (ms)", duration_ms}
});
2019-08-09 19:02:42 +03:00
inferRequestsQueue.resetTimes();
2020-02-11 22:48:49 +03:00
auto startTime = Time::now();
2019-08-09 19:02:42 +03:00
auto execTime = std::chrono::duration_cast<ns>(Time::now() - startTime).count();
/** Start inference & calculate performance **/
/** to align number if iterations to guarantee that last infer requests are executed in the same conditions **/
ProgressBar progressBar(progressBarTotalCount, FLAGS_stream_output, FLAGS_progress);
while ((niter != 0LL && iteration < niter) ||
(duration_nanoseconds != 0LL && (uint64_t)execTime < duration_nanoseconds) ||
(FLAGS_api == "async" && iteration % nireq != 0)) {
inferRequest = inferRequestsQueue.getIdleRequest();
if (!inferRequest) {
THROW_IE_EXCEPTION << "No idle Infer Requests!";
}
2019-08-09 19:02:42 +03:00
if (FLAGS_api == "sync") {
2019-04-12 18:25:53 +03:00
inferRequest->infer();
2018-11-23 16:19:43 +03:00
} else {
2020-02-11 22:48:49 +03:00
// As the inference request is currently idle, the wait() adds no additional overhead (and should return immediately).
// The primary reason for calling the method is exception checking/re-throwing.
// Callback, that governs the actual execution can handle errors as well,
// but as it uses just error codes it has no details like what() method of `std::exception`
// So, rechecking for any exceptions here.
inferRequest->wait();
2019-08-09 19:02:42 +03:00
inferRequest->startAsync();
2018-11-23 16:19:43 +03:00
}
2019-08-09 19:02:42 +03:00
iteration++;
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
execTime = std::chrono::duration_cast<ns>(Time::now() - startTime).count();
2019-04-12 18:25:53 +03:00
2019-08-09 19:02:42 +03:00
if (niter > 0) {
progressBar.addProgress(1);
} else {
// calculate how many progress intervals are covered by current iteration.
// depends on the current iteration time and time of each progress interval.
// Previously covered progress intervals must be skipped.
auto progressIntervalTime = duration_nanoseconds / progressBarTotalCount;
size_t newProgress = execTime / progressIntervalTime - progressCnt;
progressBar.addProgress(newProgress);
progressCnt += newProgress;
2018-11-23 16:19:43 +03:00
}
2019-08-09 19:02:42 +03:00
}
2018-11-23 16:19:43 +03:00
2019-08-09 19:02:42 +03:00
// wait the latest inference executions
inferRequestsQueue.waitAll();
2019-10-04 19:26:43 +03:00
double latency = getMedianValue<double>(inferRequestsQueue.getLatencies());
double totalDuration = inferRequestsQueue.getDurationInMilliseconds();
double fps = (FLAGS_api == "sync") ? batchSize * 1000.0 / latency :
batchSize * 1000.0 * iteration / totalDuration;
2019-10-04 19:26:43 +03:00
if (statistics) {
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"total execution time (ms)", double_to_string(totalDuration)},
{"total number of iterations", std::to_string(iteration)},
2019-10-04 19:26:43 +03:00
});
if (device_name.find("MULTI") == std::string::npos) {
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"latency (ms)", double_to_string(latency)},
2019-10-04 19:26:43 +03:00
});
2018-11-23 16:19:43 +03:00
}
2019-10-04 19:26:43 +03:00
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"throughput", double_to_string(fps)}
2019-10-04 19:26:43 +03:00
});
2019-04-12 18:25:53 +03:00
}
2019-08-09 19:02:42 +03:00
progressBar.finish();
// ----------------- 11. Dumping statistics report -------------------------------------------------------------
next_step();
2018-11-23 16:19:43 +03:00
#ifdef USE_OPENCV
if (!FLAGS_dump_config.empty()) {
dump_config(FLAGS_dump_config, config);
slog::info << "Inference Engine configuration settings were dumped to " << FLAGS_dump_config << slog::endl;
}
#endif
2019-04-12 18:25:53 +03:00
if (!FLAGS_exec_graph_path.empty()) {
2019-08-09 19:02:42 +03:00
try {
CNNNetwork execGraphInfo = exeNetwork.GetExecGraphInfo();
execGraphInfo.serialize(FLAGS_exec_graph_path);
slog::info << "executable graph is stored to " << FLAGS_exec_graph_path << slog::endl;
} catch (const std::exception & ex) {
slog::err << "Can't get executable graph: " << ex.what() << slog::endl;
}
}
2019-10-04 19:26:43 +03:00
if (perf_counts) {
std::vector<std::map<std::string, InferenceEngine::InferenceEngineProfileInfo>> perfCounts;
2019-08-09 19:02:42 +03:00
for (size_t ireq = 0; ireq < nireq; ireq++) {
2019-10-04 19:26:43 +03:00
auto reqPerfCounts = inferRequestsQueue.requests[ireq]->getPerformanceCounts();
if (FLAGS_pc) {
2021-02-16 13:08:54 +09:00
slog::info << "Performance counts for " << ireq << "-th infer request:" << slog::endl;
2019-10-04 19:26:43 +03:00
printPerformanceCounts(reqPerfCounts, std::cout, getFullDeviceName(ie, FLAGS_d), false);
}
perfCounts.push_back(reqPerfCounts);
}
if (statistics) {
statistics->dumpPerformanceCounters(perfCounts);
2019-08-09 19:02:42 +03:00
}
2018-11-23 16:19:43 +03:00
}
2019-04-12 18:25:53 +03:00
2019-10-04 19:26:43 +03:00
if (statistics)
statistics->dump();
2019-08-09 19:02:42 +03:00
std::cout << "Count: " << iteration << " iterations" << std::endl;
2020-02-11 22:48:49 +03:00
std::cout << "Duration: " << double_to_string(totalDuration) << " ms" << std::endl;
2019-10-04 19:26:43 +03:00
if (device_name.find("MULTI") == std::string::npos)
2020-02-11 22:48:49 +03:00
std::cout << "Latency: " << double_to_string(latency) << " ms" << std::endl;
std::cout << "Throughput: " << double_to_string(fps) << " FPS" << std::endl;
2018-11-23 16:19:43 +03:00
} catch (const std::exception& ex) {
slog::err << ex.what() << slog::endl;
2019-10-04 19:26:43 +03:00
if (statistics) {
statistics->addParameters(StatisticsReport::Category::EXECUTION_RESULTS,
{
{"error", ex.what()},
2019-10-04 19:26:43 +03:00
});
statistics->dump();
}
2018-11-23 16:19:43 +03:00
return 3;
}
return 0;
}