Files
openvino/inference-engine/samples/style_transfer_sample/style_transfer_sample.h
Kate Generalova 99df448669 refactor: add clang style check for samples (#5306)
* refactor: add clang style check for samples

* fix: add .clang-format for ie

* fix: style check for missing headers

* refactor: remove cpplint for IE samples

* fix: setw is not a member of std for classification_results.hpp

* feat: add indent after ifdefine

* feat: set up google style for IE samples

* fix indents for w_dirent headers

* fix: include issues for utils.cpp due to clang-format
2021-04-22 14:02:54 +03:00

89 lines
3.8 KiB
C++

// Copyright (C) 2018-2021 Intel Corporation
// SPDX-License-Identifier: Apache-2.0
//
#pragma once
#include <gflags/gflags.h>
#include <iostream>
#include <string>
#include <vector>
/// @brief message for help argument
static const char help_message[] = "Print a usage message.";
/// @brief message for images argument
static const char image_message[] = "Path to a folder with images or paths to image files.";
/// @brief message for model argument
static const char model_message[] = "Required. Path to an .xml file with a trained model.";
/// @brief message for assigning cnn calculation to device
static const char target_device_message[] = "Optional. Specify the target device to infer on (the list of available devices is shown "
"below). "
"Default value is CPU. Use \"-d HETERO:<comma_separated_devices_list>\" format to specify "
"HETERO plugin. "
"Sample will look for a suitable plugin for device specified";
/// @brief message for plugin custom kernels desc
static const char custom_plugin_cfg_message[] = "Required for GPU, MYRIAD, HDDL custom kernels. "
"Absolute path to the .xml config file with the kernels descriptions.";
/// @brief message for user library argument
static const char custom_ex_library_message[] = "Required for CPU plugin custom layers. "
"Absolute path to a shared library with the kernels implementations.";
/// @brief message for mean values arguments
static const char preprocess_data_message[] = "Mean values. Required if the model needs mean values for preprocessing and postprocessing.";
/// @brief Define flag for showing help message <br>
DEFINE_bool(h, false, help_message);
/// @brief Define parameter for set image file or folder with images <br>
/// It is a required parameter
DEFINE_string(i, "", image_message);
/// @brief Define parameter for set model file <br>
/// It is a required parameter
DEFINE_string(m, "", model_message);
/// @brief device the target device to infer on <br>
/// It is an optional parameter
DEFINE_string(d, "CPU", target_device_message);
/// @brief Define parameter for plugin custom kernels path <br>
/// It is an optional parameter
DEFINE_string(c, "", custom_plugin_cfg_message);
/// @brief Absolute path to CPU extension library with user layers <br>
/// It is an optional parameter
DEFINE_string(l, "", custom_ex_library_message);
/// @brief Define parameter preprocess_data for rgb channels <br>
/// (default 0) for each channels
/// It is an optional parameter
DEFINE_double(mean_val_r, 0.0, preprocess_data_message);
DEFINE_double(mean_val_g, 0.0, preprocess_data_message);
DEFINE_double(mean_val_b, 0.0, preprocess_data_message);
/**
* @brief This function show a help message
*/
static void showUsage() {
std::cout << std::endl;
std::cout << "style_transfer_sample [OPTION]" << std::endl;
std::cout << "Options:" << std::endl;
std::cout << std::endl;
std::cout << " -h " << help_message << std::endl;
std::cout << " -i \"<path>\" " << image_message << std::endl;
std::cout << " -m \"<path>\" " << model_message << std::endl;
std::cout << " -l \"<absolute_path>\" " << custom_ex_library_message << std::endl;
std::cout << " Or" << std::endl;
std::cout << " -c \"<absolute_path>\" " << custom_plugin_cfg_message << std::endl;
std::cout << " -d \"<device>\" " << target_device_message << std::endl;
std::cout << " -mean_val_r," << std::endl;
std::cout << " -mean_val_g," << std::endl;
std::cout << " -mean_val_b " << preprocess_data_message << std::endl;
}