* Renamed hetero md * Renamed some guides * Updated OpenVINO_Runtime_User_Guide.md * Updated plugin's page * More updates * Fixed links * Updated link names * Fixed links * Fixed docs build * Self-review * Fixed issues in doc snippets
68 lines
2.0 KiB
C++
68 lines
2.0 KiB
C++
#include <openvino/runtime/core.hpp>
|
|
|
|
int main() {
|
|
//! [get_available_devices]
|
|
ov::Core core;
|
|
std::vector<std::string> available_devices = core.get_available_devices();
|
|
//! [get_available_devices]
|
|
|
|
//! [hetero_priorities]
|
|
auto device_priorites = core.get_property("HETERO", ov::device::priorities);
|
|
//! [hetero_priorities]
|
|
|
|
//! [cpu_device_name]
|
|
auto cpu_device_name = core.get_property("CPU", ov::device::full_name);
|
|
//! [cpu_device_name]
|
|
|
|
auto model = core.read_model("sample.xml");
|
|
{
|
|
//! [compile_model_with_property]
|
|
auto compiled_model = core.compile_model(model, "CPU",
|
|
ov::hint::performance_mode(ov::hint::PerformanceMode::THROUGHPUT),
|
|
ov::hint::inference_precision(ov::element::f32));
|
|
//! [compile_model_with_property]
|
|
}
|
|
|
|
{
|
|
//! [optimal_number_of_infer_requests]
|
|
auto compiled_model = core.compile_model(model, "CPU");
|
|
auto nireq = compiled_model.get_property(ov::optimal_number_of_infer_requests);
|
|
//! [optimal_number_of_infer_requests]
|
|
}
|
|
{
|
|
//! [core_set_property_then_compile]
|
|
// set letency hint is a default for CPU
|
|
core.set_property("CPU", ov::hint::performance_mode(ov::hint::PerformanceMode::LATENCY));
|
|
// compiled with latency configuration hint
|
|
auto compiled_model_latency = core.compile_model(model, "CPU");
|
|
// compiled with overriden ov::hint::performance_mode value
|
|
auto compiled_model_thrp = core.compile_model(model, "CPU",
|
|
ov::hint::performance_mode(ov::hint::PerformanceMode::THROUGHPUT));
|
|
//! [core_set_property_then_compile]
|
|
}
|
|
|
|
{
|
|
//! [device_thermal]
|
|
auto compiled_model = core.compile_model(model, "MYRIAD");
|
|
float temperature = compiled_model.get_property(ov::device::thermal);
|
|
//! [device_thermal]
|
|
}
|
|
|
|
{
|
|
//! [inference_num_threads]
|
|
auto compiled_model = core.compile_model(model, "CPU");
|
|
auto nthreads = compiled_model.get_property(ov::inference_num_threads);
|
|
//! [inference_num_threads]
|
|
}
|
|
|
|
{
|
|
//! [multi_device]
|
|
auto compiled_model = core.compile_model(model, "MULTI",
|
|
ov::device::priorities("CPU", "GPU"));
|
|
// change the order of priorities
|
|
compiled_model.set_property(ov::device::priorities("GPU", "CPU"));
|
|
//! [multi_device]
|
|
}
|
|
return 0;
|
|
}
|