#include int main() { { //! [part0] ov::Core core; // Read a network in IR, PaddlePaddle, or ONNX format: std::shared_ptr model = core.read_model("sample.xml"); // compile a model on AUTO using the default list of device candidates. // The following lines are equivalent: ov::CompiledModel model0 = core.compile_model(model); ov::CompiledModel model1 = core.compile_model(model, "AUTO"); // Optional // You can also specify the devices to be used by AUTO. // The following lines are equivalent: ov::CompiledModel model3 = core.compile_model(model, "AUTO:GPU,CPU"); ov::CompiledModel model4 = core.compile_model(model, "AUTO", ov::device::priorities("GPU,CPU")); //Optional // the AUTO plugin is pre-configured (globally) with the explicit option: core.set_property("AUTO", ov::device::priorities("GPU,CPU")); //! [part0] } return 0; }