[Docs] Update python snippets with new properties imports (#19872)
This commit is contained in:
parent
2ec80439d7
commit
68ba8873a2
@ -1,8 +1,10 @@
|
||||
# Copyright (C) 2022 Intel Corporation
|
||||
# Copyright (C) 2018-2023 Intel Corporation
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
|
||||
|
||||
import openvino as ov
|
||||
import openvino.properties.device as device
|
||||
import openvino.properties.hint as hints
|
||||
|
||||
from snippets import get_model
|
||||
|
||||
@ -10,16 +12,16 @@ model = get_model()
|
||||
|
||||
#! [part0]
|
||||
core = ov.Core()
|
||||
cpu_optimization_capabilities = core.get_property("CPU", ov.properties.device.capabilities())
|
||||
cpu_optimization_capabilities = core.get_property("CPU", device.capabilities())
|
||||
#! [part0]
|
||||
|
||||
#! [part1]
|
||||
core = ov.Core()
|
||||
compiled_model = core.compile_model(model, "CPU")
|
||||
inference_precision = core.get_property("CPU", ov.properties.hint.inference_precision())
|
||||
inference_precision = core.get_property("CPU", hints.inference_precision())
|
||||
#! [part1]
|
||||
|
||||
#! [part2]
|
||||
core = ov.Core()
|
||||
core.set_property("CPU", {ov.properties.hint.inference_precision(): ov.Type.f32})
|
||||
core.set_property("CPU", {hints.inference_precision(): ov.Type.f32})
|
||||
#! [part2]
|
||||
|
@ -1,18 +1,19 @@
|
||||
# Copyright (C) 2023 Intel Corporation
|
||||
# Copyright (C) 2018-2023 Intel Corporation
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
|
||||
#! [ov:execution_mode:part0]
|
||||
import openvino as ov
|
||||
import openvino.properties.hint as hints
|
||||
|
||||
core = ov.Core()
|
||||
# in case of Accuracy
|
||||
core.set_property(
|
||||
"CPU",
|
||||
{ov.properties.hint.execution_mode(): ov.properties.hint.ExecutionMode.ACCURACY},
|
||||
{hints.execution_mode(): hints.ExecutionMode.ACCURACY},
|
||||
)
|
||||
# in case of Performance
|
||||
core.set_property(
|
||||
"CPU",
|
||||
{ov.properties.hint.execution_mode(): ov.properties.hint.ExecutionMode.PERFORMANCE},
|
||||
{hints.execution_mode(): hints.ExecutionMode.PERFORMANCE},
|
||||
)
|
||||
#! [ov:execution_mode:part0]
|
||||
|
@ -10,8 +10,10 @@ model = get_model()
|
||||
device_name = "CPU"
|
||||
xml_path = "model.xml"
|
||||
# ! [ov:intel_cpu:sparse_weights_decompression:part0]
|
||||
import openvino.properties.intel_cpu as intel_cpu
|
||||
|
||||
core = ov.Core()
|
||||
core.set_property("CPU", ov.properties.intel_cpu.sparse_weights_decompression_rate(0.8))
|
||||
core.set_property("CPU", intel_cpu.sparse_weights_decompression_rate(0.8))
|
||||
compiled_model = core.compile_model(model=model, device_name=device_name)
|
||||
# ! [ov:intel_cpu:sparse_weights_decompression:part0]
|
||||
assert compiled_model
|
||||
|
@ -39,12 +39,14 @@ def main():
|
||||
#! [compile_model_batch_plugin]
|
||||
|
||||
#! [compile_model_auto_batch]
|
||||
import openvino.properties.hint as hints
|
||||
|
||||
core = ov.Core()
|
||||
compiled_model = core.compile_model(
|
||||
model,
|
||||
"GPU",
|
||||
{
|
||||
ov.properties.hint.performance_mode(): ov.properties.hint.PerformanceMode.THROUGHPUT,
|
||||
hints.performance_mode(): hints.PerformanceMode.THROUGHPUT,
|
||||
},
|
||||
)
|
||||
#! [compile_model_auto_batch]
|
||||
|
@ -1,5 +1,13 @@
|
||||
import sys
|
||||
# Copyright (C) 2018-2023 Intel Corporation
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
#
|
||||
|
||||
import openvino as ov
|
||||
import openvino.properties as properties
|
||||
import openvino.properties.device as device
|
||||
import openvino.properties.hint as hints
|
||||
import openvino.properties.log as log
|
||||
|
||||
from openvino.inference_engine import IECore
|
||||
from utils import get_model, get_ngraph_model
|
||||
|
||||
@ -23,13 +31,13 @@ def part0():
|
||||
compiled_model = core.compile_model(
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={ov.properties.device.priorities(): "GPU,CPU"},
|
||||
config={device.priorities(): "GPU,CPU"},
|
||||
)
|
||||
|
||||
# Optional
|
||||
# the AUTO plugin is pre-configured (globally) with the explicit option:
|
||||
core.set_property(
|
||||
device_name="AUTO", properties={ov.properties.device.priorities(): "GPU,CPU"}
|
||||
device_name="AUTO", properties={device.priorities(): "GPU,CPU"}
|
||||
)
|
||||
#! [part0]
|
||||
|
||||
@ -52,13 +60,13 @@ def part1():
|
||||
exec_net = ie.load_network(
|
||||
network=net,
|
||||
device_name="AUTO",
|
||||
config={ov.properties.device.priorities(): "GPU,CPU"},
|
||||
config={device.priorities(): "GPU,CPU"},
|
||||
)
|
||||
|
||||
# Optional
|
||||
# the AUTO plugin is pre-configured (globally) with the explicit option:
|
||||
ie.set_config(
|
||||
config={ov.properties.device.priorities(): "GPU,CPU"}, device_name="AUTO"
|
||||
config={device.priorities(): "GPU,CPU"}, device_name="AUTO"
|
||||
)
|
||||
#! [part1]
|
||||
|
||||
@ -73,7 +81,7 @@ def part3():
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={
|
||||
ov.properties.hint.performance_mode(): ov.properties.hint.PerformanceMode.THROUGHPUT
|
||||
hints.performance_mode(): hints.PerformanceMode.THROUGHPUT
|
||||
},
|
||||
)
|
||||
# To use the “LATENCY” mode:
|
||||
@ -81,7 +89,7 @@ def part3():
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={
|
||||
ov.properties.hint.performance_mode(): ov.properties.hint.PerformanceMode.LATENCY
|
||||
hints.performance_mode(): hints.PerformanceMode.LATENCY
|
||||
},
|
||||
)
|
||||
# To use the “CUMULATIVE_THROUGHPUT” mode:
|
||||
@ -89,7 +97,7 @@ def part3():
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={
|
||||
ov.properties.hint.performance_mode(): ov.properties.hint.PerformanceMode.CUMULATIVE_THROUGHPUT
|
||||
hints.performance_mode(): hints.PerformanceMode.CUMULATIVE_THROUGHPUT
|
||||
},
|
||||
)
|
||||
#! [part3]
|
||||
@ -103,19 +111,19 @@ def part4():
|
||||
compiled_model0 = core.compile_model(
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={ov.properties.hint.model_priority(): ov.properties.hint.Priority.HIGH},
|
||||
config={hints.model_priority(): hints.Priority.HIGH},
|
||||
)
|
||||
compiled_model1 = core.compile_model(
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={
|
||||
ov.properties.hint.model_priority(): ov.properties.hint.Priority.MEDIUM
|
||||
hints.model_priority(): hints.Priority.MEDIUM
|
||||
},
|
||||
)
|
||||
compiled_model2 = core.compile_model(
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={ov.properties.hint.model_priority(): ov.properties.hint.Priority.LOW},
|
||||
config={hints.model_priority(): hints.Priority.LOW},
|
||||
)
|
||||
# Assume that all the devices (CPU and GPUs) can support all the networks.
|
||||
# Result: compiled_model0 will use GPU.1, compiled_model1 will use GPU.0, compiled_model2 will use CPU.
|
||||
@ -124,19 +132,19 @@ def part4():
|
||||
compiled_model3 = core.compile_model(
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={ov.properties.hint.model_priority(): ov.properties.hint.Priority.HIGH},
|
||||
config={hints.model_priority(): hints.Priority.HIGH},
|
||||
)
|
||||
compiled_model4 = core.compile_model(
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={
|
||||
ov.properties.hint.model_priority(): ov.properties.hint.Priority.MEDIUM
|
||||
hints.model_priority(): hints.Priority.MEDIUM
|
||||
},
|
||||
)
|
||||
compiled_model5 = core.compile_model(
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={ov.properties.hint.model_priority(): ov.properties.hint.Priority.LOW},
|
||||
config={hints.model_priority(): hints.Priority.LOW},
|
||||
)
|
||||
# Assume that all the devices (CPU ang GPUs) can support all the networks.
|
||||
# Result: compiled_model3 will use GPU.1, compiled_model4 will use GPU.1, compiled_model5 will use GPU.0.
|
||||
@ -161,12 +169,12 @@ def part6():
|
||||
compiled_model = core.compile_model(
|
||||
model=model,
|
||||
device_name="AUTO",
|
||||
config={ov.properties.log.level(): ov.properties.log.Level.DEBUG},
|
||||
config={log.level(): log.Level.DEBUG},
|
||||
)
|
||||
# set log level with set_property and compile model
|
||||
core.set_property(
|
||||
device_name="AUTO",
|
||||
properties={ov.properties.log.level(): ov.properties.log.Level.DEBUG},
|
||||
properties={log.level(): log.Level.DEBUG},
|
||||
)
|
||||
compiled_model = core.compile_model(model=model, device_name="AUTO")
|
||||
#! [part6]
|
||||
@ -179,7 +187,7 @@ def part7():
|
||||
# compile a model on AUTO and set log level to debug
|
||||
compiled_model = core.compile_model(model=model, device_name="AUTO")
|
||||
# query the runtime target devices on which the inferences are being executed
|
||||
execution_devices = compiled_model.get_property(ov.properties.execution_devices())
|
||||
execution_devices = compiled_model.get_property(properties.execution_devices())
|
||||
#! [part7]
|
||||
|
||||
|
||||
|
@ -15,8 +15,8 @@ def main():
|
||||
return 0
|
||||
|
||||
# [compile_model]
|
||||
import openvino.runtime.properties as props
|
||||
import openvino.runtime.properties.hint as hints
|
||||
import openvino.properties as props
|
||||
import openvino.properties.hint as hints
|
||||
|
||||
config = {hints.performance_mode(): hints.PerformanceMode.THROUGHPUT}
|
||||
compiled_model = core.compile_model(model, "GPU", config)
|
||||
|
@ -5,7 +5,7 @@
|
||||
from utils import get_path_to_model, get_temp_dir
|
||||
import openvino as ov
|
||||
|
||||
import openvino.runtime.properties as props
|
||||
import openvino.properties as props
|
||||
|
||||
device_name = 'GNA'
|
||||
model_path = get_path_to_model()
|
||||
@ -35,6 +35,8 @@ compiled_model = core.compile_model(model=model_path, device_name=device_name)
|
||||
assert compiled_model
|
||||
|
||||
# ! [ov:caching:part3]
|
||||
import openvino.properties.device as device
|
||||
|
||||
# Find 'EXPORT_IMPORT' capability in supported capabilities
|
||||
caching_supported = 'EXPORT_IMPORT' in core.get_property(device_name, props.device.capabilities())
|
||||
caching_supported = 'EXPORT_IMPORT' in core.get_property(device_name, device.capabilities())
|
||||
# ! [ov:caching:part3]
|
||||
|
@ -9,8 +9,10 @@ device_name = 'CPU'
|
||||
model = get_model()
|
||||
|
||||
# ! [ov:intel_cpu:denormals_optimization:part0]
|
||||
import openvino.properties.intel_cpu as intel_cpu
|
||||
|
||||
core = ov.Core()
|
||||
core.set_property("CPU", ov.properties.intel_cpu.denormals_optimization(True))
|
||||
core.set_property("CPU", intel_cpu.denormals_optimization(True))
|
||||
compiled_model = core.compile_model(model=model, device_name=device_name)
|
||||
# ! [ov:intel_cpu:denormals_optimization:part0]
|
||||
assert compiled_model
|
||||
|
@ -1,4 +1,5 @@
|
||||
import openvino as ov
|
||||
import openvino as properties
|
||||
from utils import get_model
|
||||
|
||||
def main():
|
||||
@ -35,16 +36,20 @@ def main():
|
||||
#! [fix_automatic_affinities]
|
||||
|
||||
#! [compile_model]
|
||||
import openvino.device as device
|
||||
|
||||
compiled_model = core.compile_model(model, device_name="HETERO:GPU,CPU")
|
||||
# device priorities via configuration property
|
||||
compiled_model = core.compile_model(
|
||||
model, device_name="HETERO", config={ov.properties.device.priorities(): "GPU,CPU"}
|
||||
model, device_name="HETERO", config={device.priorities(): "GPU,CPU"}
|
||||
)
|
||||
#! [compile_model]
|
||||
|
||||
#! [configure_fallback_devices]
|
||||
core.set_property("HETERO", {ov.properties.device.priorities(): "GPU,CPU"})
|
||||
core.set_property("GPU", {ov.properties.enable_profiling(): True})
|
||||
core.set_property("CPU", {ov.properties.hint.inference_precision(): ov.Type.f32})
|
||||
import openvino.hint as hints
|
||||
|
||||
core.set_property("HETERO", {device.priorities(): "GPU,CPU"})
|
||||
core.set_property("GPU", {properties.enable_profiling(): True})
|
||||
core.set_property("CPU", {hints.inference_precision(): ov.Type.f32})
|
||||
compiled_model = core.compile_model(model=model, device_name="HETERO")
|
||||
#! [configure_fallback_devices]
|
||||
|
@ -1,5 +1,9 @@
|
||||
import sys
|
||||
# Copyright (C) 2018-2023 Intel Corporation
|
||||
# SPDX-License-Identifier: Apache-2.0
|
||||
|
||||
import openvino as ov
|
||||
import openvino.properties as properties
|
||||
import openvino.properties.device as device
|
||||
from utils import get_model
|
||||
|
||||
model = get_model()
|
||||
@ -13,7 +17,7 @@ def MULTI_0():
|
||||
# Pre-configure MULTI globally with explicitly defined devices,
|
||||
# and compile the model on MULTI using the newly specified default device list.
|
||||
core.set_property(
|
||||
device_name="MULTI", properties={ov.properties.device.priorities(): "GPU,CPU"}
|
||||
device_name="MULTI", properties={device.priorities(): "GPU,CPU"}
|
||||
)
|
||||
compiled_model = core.compile_model(model=model, device_name="MULTI")
|
||||
|
||||
@ -24,7 +28,7 @@ def MULTI_0():
|
||||
compiled_model = core.compile_model(
|
||||
model=model,
|
||||
device_name="MULTI",
|
||||
config={ov.properties.device.priorities(): "GPU,CPU"},
|
||||
config={device.priorities(): "GPU,CPU"},
|
||||
)
|
||||
#! [MULTI_0]
|
||||
|
||||
@ -34,22 +38,22 @@ def MULTI_1():
|
||||
core = ov.Core()
|
||||
|
||||
core.set_property(
|
||||
device_name="MULTI", properties={ov.properties.device.priorities(): "CPU,GPU"}
|
||||
device_name="MULTI", properties={device.priorities(): "CPU,GPU"}
|
||||
)
|
||||
# Once the priority list is set, you can alter it on the fly:
|
||||
# reverse the order of priorities
|
||||
core.set_property(
|
||||
device_name="MULTI", properties={ov.properties.device.priorities(): "GPU,CPU"}
|
||||
device_name="MULTI", properties={device.priorities(): "GPU,CPU"}
|
||||
)
|
||||
|
||||
# exclude some devices (in this case, CPU)
|
||||
core.set_property(
|
||||
device_name="MULTI", properties={ov.properties.device.priorities(): "GPU"}
|
||||
device_name="MULTI", properties={device.priorities(): "GPU"}
|
||||
)
|
||||
|
||||
# bring back the excluded devices
|
||||
core.set_property(
|
||||
device_name="MULTI", properties={ov.properties.device.priorities(): "GPU,CPU"}
|
||||
device_name="MULTI", properties={device.priorities(): "GPU,CPU"}
|
||||
)
|
||||
|
||||
# You cannot add new devices on the fly!
|
||||
@ -105,7 +109,7 @@ def MULTI_4():
|
||||
|
||||
# Optionally, query the optimal number of requests:
|
||||
nireq = compiled_model.get_property(
|
||||
ov.properties.optimal_number_of_infer_requests()
|
||||
properties.optimal_number_of_infer_requests()
|
||||
)
|
||||
#! [MULTI_4]
|
||||
|
||||
|
@ -3,8 +3,9 @@
|
||||
#
|
||||
|
||||
import openvino as ov
|
||||
import openvino.runtime.properties as props
|
||||
import openvino.runtime.properties.hint as hints
|
||||
import openvino.properties as props
|
||||
import openvino.properties.hint as hints
|
||||
import openvino.properties.device as device
|
||||
|
||||
from utils import get_model
|
||||
|
||||
@ -16,11 +17,11 @@ def main():
|
||||
# [get_available_devices]
|
||||
|
||||
# [hetero_priorities]
|
||||
device_priorites = core.get_property("HETERO", props.device.priorities())
|
||||
device_priorites = core.get_property("HETERO", device.priorities())
|
||||
# [hetero_priorities]
|
||||
|
||||
# [cpu_device_name]
|
||||
cpu_device_name = core.get_property("CPU", props.device.full_name())
|
||||
cpu_device_name = core.get_property("CPU", device.full_name())
|
||||
# [cpu_device_name]
|
||||
|
||||
model = get_model()
|
||||
@ -56,8 +57,8 @@ def main():
|
||||
return 0
|
||||
|
||||
# [multi_device]
|
||||
config = {props.device.priorities(): "CPU,GPU"}
|
||||
config = {device.priorities(): "CPU,GPU"}
|
||||
compiled_model = core.compile_model(model, "MULTI", config)
|
||||
# change the order of priorities
|
||||
compiled_model.set_property({props.device.priorities(): "GPU,CPU"})
|
||||
compiled_model.set_property({device.priorities(): "GPU,CPU"})
|
||||
# [multi_device]
|
||||
|
@ -3,8 +3,10 @@
|
||||
#
|
||||
|
||||
import openvino as ov
|
||||
import openvino.runtime.properties as props
|
||||
import openvino.runtime.properties.hint as hints
|
||||
import openvino.properties as props
|
||||
import openvino.properties.hint as hints
|
||||
import openvino.properties.device as device
|
||||
import openvino.properties.streams as streams
|
||||
|
||||
from utils import get_model
|
||||
|
||||
@ -23,7 +25,7 @@ def main():
|
||||
# ! [core_compile_model]
|
||||
compiled_model = core.compile_model(model=model, device_name="MULTI", config=
|
||||
{
|
||||
props.device.priorities(): "GPU,CPU",
|
||||
device.priorities(): "GPU,CPU",
|
||||
hints.performance_mode(): hints.PerformanceMode.THROUGHPUT,
|
||||
hints.inference_precision(): ov.Type.f32
|
||||
})
|
||||
@ -31,15 +33,15 @@ def main():
|
||||
|
||||
# ! [compiled_model_set_property]
|
||||
# turn CPU off for multi-device execution
|
||||
compiled_model.set_property(properties={props.device.priorities(): "GPU"})
|
||||
compiled_model.set_property(properties={device.priorities(): "GPU"})
|
||||
# ! [compiled_model_set_property]
|
||||
|
||||
# ! [core_get_rw_property]
|
||||
num_streams = core.get_property("CPU", props.streams.num())
|
||||
num_streams = core.get_property("CPU", streams.num())
|
||||
# ! [core_get_rw_property]
|
||||
|
||||
# ! [core_get_ro_property]
|
||||
full_device_name = core.get_property("CPU", props.device.full_name())
|
||||
full_device_name = core.get_property("CPU", device.full_name())
|
||||
# ! [core_get_ro_property]
|
||||
|
||||
# ! [compiled_model_get_rw_property]
|
||||
|
Loading…
Reference in New Issue
Block a user