autoplugin_accelerate (#14838)
This commit is contained in:
parent
b361c132d6
commit
29b17e1c23
@ -54,7 +54,7 @@ For example, if you use a CPU and a GPU, the first-inference latency of AUTO wil
|
||||
|
||||
Note that if you choose to exclude CPU from the priority list, it will be unable to support the initial model compilation stage.
|
||||
|
||||
![autoplugin_accelerate]
|
||||

|
||||
|
||||
This mechanism can be easily observed in the [Using AUTO with Benchmark app sample](#using-auto-with-openvino-samples-and-benchmark-app) section, showing how the first-inference latency (the time it takes to compile the model and perform the first inference) is reduced when using AUTO. For example:
|
||||
|
||||
@ -318,11 +318,9 @@ For more information, refer to the [C++](../../samples/cpp/benchmark_app/README.
|
||||
No demos are yet fully optimized for AUTO, by means of selecting the most suitable device, using the GPU streams/throttling, and so on.
|
||||
@endsphinxdirective
|
||||
|
||||
## See Also
|
||||
## Additional Resources
|
||||
|
||||
- [Debugging AUTO](AutoPlugin_Debugging.md)
|
||||
- [Running on Multiple Devices Simultaneously](./multi_device.md)
|
||||
- [Supported Devices](supported_plugins/Supported_Devices.md)
|
||||
|
||||
|
||||
[autoplugin_accelerate]: ../img/autoplugin_accelerate.png
|
||||
|
@ -1,3 +0,0 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:ba092e65d9c5c6fb585c4a394ba3e6a913bf4f129a386b7a8664b94aeb47878b
|
||||
size 61218
|
1073
docs/img/autoplugin_accelerate.svg
Normal file
1073
docs/img/autoplugin_accelerate.svg
Normal file
File diff suppressed because it is too large
Load Diff
After Width: | Height: | Size: 124 KiB |
Loading…
Reference in New Issue
Block a user