139 lines
4.9 KiB
ReStructuredText
139 lines
4.9 KiB
ReStructuredText
============================
|
|
OpenVINO 2023.2
|
|
============================
|
|
|
|
.. meta::
|
|
:google-site-verification: _YqumYQ98cmXUTwtzM_0WIIadtDc6r_TMYGbmGgNvrk
|
|
|
|
.. raw:: html
|
|
|
|
<link rel="stylesheet" type="text/css" href="_static/css/homepage_style.css">
|
|
|
|
|
|
|
|
.. container::
|
|
:name: ov-homepage-banner
|
|
|
|
OpenVINO 2023.2
|
|
|
|
.. raw:: html
|
|
|
|
<div class="line-block">
|
|
<section class="splide" aria-label="Splide Banner Carousel">
|
|
<div class="splide__track">
|
|
<ul class="splide__list">
|
|
<li class="splide__slide">An open-source toolkit for optimizing and deploying deep learning models.<br>Boost your AI deep-learning inference performance!</li>
|
|
|
|
<li class="splide__slide"Better OpenVINO integration with PyTorch!<br>Use PyTorch models directly, without converting them first.<br>
|
|
<a href="https://docs.openvino.ai/2023.2/openvino_docs_MO_DG_prepare_model_convert_model_Convert_Model_From_PyTorch.html">Learn more...</a>
|
|
</li>
|
|
<li class="splide__slide">OpenVINO via PyTorch 2.0 torch.compile()<br>Use OpenVINO directly in PyTorch-native applications!<br>
|
|
<a href="https://docs.openvino.ai/2023.2/pytorch_2_0_torch_compile.html">Learn more...</a>
|
|
</li>
|
|
<li class="splide__slide">Do you like Generative AI? You will love how it performs with OpenVINO!<br>
|
|
<a href="https://docs.openvino.ai/2023.2/tutorials.html">Check out our new notebooks...</a>
|
|
</ul>
|
|
</div>
|
|
</section>
|
|
</div>
|
|
|
|
.. button-ref:: get_started
|
|
:ref-type: doc
|
|
:class: ov-homepage-banner-btn
|
|
:color: primary
|
|
:outline:
|
|
|
|
Get started
|
|
|
|
.. rst-class:: openvino-diagram
|
|
|
|
.. image:: _static/images/ov_homepage_diagram.png
|
|
:align: center
|
|
|
|
|
|
.. grid:: 2 2 3 3
|
|
:class-container: ov-homepage-higlight-grid
|
|
|
|
.. grid-item-card:: Performance Benchmarks
|
|
:link: openvino_docs_performance_benchmarks
|
|
:link-alt: performance benchmarks
|
|
:link-type: doc
|
|
|
|
See latest benchmark numbers for OpenVINO and OpenVINO Model Server
|
|
|
|
.. grid-item-card:: Flexible Workflow
|
|
:link: Supported_Model_Formats
|
|
:link-alt: Supported Model Formats
|
|
:link-type: doc
|
|
|
|
Load models directly (for TensorFlow, ONNX, PaddlePaddle) or convert to the OpenVINO format.
|
|
|
|
.. grid-item-card:: Deploy at Scale With OpenVINO Model Server
|
|
:link: ovms_what_is_openvino_model_server
|
|
:link-alt: model server
|
|
:link-type: doc
|
|
|
|
Cloud-ready deployments for microservice applications
|
|
|
|
.. grid-item-card:: Model Optimization
|
|
:link: openvino_docs_model_optimization_guide
|
|
:link-alt: model optimization
|
|
:link-type: doc
|
|
|
|
Reach for performance with post-training and training-time compression with NNCF
|
|
|
|
.. grid-item-card:: PyTorch 2.0 - torch.compile() backend
|
|
:link: pytorch_2_0_torch_compile
|
|
:link-alt: torch.compile
|
|
:link-type: doc
|
|
|
|
Optimize generation of the graph model with PyTorch 2.0 torch.compile() backend
|
|
|
|
.. grid-item-card:: Generative AI optimization and deployment
|
|
:link: gen_ai_guide
|
|
:link-alt: gen ai
|
|
:link-type: doc
|
|
|
|
Generative AI optimization and deployment
|
|
|
|
|
|
Feature Overview
|
|
##############################
|
|
|
|
.. grid:: 1 2 2 2
|
|
:class-container: ov-homepage-feature-grid
|
|
|
|
.. grid-item-card:: Local Inference & Model Serving
|
|
|
|
You can either link directly with OpenVINO Runtime to run inference locally or use OpenVINO Model Server
|
|
to serve model inference from a separate server or within Kubernetes environment
|
|
|
|
.. grid-item-card:: Improved Application Portability
|
|
|
|
Write an application once, deploy it anywhere, achieving maximum performance from hardware. Automatic device
|
|
discovery allows for superior deployment flexibility. OpenVINO Runtime supports Linux, Windows and MacOS and
|
|
provides Python, C++ and C API. Use your preferred language and OS.
|
|
|
|
.. grid-item-card:: Minimal External Dependencies
|
|
|
|
Designed with minimal external dependencies reduces the application footprint, simplifying installation and
|
|
dependency management. Popular package managers enable application dependencies to be easily installed and
|
|
upgraded. Custom compilation for your specific model(s) further reduces final binary size.
|
|
|
|
.. grid-item-card:: Enhanced App Start-Up Time
|
|
|
|
In applications where fast start-up is required, OpenVINO significantly reduces first-inference latency by using the
|
|
CPU for initial inference and then switching to another device once the model has been compiled and loaded to memory.
|
|
Compiled models are cached improving start-up time even more.
|
|
|
|
|
|
|
|
.. toctree::
|
|
:maxdepth: 2
|
|
:hidden:
|
|
|
|
GET STARTED <get_started>
|
|
LEARN OPENVINO <learn_openvino>
|
|
OPENVINO WORKFLOW <openvino_workflow>
|
|
DOCUMENTATION <documentation>
|
|
ABOUT OPENVINO <about_openvino> |