Skip to main content

OpenVINO™ integration with TensorFlow

Project description

OpenVINO™ integration with TensorFlow

OpenVINO™ integration with TensorFlow is a product designed for TensorFlow* developers who want to get started with OpenVINO™ in their inferencing applications. This product delivers OpenVINO™ inline optimizations which enhance inferencing performance with minimal code modifications. OpenVINO™ integration with TensorFlow accelerates inference across many AI models on a variety of Intel® silicon such as:

  • Intel® CPUs
  • Intel® integrated GPUs
  • Intel® Movidius™ Vision Processing Units - referred to as VPU
  • Intel® Vision Accelerator Design with 8 Intel Movidius™ MyriadX VPUs - referred to as VAD-M or HDDL

[Note: For maximum performance, efficiency, tooling customization, and hardware control, we recommend the developers to adopt native OpenVINO™ APIs and its runtime.]

Installation

Requirements

  • Ubuntu 18.04, 20.04 or macOS 11.2.3
  • Python* 3.6, 3.7, 3.8 or 3.9
  • TensorFlow* v2.5.1

This OpenVINO™ integration with TensorFlow package comes with pre-built libraries of OpenVINO™ version 2021.4.1 meaning you do not have to install OpenVINO™ separately. This package supports:

  • Intel® CPUs

  • Intel® integrated GPUs

  • Intel® Movidius™ Vision Processing Units (VPUs)

      pip3 install pip==21.0.1
      pip3 install tensorflow==2.5.1
      pip3 install -U openvino-tensorflow
    

To leverage Intel® Vision Accelerator Design with Movidius™ (VAD-M) for inference, please refer to: OpenVINO™ integration with TensorFlow alongside the Intel® Distribution of OpenVINO™ Toolkit.

For more details on installation please refer to INSTALL.md, and for build from source options please refer to BUILD.md

Verify Installation

Once you have installed OpenVINO™ integration with TensorFlow, you can use TensorFlow to run inference using a trained model.

To check if OpenVINO™ integration with TensorFlow is properly installed, run

python3 -c "import tensorflow as tf; print('TensorFlow version: ',tf.__version__);\
            import openvino_tensorflow; print(openvino_tensorflow.__version__)"

This should produce an output like:

    TensorFlow version:  2.5.1
    OpenVINO integration with TensorFlow version: b'1.0.1'
    OpenVINO version used for this build: b'2021.4.1'
    TensorFlow version used for this build: v2.5.1
    CXX11_ABI flag used for this build: 0

Usage

By default, Intel® CPU is used to run inference. However, you can change the default option to either Intel® integrated GPU or Intel® VPU for AI inferencing. Invoke the following function to change the hardware on which inferencing is done.

openvino_tensorflow.set_backend('<backend_name>')

Supported backends include 'CPU', 'GPU', 'GPU_FP16', and 'MYRIAD'.

To determine what processing units are available on your system for inference, use the following function:

openvino_tensorflow.list_backends()

For more API calls and environment variables, see USAGE.md.

[Note: For the best results with TensorFlow, it is advised to enable oneDNN Deep Neural Network Library (oneDNN) by setting the environment variable TF_ENABLE_ONEDNN_OPTS=1]

[Note: If a CUDA capable device is present in the system then set the environment variable CUDA_VISIBLE_DEVICES to -1]

Examples

To see what you can do with OpenVINO™ integration with TensorFlow, explore the demos located in the examples repository.

License

OpenVINO™ integration with TensorFlow is licensed under Apache License Version 2.0. By contributing to the project, you agree to the license and copyright terms therein and release your contribution under these terms.

Support

Please submit your questions, feature requests and bug reports via GitHub issues.

How to Contribute

We welcome community contributions to OpenVINO™ integration with TensorFlow. If you have an idea for improvement:

We will review your contribution as soon as possible. If any additional fixes or modifications are necessary, we will guide you and provide feedback. Before you make your contribution, make sure you can build OpenVINO™ integration with TensorFlow and run all the examples with your fix/patch. If you want to introduce a large feature, create test cases for your feature. Upon the verification of your pull request, we will merge it to the repository provided that the pull request has met the above mentioned requirements and proved acceptable.


* Other names and brands may be claimed as the property of others.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distributions

No source distribution files available for this release.See tutorial on generating distribution archives.

Built Distributions

openvino_tensorflow-1.0.1-cp39-cp39-macosx_11_0_x86_64.whl (22.9 MB view hashes)

Uploaded CPython 3.9 macOS 11.0+ x86-64

openvino_tensorflow-1.0.1-cp38-cp38-macosx_11_0_x86_64.whl (22.9 MB view hashes)

Uploaded CPython 3.8 macOS 11.0+ x86-64

openvino_tensorflow-1.0.1-cp37-cp37m-macosx_11_0_x86_64.whl (22.9 MB view hashes)

Uploaded CPython 3.7m macOS 11.0+ x86-64

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page