Skip to main content

OpenVINO™ Training Extensions: Train, Evaluate, Optimize, Deploy Computer Vision Models via OpenVINO™

Project description

OpenVINO™ Training Extensions


Key FeaturesInstallationDocumentationLicense

PyPI

python pytorch openvino

Codecov OpenSSF Scorecard Pre-Merge Test Build Docs License Downloads


Introduction

OpenVINO™ Training Extensions is a low-code transfer learning framework for Computer Vision. The API & CLI commands of the framework allows users to train, infer, optimize and deploy models easily and quickly even with low expertise in the deep learning field. OpenVINO™ Training Extensions offers diverse combinations of model architectures, learning methods, and task types based on PyTorch and OpenVINO™ toolkit.

OpenVINO™ Training Extensions provides a "recipe" for every supported task type, which consolidates necessary information to build a model. Model templates are validated on various datasets and serve one-stop shop for obtaining the best models in general. If you are an experienced user, you can configure your own model based on torchvision, mmcv and OpenVINO Model Zoo (OMZ).

Furthermore, OpenVINO™ Training Extensions provides automatic configuration for ease of use. The framework will analyze your dataset and identify the most suitable model and figure out the best input size setting and other hyper-parameters. The development team is continuously extending this Auto-configuration functionalities to make training as simple as possible so that single CLI command can obtain accurate, efficient and robust models ready to be integrated into your project.

Key Features

OpenVINO™ Training Extensions supports the following computer vision tasks:

  • Classification, including multi-class, multi-label and hierarchical image classification tasks.
  • Object detection including rotated bounding box support
  • Semantic segmentation
  • Instance segmentation including tiling algorithm support
  • Action recognition including action classification and detection
  • Anomaly recognition tasks including anomaly classification, detection and segmentation
  • Visual Prompting tasks including segment anything model, zero-shot visual prompting

OpenVINO™ Training Extensions supports the following learning methods:

  • Supervised, incremental training, which includes class incremental scenario.

OpenVINO™ Training Extensions provides the following usability features:

  • Auto-configuration. OpenVINO™ Training Extensions analyzes provided dataset and selects the proper task and model to provide the best accuracy/speed trade-off.
  • Datumaro data frontend: OpenVINO™ Training Extensions supports the most common academic field dataset formats for each task. We are constantly working to extend supported formats to give more freedom of datasets format choice.
  • Distributed training to accelerate the training process when you have multiple GPUs
  • Mixed-precision training to save GPUs memory and use larger batch sizes
  • Integrated, efficient hyper-parameter optimization module (HPO). Through dataset proxy and built-in hyper-parameter optimizer, you can get much faster hyper-parameter optimization compared to other off-the-shelf tools. The hyperparameter optimization is dynamically scheduled based on your resource budget.

Installation

Please refer to the installation guide. If you want to make changes to the library, then a local installation is recommended.

Install from PyPI Installing the library with pip is the easiest way to get started with otx.
pip install otx[base]

Alternatively, for zsh users:

pip install 'otx[base]'
Install from source To install from source, you need to clone the repository and install the library using pip via editable mode.
# Use of virtual environment is highy recommended
# Using conda
yes | conda create -n otx_env python=3.10
conda activate otx_env

# Or using your favorite virtual environment
# ...

# Clone the repository and install in editable mode
git clone https://github.com/openvinotoolkit/training_extensions.git
cd training_extensions
pip install -e .[base]  # for zsh: pip install -e '.[base]'

Quick-Start

OpenVINO™ Training Extensions supports both API and CLI-based training. The API is more flexible and allows for more customization, while the CLI training utilizes command line interfaces, and might be easier for those who would like to use OpenVINO™ Training Extensions off-the-shelf.

For the CLI, the commands below provide subcommands, how to use each subcommand, and more:

# See available subcommands
otx --help

# Print help messages from the train subcommand
otx train --help

# Print help messages for more details
otx train --help -v   # Print required parameters
otx train --help -vv  # Print all configurable parameters

You can find details with examples in the CLI Guide. and API Quick-Guide.

Below is how to train with auto-configuration, which is provided to users with datasets and tasks:

Training via API
# Training with Auto-Configuration via Engine
from otx.engine import Engine

engine = Engine(data_root="data/wgisd", task="DETECTION")
engine.train()

For more examples, see documentation: CLI Guide

Training via CLI
otx train --data_root data/wgisd --task DETECTION

For more examples, see documentation: API Quick-Guide

In addition to the examples above, please refer to the documentation for tutorials on using custom models, training parameter overrides, and tutorial per task types, etc.


Updates - v2.2.0 (3Q24)

New features

  • Add RT-DETR model for Object Detection
  • Add Multi-Label & H-label Classification with torchvision models
  • Add Hugging-Face Model Wrapper for Classification
  • Add LoRA finetuning capability for ViT Architectures
  • Add Hugging-Face Model Wrapper for Object Detection
  • Add Hugging-Face Model Wrapper for Semantic Segmentation
  • Enable torch.compile to work with classification
  • Add otx benchmark subcommand
  • Add RTMPose for Keypoint Detection Task
  • Add Semi-SL MeanTeacher algorithm for Semantic Segmentation
  • Update head and h-label format for hierarchical label classification
  • Support configurable input size

Enhancements

  • Reimplement of ViT Architecture following TIMM
  • Enable to override data configurations
  • Enable to use input_size at transforms in recipe
  • Enable to use polygon and bitmap mask as prompt inputs for zero-shot learning
  • Refactoring ConvModule by removing conv_cfg, norm_cfg, and act_cfg
  • Support ImageFromBytes
  • enable model export
  • Move templates from OTX1.X to OTX2.X
  • Include Geti arrow dataset subset names
  • Include full image with anno in case there's no tile in tile dataset
  • Add type checker in converter for callable functions (optimizer, scheduler)
  • Change sematic segmentation to consider bbox only annotations
  • Relieve memory usage criteria on batch size 2 during adaptive batch size
  • Remove background label from RT Info for segmentation task
  • Prevent using too low confidence thresholds in detection

Bug fixes

  • Fix Combined Dataloader & unlabeled warmup loss in Semi-SL
  • Revert #3579 to fix issues with replacing coco_instance with a different format in some dataset
  • Add num_devices in Engine for multi-gpu training
  • Add missing tile recipes and various tile recipe changes
  • Change categories mapping logic
  • Fix config converter for tiling
  • Fix num_trials calculation on dataset length less than num_class
  • Fix out_features in HierarchicalCBAMClsHead
  • Fix multilabel_accuracy of MixedHLabelAccuracy
  • Fix wrong indices setting in HLabelInfo

Known issues

  • Post-Training Quantization (PTQ) optimization applied to maskrcnn_swint in the instance segmentation task may result in significantly reduced accuracy. This issue is expected to be addressed with an upgrade to OpenVINO and NNCF in a future release.

Release History

Please refer to the CHANGELOG.md


Branches

  • develop
    • Mainly maintained branch for developing new features for the future release
  • misc
    • Previously developed models can be found on this branch

License

OpenVINO™ Toolkit is licensed under Apache License Version 2.0. By contributing to the project, you agree to the license and copyright terms therein and release your contribution under these terms.


Issues / Discussions

Please use Issues tab for your bug reporting, feature requesting, or any questions.


Known limitations

  • misc branch contains training, evaluation, and export scripts for models based on TensorFlow and PyTorch. These scripts are not ready for production. They are exploratory and have not been validated.

Disclaimer

Intel is committed to respecting human rights and avoiding complicity in human rights abuses. See Intel's Global Human Rights Principles. Intel's products and software are intended only to be used in applications that do not cause or contribute to a violation of an internationally recognized human right.


Contributing

For those who would like to contribute to the library, see CONTRIBUTING.md for details.

Thank you! we appreciate your support!


Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

otx-2.2.0rc14.tar.gz (9.6 MB view details)

Uploaded Source

Built Distribution

otx-2.2.0rc14-py3-none-any.whl (1.0 MB view details)

Uploaded Python 3

File details

Details for the file otx-2.2.0rc14.tar.gz.

File metadata

  • Download URL: otx-2.2.0rc14.tar.gz
  • Upload date:
  • Size: 9.6 MB
  • Tags: Source
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/5.1.0 CPython/3.12.7

File hashes

Hashes for otx-2.2.0rc14.tar.gz
Algorithm Hash digest
SHA256 136c664ff6e3d11426c492f9ef48ff4d6e829fa5a5fc03b3fad87e9efa6bec94
MD5 04b8363892cfc816a99f2c8f0e055af8
BLAKE2b-256 d1cf91f4df974ffc0f4e8443b64d5a11dfa98a9c9180ffbb5fddd88e66113b99

See more details on using hashes here.

File details

Details for the file otx-2.2.0rc14-py3-none-any.whl.

File metadata

  • Download URL: otx-2.2.0rc14-py3-none-any.whl
  • Upload date:
  • Size: 1.0 MB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? Yes
  • Uploaded via: twine/5.1.0 CPython/3.12.7

File hashes

Hashes for otx-2.2.0rc14-py3-none-any.whl
Algorithm Hash digest
SHA256 8e8923ce96928003e925d71769a5ff8843dd634edf844211fef4c066e115ab9a
MD5 09f26a8fd013b5e8c1d2e9d1d36ff510
BLAKE2b-256 8047aca88760fc1a1467286ecb1ae82b867c3937611970e1c6c888cdcfd00c06

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page