Skip to main content

Optimum Library is an extension of the Hugging Face Transformers library, providing a framework to integrate third-party libraries from Hardware Partners and interface with their specific functionality.

Project description

ONNX Runtime neural_compressor

Hugging Face - Optimum

🤗 Optimum is an extension of 🤗 Transformers, providing a set of performance optimization tools enabling maximum efficiency to train and run models on targeted hardware.

The AI ecosystem evolves quickly and more and more specialized hardware along with their own optimizations are emerging every day. As such, Optimum enables users to efficiently use any of these platforms with the same ease inherent to transformers.

Integration with Hardware Partners

🤗 Optimum aims at providing more diversity towards the kind of hardware users can target to train and finetune their models.

To achieve this, we are collaborating with the following hardware manufacturers in order to provide the best transformers integration:

Optimizing models towards inference

Along with supporting dedicated AI hardware for training, Optimum also provides inference optimizations towards various frameworks and platforms.

We currently support ONNX runtime along with Intel Neural Compressor (INC).

Features ONNX Runtime Intel Neural Compressor
Post-training Dynamic Quantization :heavy_check_mark: :heavy_check_mark:
Post-training Static Quantization Stay tuned! :star: :heavy_check_mark:
Quantization Aware Training (QAT) :x: :heavy_check_mark:
Pruning N/A :heavy_check_mark:

Install

🤗 Optimum can be installed using pip as follows:

pip install optimum

🤗 Optimum with Intel Neural Compressor (INC) or ONNX runtime dependencies can be installed respectively using pip as follows:

pip install optimum[intel]

pip install optimum[onnxruntime]

If you'd like to play with the examples or need the bleeding edge of the code and can't wait for a new release, you must install the library from source:

pip install git+https://github.com/huggingface/optimum.git

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

optimum-0.1.3.tar.gz (34.2 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

optimum-0.1.3-py3-none-any.whl (41.8 kB view details)

Uploaded Python 3

File details

Details for the file optimum-0.1.3.tar.gz.

File metadata

  • Download URL: optimum-0.1.3.tar.gz
  • Upload date:
  • Size: 34.2 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.8.1 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.8.10

File hashes

Hashes for optimum-0.1.3.tar.gz
Algorithm Hash digest
SHA256 8b480bb8c80af5df16e55b1f71fa7f0975c22e7a5b961f9fcce957de06e7cc92
MD5 bc81f67dd5b15e2c0e522f54a0b5773a
BLAKE2b-256 33e50f5d3bd86fb56c645a2f51d8f64517ab658ffd26cee6bda3628b6d4c5133

See more details on using hashes here.

File details

Details for the file optimum-0.1.3-py3-none-any.whl.

File metadata

  • Download URL: optimum-0.1.3-py3-none-any.whl
  • Upload date:
  • Size: 41.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.2 importlib_metadata/4.8.1 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.3 CPython/3.8.10

File hashes

Hashes for optimum-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 535f2f843bf128d3530ff4b285d3d14080d276643ef0a4063a3c80566f563425
MD5 491204563f00c3175835c18e161c65a7
BLAKE2b-256 a65b2581e94dbdf0d66b61c3007ccd645e1162874d4afe7902c60f2772dbf8fd

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page