Skip to main content

Optimum Habana is the interface between the Hugging Face Transformers and Diffusers libraries and Habana's Gaudi processor (HPU). It provides a set of tools enabling easy model loading, training and inference on single- and multi-HPU settings for different downstream tasks.

Project description

Optimum Habana

🤗 Optimum Habana is the interface between the 🤗 Transformers and Diffusers libraries and Habana's Gaudi processor (HPU). It provides a set of tools enabling easy model loading, training and inference on single- and multi-HPU settings for different downstream tasks. The list of officially validated models and tasks is available here. Users can try other models and tasks with only few changes.

What is a Habana Processing Unit (HPU)?

HPUs offer fast model training and inference as well as a great price-performance ratio. Check out this blog post about BERT pre-training and this article benchmarking Habana Gaudi2 versus Nvidia A100 GPUs for concrete examples. If you are not familiar with HPUs and would like to know more about them, we recommend you take a look at our conceptual guide.

Install

To install the latest release of this package:

pip install optimum[habana]

To use DeepSpeed on HPUs, you also need to run the following command:

pip install git+https://github.com/HabanaAI/DeepSpeed.git@1.8.0

Optimum Habana is a fast-moving project, and you may want to install it from source:

pip install git+https://github.com/huggingface/optimum-habana.git

Alternatively, you can install the package without pip as follows:

git clone https://github.com/huggingface/optimum-habana.git
cd optimum-habana
python setup.py install

Last but not least, don't forget to install the requirements for every example:

cd <example-folder>
pip install -r requirements.txt

How to use it?

Quick Start

🤗 Optimum Habana was designed with one goal in mind: to make training and evaluation straightforward for any 🤗 Transformers and 🤗 Diffusers user while leveraging the complete power of Gaudi processors.

Transformers Interface

There are two main classes one needs to know:

  • GaudiTrainer: the trainer class that takes care of compiling (lazy or eager mode) and distributing the model to run on HPUs, and performing training and evaluation.
  • GaudiConfig: the class that enables to configure Habana Mixed Precision and to decide whether optimized operators and optimizers should be used or not.

The GaudiTrainer is very similar to the 🤗 Transformers Trainer, and adapting a script using the Trainer to make it work with Gaudi will mostly consist in simply swapping the Trainer class for the GaudiTrainer one. That's how most of the example scripts were adapted from their original counterparts.

Original script:

from transformers import Trainer, TrainingArguments

training_args = TrainingArguments(
  # training arguments...
)

# A lot of code here

# Initialize our Trainer
trainer = Trainer(
    model=model,
    args=training_args,  # Original training arguments.
    train_dataset=train_dataset if training_args.do_train else None,
    eval_dataset=eval_dataset if training_args.do_eval else None,
    compute_metrics=compute_metrics,
    tokenizer=tokenizer,
    data_collator=data_collator,
)

Transformed version that can run on Gaudi:

from optimum.habana import GaudiConfig, GaudiTrainer, GaudiTrainingArguments

training_args = GaudiTrainingArguments(
  # same training arguments...
  use_habana=True,
  use_lazy_mode=True,  # whether to use lazy or eager mode
  use_hpu_graphs=True,  # whether to use HPU graphs for inference
  gaudi_config_name=path_to_gaudi_config,
)

# A lot of the same code as the original script here

# Initialize our Trainer
trainer = GaudiTrainer(
    model=model,
    # You can manually specify the Gaudi configuration to use with
    # gaudi_config=my_gaudi_config
    args=training_args,
    train_dataset=train_dataset if training_args.do_train else None,
    eval_dataset=eval_dataset if training_args.do_eval else None,
    compute_metrics=compute_metrics,
    tokenizer=tokenizer,
    data_collator=data_collator,
)

where gaudi_config_name is the name of a model from the Hub (Gaudi configurations are stored in model repositories). You can also give the path to a custom Gaudi configuration written in a JSON file such as this one:

{
  "use_habana_mixed_precision": true,
  "hmp_opt_level": "O1",
  "hmp_is_verbose": false,
  "use_fused_adam": true,
  "use_fused_clip_norm": true,
  "hmp_bf16_ops": [
    "add",
    "addmm",
    "bmm",
    "div",
    "dropout",
    "gelu",
    "iadd",
    "linear",
    "layer_norm",
    "matmul",
    "mm",
    "rsub",
    "softmax",
    "truediv"
  ],
  "hmp_fp32_ops": [
    "embedding",
    "nll_loss",
    "log_softmax"
  ]
}

If you prefer to instantiate a Gaudi configuration to work on it before giving it to the trainer, you can do it as follows:

gaudi_config = GaudiConfig.from_pretrained(
    gaudi_config_name,
    cache_dir=model_args.cache_dir,
    revision=model_args.model_revision,
    use_auth_token=True if model_args.use_auth_token else None,
)

Diffusers Interface

You can generate images from prompts using Stable Diffusion on Gaudi using the GaudiStableDiffusionPipeline class and the [GaudiDDIMScheduler] which have been both optimized for HPUs. Here is how to use them and the differences with the 🤗 Diffusers library:

- from diffusers import DDIMScheduler, StableDiffusionPipeline
+ from optimum.habana.diffusers import GaudiDDIMScheduler, GaudiStableDiffusionPipeline


model_name = "CompVis/stable-diffusion-v1-4"

- scheduler = DDIMScheduler.from_pretrained(model_name, subfolder="scheduler")
+ scheduler = GaudiDDIMScheduler.from_pretrained(model_name, subfolder="scheduler")

- pipeline = StableDiffusionPipeline.from_pretrained(
+ pipeline = GaudiStableDiffusionPipeline.from_pretrained(
    model_name,
    scheduler=scheduler,
+   use_habana=True,
+   use_hpu_graphs=True,
+   gaudi_config="Habana/stable-diffusion",
)

outputs = generator(
    ["An image of a squirrel in Picasso style"],
    num_images_per_prompt=16,
+   batch_size=4,
)

Documentation

Check out the documentation of Optimum Habana for more advanced usage.

Validated Models

The following model architectures, tasks and device distributions have been validated for 🤗 Optimum Habana:

Architecture Single Card Multi Card DeepSpeed Tasks
BERT :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • text classification
  • question answering
  • language modeling
  • RoBERTa :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • question answering
  • language modeling
  • ALBERT :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • question answering
  • language modeling
  • DistilBERT :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • question answering
  • language modeling
  • GPT2 :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • language modeling
  • T5 :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • summarization
  • translation
  • ViT :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • image classification
  • Swin :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • image classification
  • Wav2Vec2 :heavy_check_mark: :heavy_check_mark: :heavy_check_mark:
  • audio classification
  • speech recognition
  • Stable Diffusion :heavy_check_mark:
  • text-to-image generation
  • Other models and tasks supported by the 🤗 Transformers library may also work. You can refer to this section for using them with 🤗 Optimum Habana. Besides, this page explains how to modify any example from the 🤗 Transformers library to make it work with 🤗 Optimum Habana.

    If you find any issue while using those, please open an issue or a pull request.

    Gaudi Setup

    Please refer to Habana Gaudi's official installation guide.

    Tests should be run in a Docker container based on Habana Docker images.

    The current version has been validated for SynapseAI 1.8.

    Project details


    Download files

    Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

    Source Distribution

    optimum-habana-1.4.2.tar.gz (87.3 kB view details)

    Uploaded Source

    Built Distribution

    If you're not sure about the file name format, learn more about wheel file names.

    optimum_habana-1.4.2-py3-none-any.whl (99.6 kB view details)

    Uploaded Python 3

    File details

    Details for the file optimum-habana-1.4.2.tar.gz.

    File metadata

    • Download URL: optimum-habana-1.4.2.tar.gz
    • Upload date:
    • Size: 87.3 kB
    • Tags: Source
    • Uploaded using Trusted Publishing? No
    • Uploaded via: twine/4.0.2 CPython/3.8.10

    File hashes

    Hashes for optimum-habana-1.4.2.tar.gz
    Algorithm Hash digest
    SHA256 1f08f902bcc603405a77d7a1f0671af6b50bc818c0df1c50f9eb1ca7ce6f7af4
    MD5 b86df53783a920323826ec3bfed90e8c
    BLAKE2b-256 6d9b907b3da5bff5be5c53e9be876032a53ccd02e3ccf505a0bc6e409ece485a

    See more details on using hashes here.

    File details

    Details for the file optimum_habana-1.4.2-py3-none-any.whl.

    File metadata

    • Download URL: optimum_habana-1.4.2-py3-none-any.whl
    • Upload date:
    • Size: 99.6 kB
    • Tags: Python 3
    • Uploaded using Trusted Publishing? No
    • Uploaded via: twine/4.0.2 CPython/3.8.10

    File hashes

    Hashes for optimum_habana-1.4.2-py3-none-any.whl
    Algorithm Hash digest
    SHA256 95d41c3ff58cba4b0fc3693c36ee7358759a5e976f0e017b1a2825684772d080
    MD5 91d0dd924b32f6932f24e316c55abc24
    BLAKE2b-256 ecca28d2d1bb24a3cc6c3e81a71a9b015f4dd0119eecd8fa7dd07d0227dafb46

    See more details on using hashes here.

    Supported by

    AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page