Skip to main content

Pretrained models for Keras.

Project description

KerasHub: Multi-framework Pretrained Models

Python contributions welcome

[!IMPORTANT] 📢 KerasNLP is now KerasHub! 📢 Read the announcement.

KerasHub is a pretrained modeling library that aims to be simple, flexible, and fast. The library provides Keras 3 implementations of popular model architectures, paired with a collection of pretrained checkpoints available on Kaggle Models. Models can be used with text, image, and audio data for generation, classification, and many other built in tasks.

KerasHub is an extension of the core Keras API; KerasHub components are provided as Layer and Model implementations. If you are familiar with Keras, congratulations! You already understand most of KerasHub.

All models support JAX, TensorFlow, and PyTorch from a single model definition and can be fine-tuned on GPUs and TPUs out of the box. Models can be trained on individual accelerators with built-in PEFT techniques, or fine-tuned at scale with model and data parallel training. See our Getting Started guide to start learning our API.

Quick Links

For everyone

For contributors

Quickstart

Choose a backend:

import os
os.environ["KERAS_BACKEND"] = "jax"  # Or "tensorflow" or "torch"!

Import KerasHub and other libraries:

import keras
import keras_hub
import numpy as np
import tensorflow_datasets as tfds

Load a resnet model and use it to predict a label for an image:

classifier = keras_hub.models.ImageClassifier.from_preset(
    "resnet_50_imagenet",
    activation="softmax",
)
url = "https://upload.wikimedia.org/wikipedia/commons/a/aa/California_quail.jpg"
path = keras.utils.get_file(origin=url)
image = keras.utils.load_img(path)
preds = classifier.predict(np.array([image]))
print(keras_hub.utils.decode_imagenet_predictions(preds))

Load a Bert model and fine-tune it on IMDb movie reviews:

classifier = keras_hub.models.TextClassifier.from_preset(
    "bert_base_en_uncased",
    activation="softmax",
    num_classes=2,
)
imdb_train, imdb_test = tfds.load(
    "imdb_reviews",
    split=["train", "test"],
    as_supervised=True,
    batch_size=16,
)
classifier.fit(imdb_train, validation_data=imdb_test)
preds = classifier.predict(["What an amazing movie!", "A total waste of time."])
print(preds)

Installation

To install the latest KerasHub release with Keras 3, simply run:

pip install --upgrade keras-hub

Our text tokenizers are based on TensorFlow Text. Hence, if you are using any model which has language as a modality, you will have to run:

pip install --upgrade keras-hub[nlp]

To install the latest nightly changes for both KerasHub and Keras, you can use our nightly package.

pip install --upgrade keras-hub-nightly

Currently, installing KerasHub will always pull in TensorFlow for use of the tf.data API for preprocessing. When pre-processing with tf.data, training can still happen on any backend.

Visit the core Keras getting started page for more information on installing Keras 3, accelerator support, and compatibility with different frameworks.

Configuring your backend

If you have Keras 3 installed in your environment (see installation above), you can use KerasHub with any of JAX, TensorFlow and PyTorch. To do so, set the KERAS_BACKEND environment variable. For example:

export KERAS_BACKEND=jax

Or in Colab, with:

import os
os.environ["KERAS_BACKEND"] = "jax"

import keras_hub

[!IMPORTANT] Make sure to set the KERAS_BACKEND before importing any Keras libraries; it will be used to set up Keras when it is first imported.

Compatibility

We follow Semantic Versioning, and plan to provide backwards compatibility guarantees both for code and saved models built with our components. While we continue with pre-release 0.y.z development, we may break compatibility at any time and APIs should not be considered stable.

Disclaimer

KerasHub provides access to pre-trained models via the keras_hub.models API. These pre-trained models are provided on an "as is" basis, without warranties or conditions of any kind. The following underlying models are provided by third parties, and subject to separate licenses: BART, BLOOM, DeBERTa, DistilBERT, GPT-2, Llama, Mistral, OPT, RoBERTa, Whisper, and XLM-RoBERTa.

Citing KerasHub

If KerasHub helps your research, we appreciate your citations. Here is the BibTeX entry:

@misc{kerashub2024,
  title={KerasHub},
  author={Watson, Matthew, and  Chollet, Fran\c{c}ois and Sreepathihalli,
  Divyashree, and Saadat, Samaneh and Sampath, Ramesh, and Rasskin, Gabriel and
  and Zhu, Scott and Singh, Varun and Wood, Luke and Tan, Zhenyu and Stenbit,
  Ian and Qian, Chen, and Bischof, Jonathan and others},
  year={2024},
  howpublished={\url{https://github.com/keras-team/keras-hub}},
}

Acknowledgements

Thank you to all of our wonderful contributors!

Project details


Release history Release notifications | RSS feed

Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

keras_hub_nightly-0.22.0.dev202506160417.tar.gz (617.5 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

File details

Details for the file keras_hub_nightly-0.22.0.dev202506160417.tar.gz.

File metadata

File hashes

Hashes for keras_hub_nightly-0.22.0.dev202506160417.tar.gz
Algorithm Hash digest
SHA256 278c1b0d99c1635e01e3ea01b916b70352af5bafcf21f8f5775ddbad4c66af6c
MD5 945f31067a79c68dfe8d765f3acfa28c
BLAKE2b-256 43cc44eb500ba8609ed0f0c9400ed085d37a1b8df8fc7e27ca53cd2270078cd2

See more details on using hashes here.

File details

Details for the file keras_hub_nightly-0.22.0.dev202506160417-py3-none-any.whl.

File metadata

File hashes

Hashes for keras_hub_nightly-0.22.0.dev202506160417-py3-none-any.whl
Algorithm Hash digest
SHA256 338b07e631920ab451670b05e4331e6dd41218ae0065e6e1d2848cffbed59c89
MD5 08af4bb8beca3dbfbce1d9a5924b462d
BLAKE2b-256 3ef7dfe7ec2f7e881afe1dd8c338944d10e06a0a43b266f67e2b8040dd715fc8

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page