Skip to main content

Visual Prompting for Large Multimodal Models (LMMs)

Project description

maestro

coming: when it's ready...

👋 hello

maestro is a tool designed to streamline and accelerate the fine-tuning process for multimodal models. It provides ready-to-use recipes for fine-tuning popular vision-language models (VLMs) such as Florence-2, PaliGemma, and Phi-3.5 Vision on downstream vision-language tasks.

💻 install

Pip install the supervision package in a Python>=3.8 environment.

pip install maestro

🔥 quickstart

CLI

VLMs can be fine-tuned on downstream tasks directly from the command line with maestro command:

maestro florence2 train --dataset='<DATASET_PATH>' --epochs=10 --batch-size=8

SDK

Alternatively, you can fine-tune VLMs using the Python SDK, which accepts the same arguments as the CLI example above:

from maestro.trainer.common import MeanAveragePrecisionMetric
from maestro.trainer.models.florence_2 import train, TrainingConfiguration

config = TrainingConfiguration(
    dataset='<DATASET_PATH>',
    epochs=10,
    batch_size=8,
    metrics=[MeanAveragePrecisionMetric()]
)

train(config)

🦸 contribution

We would love your help in making this repository even better! We are especially looking for contributors with experience in fine-tuning vision-language models (VLMs). If you notice any bugs or have suggestions for improvement, feel free to open an issue or submit a pull request.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

maestro-0.2.0rc3.tar.gz (28.4 kB view hashes)

Uploaded Source

Built Distribution

maestro-0.2.0rc3-py3-none-any.whl (36.4 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page