Skip to main content

Automation scripts for running ML applications using MLC interface

Project description

MLPerf Automations and Scripts

License Downloads MLC script automation features test MLPerf Inference ABTF POC Test

Welcome to the MLPerf Automations and Scripts repository! This repository is your go-to resource for tools, automations, and scripts designed to streamline the execution of MLPerf benchmarks—with a strong emphasis on MLPerf Inference benchmarks.

Starting January 2025, MLPerf automation scripts are built on the powerful MLCFlow automation interface. This modern interface replaces the earlier Collective Mind (CM), offering a more robust and efficient framework for benchmarking workflows.


🚀 Key Features

  • Automated Benchmarking – Simplifies running MLPerf Inference benchmarks with minimal manual intervention.
  • Modular and Extensible – Easily extend the scripts to support additional benchmarks and configurations.
  • Seamless Integration – Compatible with Docker, cloud environments, and local machines.
  • MLCFlow (MLC) Integration – Utilizes the MLC framework to enhance reproducibility and automation.

🧰 MLCFlow (MLC) Automations

Building upon the robust foundation of its predecessor, the Collective Mind (CM) framework, MLCFlow elevates machine learning workflows by simplifying complex tasks such as Docker container management and caching. Written in Python, the mlcflow package offers a versatile interface, supporting both a user-friendly command-line interface (CLI) and a flexible API for effortless automation script management.

At its core, MLCFlow relies on a single powerful automation, the Script, which is extended by two actions: CacheAction and DockerAction. Together, these components provide streamlined functionality to optimize and enhance your ML workflow automation experience.


🤝 Contributing

We welcome contributions from the community! To contribute:

  1. Submit pull requests (PRs) to the dev branch.
  2. Review our CONTRIBUTORS.md for guidelines and best practices.
  3. Explore more about MLPerf Inference automation in the official MLPerf Inference Documentation.

Your contributions help drive the project forward!


💬 Join the Discussion

Connect with us on the MLCommons Benchmark Infra Discord channel to engage in discussions about MLCFlow and MLPerf Automations. We’d love to hear your thoughts, questions, and ideas!


📰 Stay Updated

Keep track of the latest development progress and tasks on our MLPerf Automations Development Board.
Stay tuned for exciting updates and announcements!


📄 License

This project is licensed under the Apache 2.0 License.


💡 Acknowledgments and Funding

This project is made possible through the generous support of:

We appreciate their contributions and sponsorship!


Thank you for your interest and support in MLPerf Automations and Scripts!

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

mlc_scripts-0.0.7.tar.gz (9.5 kB view details)

Uploaded Source

File details

Details for the file mlc_scripts-0.0.7.tar.gz.

File metadata

  • Download URL: mlc_scripts-0.0.7.tar.gz
  • Upload date:
  • Size: 9.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.0.1 CPython/3.12.3

File hashes

Hashes for mlc_scripts-0.0.7.tar.gz
Algorithm Hash digest
SHA256 fa581600cbce611f6f20f5f3239f9b242a068e2c3f9f219c95629cf287e15fa8
MD5 1c48b4427920b595d853ace5ec488401
BLAKE2b-256 4bbd5acb49ff0dbf00eeacea26a82a65f96fd8e0b47940d170c396856659cad5

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page