d-Matrix Compressor
Project description
This project contains tools for deep neural net co-design for custom hardware accelerators.
Overview
In essence this is an extension of the PyTorch framework that implements hardware-efficient features, including
- custom low-precision numerical formats and arithmetic,
- fine-grain structured weight sparsity, and
- custom operator approximation logic.
In addition, the project provides a set of optimization tools for co-design using the above features.
Getting started
pip install dmx-compressor
Usage
Basic API
Given a PyTorch model, e.g. Net(), wrap it in a DmxModel container:
from dmx.compressor.modeling import DmxModel
model = DmxModel.from_torch(Net())
Here model is functionally equivalent to Net(), and all torch functionalities are still available, but model is equipped with d-Matrix specific features, making it ready for co-design configuration and/or optimization, at training time or post-training.
See advanced topics for further details.
model.dmx_config is a dictionary that contains all, and only those, configurations that affect the functional behavior of the model, different from the behavior of the original Net().
Use method model.transform() to set these configurations, through application of configuration rules.
See advanced topics for engineering of configuration rules.
There are two predefined special rule sets config_rules.BASELINE and config_rules.BASIC; the former is a dummy that does not change the original model's functional behavior, whereas the latter brings the model to a functional state that is equivalent to basic-mode execution on d-Matrix's hardware, e.g.
from dmx.compressor import config_rules
model = model.transform(
model.dmx_config,
*config_rules.BASIC,
)
Hugging Face pipeline API
To leverage the popularity of Hugging Face's pipeline API for inference, we extend transformers.pipeline() to dmx.compressor.modeling.hf.pipeline(), which retains all existing functionality of pipelines while enabling model transformation and configuration for deployment on d-Matrix hardware.
from dmx.compressor.modeling.hf import pipeline
pipe = pipeline(
task="text-generation",
model="facebook/opt-125m",
dmx_config="BASIC", # make the model deployable on d-Matrix backend
...
)
# Deploy pipe the same way as Hugging Face provides.
Next steps
For more detailed information, go over the following documents on specific topics. Find more usage examples here.
- Configurations
- Numerics
- Weight sparsity
- Custom approximation logic
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file dmx_compressor-0.1.10.tar.gz.
File metadata
- Download URL: dmx_compressor-0.1.10.tar.gz
- Upload date:
- Size: 113.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/2.1.3 CPython/3.10.15 Linux/5.15.0-1057-azure
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
0fc0d4b5e9ae9ecd3349c2ddf4bbc95403f92750fc72b69c99cd8f1da15ee834
|
|
| MD5 |
69601db41b4eea1149837754e6c1ab30
|
|
| BLAKE2b-256 |
00ef4ecd9fe12237da42f53ac09612c02a36431496ead3b4650d2fe9b59e53fc
|
File details
Details for the file dmx_compressor-0.1.10-py3-none-any.whl.
File metadata
- Download URL: dmx_compressor-0.1.10-py3-none-any.whl
- Upload date:
- Size: 121.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: poetry/2.1.3 CPython/3.10.15 Linux/5.15.0-1057-azure
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
56b878c63b673f1b4ca6451f7ed3b947679d814da1645e91f8fdb8e01c3acd78
|
|
| MD5 |
3d093d3cd9c4a28693dd7835957c8292
|
|
| BLAKE2b-256 |
9bf7c14f6c65c48e77b1f66a6758a5ae12f20ebc6ba5914c8ef68a688693685a
|