Skip to main content

High Granularity Quantizarion

Project description

HGQ-logo

Notice

HGQ 2 is now available! Please check out https://github.com/calad0i/HGQ2 for more details.

High Granularity Quantization

License Apache 2.0 Documentation PyPI version ArXiv

HGQ is an gradient-based automatic bitwidth optimization and quantization-aware training algorithm for neural networks to be deployed on FPGAs, By laveraging gradients, it allows for bitwidth optimization at arbitrary granularity, up to per-weight and per-activation level.

HGQ-overview

Compare to the other heterogeneous quantization approach, like the QKeras counterpart, HGQ provides the following advantages:

  • High Granularity: HGQ supports per-weight and per-activation bitwidth optimization, or any other lower granularity.
  • Automatic Quantization: By setting a resource regularization term, HGQ could automatically optimize the bitwidth of all parameters during training. Pruning is performed naturally when a bitwidth is reduced to 0.
  • Bit-accurate conversion to hls4ml: You get exactly what you get from Keras models from hls4ml models. HGQ provides a bit-accurate conversion interface, proxy models, for bit-accurate conversion to hls4ml models.
    • still subject to machine float precision limitation.
  • Accurate Resource Estimation: BOPs estimated by HGQ is roughly #LUTs + 55#DSPs for actual (post place & route) FPGA resource consumption. This metric is available during training, and one can estimate the resource consumption of the final model in a very early stage.

Depending on the specific application, HGQ could achieve up to 20x resource reduction compared to the AutoQkeras approach, while maintaining the same accuracy. For some more challenging tasks, where the model is already under-fitted, HGQ could still improve the performance under the same on-board resource consumption. For more details, please refer to our paper here.

Installation

You will need python>=3.10 and tensorflow>=2.13 to run this framework. You can install it via pip:

pip install hgq

Usage

Please refer to the documentation for more details.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

hgq-0.2.6.tar.gz (97.8 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

hgq-0.2.6-py3-none-any.whl (49.5 kB view details)

Uploaded Python 3

File details

Details for the file hgq-0.2.6.tar.gz.

File metadata

  • Download URL: hgq-0.2.6.tar.gz
  • Upload date:
  • Size: 97.8 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.9.23

File hashes

Hashes for hgq-0.2.6.tar.gz
Algorithm Hash digest
SHA256 13d3074f634fdcb73ae43e179ef49f3aa44bd9d84ef622685d0446404fcd973b
MD5 c83c83f9df5317835a8c3e42e817ff0c
BLAKE2b-256 8e0081f5e5f4bc0689a826908d1e40d112bb15e475790d02e37310fbfa8a12a6

See more details on using hashes here.

File details

Details for the file hgq-0.2.6-py3-none-any.whl.

File metadata

  • Download URL: hgq-0.2.6-py3-none-any.whl
  • Upload date:
  • Size: 49.5 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.9.23

File hashes

Hashes for hgq-0.2.6-py3-none-any.whl
Algorithm Hash digest
SHA256 56260b16a8ead7d9bcfd75ff0b128fc58224cc047ba6c701e6fd54a81cdb0af9
MD5 efafc216f08395770d93d27777a9def6
BLAKE2b-256 346b82de1b7db922a694e386457315c0210f13b9892f5c4941a95b0676d7a2b2

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page