A PyTorch class for end-to-end training of vector quantization codebook in a deep neural network.
Project description
Welcome to diveq
diveq (short for differentiable vector quantization) is a tool for implementing and training vector quantization (VQ) in deep neural networks (DNNs), such as a VQ-VAE. It allows end-to-end training of DNNs that contain the non-differentiable VQ module, without any auxiliary losses and hyperparameter tunings. diveq is implemented via PyTorch, and it requires python >= 3.11 and torch >= 2.0.0 .
diveq method is published as a research paper entitled "DiVeQ: Differentiable Vector Quantization Using the Reparameterization Trick" in the International Conference on Learning Representations (ICLR) in 2026. You can find the original GitHub repository of the paper at https://github.com/AaltoML/DiVeQ.
diveq package includes eight different vector quantization (VQ) methods:
from diveq import DIVEQoptimizes the VQ codebook via DiVeQ technique. DiVeQ is the first proposed method in the paper that works as an ordinary VQ by mapping the input to codebook vectors.from diveq import SFDIVEQoptimizes the VQ codebook via Space-Filling DiVeQ technique. SF-DiVeQ is the second proposed method in the paper, different from ordinary VQ in a way that it maps the input to a space-filling curve constructed from codebook vectors.
VQ variants that use multiple codebooks for vector quantization, i.e., Residual VQ and Product VQ:
from diveq import ResidualDIVEQoptimizes the Residual VQ codebooks via DiVeQ technique.from diveq import ResidualSFDIVEQoptimizes the Residual VQ codebooks via SF-DiVeQ technique.from diveq import ProductDIVEQoptimizes the Product VQ codebooks via DiVeQ technique.from diveq import ProductSFDIVEQoptimizes the Product VQ codebooks via SF-DiVeQ technique.
Variants of DiVeQ and SF-DiVeQ techniques that use deterministic quantization instead of stochastic quantization:
from diveq import DIVEQDetachoptimizes the VQ codebook via DiVeQ_Detach technique.from diveq import SFDIVEQDetachoptimizes the VQ codebook via SF-DiVeQ_Detach technique.
For more details on these eight different VQ methods, please see the paper.
Installation
You can install diveq through pip by running:
pip install diveq
After installing diveq, you can verify the installation and package details by running:
python -m pip show diveq
Usage Example
Before using diveq, you have to install it using pip install diveq.
Below you see a minimal example of how to import and use the DIVEQ optimization method as a vector quantizer in a model.
from diveq import DIVEQ
vector_quantizer = DIVEQ(num_embeddings, embedding_dim)
vector_quantizeris the vector quantization module that will be used for building the model.num_embeddingsandembedding_dimare the codebook size and dimension of each codebook entry, respectively. In the following, you can find the list of all parameters used in different vector quantization modules incorporated indiveqpackage.
In the example directory of the GitHub for diveq package, we provide a code example of how vector quantization modules in diveq can be used in a vector quantized variational autoencoder (VQ-VAE). You can create the required environment to run the code by running:
cd example #change directory to the example folder
conda create --name diveq_example python=3.11
conda activate diveq_example
pip install -r requirements.txt
Then, you can train the VQ-VAE model by running:
python train.py
List of Parameters
Here, we provide the list of parameters that are used as inputs to eight different vector quantization methods included in diveq package.
num_embeddings(integer): Codebook size or the number of codewords in the codebook.embedding_dim(integer): Dimensionality of each codebook entry or codeword.noise_var(float): Variance of the directional noise for stochastic DiVeQ- and SF-DiVeQ-based methods.replacement_iters(integer): Number of training iterations to apply codebook replacement.discard_threshold(float): Threshold to discard the codebook entries that are used less than this threshold after replacement_iters iterations.perturb_eps(float): Adjusts perturbation/shift magnitude from used codewords for codebook replacement.uniform_init(bool): Whether to use uniform initialization. If False, codebook is initialized from a normal distribution.verbose(bool): Whether to print codebook replacement status, i.e., to print how many unused codewords are replaced.skip_iters(integer): Number of training iterations to skip quantization (for SF-DiVeQ and SF-DiVeQ_Detach) or to use DiVeQ quantization (for Residual_SF-DiVeQ and Product_SF-DiVeQ) in the custom initialization.avg_iters(integer): Number of recent training iterations to extract latents for custom codebook initialization in Space-Filling Versions.latents_on_cpu(bool): Whether to collect latents for custom initialization on CPU. If running out of CUDA memory, set it to True.allow_warning(bool): Whether to print the warnings. The warnings will warn if the user inserts unusual values for the parameters.num_codebooks(integer): Number of codebooks to be used for quantization in VQ variants of Residual VQ and Product VQ. All the codebooks will have the same size and dimensionality.
Important Notes about Parameters
-
Codebook Replacement: Note that to prevent codebook collapse, we include a codebook replacement function (in cases where it is required) inside different quantization modules. Codebook replacement function is called after each
replacement_iterstraining iterations, and it replaces the codewords which are used less thandiscard_thresholdwith perturbation of actively used codewords which are shifted byperturb_epsmagnitude. Ifverbose=True, the status of how many unused codewords are replaced will be printed by the module. Note that the number of unused codewords should decrease over training, and it might take a while. -
Variants of Vector Quantization: Residual VQ and Product VQ are two variants of vector quantization, which are included in the
diveqpackage. These variants utilize multiple codebooks for quantization, wherenum_codebooksdetermines the number of codebooks used in these VQ variants. -
Space-Filling Methods: Quantization methods based on Space-Filling (i.e., SF-DiVeQ, SF-DiVeQ_Detach, Residual_SF-DiVeQ, Product_SF-DiVeQ) use a custom initilization. SF-DiVeQ and SF-DiVeQ_Detach skip quantizing the latents for
skip_iterstraining iterations, and initialize the codebook with an average of latents captured fromavg_itersrecent training iterations. After this custom initialization, they start to quantize the latents. Residual_SF-DiVeQ and Product_SF-DiVeQ work in the same way, but they apply DiVeQ for the firstskip_iterstraining iterations. Note that ifavg_itersvalue is set to a large value, CUDA might run out of memory, as there should be a large pull of latents to be stored for custom initialization. Therefore, the user can setlatents_on_cpu=Trueto store the latents on CPU, or set a smaller value foravg_iters. -
Detach Methods: DiVeQ_Detach and SF-DiVeQ_Detach methods do not use directional noise. Therefore, they do not need to set the
noise_varparameter.
For further details about different vector quantization methods in the diveq package and their corresponding parameters, please see the details provided in the Python codes in src directory of the diveq package.
Citation
If this package contributed to your work, please consider citing it:
@InProceedings{vali2026diveq,
title={{DiVeQ}: {D}ifferentiable {V}ector {Q}uantization {U}sing the {R}eparameterization {T}rick},
author={Vali, Mohammad Hassan and Bäckström, Tom and Solin, Arno},
booktitle={International Conference on Learning Representations (ICLR)},
year={2026}
}
License
diveq was developed by Mohammad Hassan Vali, part of the AaltoML research group from Aalto University and is licensed under MIT license.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file diveq-0.1.0.tar.gz.
File metadata
- Download URL: diveq-0.1.0.tar.gz
- Upload date:
- Size: 14.1 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.8
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
0506fcd7f7a77df2fdccf9b56a60e35d0eddfd11c6fd24489514fcc0d31ed316
|
|
| MD5 |
8bde4c8f493b3753a6175969b7fa91fc
|
|
| BLAKE2b-256 |
547ab3112541e92f6b45bd15580aaf73635f0a23935418d7f428a25fd06d5681
|
File details
Details for the file diveq-0.1.0-py3-none-any.whl.
File metadata
- Download URL: diveq-0.1.0-py3-none-any.whl
- Upload date:
- Size: 34.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/6.2.0 CPython/3.12.8
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
7d1e65ab26a2f8937cb92a4785d1e7c68a3e67d4bcf746629297ca96276bb86a
|
|
| MD5 |
4b8191d5df8831f5b13dbc6d4dd85a39
|
|
| BLAKE2b-256 |
0fa5fc41b95e905c63ce4d31d90cadbc61488bdb4a79064e82202830cd8d5849
|