A Compact Recurrent-Invariant Eigenvalue Network for Portfolio Optimization
Project description
RIEnet: A Rotational Invariant Estimator Network for GMV Optimization
This library implements the neural estimators introduced in:
- Bongiorno, C., Manolakis, E., & Mantegna, R. N. (2026). End-to-End Large Portfolio Optimization for Variance Minimization with Neural Networks through Covariance Cleaning. The Journal of Finance and Data Science: 100179. 10.1016/j.jfds.2026.100179
- Bongiorno, C., Manolakis, E., & Mantegna, R. N. (2025). Neural Network-Driven Volatility Drag Mitigation under Aggressive Leverage. In Proceedings of the 6th ACM International Conference on AI in Finance (ICAIF ’25). 10.1145/3768292.3770370
RIEnet is a TensorFlow/Keras research implementation for end-to-end global minimum-variance portfolio construction.
Given a tensor of asset returns, the model estimates a structured covariance / precision representation and produces analytic GMV portfolio weights in a single forward pass.
This repository is intended for:
- research and methodological replication,
- experimentation on large equity universes,
- integration into quantitative portfolio construction workflows.
For a pyTorch implementation, please refer to the RIEnet-torch repository.
What this package provides
- End-to-end training on a realized-variance objective for GMV portfolios
- Access to portfolio weights, cleaned covariance matrices, and precision matrices
- A dimension-agnostic architecture suitable for large cross-sectional universes
- A TensorFlow/Keras implementation aligned with the published methodology
Evidence in published experiments
The empirical properties of the method are documented in the associated papers.
In particular, the published experiments evaluate the model on large equity universes under a global minimum-variance objective and compare it against standard covariance-based benchmarks.
For details on datasets, training protocol, benchmark definitions, and evaluation metrics, please refer to the papers listed above.
Module Organization
rienet.trainable_layers: layers with trainable parameters (RIEnetLayer,LagTransformLayer,DeepLayer,DeepRecurrentLayer,CorrelationEigenTransformLayer).rienet.ops_layers: deterministic operation layers (statistics, normalization, eigensystem algebra, weight post-processing).
Installation
Install from PyPI:
pip install rienet
Or install from source:
git clone https://github.com/bongiornoc/RIEnet.git
cd RIEnet
pip install -e .
Quick Start
Basic Usage
import tensorflow as tf
from rienet import RIEnetLayer, variance_loss_function
# Defaults reproduce the compact GMV architecture (bidirectional GRU with 16 units)
rienet_layer = RIEnetLayer(output_type=['weights', 'precision'])
# Sample data: (batch_size, n_stocks, n_days)
returns = tf.random.normal((32, 10, 60), stddev=0.02)
# Retrieve GMV weights and cleaned precision in one pass
outputs = rienet_layer(returns)
weights = outputs['weights'] # (32, 10, 1)
precision = outputs['precision'] # (32, 10, 10)
# GMV training objective
covariance = tf.random.normal((32, 10, 10))
covariance = tf.matmul(covariance, covariance, transpose_b=True)
loss = variance_loss_function(covariance, weights)
print(loss.shape) # (32, 1, 1)
Training with the GMV Variance Loss
import tensorflow as tf
from rienet import RIEnetLayer, variance_loss_function
def create_portfolio_model():
inputs = tf.keras.Input(shape=(None, None))
weights = RIEnetLayer(output_type='weights')(inputs)
return tf.keras.Model(inputs=inputs, outputs=weights)
model = create_portfolio_model()
# Synthetic training data
X_train = tf.random.normal((1000, 10, 60), stddev=0.02)
Sigma_train = tf.random.normal((1000, 10, 10))
Sigma_train = tf.matmul(Sigma_train, Sigma_train, transpose_b=True)
optimizer = tf.keras.optimizers.Adam(learning_rate=1e-4, clipnorm=1.0)
model.compile(optimizer=optimizer, loss=variance_loss_function)
model.fit(X_train, Sigma_train, epochs=10, batch_size=32, verbose=True)
Tip: When you intend to deploy RIEnet on portfolios of varying size, train on batches that span different asset universes. The RIE-based architecture is dimension agnostic and benefits from heterogeneous training shapes.
Using Different Output Types
# GMV weights only
weights = RIEnetLayer(output_type='weights')(returns)
# Precision matrix only
precision_matrix = RIEnetLayer(output_type='precision')(returns)
# Precision, covariance, lag-transformed inputs, and their z-scores in one pass
outputs = RIEnetLayer(
output_type=['precision', 'covariance', 'input_transformed', 'input_zscores']
)(returns)
precision_matrix = outputs['precision']
covariance_matrix = outputs['covariance']
lagged_inputs = outputs['input_transformed']
lagged_zscores = outputs['input_zscores']
# Spectral components (non-inverse)
spectral = RIEnetLayer(
output_type=['eigenvalues', 'eigenvectors', 'transformed_std']
)(returns)
cleaned_eigenvalues = spectral['eigenvalues'] # (batch, n_stocks, 1)
eigenvectors = spectral['eigenvectors'] # (batch, n_stocks, n_stocks)
transformed_std = spectral['transformed_std'] # (batch, n_stocks, 1)
# Just the standardized lag-transformed inputs, right before correlation estimation
lagged_zscores_only = RIEnetLayer(
output_type='input_zscores'
)(returns)
# Optional: disable variance normalisation (do not use it with end-to-end GMV training)
raw_covariance = RIEnetLayer(
output_type='covariance',
normalize_transformed_variance=False
)(returns)
⚠️ When RIEnet is trained end-to-end on the GMV variance loss, leave
normalize_transformed_variance=True(the default). The loss is invariant to global covariance rescalings and the layer keeps the implied variance scale centred around one. Disable the normalisation only when using alternative objectives where the absolute volatility scale must be preserved.
input_zscores is the lag-transformed input after de-meaning and division by the
sample standard deviation along the lookback axis, i.e. the exact tensor used to
build the correlation matrix.
Using LagTransformLayer Directly
LagTransformLayer is exposed both at package root and in the dedicated module:
import tensorflow as tf
from rienet import LagTransformLayer
# or: from rienet.lag_transform import LagTransformLayer
# Dynamic lookback (T can change call-by-call)
compact = LagTransformLayer(variant="compact")
y1 = compact(tf.random.normal((4, 12, 20)))
y2 = compact(tf.random.normal((4, 12, 40)))
# Fixed lookback inferred at first build/call (requires static T)
per_lag = LagTransformLayer(variant="per_lag")
z1 = per_lag(tf.random.normal((4, 12, 20)))
z2 = per_lag(tf.random.normal((4, 8, 20))) # n_assets can change
Using EigenWeightsLayer Directly
EigenWeightsLayer is part of the public API and can be imported directly:
import tensorflow as tf
from rienet import EigenWeightsLayer
layer = EigenWeightsLayer(name="gmv_weights")
# Inputs
eigenvectors = tf.random.normal((8, 20, 20)) # (..., n_assets, n_assets)
inverse_eigenvalues = tf.random.uniform((8, 20, 1)) # (..., n_assets) or (..., n_assets, 1)
inverse_std = tf.random.uniform((8, 20, 1)) # optional
# Full GMV-like branch (includes inverse_std scaling)
weights = layer(eigenvectors, inverse_eigenvalues, inverse_std)
# Covariance-eigensystem branch (inverse_std omitted)
weights_cov = layer(eigenvectors, inverse_eigenvalues)
Notes:
inverse_stdis optional by design.- If
inverse_stdis omitted, the layer uses a dedicated branch with fewer operations (it does not materialize a vector of ones). - Output shape is always
(..., n_assets, 1), normalized to sum to one along assets.
Using CorrelationEigenTransformLayer Directly
import tensorflow as tf
from rienet import CorrelationEigenTransformLayer
layer = CorrelationEigenTransformLayer(name="corr_cleaner")
# Correlation matrix: (batch, n_assets, n_assets)
corr = tf.eye(6, batch_shape=[4])
# Optional attributes: (batch, k) e.g. q, lookback, regime flags, etc.
attrs = tf.constant([
[0.5, 60.0],
[0.7, 60.0],
[1.2, 30.0],
[0.9, 90.0],
], dtype=tf.float32)
# With attributes (default output_type='correlation')
cleaned_corr = layer(corr, attributes=attrs)
# Request multiple outputs
details = layer(
corr,
attributes=attrs,
output_type=[
'correlation',
'inverse_correlation',
'eigenvalues',
'eigenvectors',
'inverse_eigenvalues',
],
)
cleaned_eigvals = details['eigenvalues'] # (batch, n_assets, 1)
cleaned_inv_eigvals = details['inverse_eigenvalues'] # (batch, n_assets, 1)
inv_corr = details['inverse_correlation'] # (batch, n_assets, n_assets)
# Without attributes
cleaned_corr_no_attr = CorrelationEigenTransformLayer(name="corr_cleaner_no_attr")(corr)
Notes:
attributesis optional and can have shape(batch, k)or(batch, n_assets, k).- The output is a cleaned correlation matrix
(batch, n_assets, n_assets). - If you change attribute width
k, use a new layer instance.
Loss Function
Variance Loss Function
from rienet import variance_loss_function
loss = variance_loss_function(
covariance_true=true_covariance, # (batch_size, n_assets, n_assets)
weights_predicted=predicted_weights # (batch_size, n_assets, 1)
)
Mathematical Formula:
Loss = n_assets × wᵀ Σ w
Where w are the portfolio weights and Σ is the realised covariance matrix.
Architecture Details
The RIEnet pipeline consists of:
- Input Scaling – Annualise returns by 252
- Lag Transformation – Five-parameter memory kernel for temporal weighting
- Covariance Estimation – Sample covariance across assets
- Eigenvalue Decomposition – Spectral analysis of the covariance matrix
- Recurrent Cleaning – Bidirectional GRU/LSTM processing of eigen spectra
- Marginal Volatility Head – Dense network forecasting inverse standard deviations
- Matrix Reconstruction – RIE-based synthesis of Σ⁻¹ and GMV weight normalisation
Paper defaults use a single bidirectional GRU layer with 16 units per direction and a marginal-volatility head with 8 hidden units, matching the compact network described in Bongiorno et al. (2025).
Requirements
- Python ≥ 3.8
- TensorFlow ≥ 2.10.0
- Keras ≥ 2.10.0
- NumPy ≥ 1.21.0
Development
git clone https://github.com/bongiornoc/RIEnet.git
cd RIEnet
pip install -e ".[dev]"
pytest tests/
Citation
Please cite the following references when using RIEnet:
@article{bongiorno2026end,
title={End-to-end large portfolio optimization for variance minimization with neural networks through covariance cleaning},
author={Bongiorno, Christian and Manolakis, Efstratios and Mantegna, Rosario Nunzio},
journal={The Journal of Finance and Data Science},
pages={100179},
year={2026},
publisher={Elsevier}
}
@inproceedings{bongiorno2025Neural,
author = {Bongiorno, Christian and Manolakis, Efstratios and Mantegna, Rosario Nunzio},
title = {Neural Network-Driven Volatility Drag Mitigation under Aggressive Leverage},
year = {2025},
isbn = {9798400722202},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3768292.3770370},
doi = {10.1145/3768292.3770370},
booktitle = {Proceedings of the 6th ACM International Conference on AI in Finance},
pages = {449–455},
numpages = {7},
location = {},
series = {ICAIF '25}
}
For software citation:
@software{rienet2025,
title={RIEnet: A Rotational Invariant Estimator Network for Global Minimum-Variance Optimisation},
author={Bongiorno, Christian},
year={2026},
version={1.0.0},
url={https://github.com/bongiornoc/RIEnet}
}
You can print citation information programmatically:
import rienet
rienet.print_citation()
Support
For questions, issues, or contributions, please:
- Open an issue on GitHub
- Check the documentation
- Contact Prof. Christian Bongiorno (christian.bongiorno@centralesupelec.fr) for calibrated model weights or collaboration requests
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file rienet-1.1.6.tar.gz.
File metadata
- Download URL: rienet-1.1.6.tar.gz
- Upload date:
- Size: 55.2 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
f9e45d9ddb093bb52ee1aec701087aa6e2c80d14a49fff176f119e16b2948201
|
|
| MD5 |
13ca4588917112c3c618c9ac59ff0831
|
|
| BLAKE2b-256 |
159f3ec64f038a1ca48a57a68258fe7aa6605a737be5ed84389d8d859004d125
|
Provenance
The following attestation bundles were made for rienet-1.1.6.tar.gz:
Publisher:
publish.yml on bongiornoc/RIEnet
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
rienet-1.1.6.tar.gz -
Subject digest:
f9e45d9ddb093bb52ee1aec701087aa6e2c80d14a49fff176f119e16b2948201 - Sigstore transparency entry: 1277402473
- Sigstore integration time:
-
Permalink:
bongiornoc/RIEnet@36174a1b522af7a8afbea6cef09a4d873525df75 -
Branch / Tag:
refs/heads/main - Owner: https://github.com/bongiornoc
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@36174a1b522af7a8afbea6cef09a4d873525df75 -
Trigger Event:
push
-
Statement type:
File details
Details for the file rienet-1.1.6-py3-none-any.whl.
File metadata
- Download URL: rienet-1.1.6-py3-none-any.whl
- Upload date:
- Size: 36.6 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/6.1.0 CPython/3.13.12
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
ad3c05bbd826279cb7bc79e432830f7e17c94f1d87b7deb416f770b5e67af4c2
|
|
| MD5 |
f2c2e38fe2fe89a1022cb4da1c26a39e
|
|
| BLAKE2b-256 |
eb15c17703dfa6bfc3cd675c87b730fec8db7647ff782fb52e83600f1df130f6
|
Provenance
The following attestation bundles were made for rienet-1.1.6-py3-none-any.whl:
Publisher:
publish.yml on bongiornoc/RIEnet
-
Statement:
-
Statement type:
https://in-toto.io/Statement/v1 -
Predicate type:
https://docs.pypi.org/attestations/publish/v1 -
Subject name:
rienet-1.1.6-py3-none-any.whl -
Subject digest:
ad3c05bbd826279cb7bc79e432830f7e17c94f1d87b7deb416f770b5e67af4c2 - Sigstore transparency entry: 1277402567
- Sigstore integration time:
-
Permalink:
bongiornoc/RIEnet@36174a1b522af7a8afbea6cef09a4d873525df75 -
Branch / Tag:
refs/heads/main - Owner: https://github.com/bongiornoc
-
Access:
public
-
Token Issuer:
https://token.actions.githubusercontent.com -
Runner Environment:
github-hosted -
Publication workflow:
publish.yml@36174a1b522af7a8afbea6cef09a4d873525df75 -
Trigger Event:
push
-
Statement type: