Skip to main content

Iterative eigendecomposition of curvature operators (Hessian, GGN, Fisher) for PyTorch models.

Project description

pytorch-hessian-eigenthings

PyPI Documentation CI License

The hessian-eigenthings module provides an efficient (and scalable!) way to compute the eigendecomposition of the Hessian, plus other curvature matrices like the Generalized Gauss-Newton and empirical Fisher, for an arbitrary PyTorch model. You get top eigenvalues and eigenvectors via Lanczos or stochastic power iteration, trace estimates via Hutch++, and the spectral density via Stochastic Lanczos Quadrature.

v1.0.0a1: alpha release. The 0.x API has been removed; pin hessian-eigenthings==0.0.2 if you depend on it.

Why use this?

The eigenvalues and eigenvectors of the Hessian have been implicated in many generalization properties of neural networks. People hypothesize that "flat minima" generalize better, that Hessians of large models are very low-rank, that certain optimizers lead to flatter minima, and so on. But the full Hessian costs memory quadratic in the number of parameters, infeasible for anything but toy models.

Iterative methods like Lanczos and power iteration only need a matrix-vector product. The Hessian-vector product (HVP) is exactly that, and it costs linear memory. This library combines the HVP with iterative algorithms to compute the eigendecomposition without the quadratic memory bottleneck, and works on real models including HuggingFace and TransformerLens transformers.

Installation

pip install hessian-eigenthings
# or with HuggingFace / TransformerLens helpers:
pip install "hessian-eigenthings[transformers,transformer-lens]"

Usage

Build a CurvatureOperator from your model, run any algorithm against it.

import torch
from torch import nn

from hessian_eigenthings import (
    HessianOperator, lanczos, trace, spectral_density, supervised_loss,
)

model = nn.Sequential(nn.Linear(20, 32), nn.Tanh(), nn.Linear(32, 1)).to(torch.float64)
x, y = torch.randn(128, 20, dtype=torch.float64), torch.randn(128, 1, dtype=torch.float64)
data = [(x[i:i+32], y[i:i+32]) for i in range(0, 128, 32)]

H = HessianOperator(model, data, supervised_loss(nn.functional.mse_loss))

eig = lanczos(H, k=5, seed=0)             # top-5 eigenvalues + eigenvectors
t = trace(H, num_matvecs=99, seed=0)      # Hutch++ trace estimate
density = spectral_density(H, num_runs=8, lanczos_steps=40, seed=0)

If you'd rather use the GGN (PSD by construction, often what's meant by "the Hessian" on classification losses), swap in GGNOperator. For per-sample-gradient outer products, EmpiricalFisherOperator. They share the same interface so all the algorithms above work on any of them.

There's a finite-difference HVP path (HessianOperator(method="finite_difference")) for when double-backward is impractical, useful with FSDP and similar setups. You can restrict to a parameter subset with param_filter=match_names("blocks.*.attn.*") for per-block analysis.

See examples/ for runnable scripts on a small MLP, HuggingFace tiny-GPT2, and a TransformerLens model. Full docs at https://noahgolmant.github.io/pytorch-hessian-eigenthings.

Working on the library

Uses uv:

git clone https://github.com/noahgolmant/pytorch-hessian-eigenthings
cd pytorch-hessian-eigenthings
uv sync --group dev --group docs --extra transformers --extra transformer-lens --extra curvlinops
uv run pytest
uv run mkdocs serve

Citing this work

If you find this repo useful and would like to cite it (as others have done, thank you!):

@misc{hessian-eigenthings,
    author       = {Noah Golmant and Zhewei Yao and Amir Gholami and Michael Mahoney and Joseph Gonzalez},
    title        = {pytorch-hessian-eigenthings: efficient PyTorch Hessian eigendecomposition},
    month        = oct,
    year         = 2018,
    version      = {1.0},
    url          = {https://github.com/noahgolmant/pytorch-hessian-eigenthings}
}

Acknowledgements

The original 2018 implementation was written with Zhewei Yao, Amir Gholami, Michael Mahoney, and Joseph Gonzalez at UC Berkeley's RISELab.

The deflated power iteration is based on code from HessianFlow (Z. Yao, A. Gholami, Q. Lei, K. Keutzer, M. Mahoney. "Hessian-based Analysis of Large Batch Training and Robustness to Adversaries", NeurIPS 2018, arXiv:1802.08241). Accelerated stochastic power iteration is from C. De Sa et al., "Accelerated Stochastic Power Iteration", PMLR 2017 (arXiv:1707.02670). The v1 refresh borrows ideas from PyHessian, curvlinops, and HessFormer.

License

MIT.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

hessian_eigenthings-1.0.0a3.tar.gz (357.9 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

hessian_eigenthings-1.0.0a3-py3-none-any.whl (32.4 kB view details)

Uploaded Python 3

File details

Details for the file hessian_eigenthings-1.0.0a3.tar.gz.

File metadata

  • Download URL: hessian_eigenthings-1.0.0a3.tar.gz
  • Upload date:
  • Size: 357.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.9.17 {"installer":{"name":"uv","version":"0.9.17","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for hessian_eigenthings-1.0.0a3.tar.gz
Algorithm Hash digest
SHA256 810ce7b304bfe5bf4062dc05ecda30406959c5bc7e27098a636e0895c59d8690
MD5 63bad35571827cc96c620c2c6bc9ad80
BLAKE2b-256 90d41d5c8864a6a02bfa06825ccfa01c0edf4e389928b04a44916f6b611c4419

See more details on using hashes here.

File details

Details for the file hessian_eigenthings-1.0.0a3-py3-none-any.whl.

File metadata

  • Download URL: hessian_eigenthings-1.0.0a3-py3-none-any.whl
  • Upload date:
  • Size: 32.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.9.17 {"installer":{"name":"uv","version":"0.9.17","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for hessian_eigenthings-1.0.0a3-py3-none-any.whl
Algorithm Hash digest
SHA256 84a7c6d69b3b4c7365b9317abfa127ad1b578558c60ccaf2563dc02944be1951
MD5 ca309e26220d30f03cad6fc0c6e24904
BLAKE2b-256 b5f3a1a6985aac33415b690ab8aa3b70be8d29aec041ccebc2fc0980c3e5b483

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page