Actuarial tail risk quantile and expectile regression for UK personal lines pricing, including EQRN extreme quantile neural nets (Pasche & Engelke 2024)
Project description
insurance-quantile
Actuarial tail risk quantile and expectile regression for UK personal lines pricing.
Wraps CatBoost's native MultiQuantile loss with the vocabulary actuaries actually use: TVaR, large loss loading, ILFs, OEP curves. Polars in, Polars out.
The problem
Tweedie GBMs estimate E[Y | X] well. But pricing teams routinely need:
- Large loss loading: how much extra to add for claims that blow past the mean
- Increased Limits Factors: what to charge for higher policy limits
- TVaR per risk: expected loss given it exceeds its VaR threshold
- OEP curves: exceedance probability for reinsurance attachment points
None of this comes out of a Tweedie model. You need the full conditional distribution, or at least a set of quantiles.
Why CatBoost quantile regression, not quantile forests or GLMs
CatBoost's MultiQuantile loss trains a single model for all quantile levels simultaneously — shared feature representations, one training pass. It outperforms separate models and is faster to fit than quantile random forests on structured tabular data. The downside is quantile crossing at prediction time (CatBoost issue #2317), which we fix with per-row isotonic regression.
For heavy-tailed lines (motor BI, liability), expectile mode is available. Expectile regression has a property quantile regression lacks: it is both coherent (satisfies subadditivity) and elicitable (has a proper scoring rule). This makes it backtestable and suitable for ORSA and Solvency II reporting.
Installation
pip install insurance-quantile
Quick start
import numpy as np
import polars as pl
from sklearn.model_selection import train_test_split
from insurance_quantile import QuantileGBM, per_risk_tvar, large_loss_loading
# Synthetic motor severity portfolio — 1,000 non-zero claims
rng = np.random.default_rng(42)
n = 1_000
vehicle_age = rng.integers(1, 15, n).astype(float)
driver_age = rng.integers(21, 75, n).astype(float)
ncd_years = rng.integers(0, 9, n).astype(float)
vehicle_group = rng.choice([1.0, 2.0, 3.0, 4.0], size=n) # encoded as float
exposure = rng.uniform(0.3, 1.0, n)
# Heteroskedastic lognormal severity: tail weight increases with vehicle group
log_mu = 6.5 + 0.03 * vehicle_age - 0.01 * ncd_years + 0.1 * vehicle_group
log_sigma = 0.5 + 0.05 * vehicle_group # tail weight varies by segment
claim_amount = np.exp(rng.normal(log_mu, log_sigma, n))
# Feature matrix
X = np.column_stack([vehicle_age, driver_age, ncd_years, vehicle_group])
y = claim_amount
idx_train, idx_val = train_test_split(np.arange(n), test_size=0.2, random_state=42)
X_train, X_val = X[idx_train], X[idx_val]
y_train, y_val = y[idx_train], y[idx_val]
exposure_train = exposure[idx_train]
# Fit quantile GBM
model = QuantileGBM(
quantiles=[0.5, 0.75, 0.9, 0.95, 0.99],
fix_crossing=True,
iterations=500,
)
model.fit(X_train, y_train, exposure=exposure_train)
# Predict quantiles — Polars DataFrame, columns: q_0.5, q_0.75, q_0.9, q_0.95, q_0.99
preds = model.predict(X_val)
# TVaR per risk
tvar = per_risk_tvar(model, X_val, alpha=0.95)
# Large loss loading: requires a fitted mean model for comparison
from catboost import CatBoostRegressor
tweedie_model = CatBoostRegressor(loss_function="Tweedie:variance_power=1.5",
iterations=200, verbose=0)
tweedie_model.fit(X_train, y_train)
loading = large_loss_loading(tweedie_model, model, X_val, alpha=0.95)
Module overview
insurance_quantile/
QuantileGBM — core class: fit/predict, quantile or expectile mode
per_risk_tvar — TVaR per risk at confidence level alpha
portfolio_tvar — aggregated portfolio TVaR
large_loss_loading — additive loading: TVaR minus mean model prediction
ilf — Increased Limits Factor: E[min(Y,L2)] / E[min(Y,L1)]
exceedance_curve — P(Y > x) averaged across portfolio
oep_curve — occurrence exceedance probability (OEP)
coverage_check — calibration: observed vs expected coverage per quantile
pinball_loss — standard scoring rule for quantile regression
insurance_quantile.eqrn/
EQRNModel — extreme quantile regression neural network (Pasche & Engelke 2024)
EQRNDiagnostics — GPD QQ, calibration, threshold stability plots
GPDNet — feedforward network for covariate-dependent GPD parameters
IntermediateQuantileEstimator — K-fold OOF intermediate quantile estimation
Expectile mode
# For motor bodily injury or other heavy-tailed lines
model = QuantileGBM(
quantiles=[0.5, 0.75, 0.9, 0.95],
use_expectile=True, # fits separate CatBoost model per alpha
)
model.fit(X_train, y_train)
Expectiles are not the same as quantiles. The e_0.9 expectile is generally different from Q(0.9). Use expectile mode when you need a coherent, backtestable tail risk measure — not when you need P(Y > x) directly.
Zero-inflated data
Most personal lines portfolios have a large mass of zero claims. There are two ways to handle this:
- Model the full distribution (including zeros): correct but quantiles below the zero-fraction level will all be zero, which is less useful for large loss loading.
- Separate frequency and severity (recommended): fit QuantileGBM only on non-zero claims (severity), with a separate frequency model. Large loss loading then applies to severity only. Document which approach you've used in your model sign-off.
Integration with insurance-conformal
QuantileGBM output feeds directly into insurance-conformal for Conformalized Quantile Regression (CQR):
import numpy as np
from sklearn.model_selection import train_test_split
from insurance_quantile import QuantileGBM
from insurance_conformal import ConformalQuantileRegressor
# Assumes X and y defined as in the quick start above.
# Split into train / calibration sets for conformal coverage guarantee.
idx_tr, idx_cal = train_test_split(np.arange(len(X_train) + len(X_val)),
test_size=0.25, random_state=0)
X_all = np.vstack([X_train, X_val])
y_all = np.concatenate([y_train, y_val])
X_tr2, X_cal2 = X_all[idx_tr], X_all[idx_cal]
y_tr2, y_cal2 = y_all[idx_tr], y_all[idx_cal]
model = QuantileGBM(quantiles=[0.05, 0.95]).fit(X_tr2, y_tr2)
preds_cal = model.predict(X_cal2)
cqr = ConformalQuantileRegressor(alpha=0.1)
cqr.fit(y_cal2, preds_cal["q_0.05"], preds_cal["q_0.95"])
# Guaranteed 90% coverage, distribution-free
Design decisions
Quantile crossing fix: isotonic regression per row at predict time. CatBoost's MultiQuantile loss can produce crossing predictions for individual risks despite enforcing correct orderings in the loss function. The fix is O(n_rows × n_quantiles) and adds negligible overhead.
Exposure as sample_weight: exposure is passed to CatBoost as sample_weight, not as an offset. This weights each row's loss contribution, which is appropriate when the target is aggregate cost. If your target is severity (cost per claim), do not pass exposure here.
TVaR approximation: we estimate TVaR by taking the mean of quantile predictions at levels above alpha. Accuracy improves with the number of high quantile levels in the model — include 0.95, 0.99 at minimum for TVaR at alpha=0.9.
ILF integration: E[min(Y, L)] = integral_0^L P(Y > x) dx, integrated numerically using the trapezoidal rule over the interpolated survival function from quantile predictions. 200 integration points is sufficient for smooth severity distributions.
Performance
Benchmarked against parametric Gamma quantiles (Gamma GLM + analytic quantile formula) on synthetic severity data with a heteroskedastic lognormal DGP where tail weight varies with a covariate. Full notebook: notebooks/benchmark.py.
| Metric | Gamma GLM quantiles | QuantileGBM (insurance-quantile) |
|---|---|---|
| Quantile calibration (90th / 95th / 99th) | systematically biased | near stated level |
| TVaR accuracy vs DGP | underestimates for high-risk | near DGP truth |
| Heteroskedastic coverage | poor (global shape parameter) | adapts per segment |
| Pinball loss (99th percentile) | higher | lower |
The key failure mode of the Gamma baseline is the global shape parameter: it cannot represent different tail weights for different risk segments. The QuantileGBM learns the conditional quantile function directly via CatBoost's MultiQuantile pinball loss — if a high-sum-insured segment genuinely has a heavier tail, the model learns that from the data.
When to use: Large loss loading in ground-up pricing where severity is genuinely heteroskedastic (tail weight varies across risk segments). Reinsurance pricing where TVaR in a layer is the deliverable. Any application where the 95th or 99th percentile is the pricing input, not just the mean.
When NOT to use: When the portfolio has only a few hundred large claims in the training period — the tail quantiles are estimated from very few data points regardless of method, and the parametric Gamma's regularisation may actually help. Also when the actuarial deliverable requires a smooth, monotone ILF curve — quantile regression is not constrained to be monotone in the limit dimension without additional work.
Related libraries
| Library | Why it's relevant |
|---|---|
| insurance-conformal | Conformalized Quantile Regression — wraps this library's output to give distribution-free coverage guarantees |
| insurance-distributional | Parametric severity distributions (Pareto, Gamma, LogNormal) — alternative approach when you need closed-form tail quantities |
| shap-relativities | Extract what's driving the tail — SHAP values on the QuantileGBM output |
| insurance-cv | Walk-forward cross-validation for time-structured insurance data |
Read more
Your Burning Cost Has a Tail Risk Problem — why Tweedie models systematically misprice tail risk and how quantile regression fills the gap.
Source repos
This package consolidates two previously separate libraries:
insurance-quantile— core CatBoost quantile/expectile GBM (v0.1.x)insurance-eqrn— archived, merged intoinsurance_quantile.eqrn
Requirements
- Python 3.10+
- catboost >= 1.2
- polars >= 1.0
- scikit-learn >= 1.3
- numpy >= 1.24
Licence
MIT
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file insurance_quantile-0.2.1.tar.gz.
File metadata
- Download URL: insurance_quantile-0.2.1.tar.gz
- Upload date:
- Size: 201.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.8 {"installer":{"name":"uv","version":"0.10.8","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
bb4e22fd29ce2c4232d34dfc978e74390f24b259864aabd3e936c39b54201b47
|
|
| MD5 |
62afc57445f49ad8aa2ef028646f3a43
|
|
| BLAKE2b-256 |
843ca2481adba053a7689c448fe1d3150b3dffc7237e26443fd236dd9380ca5a
|
File details
Details for the file insurance_quantile-0.2.1-py3-none-any.whl.
File metadata
- Download URL: insurance_quantile-0.2.1-py3-none-any.whl
- Upload date:
- Size: 44.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.10.8 {"installer":{"name":"uv","version":"0.10.8","subcommand":["publish"]},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"Ubuntu","version":"24.04","id":"noble","libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
fd398cfc08df96bce4b0ac1c9172dbd7b2cca9cd8b9817034ac40975f9a7ed22
|
|
| MD5 |
2043f894365c6b941b8bd878a0493a4e
|
|
| BLAKE2b-256 |
de1790ed07bf67d60cd58bb0ce0d3770212b40b4c68c3bc4b526c6dbb565eb2c
|