Skip to main content

Python toolkit for competing risks: forest (RSF) today; Fine-Gray + Aalen-Johansen + Gray's test + cause-specific Cox in v0.4. Scales to n=10⁶ in ~1 min, 10–22× faster than randomForestSRC on real EHR data, scikit-learn-compatible.

Project description

comprisk

PyPI version CI DOI

comprisk — a Python toolkit for competing risks. v0.3 ships a scalable, scikit-learn-compatible competing-risks random survival forest; v0.4 adds Fine-Gray subdistribution-hazard regression, a stand-alone Aalen-Johansen cumulative-incidence estimator, Gray's K-sample test, and cause-specific Cox PH (see Roadmap). Designed to remove the Python → R workflow split that applied researchers currently endure for competing-risks survival analysis.

Status: alpha. API and internals may change before v1.0. Renamed from crforest in 0.3.1pip install comprisk, from comprisk import CompetingRiskForest (see Migrating from crforest).

Highlights

  • Forest today, regression next. v0.3 ships the only native Python competing-risks RSF (cause-specific log-rank splitting + composite CR log-rank, Aalen-Johansen CIF, Nelson-Aalen CHF, Wolbers + Uno IPCW concordance, OOB Breiman VIMP, Ishwaran minimal-depth variable selection). v0.4 adds Fine-Gray regression and the rest of the canonical CR toolbox (see Roadmap).
  • 10–22× faster than randomForestSRC on real EHR data (CHF 14–22×, SEER 11.6×; full tables in docs/benchmarks.md), with C ≈ 0.85 on both libraries. ~95× faster than rfSRC built without OpenMP (default R-on-macOS).
  • Order-of-magnitude faster than scikit-survival (16.6× at n = 5k, 544× at n = 50k), without disabling CIF/CHF outputs.
  • Bit-identical to randomForestSRC with equivalence="rfsrc" — reproduces the per-tree mtry/nsplit RNG stream for paper-grade reproducibility, sensitivity checks, and rfSRC-baseline migrations.

comprisk vs alternatives

comprisk randomForestSRC scikit-survival
Language Python R Python
Native competing risks ✗ (single-event only)
Aalen–Johansen CIF output n/a
Cumulative hazard at scale ✗¹
OOB permutation VIMP
Bit-identical reproducibility mode ✓ (equivalence="rfsrc") n/a
Scales to n = 10⁶ ✓ (63 s on i7) memory-bound past n ≈ 500 000 on consumer hardware ✗¹ / OOM²
Default parallelism ✓ (n_jobs=-1) OpenMP (build-dependent; macOS Apple clang lacks it)
GPU preview ✓ (CUDA 12)

¹ sksurv RandomSurvivalForest(low_memory=True) is the only mode that scales beyond ~10k samples, but it disables predict_cumulative_hazard_function and predict_survival_function (raises NotImplementedError). ² sksurv low_memory=False exposes CHF / survival outputs but stores per-leaf full CHF arrays; peak RSS reaches 16.8 GB at n = 5k on synthetic, OOMs (> 21.5 GB) at n = 10k on a 24 GB host.

Install

pip install comprisk          # or:  uv add comprisk
pip install "comprisk[gpu]"   # or:  uv add 'comprisk[gpu]'

Requires Python ≥ 3.10. Core dependencies: numpy, scipy, pandas, joblib, numba, scikit-learn. GPU extra adds cupy + CUDA 12 runtime libs (preview; faster only at low feature count today, full rewrite scheduled for v1.1).

Quickstart

import numpy as np
from comprisk import CompetingRiskForest

# Toy competing-risks data: 500 subjects, 6 features, 2 causes (+ censoring).
rng = np.random.default_rng(42)
n = 500
X = rng.normal(size=(n, 6))
time = rng.exponential(2.0, size=n) + 0.1
event = rng.choice([0, 1, 2], size=n, p=[0.4, 0.4, 0.2])  # 0 = censored

# Fit. Defaults: n_estimators=100, max_features="sqrt", logrankCR, n_jobs=-1.
forest = CompetingRiskForest(n_estimators=100, random_state=42).fit(X, time, event)

# Per-subject risk score for cause 1 (suitable for Wolbers C-index).
risk = forest.predict_risk(X[:5], cause=1)

# Aalen-Johansen cumulative incidence over the forest's chosen time grid.
cif = forest.predict_cif(X[:5])                       # (5, n_causes, n_times)
cif_at = forest.predict_cif(X[:5], times=[1.0, 2.0, 5.0])

# Cause-specific Wolbers concordance.
print("C-index, cause 1:", forest.score(X, time, event, cause=1))

# OOB permutation VIMP, scored with Uno IPCW.
vimp = forest.compute_importance(random_state=42)
print(vimp.sort_values("composite_vimp", ascending=False).head())

Variable selection

Rank features by Ishwaran's minimal-depth criterion and apply the forest- averaged null-distribution threshold from Ishwaran et al. (2010, JASA, Theorem 1 + Section 3):

forest = CompetingRiskForest(n_estimators=200, random_state=0).fit(X, time, event)
vs = forest.minimal_depth()
selected = vs.loc[vs["selected"], "feature"].tolist()

Variables with mean minimal depth below the threshold are flagged as informative. Pass return_extra=True to additionally inspect quartiles and per-feature usage rates across trees.

Note on rfSRC compatibility: this implements the paper's forest-averaged threshold (Section 3); randomForestSRC::max.subtree defaults to a tree-averaged threshold, so the threshold scalar differs. Per-feature mean minimal depth values are bit-equivalent under matched fit config (equivalence='rfsrc', bootstrap=False, min_samples_split=2*nodesize, min_samples_leaf=1, max_depth=None).

See docs/quickstart.md for the full walkthrough — data format, prediction shapes, cross-validation, GPU, and migrating from rfSRC.

scikit-learn drop-in. CompetingRiskForest is a real sklearn estimator (BaseEstimator, clone()-friendly, picklable). cross_val_score, KFold, Pipeline work without a wrapper — pass Surv.from_arrays(event, time) as the y argument, or use the legacy 3-arg fit(X, time, event) form. Full example in docs/quickstart.md § Cross-validation.

Roadmap

comprisk is positioned as a complete, Python-native CR toolkit. The 12-month scope is locked to competing-risks methods only — generalist survival methods (general Cox PH, AFT, parametric, deep-survival, Kaplan-Meier as a standalone API) are out of scope; use lifelines or scikit-survival for those.

Version Module Status
v0.3 CompetingRiskForest (CR-RSF) Shipped
v0.4 FineGrayRegression (subdistribution hazard) Planned (Q3-Q4 2026)
v0.4 CumulativeIncidence (stand-alone Aalen-Johansen) Planned (Q3-Q4 2026)
v0.4 gray_test (Gray's K-sample log-rank) Planned (Q3-Q4 2026)
v0.4 CauseSpecificCox (CR-aware censoring) Planned (Q3-Q4 2026)
v1.0 API freeze + JMLR MLOSS submission Planned
v1.1 Full GPU rewrite Planned

Benchmarks

Headline numbers — full tables, methodology, and reproducibility scripts in docs/benchmarks.md.

vs randomForestSRC, matched-pair on real EHR data:

Cohort n × p Hardware comprisk rfSRC OMP-on Speedup
CHF (cardio) 75k × 58 Apple M4 / i7-14700K / HPC 5.6–9.4 s 84.8–207.3 s 14–22×
SEER breast (oncology) 238k × 17 HPC Xeon Gold 6148 7.0 s 81.6 s 11.6×

Both libraries fit similarly well at every tested workload (HF / cancer-specific C ≈ 0.85). The 10–22× cross-dataset band tracks feature count: rfSRC's per-split exhaustive scan scales with p, so the gap narrows on lower-p cohorts. ~95× speedup vs rfSRC built without OpenMP (default R-on-macOS install).

vs scikit-survival, paired on i7-14700K — synthetic 2-cause Weibull, p = 58, both libraries at their best config:

n sksurv low_memory=True comprisk speedup
5 000 18.2 s 1.10 s 16.6×
50 000 2935 s (49 min) 5.40 s 544×

The gap widens super-linearly (sksurv ≈ n^2.2; comprisk ≈ n^0.7). comprisk also provides Aalen-Johansen CIF + Nelson-Aalen CHF that sksurv low_memory=True raises NotImplementedError for.

Scaling on a consumer desktop: n = 10⁶ in 63 s on i7-14700K, 14.5 GB RSS. Reproducible via validation/spikes/lambda/exp5_paper_scale_bench.py.

API

Full parameter list in src/comprisk/forest.py; usage by task in docs/quickstart.md. Two splitrules are available: logrankCR (composite competing-risks log-rank, default) and logrank (cause-specific).

Migrating from crforest

comprisk 0.3.1 is the same codebase as crforest 0.3.0 under a new name and a slightly broader scope. Update one line:

# before
from crforest import CompetingRiskForest

# after
from comprisk import CompetingRiskForest

The package is otherwise identical — same CompetingRiskForest API, same equivalence="rfsrc" mode, same minimal-depth feature selection, same GPU preview, same metrics module. Pin to comprisk>=0.3.1 (or crforest==0.3.0 on the legacy name); the crforest PyPI package will emit a deprecation pointer and stop receiving new releases. The GitHub URL github.com/sunnyadn/crforest auto-redirects to github.com/sunnyadn/comprisk.

Documentation

  • Quickstart — common tasks with runnable code
  • PRD — what comprisk aims to be at v1.0
  • Equivalence vs rfSRC — cross-library validation methodology
  • References — algorithmic provenance (Park-Miller, Bays-Durham, Wolbers 2009, Uno 2011, Cole & Hernán 2008, Breiman 2001, Ishwaran 2008/2014, etc.)

Development

Requires uv.

uv venv
uv pip install -e ".[dev]"
uv run pre-commit install
uv run pytest
uv run ruff check .
uv run ruff format --check .

License

Apache-2.0. See LICENSE and NOTICE.

Citation

@software{yang_comprisk_2026,
  author    = {Yang, Sunny and Zhao, Wanqi},
  title     = {{comprisk: a Python toolkit for competing risks}},
  year      = {2026},
  publisher = {Zenodo},
  version   = {0.3.1},
  doi       = {10.5281/zenodo.19876282},
  url       = {https://doi.org/10.5281/zenodo.19876282},
}

DOI is concept-level (always resolves to the latest version). GitHub's "Cite this repository" button generates a version-specific record from CITATION.cff. Algorithmic references in docs/REFERENCES.md.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

comprisk-0.3.1.tar.gz (144.7 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

comprisk-0.3.1-py3-none-any.whl (89.0 kB view details)

Uploaded Python 3

File details

Details for the file comprisk-0.3.1.tar.gz.

File metadata

  • Download URL: comprisk-0.3.1.tar.gz
  • Upload date:
  • Size: 144.7 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.9.10 {"installer":{"name":"uv","version":"0.9.10"},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for comprisk-0.3.1.tar.gz
Algorithm Hash digest
SHA256 267bd1197ac755d2f61451e08546a8c94ce4a5719ec13df5d22623a3334fc440
MD5 292ad6adf8fa70dc88f687bd17bd9a65
BLAKE2b-256 8233436456178f7f94a145b858878ea47845c6b5b716df0c2500e7ef1a2459db

See more details on using hashes here.

File details

Details for the file comprisk-0.3.1-py3-none-any.whl.

File metadata

  • Download URL: comprisk-0.3.1-py3-none-any.whl
  • Upload date:
  • Size: 89.0 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: uv/0.9.10 {"installer":{"name":"uv","version":"0.9.10"},"python":null,"implementation":{"name":null,"version":null},"distro":{"name":"macOS","version":null,"id":null,"libc":null},"system":{"name":null,"release":null},"cpu":null,"openssl_version":null,"setuptools_version":null,"rustc_version":null,"ci":null}

File hashes

Hashes for comprisk-0.3.1-py3-none-any.whl
Algorithm Hash digest
SHA256 d36e8577b491f9cdd63c44eff13c6f688544b73cca76f2ca2796a59854e7af65
MD5 02965d6869a43308c33e94f7b068b8e1
BLAKE2b-256 0faf8df2f948fdfa480318e2ea4873757f572e98ec652cdbd4ddaa9f573b7639

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page