Skip to main content

Deep survival analysis made easy with pytorch

Project description

Deep survival analysis made easy

CodeQC Docs PyPI - Version Conda arXiv Documentation PyPI Downloads Conda Downloads

TorchSurv is a Python package that serves as a companion tool to perform deep survival modeling within the PyTorch environment. Unlike existing libraries that impose specific parametric forms on users, TorchSurv enables the use of custom PyTorch-based deep survival models. With its lightweight design, minimal input requirements, full PyTorch backend, and freedom from restrictive survival model parameterizations, TorchSurv facilitates efficient survival model implementation, particularly beneficial for high-dimensional input data scenarios.

TL;DR

Our idea is to keep things simple. You are free to use any model architecture you want! Our code has 100% PyTorch backend and behaves like any other functions (losses or metrics) you may be familiar with.

Our functions are designed to support you, not to make you jump through hoops. Here's a pseudo code illustrating how easy is it to use TorchSurv to fit and evaluate a Cox proportional hazards model:

from torchsurv.loss import cox
from torchsurv.metrics.cindex import ConcordanceIndex

# Pseudo training loop
for data in dataloader:
    x, event, time = data
    estimate = model(x)  # shape = torch.Size([64, 1]), if batch size is 64
    loss = cox.neg_partial_log_likelihood(estimate, event, time)
    loss.backward()  # native torch backend

# You can check model performance using our evaluation metrics, e.g, the concordance index with
cindex = ConcordanceIndex()
cindex(estimate, event, time)

# You can obtain the confidence interval of the c-index
cindex.confidence_interval()

# You can test whether the observed c-index is greater than 0.5 (random estimator)
cindex.p_value(method="noether", alternative="two_sided")

# You can even compare the metrics between two models (e.g., vs. model B)
cindex.compare(cindexB)

Installation and dependencies

First, install the package using either PyPI or Conda

  • Using conda (recommended)
conda install conda-forge::torchsurv
  • Using PyPI
pip install torchsurv
  • Using for local installation (latest version)
git clone <repo>
cd <repo>
pip install -e .

Additionally, to build the documentation (notebooks, sphinx) and for package development (tests), please see the development notes and dev/environment.yml. This step is not required to use TorchSurv in your projects but only for optional features.

Getting started

We recommend starting with the introductory guide, where you'll find an overview of the package's functionalities.

Survival data

We simulate a random batch of 64 subjects. Each subject is associated with a binary event status (= True if event occured), a time-to-event or censoring and 16 covariates.

>>> import torch
>>> _ = torch.manual_seed(52)
>>> n = 64
>>> x = torch.randn((n, 16))
>>> event = torch.randint(low=0, high=2, size=(n,)).bool()
>>> time = torch.randint(low=1, high=100, size=(n,)).float()

Cox proportional hazards model

The user is expected to have defined a model that outputs the estimated log relative hazard for each subject. For illustrative purposes, we define a simple linear model that generates a linear combination of the covariates.

>>> from torch import nn
>>> model_cox = nn.Sequential(nn.Linear(16, 1))
>>> log_hz = model_cox(x)
>>> print(log_hz.shape)
torch.Size([64, 1])

Given the estimated log relative hazard and the survival data, we calculate the current loss for the batch with:

>>> from torchsurv.loss.cox import neg_partial_log_likelihood
>>> loss = neg_partial_log_likelihood(log_hz, event, time)
>>> print(loss)
tensor(4.1723, grad_fn=<DivBackward0>)

We obtain the concordance index for this batch with:

>>> from torchsurv.metrics.cindex import ConcordanceIndex
>>> with torch.no_grad(): log_hz = model_cox(x)
>>> cindex = ConcordanceIndex()
>>> print(cindex(log_hz, event, time))
tensor(0.4872)

We obtain the Area Under the Receiver Operating Characteristic Curve (AUC) at a new time t = 50 for this batch with:

>>> from torchsurv.metrics.auc import Auc
>>> new_time = torch.tensor(50.)
>>> auc = Auc()
>>> print(auc(log_hz, event, time, new_time=50))
tensor([0.4737])

Weibull accelerated failure time (AFT) model

The user is expected to have defined a model that outputs for each subject the estimated log scale and optionally the log shape of the Weibull distribution that the event density follows. In case the model has a single output, TorchSurv assume that the shape is equal to 1, resulting in the event density to be an exponential distribution solely parametrized by the scale.

For illustrative purposes, we define a simple linear model that estimate two linear combinations of the covariates (log scale and log shape parameters).

>>> from torch import nn
>>> model_weibull = nn.Sequential(nn.Linear(16, 2))
>>> log_params = model_weibull(x)
>>> print(log_params.shape)
torch.Size([64, 2])

Given the estimated log scale and log shape and the survival data, we calculate the current loss for the batch with:

>>> from torchsurv.loss.weibull import neg_log_likelihood
>>> loss = neg_log_likelihood(log_params, event, time)
>>> print(loss)
tensor(82931.5078, grad_fn=<DivBackward0>)

To evaluate the predictive performance of the model, we calculate subject-specific log hazard and survival function evaluated at all times with:

>>> from torchsurv.loss.weibull import log_hazard
>>> from torchsurv.loss.weibull import survival_function
>>> with torch.no_grad(): log_params = model_weibull(x)
>>> log_hz = log_hazard(log_params, time)
>>> print(log_hz.shape)
torch.Size([64, 64])
>>> surv = survival_function(log_params, time)
>>> print(surv.shape)
torch.Size([64, 64])

We obtain the concordance index for this batch with:

>>> from torchsurv.metrics.cindex import ConcordanceIndex
>>> cindex = ConcordanceIndex()
>>> print(cindex(log_hz, event, time))
tensor(0.4062)

We obtain the AUC at a new time t = 50 for this batch with:

>>> from torchsurv.metrics.auc import Auc
>>> new_time = torch.tensor(50.)
>>> log_hz_t = log_hazard(log_params, time=new_time)
>>> auc = Auc()
>>> print(auc(log_hz_t, event, time, new_time=new_time))
tensor([0.3509])

We obtain the integrated brier-score with:

>>> from torchsurv.metrics.brier_score import BrierScore
>>> brier_score = BrierScore()
>>> bs = brier_score(surv, event, time)
>>> print(brier_score.integral())
tensor(0.4447)

Related Packages

The table below compares the functionalities of TorchSurv with those of auton-survival, pycox, torchlife, scikit-survival, lifelines, and deepsurv. While several libraries offer survival modelling functionalities, no existing library provides the flexibility to use a custom PyTorch-based neural networks to define the survival model parameters.

The outputs of both the log-likelihood functions and the evaluation metrics functions have undergone thorough comparison with benchmarks generated using Python and R packages. The comparisons (at time of publication) are summarised in the Related packages summary.

Survival analysis libraries in Python Survival analysis libraries in Python

Survival analysis libraries in R. For obtaining the evaluation metrics, packages survival, riskRegression, SurvMetrics and pec require the fitted model object as input (a specific object format) and RisksetROC imposes a smoothing method. Packages timeROC, riskRegression and pec force the user to choose a form for subject-specific weights (e.g., inverse probability of censoring weighting (IPCW)). Packages survcomp and SurvivalROC do not implement the general AUC but the censoring-adjusted AUC estimator proposed by Heagerty et al. (2000).

Survival analysis libraries in R

Contributing

We value contributions from the community to enhance and improve this project. If you'd like to contribute, please consider the following:

  1. Create Issues: If you encounter bugs, have feature requests, or want to suggest improvements, please create an issue in the GitHub repository. Make sure to provide detailed information about the problem, including code for reproducibility, or enhancement you're proposing.

  2. Fork and Pull Requests: If you're willing to address an existing issue or contribute a new feature, fork the repository, create a new branch, make your changes, and then submit a pull request. Please ensure your code follows our coding conventions and include tests for any new functionality.

By contributing to this project, you agree to license your contributions under the same license as this project.

Contacts

If you have any questions, suggestions, or feedback, feel free to reach out the developement team us.

Cite

If you use this project in academic work or publications, we appreciate citing it using the following BibTeX entry:

@misc{monod2024torchsurv,
      title={TorchSurv: A Lightweight Package for Deep Survival Analysis}, 
      author={M{\'{e}}lodie Monod and Peter Krusche and Qian Cao and Berkman Sahiner and Nicholas Petrick and David Ohlssen and Thibaud Coroller},
      year={2024},
      eprint={2404.10761},
      archivePrefix={arXiv},
      primaryClass={cs.LG},
      doi={https://doi.org/10.48550/arXiv.2404.10761}
}

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

torchsurv-0.1.3.tar.gz (59.9 kB view details)

Uploaded Source

Built Distribution

torchsurv-0.1.3-py3-none-any.whl (51.9 kB view details)

Uploaded Python 3

File details

Details for the file torchsurv-0.1.3.tar.gz.

File metadata

  • Download URL: torchsurv-0.1.3.tar.gz
  • Upload date:
  • Size: 59.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.9.6

File hashes

Hashes for torchsurv-0.1.3.tar.gz
Algorithm Hash digest
SHA256 da67f1e229e63b9c94b48a886c6a79d1bed687da0e59c770f1030f9d837911bd
MD5 3ff3dec7d3a44df2a6f3e2071f578aa4
BLAKE2b-256 f2619942fb54e8defbf0adbba224c9995c491b4dc67082099fe13f7c82f3341b

See more details on using hashes here.

File details

Details for the file torchsurv-0.1.3-py3-none-any.whl.

File metadata

  • Download URL: torchsurv-0.1.3-py3-none-any.whl
  • Upload date:
  • Size: 51.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.0.0 CPython/3.9.6

File hashes

Hashes for torchsurv-0.1.3-py3-none-any.whl
Algorithm Hash digest
SHA256 d85c41f30b90bc7381c982e352ee93d0b9e8c02b1c2fc2b4065148dddc6717a5
MD5 3d87995ea14888c58e417a2ee8e7ea65
BLAKE2b-256 81978eabf55c0aad06bd798725d4b25ab520e73153f5930c1696c43b41985fba

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page