Skip to main content

Python implementation of integrated path stability selection (IPSS)

Project description

Integrated path stability selection (IPSS)

Integrated path stability selection (IPSS) is a general method for improving feature selection algorithms. Given an n-by-p data matrix X (n = number of samples, p = number of features), and an n-dimensional response variable y, IPSS applies a base selection algorithm to subsamples of the data to select features (columns of X) that are most related to y. This package includes IPSS for gradient boosting (IPSSGB), random forests (IPSSRF), and L1-regularized linear models (IPSSL). The final outputs are efp scores and q-values for each feature in X.

  • The efp score of feature j is the expected number of false positives, E(FP), selected when j is selected.
    • So to control the E(FP) at target_fp, select the features with efp scores at most target_fp.
  • The q-value of feature j is the false discovery rate (FDR) when feature j is selected.
    • So to control the FDR at target_fdr, select the features with q-values at most target_fdr.

Key attributes

  • Error control: IPSS controls the number of false positives and the FDR.
  • Generality: IPSSGB and IPSSRF are nonlinear, nonparametric methods. IPSSL is linear.
  • Speed: IPSS is efficient. For example, IPSSGB runs in <20 seconds when n = 500 and p = 5000.
  • Simplicity: The only required inputs are X and y. Users can also specify the base method (IPSSGB, IPSSRF, or IPSSL), and the target number of false positives or the target FDR.

Associated papers

IPSSL: https://arxiv.org/abs/2403.15877
IPSSGB and IPSSRF: https://arxiv.org/abs/2410.02208v1

Installation

Install from PyPI:

pip install ipss

Tests

Basic test (test_basic.py)

  • Run the test:
python3 test_basic.py
  • Expected output: "All tests passed."

Ovarian cancer: microRNAs and tumor purity (test_oc.py)

  • Identify microRNAs related to tumor purity in tumor samples from ovarian cancer patients
  • Data are from LinkedOmics and located in examples/cancer/ovarian
  • Inputs:
    • Features: matrix of microRNA expression levels for p = 585 microRNAs from n = 451 patients
    • Response: tumor purity (proportion of cancerous cells in a tissue sample) from n = 451 patients
  • Run the test:
python3 test_oc.py
  • Expected output: q-values and efp scores for the top ranked microRNAs

Usage

from ipss import ipss

# load n-by-p feature matrix X and n-by-1 response vector y

# run ipss:
ipss_output = ipss(X, y)

# select features based on target number of false positives
target_fp = 1
efp_scores = ipss_output['efp_scores']
selected_features = [idx for idx, efp_score in efp_scores.items() if efp_score <= target_fp]
print(f'Selected features (target E(FP) = {target_fp}): {selected_features}')

# select features based on target FDR
target_fdr = 0.1
q_values = ipss_output['q_values']
selected_features = [idx for idx, q_value in q_values.items() if q_value <= target_fdr]
print(f'Selected features (target FDR = {target_fdr}): {selected_features}')

Results

ipss_output = ipss(X, y) is a dictionary containing:

  • efp_scores: Dictionary where keys are feature indices and values are their efp scores (dict of length p).
  • q_values: Dictionary where keys are feature indices and values are their q-values (dict of length p).
  • runtime: Runtime of the algorithm in seconds (float).
  • selected_features: List of indices of features selected by IPSS; empty list if target_fp and target_fdr are not specified (list of ints).
  • stability_paths: Estimated selection probabilities at each parameter value (array of shape (n_alphas, p))

Examples

Additional examples are available in the examples folder. These include

Full list of ipss arguments

Required arguments:

  • X: Features (array of shape (n, p)), where n is the number of samples and p is the number of features.
  • y: Response (array of shape (n,) or (n, 1)). ipss automatically detects if y is continuous or binary.

Optional arguments:

  • selector: Base algorithm to use (str; default 'gb'). Options are:
    • 'gb': Gradient boosting with XGBoost.
    • 'l1': L1-regularized linear or logistic regression.
    • 'rf': Random forest with sci-kit learn.
  • selector_args: Arguments for the base algorithm (dict; default None).
  • target_fp: Target number of false positives to control (positive float; default None).
  • target_fdr: Target false discovery rate (FDR) (positive float; default None).
  • B: Number of subsampling steps (int; default 100 for IPSSGB, 50 otherwise).
  • n_alphas: Number of values in the regularization or threshold grid (int; default 100).
  • ipss_function: Function to apply to selection probabilities (str; default 'h3'). Options are:
    • 'h1': Linear function, h1(x) = 2x - 1 if x >= 0.5 else 0.
    • 'h2': Quadratic function, h2(x) = (2x - 1)**2 if x >= 0.5 else 0.
    • 'h3': Cubic function, h3(x) = (2x - 1)**3 if x >= 0.5 else 0.
  • preselect: Number (if int) or percentage (if float) of features to preselect. False for no preselection (default 0.05).
  • preselect_min: Minimum number of features to keep in the preselection step (int; default 200).
  • preselect_args: Arguments for the preselection algorithm (dict; default None).
  • cutoff: Maximum value of the theoretical integral bound I(Lambda) (positive float; default 0.05).
  • delta: Defines probability measure; see Associated papers (float; default 1).
  • standardize_X: Scale features to have mean 0, standard deviation 1 (bool; default None).
  • center_y: Center response to have mean 0 (bool; default None).
  • true_features: List of true feature indices when known, e.g., in simulations (list; default None).
  • n_jobs: Number of jobs to run in parallel (int; default 1).

General observations/recommendations:

  • IPSSGB is usually best for capturing nonlinear relationships between features and response
  • IPSSL is usually best for capturing linear relationships between features and response
  • target_fp or target_fdr (at most one is specified) are problem specific/left to the user
  • In general, all other parameters should not changed
    • selector_args include, e.g., decision tree parameters for tree-based models
    • Results are robust to B provided it is bigger than 25
    • Results are robust to n_alphas provided it is bigger than 50
    • 'h3' yields the most true positives. 'h2' is more conservative, and 'h1' even more so.
    • Preselection can significantly reduce computation time. Results are robust otherwise.
    • Results are robust to cutoff provided it is between 0.025 and 0.1.
    • Results are robust to delta provided it is between 0 and 1.5.
    • Standardization is automatically applied for IPSSL. IPSSGB and IPSSRF are unaffected by this.
    • Centering y is automatically applied for IPSSL. IPSSGB and IPSSRF are unaffected by this.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

ipss-1.0.9.tar.gz (17.9 kB view details)

Uploaded Source

Built Distribution

ipss-1.0.9-py3-none-any.whl (19.9 kB view details)

Uploaded Python 3

File details

Details for the file ipss-1.0.9.tar.gz.

File metadata

  • Download URL: ipss-1.0.9.tar.gz
  • Upload date:
  • Size: 17.9 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.0

File hashes

Hashes for ipss-1.0.9.tar.gz
Algorithm Hash digest
SHA256 7cb37bc675605f1dd7ea82ab6c33ba2d4e117aa3bb2003b99bc144c985b6a74c
MD5 ab32fba5f0af231f49491eac381e72ea
BLAKE2b-256 61f5d54a11c64dbd738ca2eec15259fd5f4a74b06314ddc636369d51b18b83a3

See more details on using hashes here.

File details

Details for the file ipss-1.0.9-py3-none-any.whl.

File metadata

  • Download URL: ipss-1.0.9-py3-none-any.whl
  • Upload date:
  • Size: 19.9 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/5.1.1 CPython/3.12.0

File hashes

Hashes for ipss-1.0.9-py3-none-any.whl
Algorithm Hash digest
SHA256 8b304900c441a1d40979da55dd4b7940767f18d71b2c106d6d695aadbb0a20f7
MD5 f4c5547ebff2882269e5c2911c26e31d
BLAKE2b-256 4ec0902b37a300c3a16abd50c0522ff83277473a89d988e7d248467af8d4a944

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page