Skip to main content

IMPACT framework, an interpretable multi-target framework for multi-class outputs

Project description

logo

An interpretable model for multi-target predictions with multi-class outputs


Welcome to the official repository for IMPACT – a novel, interpretable model designed for Multi-Target Predictions (MTP) with multi-class outputs. IMPACT extends the Cognitive Diagnosis Bayesian Personalized Ranking framework to effectively handle multi-class prediction tasks. Built primarily in Python using PyTorch, this repository includes the four datasets featured in our publication, along with Jupyter notebooks that allow you to reproduce our experiments and conduct your own analyses.

Installing IMPACT

Directly from pip

  • with dependencies :
pip install lirisimpact
import IMPACT
  • without dependencies :
SKIP_INSTALL_REQUIRES=1 pip install lirisimpact
import IMPACT

Or from source for developers

git clone https://github.com/arthur-batel/IMPACT.git
cd IMPACT
make install
conda activate impact-env
# open one of the notebooks in the experiments/notebook_examples folder

Requirements

IMPACT in few lines of code

from IMPACT import utils, model, dataset

# Set all the required parameters ---------------
config = utils.generate_eval_config(dataset_name="postcovid", learning_rate=0.02026, lambda_=1.2e-5, batch_size=2048, num_epochs=200,
                                    valid_metric='rmse', pred_metrics=['rmse', 'mae'], profile_metrics=['doa', 'pc-er'])

# Read the dataset and the metadata -------------
concept_map, train_data, valid_data, test_data = utils.prepare_dataset(config, i_fold=0)

# Train the model --------------------------------
algo = model.IMPACT(**config)
algo.init_model(train_data, valid_data)
algo.train(train_data, valid_data)

# Test the model --------------------------------
eval_preds = algo.evaluate_predictions(test_data)
eval_profiles = algo.evaluate_profiles(test_data)

print("Evaluation of the predictions :",eval_preds)
print("Evaluation of the profiles :",eval_profiles)

Repository map

  • experiments/ : Contains the jupyter notebooks and datasets to run the experiments of the scientific paper.
    • experiments/ckpt/ : Folder for models parameter saving
    • experiments/datasets/ : Contains the raw and pre-processed datasets, as well as there pre-processing jupyter notebook
    • experiments/embs/ : Folder for user embeddings saving
    • experiments/hyperparam_search/ : Contains the csv files of the optimal hyperparameter for each method (obtained with Tree-structured Parzen Estimator (TPE) sampler)
    • experiments/logs/ : Folder for running logs saving
    • experiments/notebook_example/ : Contains the jupyter notebooks to run the experiments of the scientific paper, including competitors.
    • experiments/preds/ : Folder for predictions saving
    • experiments/tensorboard/ : Folder for tensorboard data saving
  • figs/ : Contains the figures of the paper
  • IMPACT/ : Contains the source code of the IMPACT model
    • IMPACT/dataset/ : Contains the code of the dataset class
    • IMPACT/models/ : Contains the code of the IMPACT model and its abstract class, handling the learning process
    • IMPACT/utils/ : Contains utility functions for logging, complex metric computations, configuration handling, etc.

Authors

Arthur Batel, arthur.batel@insa-lyon.fr, INSA Lyon, LIRIS UMR 5205 FR

Marc Plantevit, marc.plantevit@epita.fr, EPITA Lyon, EPITA Research Laboratory (LRE) FR

Idir Benouaret, idir.benouaret@epita.fr, EPITA Lyon, EPITA Research Laboratory (LRE) FR

Céline Robardet, celine.robardet@insa-lyon.fr, INSA Lyon, LIRIS UMR 5205 FR

Contributor

Lucas Michaëli

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

lirisimpact-2.0.25.tar.gz (48.6 kB view details)

Uploaded Source

Built Distribution

If you're not sure about the file name format, learn more about wheel file names.

lirisimpact-2.0.25-py3-none-any.whl (53.8 kB view details)

Uploaded Python 3

File details

Details for the file lirisimpact-2.0.25.tar.gz.

File metadata

  • Download URL: lirisimpact-2.0.25.tar.gz
  • Upload date:
  • Size: 48.6 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.7

File hashes

Hashes for lirisimpact-2.0.25.tar.gz
Algorithm Hash digest
SHA256 db29e9e14302a89798039dc957422d8cac95dfe331bc5a58c9297b12fd339389
MD5 ca7c424180fa3b01e01c56a0c9f6abc0
BLAKE2b-256 c80222c017f791ae4744fa9c1cc2543f67ef4c8abd8a0bbe809f0d49761bcf00

See more details on using hashes here.

File details

Details for the file lirisimpact-2.0.25-py3-none-any.whl.

File metadata

  • Download URL: lirisimpact-2.0.25-py3-none-any.whl
  • Upload date:
  • Size: 53.8 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/6.1.0 CPython/3.11.7

File hashes

Hashes for lirisimpact-2.0.25-py3-none-any.whl
Algorithm Hash digest
SHA256 550a3ab9222967e52df90819709211dbccd146b955245b685b76013acaa86907
MD5 82d2ff6944010b709c796377eb61d1c6
BLAKE2b-256 2a9e615f851c77a9e1cfc0348094b814a16a778b7afc7b6baa5878a2b2fc2255

See more details on using hashes here.

Supported by

AWS Cloud computing and Security Sponsor Datadog Monitoring Depot Continuous Integration Fastly CDN Google Download Analytics Pingdom Monitoring Sentry Error logging StatusPage Status page