Train-test splitter for Lead Optimization and Hit Identification tasks.
Project description
Lo-Hi Splitter
Description
There are two distinct tasks in drug discovery: Lead optimization (Lo) and Hit identification (Hi). The Lo-Hi Splitter enables users to partition molecular datasets based on these tasks and helps choose the most appropriate models. There is a paper related to this repository: @TODO
Hit Identification
The objective of the Hit Identification task is to discover new molecules with a desired property that are distinct from molecules with known activity. The Hi splitter divides the dataset into training and test sets such that no molecule in the test set resembles any molecule in the training set based on ECFP4 Tanimoto similarity. In some instances, an exact split might not be achievable, resulting in the discarding of certain molecules. To reduce the number of discarded molecules, the Hi splitter employs the minimal vertex k-cut method.
Lead Optimization
The goal of the Lead Optimization task is to predict how minor modifications of a molecule affect its properties. The Lo splitter divides the dataset into training and test sets in such a way that:
- The test set consists of clusters of similar molecules.
- Clusters have a variation of the property larger than the experimental noise.
- In each test cluster, exactly one molecule is moved to train to imitate a known hit. This scenario imitates goal-directed optimization tasks in generative models, which employ a predictive model to guide the optimization.
Installation
Pip
From Source
python -m pip install --upgrade pip build
python -m build
pip install dist/lohi_splitter-1.0-py3-none-any.whl
Tutorial
Here are some quick tutorials to help you get started. To run the notebooks located in tutorial/, install the required packages:
pip install -r requirements.in
00_hi_split_quick.ipynb
- A brief introduction to the Hi split.01_hi_split_coarsening.ipynb
- How to accelerate the Hi splitter.02_lo_split.ipynb
- A brief introduction to the Lo split.03_hi_under_the_hood.ipynb
- Delving into the details of the Hi splitter implementation.
Usage
Import the library:
import lohi_splitter as lohi
Hi Splitter, train-test:
smiles = ['C', 'CCC', 'CCCCC']
# Note that train_min_frac and test_min_frac are less than 1.0.
# It is made to allow solutions that discard some molecules.
train_test_partition = lohi.hi_train_test_split(smiles=smiles,
similarity_threshold=0.4,
train_min_frac=0.7,
test_min_frac=0.1)
Hi Splitter, k-fold:
k_fold_partition = lohi.hi_k_fold_split(smiles=smiles,
similarity_threshold=0.4,
k=3)
Hi Splitter, faster, but more discarded molecules:
partition = lohi.hi_train_test_split(smiles=smiles,
similarity_threshold=0.4,
train_min_frac=0.65,
test_min_frac=0.1,
coarsening_threshold=0.4,
max_mip_gap=0.3)
Lo Splitter:
smiles = ['C', 'CCC', 'CCCCC']
values = [0.3, 0.6, 0.1]
cluster_smiles, train_smiles = lohi.lo_train_test_split(smiles=smiles,
threshold=0.4,
min_cluster_size=5,
max_clusters=50,
values=values,
std_threshold=0.60)
Docs
Description of arguments
def hi_train_test_split(
smiles,
similarity_threshold=0.4,
train_min_frac=0.7,
test_min_frac=0.1,
coarsening_threshold=None,
verbose=True,
max_mip_gap=0.1,
):
"""
Splits a list of smiles into train and test sets such that no molecule in the test
has ECFP4 Tanimoto similarity to the train > similarity_threshold.
Parameters:
smiles -- List of smiles to split.
similarity_threshold -- ECFP4 Tanimoto threshold. Molecules in the test set won't
have a similarity greater than similarity_threshold to those in the train set.
train_min_frac -- Minimum fraction for the train set, e.g., 0.7 of the entire dataset.
test_min_frac -- Minimum fraction for the test set, e.g., 0.1 of the entire dataset.
It's possible that the k-cut might not be feasible without discarding some molecules,
so ensure that the sum of train_min_frac and test_min_frac is less than 1.0.
coarsening_threshold -- Molecules with a similarity greater than the coarsening_threshold will be
clustered together. It speeds up execution, but makes the solution less optimal.
None -- Disables clustering (default value).
1.0 -- Won't do anything
0.90 -- will cluster molecules with similarity > 0.90 together
verbose -- If set to False, suppresses status messages.
max_mip_gap -- Determines when to halt optimization based on proximity to the optimal solution.
For example, setting it to 0.5 yields a faster but less optimal solution, while 0.01 aims for a more
optimal solution, potentially at the cost of more computation time. See more in MIP Python documentation.
Returns:
partitions - list of two lists. The first contains indices of train smiles, and the second contains indices of test smiles.
"""
def hi_k_fold_split(
smiles,
similarity_threshold=0.4,
fold_min_frac=None,
k=3,
coarsening_threshold=None,
verbose=True,
max_mip_gap=0.1,
):
"""
Splits the list of smiles into k folds such that no molecule in any fold has an ECFP4 Tanimoto
similarity greater than similarity_threshold when compared to molecules in another fold.
Parameters:
smiles -- List of smiles to split
similarity_threshold -- ECFP4 Tanimoto threshold. Molecules in one fold won't
have a similarity greater than similarity_threshold to those in another fold.
fold_min_frac -- Minimum fraction of a fold (e.g., 0.2 of the entire dataset).
If not specified (None), it defaults to 0.9 / k.
k -- number of folds
coarsening_threshold -- Molecules with a similarity greater than the coarsening_threshold will be
clustered together. It speeds up execution, but makes the solution less optimal.
None -- Disables clustering (default value).
1.0 -- Won't do anything
0.90 -- will cluster molecules with similarity > 0.90 together
verbose -- If set to False, suppresses status messages.
max_mip_gap -- Determines when to halt optimization based on proximity to the optimal solution.
For example, setting it to 0.5 yields a faster but less optimal solution, while 0.01 aims for a more
optimal solution, potentially at the cost of more computation time. See more in MIP Python documentation.
Returns:
partitions - list of lists. Each list contains the indices of smiles in that fold.
"""
def lo_train_test_split(
smiles, threshold, min_cluster_size, max_clusters, values, std_threshold
):
"""
Lo splitter. Refer to tutorial 02_lo_split.ipynb and the paper by Simon Steshin titled "Lo-Hi: Practical ML Drug Discovery Benchmark", 2023.
Parameters:
smiles -- list of smiles
threshold -- molecules with similarity larger than this number are considered similar
min_cluster_size -- number of molecules per cluster
max_clusters -- maximum number of selected clusters. The remaining molecules go to the training set.
values -- values of the smiles
std_threshold -- Lower bound of the acceptable standard deviation for a cluster. It should be greater than measurement noise.
If you're using ChEMBL-like data, set it to 0.60 for logKi and 0.70 for logIC50.
Set it lower if you have a high-quality dataset. Refer to the paper, Appendix B.
Returns:
clusters -- list of lists of smiles.
train_smiles -- list of train smiles
"""
FAQ
Isn't a scaffold split sufficient?
A scaffold split often leads to a considerable number of highly similar molecules between the training and testing sets. Refer to tutorial/00_hi_split_quick.ipynb for examples.
How do I choose the appropriate similarity threshold?
Your selection will depend on the desired generality of your model. Many prefer a 0.40 ECFP4 Tanimoto similarity threshold. For a detailed answer, consult the associated paper, specifically chapter "3. Novelty."
How should I determine the std_threshold for the Lo splitter?
The threshold choice is contingent on the noise level in your data. For ChEMBL-like data, consider using 0.70 for pIC50 and 0.60 for pKi. For high-quality in-house datasets, a threshold of 0.20 for pIC50 might suffice. For a comprehensive answer, refer to the associated paper, specifically Appendix B titled "Lo dataset is not just noise."
Development
Development environment
We use flake8 for linting and black for code formatting.
conda create --name lohi_splitter_dev python=3.7
conda activate lohi_splitter_dev
python -m pip install --upgrade pip build
python -m build
pip install dist/lohi_splitter-1.0-py3-none-any.whl
pip install -r requirements.in
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file lohi_splitter-1.0.tar.gz
.
File metadata
- Download URL: lohi_splitter-1.0.tar.gz
- Upload date:
- Size: 951.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.7.16
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 12fb30fadf43a2295e62acafecb58871297b92207f99939c7766197d21d6b0f3 |
|
MD5 | 592b5d728c8c13f4f2c674e9937bc4e2 |
|
BLAKE2b-256 | 77e90f9d02f27efda5c369437e8056216d96f35ff51d54b9c75023a2c27122c8 |
File details
Details for the file lohi_splitter-1.0-py3-none-any.whl
.
File metadata
- Download URL: lohi_splitter-1.0-py3-none-any.whl
- Upload date:
- Size: 12.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.2 CPython/3.7.16
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 2900eff584f6c30ad02643948cbe70616e6dc1e7462329fab6de9e11626bcf72 |
|
MD5 | 1b6ad8e969946ce8583827ced65abfb6 |
|
BLAKE2b-256 | b19afef21b199bc4c2a783ab3aad506f1be602008d0d07676cee9d8161b6dccb |