Diverse Genomic Embedding Benchmark
Project description
title: DGEB app_file : leaderboard/app.py sdk: docker sdk_version: 4.36.1
Diverse Genomic Embedding Benchmark
Installation | Usage | Leaderboard | Citing
DGEB is a benchmark for evaluating biological sequence models on functional and evolutionary information.
DGEB is designed to evaluate model embeddings using:
- Diverse sequences accross the tree of life.
- Diverse tasks that capture different aspects of biological function.
- Both amino acid and nucleotide sequences.
The current version of DGEB consists of 18 datasets covering all three domains of life (Bacteria, Archaea and Eukarya). DGEB evaluates embeddings using six different embedding tasks: Classification, BiGene mining, Evolutionary Distance Similarity (EDS), Pair Classification, Clustering, and Retrieval.
We welcome contributions of new tasks and datasets.
Installation
Install DGEB using pip.
pip install dgeb
Usage
- Launch evaluation using the python script (see cli.py):
dgeb --model facebook/esm2_t6_8M_UR50D
- To see all supported models and tasks:
dgeb --help
- Using the python API:
import dgeb
model = dgeb.get_model("facebook/esm2_t6_8M_UR50D")
tasks = dgeb.get_tasks_by_modality(dgeb.Modality.PROTEIN)
evaluation = dgeb.DGEB(tasks=tasks)
evaluation.run(model, output_folder="results")
Using a custom model
Custom models should be wrapped with the dgeb.models.BioSeqTransformer
abstract class, and specify the modality, number of layers, and embedding dimension. See models.py for additional examples on custom model loading and inference.
import dgeb
from dgeb.models import BioSeqTransformer
from dgeb.tasks.tasks import Modality
class MyModel(BioSeqTransformer):
@property
def modality(self) -> Modality:
return Modality.PROTEIN
@property
def num_layers(self) -> int:
return self.config.num_hidden_layers
@property
def embed_dim(self) -> int:
return self.config.hidden_size
model = MyModel(model_name='path_to/huggingface_model')
tasks = dgeb.get_tasks_by_modality(model.modality)
evaluation = dgeb.DGEB(tasks=tasks)
evaluation.run(model)
Evaluating on a custom dataset
We strongly encourage users to contribute their custom datasets to DGEB. Please open a PR adding your dataset so that the community can benefit!
To evaluate on a custom dataset, first upload your dataset to the Huggingface Hub. Then define a Task
subclass with TaskMetadata
that points to your huggingface dataset. For example, a classification task on a custom dataset can be defined as follows:
import dgeb
from dgeb.models import BioSeqTransformer
from dgeb.tasks import Dataset, Task, TaskMetadata, TaskResult
from dgeb.tasks.classification_tasks import run_classification_task
class MyCustomTask(Task):
metadata = TaskMetadata(
id="my_custom_classification",
display_name="...",
description="...",
type="classification",
modality=Modality.PROTEIN,
datasets=[
Dataset(
path="path_to/huggingface_dataset",
revision="...",
)
],
primary_metric_id="f1",
)
def run(self, model: BioSeqTransformer) -> TaskResult:
return run_classification_task(model, self.metadata)
model = dgeb.get_model("facebook/esm2_t6_8M_UR50D")
evaluation = dgeb.DGEB(tasks=[MyCustomTask])
evaluation.run(model)
Leaderboard
To add your submission to the DGEB leaderboard, proceed through the following instructions.
-
Fork the DGEB repository by following GitHub's instruction Forking Workflow.
-
Add your submission .json file to the leaderboard/submissions/<HF_MODEL_NAME>/ directory.
mv /path/to/<SUBMISSION_FILE>.json /path/to/DGEB/leaderboard/submissions/<HF_MODEL_NAME>/
- Update your fork with the new submission:
git add leaderboard/submissions/<HF_MODEL_NAME>/<SUBMISSION_FILE>.json
git commit -m "Add submission for <HF_MODEL_NAME>"
git push
-
Open a pull request to the main branch of the repository via the Github interface.
-
Once the PR is review and merged, your submission will be added to the leaderboard!
Acknowledgements
DGEB follows the design of text embedding bechmark MTEB developed by Huggingface 🤗. The evaluation code is adapted from the MTEB codebase.
Citing
DGEB was introduced in "Diverse Genomic Embedding Benchmark for Functional Evaluation Across the Tree of Life", feel free to cite:
@article{WestRoberts2024,
title = {Diverse Genomic Embedding Benchmark for functional evaluation across the tree of life},
url = {http://dx.doi.org/10.1101/2024.07.10.602933},
DOI = {10.1101/2024.07.10.602933},
publisher = {Cold Spring Harbor Laboratory},
author = {West-Roberts, Jacob and Kravitz, Joshua and Jha, Nishant and Cornman, Andre and Hwang, Yunha},
year = {2024},
month = jul
}
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file dgeb-0.1.1.tar.gz
.
File metadata
- Download URL: dgeb-0.1.1.tar.gz
- Upload date:
- Size: 162.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.0 CPython/3.12.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 0631ae65ce78d4d8dbf3e5340da8e4244a9537d73bae18c09c1382cea87e682f |
|
MD5 | 3557fa023bb9111336459848435aea05 |
|
BLAKE2b-256 | 27535e7bba5d2478089359efa8a06773b3ab49da1a8329296423d5dd9b58750f |
File details
Details for the file dgeb-0.1.1-py3-none-any.whl
.
File metadata
- Download URL: dgeb-0.1.1-py3-none-any.whl
- Upload date:
- Size: 296.3 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.1.0 CPython/3.12.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | aefad341f9ab68f6c63e01747ac3a27ef1ae39c322530e3a2ba96840145d23f2 |
|
MD5 | ad94a4b4b325b70343f0294878f35659 |
|
BLAKE2b-256 | a06ae1d55c03c24397ead09c8151ea23af9d3fbc19dbc49cbd76aa7663e67e47 |