Efficient and Effective Passage Search via Contextualized Late Interaction over BERT and XLM-RoBERTa
Project description
PLAID-X
This is a generalized version of PLAID and the previous ColBERT-X for CLIR. The codebase supports models trained with the original ColBERT-X scripts, which are not compatible with the PLAID codebase released from the Stadford Futuredata Group.
Installation
PLAID-X is available on PyPi. You can install it through
pip install PLAID-X
Make sure your gcc and gxx version is >=9.4.0
, which is the requirement for ninja
to work properly.
We recommand using a conda
environment to control it.
Usage
We have published a tutorial on CLIR with notebooks to run various models. Please refer to the PLAID-X notebook there for a simple working example in Python.
The following provides a series of CLI commands for running a larger scale.
Training Models
The following command starts the training process using the t53b-monot5-msmarco-engeng.jsonl.gz
triple file on the Huggingface Dataset repository hltcoe/tdist-msmarco-scores
with English queries and translated Chinese passages from neuMARCO.
python -m colbert.scripts.train \
--model_name xlm-roberta-large \
--training_triples hltcoe/tdist-msmarco-scores:t53b-monot5-msmarco-engeng.jsonl.gz \
--training_irds_id neumarco/zh/train \
--maxsteps 200000 \
--learning_rate 5e-6 \
--kd_loss KLD \
--only_top \
--per_device_batch_size 8 \
--nway 6 \
--run_tag test \
--experiment test
Indexing
Since PLAID-X is a passage retrieval engine, you need to create passage collections if you are intended to search a document collection. The following command creates a passage collection for the NeuCLIR1 Chinese corpus (file implicitly downloaded from Huggingface).
python -m colbert.scripts.collection_utils create_passage_collection \
--root ./test_coll/ --corpus neuclir/neuclir1:data/zho-00000-of-00001.jsonl.gz
The indexing processes is broken into three steps. This is changed from the last version where we have two steps and also different from the original Stanford codebase where they combines everything into one Python call. Separating the steps provides better allocation for the computation resources and avoid bad GPU reservation deadlocks between Pytorch and FAISS.
for step in prepare encode finalize; do
python -m colbert.scripts.index \
--coll_dir ./test_coll \
--index_name test_index \
--dataset_name test_coll \
--nbits 1 \
--step $step \
--checkpoint eugene-yang/plaidx-xlmr-large-mlir-neuclir \
--experiment test
done
Note that the --checkpoint
flag accept ColBERT-X and ColBERT models stored on Huggingface Models.
Searching
Finally, the following command searches the collection with a query .tsv
file where the first column is the query id and the second column contains the query text.
python -m colbert.scripts.search \
--index_name neuclir-zho.1bits \
--passage_mapping ./test_coll/mapping.tsv \
--query_file query.tsv \
--metrics nDCG@20 MAP R@100 R@1000 Judged@10 \
--qrel qrels.txt \
--experiment test
Citation and Credit
Please cite the following paper if you use the CLIR generalization of ColBERT.
@inproceedings{ecir2022colbert-x,
author = {Suraj Nair and Eugene Yang and Dawn Lawrie and Kevin Duh and Paul McNamee and Kenton Murray and James Mayfield and Douglas W. Oard},
title = {Transfer Learning Approaches for Building Cross-Language Dense Retrieval Models},
booktitle = {Proceedings of the 44th European Conference on Information Retrieval (ECIR)},
year = {2022},
url = {https://arxiv.org/abs/2201.08471}
}
Please cite the following paper if you use the MLIR generalization.
@inproceedings{ecir2023mlir,
title = {Neural Approaches to Multilingual Information Retrieval},
author = {Dawn Lawrie and Eugene Yang and Douglas W Oard and James Mayfield},
booktitle = {Proceedings of the 45th European Conference on Information Retrieval (ECIR)},
year = {2023},
url = {https://arxiv.org/abs/2209.01335}
}
Please cite the following paper if you use the PLAID-X updated implemention or the translate-distil capability of the codebase.
@inproceedings{ecir2024translate-distill,
author = {Eugene Yang and Dawn Lawrie and James Mayfield and Douglas W. Oard and Scott Miller},
title = {Translate-Distill: Learning Cross-Language Dense Retrieval by Translation and Distillation},
booktitle = {Proceedings of the 46th European Conference on Information Retrieval (ECIR)},
year = {2024},
url = {https://arxiv.org/abs/2401.04810}
}
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file PLAID-X-0.3.1.tar.gz
.
File metadata
- Download URL: PLAID-X-0.3.1.tar.gz
- Upload date:
- Size: 80.3 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.10.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 83faf8ec8cf01341686ca342440d93877f5093bad5e640b5c8376479bf979e33 |
|
MD5 | 995f8a1a365eec4988f4e8e6db2188c5 |
|
BLAKE2b-256 | 15fc23e3bdebe135f67fc98493e9810b77bef1bbc437b80231a9a2200ad31f96 |
File details
Details for the file PLAID_X-0.3.1-py3-none-any.whl
.
File metadata
- Download URL: PLAID_X-0.3.1-py3-none-any.whl
- Upload date:
- Size: 107.5 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.0.0 CPython/3.10.5
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | cc887c0d70632d61a9e7bfe43becc493c0ffb50e5af25db4c11ef4def867ac53 |
|
MD5 | 2bc66662d67a116491d63991a6028b0c |
|
BLAKE2b-256 | 8dd3fcb2a45af7a44d6de4ffcebae7bf7e260969867489432772f1ebe6aba9c3 |