Auto metrics for evaluating generated questions
Project description
How to use
Our codes provide the ability to evaluate automatic metrics
which concludes the ability to calculate automatic metrics
. Please follow these steps to calculate automatic QG metrics and evaluate automatic metrics on our benchmark.
Enviroment
run pip install -r requirements.txt
to install the required packages.
Calculate Automatic Metrics
-
Prepare data
Use the data we provided at ../data/scores.xlsx, or use your own data, which should provide passages, answers, and references.
-
Calculate automatic metrics.
-
Download models at
coming soon
for metrics. -
Update model path inside the codes. See
QRelScore
as an example.# update the path of mlm_model and clm_model def corpus_qrel(preds, contexts, device='cuda'): assert len(contexts) == len(preds) mlm_model = 'model/bert-base-cased' clm_model = 'model/gpt2' scorer = QRelScore(mlm_model=mlm_model, clm_model=clm_model, batch_size=16, nthreads=4, device=device) scores = scorer.compute_score_flatten(contexts, preds) return scores
-
Run
python metrics.py
to calculate your assigned metrics results by changingscore_names
inmetrics.py
. (data_path
in each file should be changed into your own data path)# Run QRelScore and RQUGE based on our dataset # load data data_path = '../data/scores.xlsx' save_path = './result/metric_result.xlsx' data = pd.read_excel(data_path) hypos = data['prediction'].tolist() refs_list = [data['reference'].tolist()] contexts = data['passage'].tolist() answers = data['answer'].tolist() # scores to use score_names = ['QRelScore', 'RQUGE'] # run metrics res = get_metrics(hypos, refs_list, contexts, answers, score_names=score_names) # handle results for k, v in res.items(): data[k] = v print(data.columns) # save results data.to_excel(save_path, index=False)
-
or run the code file for specific metric to calculate. For example, run
python qrel.py
to calculate QRelScore results.
-
Evaluate Automatic Metrics
Run python coeff.py
to obtain the Pearson, Spearman, and Kendall correlation coefficient between the generated results and the labeled results. For detailed process, please refer to readme of QGEval.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file QGEval_metrics-1.0.9.tar.gz
.
File metadata
- Download URL: QGEval_metrics-1.0.9.tar.gz
- Upload date:
- Size: 11.8 MB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.9.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ff61e4984a73ff82d516302662c91e179963e50e0cf3908db55dfd049f5f0fb7 |
|
MD5 | c7515e36e9662819e61ee6d88436be8d |
|
BLAKE2b-256 | 75200f0df4654af51aa2fecdd3e1f64b132bbe936e8a8ec0cd8998afaabd5299 |
File details
Details for the file QGEval_metrics-1.0.9-py3-none-any.whl
.
File metadata
- Download URL: QGEval_metrics-1.0.9-py3-none-any.whl
- Upload date:
- Size: 12.0 MB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/5.1.1 CPython/3.9.10
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 2a22d0abf980406126baace9523a6ecb6208e48dc9814ecd3592086ac2f13131 |
|
MD5 | e31578a8fca64b95f942fa94fbc66b52 |
|
BLAKE2b-256 | 98c4345dee182c6a15aea4ea14cfc1c1b541cd4d79cc0c8ec602b359766c72c5 |