Evaluation as a Service for Natural Language Processing
Project description
Evaluation-as-a-Service for NLP
Usage
Before using EaaS, please see the terms of use. Detailed documentation can be found here. To install the EaaS, simply run
pip install eaas
Run your "Hello, world"
A minimal EaaS application looks something like this:
from eaas import Config, Client
client = Client(Config())
inputs = [{
"source": "Hello, my world",
"references": ["Hello, world", "Hello my world"],
"hypothesis": "Hi, my world"
}]
metrics = ["rouge1", "bleu", "chrf"]
score_dic = client.score(inputs, metrics=metrics)
If eaas has been installed successfully, you should get the results
below by printing score_dic. Each entry corresponds to the metrics passed
to metrics (in the same order). The corpus entry indicates the corpus-level
score, sample entry is a list of sample-level scores:
score_dic = {'scores':
[
{'corpus': 0.6666666666666666, 'sample': [0.6666666666666666]},
{'corpus': 0.35355339059327373, 'sample': [0.35355339059327373]},
{'corpus': 0.4900623006253688, 'sample': [0.4900623006253688]}
]
}
Notably:
- To use this API for scoring, you need to format your input as list of dictionary.
- Each dictionary consists of
source(string, optional),references(list of string, optional) andhypothesis(string, required).sourceandreferencesare optional based on the metrics you want to use. - Please do not conduct any preprocessing on
source,referencesorhypothesis. - We expect normal-cased detokenized texts. All the preprocessing steps are taken by the metrics.
Supported Metrics
Currently, EaaS supports the following metrics:
bart_score_en_ref: BARTScore is a sequence to sequence framework based on pre-trained language model BART.bart_score_cnn_hypo_refuses the CNNDM finetuned BART. It calculates the average generation score ofScore(hypothesis|reference)andScore(reference|hypothesis).bart_score_en_src: BARTScore using the CNNDM finetuned BART. It calculatesScore(hypothesis|source).bert_score_p: BERTScore is a metric designed for evaluating translated text using BERT-based matching framework.bert_score_pcalculates the BERTScore precision.bert_score_r: BERTScore recall.bert_score_f: BERTScore f score.bleu: BLEU measures modified ngram matches between each candidate translation and the reference translations.chrf: CHRF measures the character-level ngram matches between hypothesis and reference.comet: COMET is a neural framework for training multilingual machine translation evaluation models.cometuses thewmt20-comet-dacheckpoint which utilizes source, hypothesis and reference.comet_qe: COMET for quality estimation.comet_qeuses thewmt20-comet-qe-dacheckpoint which utilizes only source and hypothesis.mover_score: MoverScore is a metric similar to BERTScore. Different from BERTScore, it uses the Earth Mover’s Distance instead of the Euclidean Distance.prism: PRISM is a sequence to sequence framework trained from scratch.prismcalculates the average generation score ofScore(hypothesis|reference)andScore(reference|hypothesis).prism_qe: PRISM for quality estimation. It calculatesScore(hypothesis| source).rouge1: ROUGE-1 refers to the overlap of unigram (each word) between the system and reference summaries.rouge2: ROUGE-2 refers to the overlap of bigrams between the system and reference summaries.rougeL: ROUGE-L refers to the longest common subsequence between the system and reference summaries.
The default configurations for each metric can refer to this doc
Asynchronous Requests
If you want to make a call to the EaaS server to calculate some metrics and continue local computation while waiting for the result, you can do so as follows:
from eaas import Config
from eaas.async_client import AsyncClient
config = Config()
client = AsyncClient(config)
inputs = ...
req = client.async_score(inputs, metrics=["bleu"])
# do some other computation
result = req.get_result()
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file eaas-0.3.10.tar.gz.
File metadata
- Download URL: eaas-0.3.10.tar.gz
- Upload date:
- Size: 13.8 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.9.15
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
0ccb555c9d029f4bc8fa3059cb3e386c48dadd40613c90a39cb148250d59cb3e
|
|
| MD5 |
2e88265e003e201bea7892d4fd5c5d91
|
|
| BLAKE2b-256 |
42cb0f03d0d2c31eb2211a764e70a5b5b239ccc9465dd85f4c4795394e023cfe
|
File details
Details for the file eaas-0.3.10-py2.py3-none-any.whl.
File metadata
- Download URL: eaas-0.3.10-py2.py3-none-any.whl
- Upload date:
- Size: 14.4 kB
- Tags: Python 2, Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/4.0.1 CPython/3.9.15
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
92f3595996b90070641149c34203634d83301665feeeee5e93a68fe72e9ab350
|
|
| MD5 |
cdbf43e26c85252abf40f9259bdcc7df
|
|
| BLAKE2b-256 |
2e60a8653ec9045849a954b99151ce8fb406b463944ab403fd29c23a8b75a947
|