Evaluation toolkit for neural language generation.
Project description
Jury
Simple tool/toolkit for evaluating NLG (Natural Language Generation) offering various automated metrics. Jury offers a smooth and easy-to-use interface. It uses datasets for underlying metric computation, and hence adding custom metric is easy as adopting datasets.Metric
.
Main advantages that Jury offers are:
- Easy to use for any NLG system.
- Calculate many metrics at once.
- Metrics calculations are handled concurrently to save processing time.
- It supports evaluating multiple predictions.
To see more, check the official Jury blog post.
Installation
Through pip,
pip install jury
or build from source,
git clone https://github.com/obss/jury.git
cd jury
python setup.py install
Usage
API Usage
It is only two lines of code to evaluate generated outputs.
from jury import Jury
jury = Jury()
# Microsoft translator translation for "Yurtta sulh, cihanda sulh." (16.07.2021)
predictions = ["Peace in the dormitory, peace in the world."]
references = ["Peace at home, peace in the world."]
scores = jury.evaluate(predictions, references)
Specify metrics you want to use on instantiation.
jury = Jury(metrics=["bleu", "meteor"])
scores = jury.evaluate(predictions, references)
Custom Metrics
You can use custom metrics with inheriting jury.metrics.Metric
, you can see current metrics on datasets/metrics. The code snippet below gives a brief explanation.
from jury.metrics import Metric
CustomMetric(Metric):
def compute(self, predictions, references):
pass
Contributing
PRs are welcomed as always :)
Installation
git clone https://github.com/obss/jury.git
cd jury
pip install -e .[develop]
Tests
To tests simply run.
python tests/run_tests.py
Code Style
To check code style,
python tests/run_code_style.py check
To format codebase,
python tests/run_code_style.py format
License
Licensed under the MIT License.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file jury-1.0.1.tar.gz
.
File metadata
- Download URL: jury-1.0.1.tar.gz
- Upload date:
- Size: 13.4 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.4.2 importlib_metadata/4.6.3 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.0 CPython/3.9.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | f3b4f1bc56c69001a14da3f6129703c80ee4fc33bb2c868e698283d4d8cc9a2f |
|
MD5 | 76302b978e9bb90d84a16a92e6ee7a78 |
|
BLAKE2b-256 | f762b9a92c32d225ee6d7f8518a4bc2bf62afff1719901dea3d9f19e43468c9f |
File details
Details for the file jury-1.0.1-py3-none-any.whl
.
File metadata
- Download URL: jury-1.0.1-py3-none-any.whl
- Upload date:
- Size: 18.4 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: twine/3.4.2 importlib_metadata/4.6.3 pkginfo/1.7.1 requests/2.26.0 requests-toolbelt/0.9.1 tqdm/4.62.0 CPython/3.9.6
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | ba4824a91cf03034ea6b6a55bf877be9e8bf45683c8fd744d4f9a214b8c1ef65 |
|
MD5 | 4921925398772c0a22931379b3d3c09b |
|
BLAKE2b-256 | 2d3a3d44f57585b4401a73238cf600ef683de8b51daa91ddc9baefc1a25a45c3 |