Unofficial CodeBLEU implementation that supports Linux, MacOS and Windows available on PyPI.
Project description
CodeBLEU
This repository contains an unofficial CodeBLEU
implementation that supports Linux
, MacOS
(incl. M-series) and Windows
. It is available through PyPI
and the evaluate
library.
Available for: Python
, C
, C#
, C++
, Java
, JavaScript
, PHP
, Go
, Ruby
, Rust
.
The code is based on the original CodeXGLUE/CodeBLEU and updated version by XLCoST/CodeBLEU. It has been refactored, tested, built for macOS and Windows, and multiple improvements have been made to enhance usability.
Metric Description
An ideal evaluation metric should consider the grammatical correctness and the logic correctness. We propose weighted n-gram match and syntactic AST match to measure grammatical correctness, and introduce semantic data-flow match to calculate logic correctness.
[from CodeXGLUE repo]
In a nutshell, CodeBLEU
is a weighted combination of n-gram match (BLEU)
, weighted n-gram match (BLEU-weighted)
, AST match
and data-flow match
scores.
The metric has shown higher correlation with human evaluation than BLEU
and accuracy
metrics.
Installation
This library requires so
file compilation with tree-sitter, so it is platform dependent.
Currently available for Linux
(manylinux), MacOS
and Windows
with Python 3.8+.
The metrics is available as pip package and can be installed as indicated above:
pip install codebleu
or directly from git repo (require internet connection to download tree-sitter):
pip install git+https://github.com/k4black/codebleu.git
Also you have to install tree-sitter language you need (e.g. python, rust, etc):
pip install tree-sitter-python
Or you can install all languages:
pip install codebleu[all]
Note: At the moment (May 2024) precompiled languages are NOT available for arm64 (M1) MacOS, so you have to install and build tree-sitter languages manually, for example:
pip install pip install git+https://github.com/tree-sitter/tree-sitter-python.git
Usage
from codebleu import calc_codebleu
prediction = "def add ( a , b ) :\n return a + b"
reference = "def sum ( first , second ) :\n return second + first"
result = calc_codebleu([reference], [prediction], lang="python", weights=(0.25, 0.25, 0.25, 0.25), tokenizer=None)
print(result)
# {
# 'codebleu': 0.5537,
# 'ngram_match_score': 0.1041,
# 'weighted_ngram_match_score': 0.1109,
# 'syntax_match_score': 1.0,
# 'dataflow_match_score': 1.0
# }
where calc_codebleu
takes the following arguments:
refarences
(list[str]
orlist[list[str]]
): reference codepredictions
(list[str]
) predicted codelang
(str
): code language, seecodebleu.AVAILABLE_LANGS
for available languages (python, c_sharp c, cpp, javascript, java, php, go and ruby at the moment)weights
(tuple[float,float,float,float]
): weights of thengram_match
,weighted_ngram_match
,syntax_match
, anddataflow_match
respectively, defaults to(0.25, 0.25, 0.25, 0.25)
tokenizer
(callable
): to split code string to tokens, defaults tos.split()
and outputs the dict[str, float]
with following fields:
codebleu
: the finalCodeBLEU
scorengram_match_score
:ngram_match
score (BLEU)weighted_ngram_match_score
:weighted_ngram_match
score (BLEU-weighted)syntax_match_score
:syntax_match
score (AST match)dataflow_match_score
:dataflow_match
score
Alternatively, you can use k4black/codebleu
from HuggingFace Spaces (codebleu
package required):
import evaluate
metric = evaluate.load("dvitel/codebleu")
prediction = "def add ( a , b ) :\n return a + b"
reference = "def sum ( first , second ) :\n return second + first"
result = metric.compute([reference], [prediction], lang="python", weights=(0.25, 0.25, 0.25, 0.25))
Feel free to check the HF Space with online example: k4black/codebleu
Contributing
Contributions are welcome!
If you have any questions, suggestions, or bug reports, please open an issue on GitHub.
Make your own fork and clone it:
git clone https://github.com/k4black/codebleu
For development, you need to install library with all
precompiled languages and test
extra:
(require internet connection to download tree-sitter)
python -m pip install -e .[all,test]
python -m pip install -e .\[all,test\] # for macos
For testing just run pytest:
python -m pytest
To perform a style check, run:
python -m isort codebleu --check
python -m black codebleu --check
python -m ruff codebleu
python -m mypy codebleu
License
This project is licensed under the terms of the MIT license.
Citation
Official CodeBLEU paper can be cited as follows:
@misc{ren2020codebleu,
title={CodeBLEU: a Method for Automatic Evaluation of Code Synthesis},
author={Shuo Ren and Daya Guo and Shuai Lu and Long Zhou and Shujie Liu and Duyu Tang and Neel Sundaresan and Ming Zhou and Ambrosio Blanco and Shuai Ma},
year={2020},
eprint={2009.10297},
archivePrefix={arXiv},
primaryClass={cs.SE}
}
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
File details
Details for the file codebleu-0.7.0.tar.gz
.
File metadata
- Download URL: codebleu-0.7.0.tar.gz
- Upload date:
- Size: 178.0 kB
- Tags: Source
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.0.0 CPython/3.12.3
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | 6f379d5cd1663e1d248ee79d9b23d043c6bc6b60bd03fb106472bae77d2285c3 |
|
MD5 | 41e9e4fadd85c7e4233b396764b3b817 |
|
BLAKE2b-256 | 8f4587298d89e206d88ce83c5d1c36d3d5bcd4eb0fcd64cfcd36adf58b844093 |
File details
Details for the file codebleu-0.7.0-py3-none-any.whl
.
File metadata
- Download URL: codebleu-0.7.0-py3-none-any.whl
- Upload date:
- Size: 31.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? Yes
- Uploaded via: twine/5.0.0 CPython/3.12.3
File hashes
Algorithm | Hash digest | |
---|---|---|
SHA256 | e664e21cf407a355a726fd5af05e0021d8a162f28999281d907d4e7b36c54873 |
|
MD5 | 3363809c700c2a34529612509e91ab5f |
|
BLAKE2b-256 | a744888354a4cc9376e3ddc26ed5ed765e2ede0dcdba37ea7b71411b9dab90d2 |