Skip to main content

A library for benchmarking AI/ML applications.

Project description

https://img.shields.io/github/license/ebu/benchmarkstt.svg GitHub Workflow Status (branch) Documentation Status

About

This is a command line tool for benchmarking Automatic Speech Recognition engines.

It is designed for non-academic production environments, and prioritises ease of use and relative benchmarking over scientific procedure and high-accuracy absolute scoring.

Because of the wide range of languages, algorithms and audio characteristics, no single STT engine can be expected to excel in all circumstances. For this reason, this tool places responsibility on the users to design their own benchmarking procedure and to decide, based on the combination of test data and metrics, which engine is best suited for their particular use case.

Usage examples

Returns the number of word insertions, deletions, replacements and matches for the hypothesis transcript compared to the reference:

benchmarkstt --reference reference.txt --hypothesis hypothesis.txt --diffcounts

Returns the Word Error Rate after lowercasing both reference and hypothesis. This normlization improves the accuracy of the Word Error Rate as it removes diffs that might otherwise be considered errors:

benchmarkstt -r reference.txt -h hypothesis.txt --wer --lowercase

Returns a visual diff after applying all the normalization rules specified in the config file:

benchmarkstt -r reference.txt -h hypothesis.txt --worddiffs --config conf

Further information

This is a collaborative project to create a library for benchmarking AI/ML applications. It was created in response to the needs of broadcasters and providers of Access Services to media organisations, but anyone is welcome to contribute. The group behind this project is the EBU’s Media Information Management & AI group.

Currently the group is focussing on Speech-to-Text, but it will consider creating benchmarking tools for other AI/ML services.

For general information about this project, including the motivations and guiding principles, please see the project wiki

To install and start using the tool, go to the documentation.

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

benchmarkstt-1.1.tar.gz (36.5 kB view details)

Uploaded Source

Built Distribution

benchmarkstt-1.1-py3-none-any.whl (51.4 kB view details)

Uploaded Python 3

File details

Details for the file benchmarkstt-1.1.tar.gz.

File metadata

  • Download URL: benchmarkstt-1.1.tar.gz
  • Upload date:
  • Size: 36.5 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/3.10.0 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.60.0 CPython/3.9.1

File hashes

Hashes for benchmarkstt-1.1.tar.gz
Algorithm Hash digest
SHA256 d55cf005b1fac94279a466bff2145b8e5c7f6097160e383c1da46582482260fc
MD5 cd532a49100489a9589a17ae24ffd401
BLAKE2b-256 d1413cbea4465eb19cb82363f70350757b6c28b90941244c8a9bf1db2bc9e267

See more details on using hashes here.

File details

Details for the file benchmarkstt-1.1-py3-none-any.whl.

File metadata

  • Download URL: benchmarkstt-1.1-py3-none-any.whl
  • Upload date:
  • Size: 51.4 kB
  • Tags: Python 3
  • Uploaded using Trusted Publishing? No
  • Uploaded via: twine/3.4.1 importlib_metadata/3.10.0 pkginfo/1.7.0 requests/2.25.1 requests-toolbelt/0.9.1 tqdm/4.60.0 CPython/3.9.1

File hashes

Hashes for benchmarkstt-1.1-py3-none-any.whl
Algorithm Hash digest
SHA256 b1ee62cc16ac593b542cfe16d0f35ce3bff26ada80b25b6f92576e68a906877a
MD5 69aa4b9cca2408dc59cf6440876b3d92
BLAKE2b-256 04e9d1e8fcde4e3a3eb656d409595eff52d0b461c7f1246a968a07b293f660cd

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page