Skip to main content

Evaluation of finetuned models.

Project description

AlexandraAI-eval

Evaluation of finetuned models

(pronounced as in "Aye aye captain")

PyPI Status Documentation License LastCommit Code Coverage

Installation

To install the package simply write the following command in your favorite terminal:

pip install alexandra-ai-eval

Quickstart

Benchmarking from the Command Line

The easiest way to benchmark pretrained models is via the command line interface. After having installed the package, you can benchmark your favorite model like so:

evaluate --model-id <model_id> --task <task>

Here model_id is the HuggingFace model ID, which can be found on the HuggingFace Hub, and task is the task you want to benchmark the model on, such as "ner" for named entity recognition. See all options by typing

evaluate --help

The specific model version to use can also be added after the suffix '@':

evaluate --model_id <model_id>@<commit>

It can be a branch name, a tag name, or a commit id. It defaults to 'main' for latest.

Multiple models and tasks can be specified by just attaching multiple arguments. Here is an example with two models:

evaluate --model_id <model_id1> --model_id <model_id2> --task ner

See all the arguments and options available for the evaluate command by typing

evaluate --help

Benchmarking from a Script

In a script, the syntax is similar to the command line interface. You simply initialise an object of the Evaluator class, and call this evaluate object with your favorite models and/or datasets:

>>> from alexandra_ai_eval import Evaluator
>>> evaluator = Evaluator()
>>> evaluator('<model_id>', '<task>')

Contributors

If you feel like this package is missing a crucial feature, if you encounter a bug or if you just want to correct a typo in this readme file, then we urge you to join the community! Have a look at the CONTRIBUTING.md file, where you can check out all the ways you can contribute to this package. :sparkles:

  • Your name here? :tada:

Maintainers

The following are the core maintainers of the alexandra_ai_eval package:

Project structure

.
├── .flake8
├── .github
│   └── workflows
│       ├── ci.yaml
│       └── docs.yaml
├── .gitignore
├── .pre-commit-config.yaml
├── LICENSE
├── README.md
├── gfx
│   └── alexandra-ai-eval-logo.png
├── makefile
├── models
├── notebooks
├── poetry.toml
├── pyproject.toml
├── src
│   ├── alexandra_ai_eval
│   │   ├── __init__.py
│   │   ├── automatic_speech_recognition.py
│   │   ├── cli.py
│   │   ├── co2.py
│   │   ├── config.py
│   │   ├── country_codes.py
│   │   ├── evaluator.py
│   │   ├── exceptions.py
│   │   ├── hf_hub.py
│   │   ├── image_to_text.py
│   │   ├── named_entity_recognition.py
│   │   ├── question_answering.py
│   │   ├── scoring.py
│   │   ├── task.py
│   │   ├── task_configs.py
│   │   ├── task_factory.py
│   │   ├── text_classification.py
│   │   └── utils.py
│   └── scripts
│       ├── fix_dot_env_file.py
│       └── versioning.py
└── tests
    ├── __init__.py
    ├── conftest.py
    ├── test_cli.py
    ├── test_co2.py
    ├── test_config.py
    ├── test_country_codes.py
    ├── test_evaluator.py
    ├── test_exceptions.py
    ├── test_hf_hub.py
    ├── test_image_to_text.py
    ├── test_named_entity_recognition.py
    ├── test_question_answering.py
    ├── test_scoring.py
    ├── test_task.py
    ├── test_task_configs.py
    ├── test_task_factory.py
    ├── test_text_classification.py
    └── test_utils.py

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

alexandra_ai_eval-0.1.0.tar.gz (55.1 kB view details)

Uploaded Source

Built Distribution

alexandra_ai_eval-0.1.0-py3-none-any.whl (66.3 kB view details)

Uploaded Python 3

File details

Details for the file alexandra_ai_eval-0.1.0.tar.gz.

File metadata

  • Download URL: alexandra_ai_eval-0.1.0.tar.gz
  • Upload date:
  • Size: 55.1 kB
  • Tags: Source
  • Uploaded using Trusted Publishing? No
  • Uploaded via: poetry/1.2.0 CPython/3.10.10 Darwin/21.6.0

File hashes

Hashes for alexandra_ai_eval-0.1.0.tar.gz
Algorithm Hash digest
SHA256 0c1cc6c6b00b69efae20a0909684eb75ed28c1042136b0521e66c82b64722a40
MD5 3e2a93cdfc5110c790ac0469eb886ed7
BLAKE2b-256 9a915bbb19df115ef0ac048aa1b21b6a70e9ea5a1218381dd2f5a2c54c015637

See more details on using hashes here.

File details

Details for the file alexandra_ai_eval-0.1.0-py3-none-any.whl.

File metadata

File hashes

Hashes for alexandra_ai_eval-0.1.0-py3-none-any.whl
Algorithm Hash digest
SHA256 cb1233b556bfdc0e34246b641bbad4eb7a3ef4447f6eeb6bb1e586e15982d553
MD5 366cf89ea27f2d98ab54ceb7243689e4
BLAKE2b-256 06b0849faefaf5d87981c93bf511a3bce63e9f37646bb8d2c3b14ad8725cfe43

See more details on using hashes here.

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page