Skip to main content

A comprehensive toolkit for large model evaluation

Project description

👋 join us on Discord and WeChat

🧭 Welcome

to OpenCompass!

Just like a compass guides us on our journey, OpenCompass will guide you through the complex landscape of evaluating large language models. With its powerful algorithms and intuitive interface, OpenCompass makes it easy to assess the quality and effectiveness of your NLP models.

🚩🚩🚩 Explore opportunities at OpenCompass! We're currently hiring full-time researchers/engineers and interns. If you're passionate about LLM and OpenCompass, don't hesitate to reach out to us via email. We'd love to hear from you!

🔥🔥🔥 We are delighted to announce that the OpenCompass has been recommended by the Meta AI, click Get Started of Llama for more information.

Attention
We launch the OpenCompass Collaboration project, welcome to support diverse evaluation benchmarks into OpenCompass! Clike Issue for more information. Let's work together to build a more powerful OpenCompass toolkit!

🚀 What's New

  • [2023.11.06] We have supported several API-based models, include ChatGLM Pro@Zhipu, ABAB-Chat@MiniMax and Xunfei. Welcome to Models section for more details. 🔥🔥🔥.
  • [2023.10.24] We release a new benchmark for evaluating LLMs’ capabilities of having multi-turn dialogues. Welcome to BotChat for more details. 🔥🔥🔥.
  • [2023.09.26] We update the leaderboard with Qwen, one of the best-performing open-source models currently available, welcome to our homepage for more details. 🔥🔥🔥.
  • [2023.09.20] We update the leaderboard with InternLM-20B, welcome to our homepage for more details. 🔥🔥🔥.
  • [2023.09.19] We update the leaderboard with WeMix-LLaMA2-70B/Phi-1.5-1.3B, welcome to our homepage for more details.
  • [2023.09.18] We have released long context evaluation guidance.
  • [2023.09.08] We update the leaderboard with Baichuan-2/Tigerbot-2/Vicuna-v1.5, welcome to our homepage for more details.
  • [2023.09.06] Baichuan2 team adpots OpenCompass to evaluate their models systematically. We deeply appreciate the community's dedication to transparency and reproducibility in LLM evaluation.
  • [2023.09.02] We have supported the evaluation of Qwen-VL in OpenCompass.

More

✨ Introduction

image

OpenCompass is a one-stop platform for large model evaluation, aiming to provide a fair, open, and reproducible benchmark for large model evaluation. Its main features include:

  • Comprehensive support for models and datasets: Pre-support for 20+ HuggingFace and API models, a model evaluation scheme of 70+ datasets with about 400,000 questions, comprehensively evaluating the capabilities of the models in five dimensions.

  • Efficient distributed evaluation: One line command to implement task division and distributed evaluation, completing the full evaluation of billion-scale models in just a few hours.

  • Diversified evaluation paradigms: Support for zero-shot, few-shot, and chain-of-thought evaluations, combined with standard or dialogue-type prompt templates, to easily stimulate the maximum performance of various models.

  • Modular design with high extensibility: Want to add new models or datasets, customize an advanced task division strategy, or even support a new cluster management system? Everything about OpenCompass can be easily expanded!

  • Experiment management and reporting mechanism: Use config files to fully record each experiment, and support real-time reporting of results.

📊 Leaderboard

We provide OpenCompass Leaderbaord for the community to rank all public models and API models. If you would like to join the evaluation, please provide the model repository URL or a standard API interface to the email address opencompass@pjlab.org.cn.

🔝Back to top

🛠️ Installation

Below are the steps for quick installation and datasets preparation.

conda create --name opencompass python=3.10 pytorch torchvision pytorch-cuda -c nvidia -c pytorch -y
conda activate opencompass
git clone https://github.com/open-compass/opencompass opencompass
cd opencompass
pip install -e .
# Download dataset to data/ folder
wget https://github.com/open-compass/opencompass/releases/download/0.1.8.rc1/OpenCompassData-core-20231110.zip
unzip OpenCompassData-core-20231110.zip

Some third-party features, like Humaneval and Llama, may require additional steps to work properly, for detailed steps please refer to the Installation Guide.

🔝Back to top

🏗️ ️Evaluation

After ensuring that OpenCompass is installed correctly according to the above steps and the datasets are prepared, you can evaluate the performance of the LLaMA-7b model on the MMLU and C-Eval datasets using the following command:

python run.py --models hf_llama_7b --datasets mmlu_ppl ceval_ppl

OpenCompass has predefined configurations for many models and datasets. You can list all available model and dataset configurations using the tools.

# List all configurations
python tools/list_configs.py
# List all configurations related to llama and mmlu
python tools/list_configs.py llama mmlu

You can also evaluate other HuggingFace models via command line. Taking LLaMA-7b as an example:

python run.py --datasets ceval_ppl mmlu_ppl \
--hf-path huggyllama/llama-7b \  # HuggingFace model path
--model-kwargs device_map='auto' \  # Arguments for model construction
--tokenizer-kwargs padding_side='left' truncation='left' use_fast=False \  # Arguments for tokenizer construction
--max-out-len 100 \  # Maximum number of tokens generated
--max-seq-len 2048 \  # Maximum sequence length the model can accept
--batch-size 8 \  # Batch size
--no-batch-padding \  # Don't enable batch padding, infer through for loop to avoid performance loss
--num-gpus 1  # Number of minimum required GPUs

Note
To run the command above, you will need to remove the comments starting from # first.

Through the command line or configuration files, OpenCompass also supports evaluating APIs or custom models, as well as more diversified evaluation strategies. Please read the Quick Start to learn how to run an evaluation task.

🔝Back to top

📖 Dataset Support

Language Knowledge Reasoning Examination
Word Definition
  • WiC
  • SummEdits
Idiom Learning
  • CHID
Semantic Similarity
  • AFQMC
  • BUSTM
Coreference Resolution
  • CLUEWSC
  • WSC
  • WinoGrande
Translation
  • Flores
  • IWSLT2017
Multi-language Question Answering
  • TyDi-QA
  • XCOPA
Multi-language Summary
  • XLSum
Knowledge Question Answering
  • BoolQ
  • CommonSenseQA
  • NaturalQuestions
  • TriviaQA
Textual Entailment
  • CMNLI
  • OCNLI
  • OCNLI_FC
  • AX-b
  • AX-g
  • CB
  • RTE
  • ANLI
Commonsense Reasoning
  • StoryCloze
  • COPA
  • ReCoRD
  • HellaSwag
  • PIQA
  • SIQA
Mathematical Reasoning
  • MATH
  • GSM8K
Theorem Application
  • TheoremQA
  • StrategyQA
  • SciBench
Comprehensive Reasoning
  • BBH
Junior High, High School, University, Professional Examinations
  • C-Eval
  • AGIEval
  • MMLU
  • GAOKAO-Bench
  • CMMLU
  • ARC
  • Xiezhi
Medical Examinations
  • CMB
Understanding Long Context Safety Code
Reading Comprehension
  • C3
  • CMRC
  • DRCD
  • MultiRC
  • RACE
  • DROP
  • OpenBookQA
  • SQuAD2.0
Content Summary
  • CSL
  • LCSTS
  • XSum
  • SummScreen
Content Analysis
  • EPRSTMT
  • LAMBADA
  • TNEWS
Long Context Understanding
  • LEval
  • LongBench
  • GovReports
  • NarrativeQA
  • Qasper
Safety
  • CivilComments
  • CrowsPairs
  • CValues
  • JigsawMultilingual
  • TruthfulQA
Robustness
  • AdvGLUE
Code
  • HumanEval
  • HumanEvalX
  • MBPP
  • APPs
  • DS1000

OpenCompass Ecosystem

🔝Back to top

📖 Model Support

Open-source Models API Models
  • InternLM
  • LLaMA
  • Vicuna
  • Alpaca
  • Baichuan
  • WizardLM
  • ChatGLM2
  • Falcon
  • TigerBot
  • Qwen
  • ...
  • OpenAI
  • Claude
  • PaLM (coming soon)
  • ……

🔝Back to top

🔜 Roadmap

  • Subjective Evaluation
    • Release CompassAreana
    • Subjective evaluation dataset.
  • Long-context
    • Long-context evaluation with extensive datasets.
    • Long-context leaderboard.
  • Coding
    • Coding evaluation leaderboard.
    • Non-python language evaluation service.
  • Agent
    • Support various agenet framework.
    • Evaluation of tool use of the LLMs.
  • Robustness
    • Support various attack method

👷‍♂️ Contributing

We appreciate all contributions to improving OpenCompass. Please refer to the contributing guideline for the best practice.

🤝 Acknowledgements

Some code in this project is cited and modified from OpenICL.

Some datasets and prompt implementations are modified from chain-of-thought-hub and instruct-eval.

🖊️ Citation

@misc{2023opencompass,
    title={OpenCompass: A Universal Evaluation Platform for Foundation Models},
    author={OpenCompass Contributors},
    howpublished = {\url{https://github.com/open-compass/opencompass}},
    year={2023}
}

🔝Back to top

Project details


Download files

Download the file for your platform. If you're not sure which to choose, learn more about installing packages.

Source Distribution

opencompass-0.1.8.tar.gz (182.2 kB view hashes)

Uploaded Source

Built Distribution

opencompass-0.1.8-py3-none-any.whl (308.1 kB view hashes)

Uploaded Python 3

Supported by

AWS AWS Cloud computing and Security Sponsor Datadog Datadog Monitoring Fastly Fastly CDN Google Google Download Analytics Microsoft Microsoft PSF Sponsor Pingdom Pingdom Monitoring Sentry Sentry Error logging StatusPage StatusPage Status page