opencompass


Nameopencompass JSON
Version 0.2.4 PyPI version JSON
download
home_pagehttps://github.com/open-compass/opencompass
SummaryA comprehensive toolkit for large model evaluation
upload_time2024-04-23 09:22:12
maintainerOpenCompass Authors
docs_urlNone
authorOpenCompass Contributors
requires_python>=3.8.0
licenseApache License 2.0
keywords ai nlp in-context learning large language model evaluation benchmark llm
VCS
bugtrack_url
requirements No requirements were recorded.
Travis-CI No Travis.
coveralls test coverage No coveralls.
            <div align="center">
  <img src="docs/en/_static/image/logo.svg" width="500px"/>
  <br />
  <br />

[![][github-release-shield]][github-release-link]
[![][github-releasedate-shield]][github-releasedate-link]
[![][github-contributors-shield]][github-contributors-link]<br>
[![][github-forks-shield]][github-forks-link]
[![][github-stars-shield]][github-stars-link]
[![][github-issues-shield]][github-issues-link]
[![][github-license-shield]][github-license-link]

<!-- [![PyPI](https://badge.fury.io/py/opencompass.svg)](https://pypi.org/project/opencompass/) -->

[🌐Website](https://opencompass.org.cn/) |
[📖CompassHub](https://hub.opencompass.org.cn/home) |
[📊CompassRank](https://rank.opencompass.org.cn/home) |
[📘Documentation](https://opencompass.readthedocs.io/en/latest/) |
[🛠️Installation](https://opencompass.readthedocs.io/en/latest/get_started/installation.html) |
[🤔Reporting Issues](https://github.com/open-compass/opencompass/issues/new/choose)

English | [简体中文](README_zh-CN.md)

[![][github-trending-shield]][github-trending-url]

</div>

<p align="center">
    👋 join us on <a href="https://discord.gg/KKwfEbFj7U" target="_blank">Discord</a> and <a href="https://r.vansin.top/?r=opencompass" target="_blank">WeChat</a>
</p>

> \[!IMPORTANT\]
>
> **Star Us**, You will receive all release notifications from GitHub without any delay ~ ⭐️

## 📣 OpenCompass 2.0

We are thrilled to introduce OpenCompass 2.0, an advanced suite featuring three key components: [CompassKit](https://github.com/open-compass), [CompassHub](https://hub.opencompass.org.cn/home), and [CompassRank](https://rank.opencompass.org.cn/home).
![oc20](https://github.com/tonysy/opencompass/assets/7881589/90dbe1c0-c323-470a-991e-2b37ab5350b2)

**CompassRank** has been significantly enhanced into the leaderboards that now incorporates both open-source benchmarks and proprietary benchmarks. This upgrade allows for a more comprehensive evaluation of models across the industry.

**CompassHub** presents a pioneering benchmark browser interface, designed to simplify and expedite the exploration and utilization of an extensive array of benchmarks for researchers and practitioners alike. To enhance the visibility of your own benchmark within the community, we warmly invite you to contribute it to CompassHub. You may initiate the submission process by clicking [here](https://hub.opencompass.org.cn/dataset-submit).

**CompassKit** is a powerful collection of evaluation toolkits specifically tailored for Large Language Models and Large Vision-language Models. It provides an extensive set of tools to assess and measure the performance of these complex models effectively. Welcome to try our toolkits for in your research and products.

<details>
  <summary><kbd>Star History</kbd></summary>
  <picture>
    <source media="(prefers-color-scheme: dark)" srcset="https://api.star-history.com/svg?repos=open-compass%2Fopencompass&theme=dark&type=Date">
    <img width="100%" src="https://api.star-history.com/svg?repos=open-compass%2Fopencompass&type=Date">
  </picture>
</details>

## 🧭	Welcome

to **OpenCompass**!

Just like a compass guides us on our journey, OpenCompass will guide you through the complex landscape of evaluating large language models. With its powerful algorithms and intuitive interface, OpenCompass makes it easy to assess the quality and effectiveness of your NLP models.

🚩🚩🚩 Explore opportunities at OpenCompass! We're currently **hiring full-time researchers/engineers and interns**. If you're passionate about LLM and OpenCompass, don't hesitate to reach out to us via [email](mailto:zhangsongyang@pjlab.org.cn). We'd love to hear from you!

🔥🔥🔥 We are delighted to announce that **the OpenCompass has been recommended by the Meta AI**, click [Get Started](https://ai.meta.com/llama/get-started/#validation) of Llama for more information.

> **Attention**<br />
> We launch the OpenCompass Collaboration project, welcome to support diverse evaluation benchmarks into OpenCompass!
> Clike [Issue](https://github.com/open-compass/opencompass/issues/248) for more information.
> Let's work together to build a more powerful OpenCompass toolkit!

## 🚀 What's New <a><img width="35" height="20" src="https://user-images.githubusercontent.com/12782558/212848161-5e783dd6-11e8-4fe0-bbba-39ffb77730be.png"></a>

- **\[2024.04.22\]** We supported the evaluation of [LLaMA3](configs/models/hf_llama/hf_llama3_8b.py) 和 [LLaMA3-Instruct](configs/models/hf_llama/hf_llama3_8b_instruct.py), welcome to try! 🔥🔥🔥
- **\[2024.02.29\]** We supported the MT-Bench, AlpacalEval and AlignBench, more information can be found [here](https://opencompass.readthedocs.io/en/latest/advanced_guides/subjective_evaluation.html)
- **\[2024.01.30\]** We release OpenCompass 2.0. Click  [CompassKit](https://github.com/open-compass), [CompassHub](https://hub.opencompass.org.cn/home), and [CompassRank](https://rank.opencompass.org.cn/home) for more information !

> [More](docs/en/notes/news.md)

## ✨ Introduction

![image](https://github.com/open-compass/opencompass/assets/22607038/f45fe125-4aed-4f8c-8fe8-df4efb41a8ea)

OpenCompass is a one-stop platform for large model evaluation, aiming to provide a fair, open, and reproducible benchmark for large model evaluation. Its main features include:

- **Comprehensive support for models and datasets**: Pre-support for 20+ HuggingFace and API models, a model evaluation scheme of 70+ datasets with about 400,000 questions, comprehensively evaluating the capabilities of the models in five dimensions.

- **Efficient distributed evaluation**: One line command to implement task division and distributed evaluation, completing the full evaluation of billion-scale models in just a few hours.

- **Diversified evaluation paradigms**: Support for zero-shot, few-shot, and chain-of-thought evaluations, combined with standard or dialogue-type prompt templates, to easily stimulate the maximum performance of various models.

- **Modular design with high extensibility**: Want to add new models or datasets, customize an advanced task division strategy, or even support a new cluster management system? Everything about OpenCompass can be easily expanded!

- **Experiment management and reporting mechanism**: Use config files to fully record each experiment, and support real-time reporting of results.

## 📊 Leaderboard

We provide [OpenCompass Leaderboard](https://rank.opencompass.org.cn/home) for the community to rank all public models and API models. If you would like to join the evaluation, please provide the model repository URL or a standard API interface to the email address `opencompass@pjlab.org.cn`.

<p align="right"><a href="#top">🔝Back to top</a></p>

## 🛠️ Installation

Below are the steps for quick installation and datasets preparation.

### 💻 Environment Setup

#### Open-source Models with GPU

```bash
conda create --name opencompass python=3.10 pytorch torchvision pytorch-cuda -c nvidia -c pytorch -y
conda activate opencompass
git clone https://github.com/open-compass/opencompass opencompass
cd opencompass
pip install -e .
```

#### API Models with CPU-only

```bash
conda create -n opencompass python=3.10 pytorch torchvision torchaudio cpuonly -c pytorch -y
conda activate opencompass
git clone https://github.com/open-compass/opencompass opencompass
cd opencompass
pip install -e .
# also please install requiresments packages via `pip install -r requirements/api.txt` for API models if needed.
```

### 📂 Data Preparation

```bash
# Download dataset to data/ folder
wget https://github.com/open-compass/opencompass/releases/download/0.2.2.rc1/OpenCompassData-core-20240207.zip
unzip OpenCompassData-core-20240207.zip
```

Some third-party features, like Humaneval and Llama, may require additional steps to work properly, for detailed steps please refer to the [Installation Guide](https://opencompass.readthedocs.io/en/latest/get_started/installation.html).

<p align="right"><a href="#top">🔝Back to top</a></p>

## 🏗️ ️Evaluation

After ensuring that OpenCompass is installed correctly according to the above steps and the datasets are prepared, you can evaluate the performance of the LLaMA-7b model on the MMLU and C-Eval datasets using the following command:

```bash
python run.py --models hf_llama_7b --datasets mmlu_ppl ceval_ppl
```

OpenCompass has predefined configurations for many models and datasets. You can list all available model and dataset configurations using the [tools](./docs/en/tools.md#list-configs).

```bash
# List all configurations
python tools/list_configs.py
# List all configurations related to llama and mmlu
python tools/list_configs.py llama mmlu
```

You can also evaluate other HuggingFace models via command line. Taking LLaMA-7b as an example:

```bash
python run.py --datasets ceval_ppl mmlu_ppl \
--hf-path huggyllama/llama-7b \  # HuggingFace model path
--model-kwargs device_map='auto' \  # Arguments for model construction
--tokenizer-kwargs padding_side='left' truncation='left' use_fast=False \  # Arguments for tokenizer construction
--max-out-len 100 \  # Maximum number of tokens generated
--max-seq-len 2048 \  # Maximum sequence length the model can accept
--batch-size 8 \  # Batch size
--no-batch-padding \  # Don't enable batch padding, infer through for loop to avoid performance loss
--num-gpus 1  # Number of minimum required GPUs
```

> **Note**<br />
> To run the command above, you will need to remove the comments starting from `# ` first.

Through the command line or configuration files, OpenCompass also supports evaluating APIs or custom models, as well as more diversified evaluation strategies. Please read the [Quick Start](https://opencompass.readthedocs.io/en/latest/get_started/quick_start.html) to learn how to run an evaluation task.

<p align="right"><a href="#top">🔝Back to top</a></p>

## 📖 Dataset Support

<table align="center">
  <tbody>
    <tr align="center" valign="bottom">
      <td>
        <b>Language</b>
      </td>
      <td>
        <b>Knowledge</b>
      </td>
      <td>
        <b>Reasoning</b>
      </td>
      <td>
        <b>Examination</b>
      </td>
    </tr>
    <tr valign="top">
      <td>
<details open>
<summary><b>Word Definition</b></summary>

- WiC
- SummEdits

</details>

<details open>
<summary><b>Idiom Learning</b></summary>

- CHID

</details>

<details open>
<summary><b>Semantic Similarity</b></summary>

- AFQMC
- BUSTM

</details>

<details open>
<summary><b>Coreference Resolution</b></summary>

- CLUEWSC
- WSC
- WinoGrande

</details>

<details open>
<summary><b>Translation</b></summary>

- Flores
- IWSLT2017

</details>

<details open>
<summary><b>Multi-language Question Answering</b></summary>

- TyDi-QA
- XCOPA

</details>

<details open>
<summary><b>Multi-language Summary</b></summary>

- XLSum

</details>
      </td>
      <td>
<details open>
<summary><b>Knowledge Question Answering</b></summary>

- BoolQ
- CommonSenseQA
- NaturalQuestions
- TriviaQA

</details>
      </td>
      <td>
<details open>
<summary><b>Textual Entailment</b></summary>

- CMNLI
- OCNLI
- OCNLI_FC
- AX-b
- AX-g
- CB
- RTE
- ANLI

</details>

<details open>
<summary><b>Commonsense Reasoning</b></summary>

- StoryCloze
- COPA
- ReCoRD
- HellaSwag
- PIQA
- SIQA

</details>

<details open>
<summary><b>Mathematical Reasoning</b></summary>

- MATH
- GSM8K

</details>

<details open>
<summary><b>Theorem Application</b></summary>

- TheoremQA
- StrategyQA
- SciBench

</details>

<details open>
<summary><b>Comprehensive Reasoning</b></summary>

- BBH

</details>
      </td>
      <td>
<details open>
<summary><b>Junior High, High School, University, Professional Examinations</b></summary>

- C-Eval
- AGIEval
- MMLU
- GAOKAO-Bench
- CMMLU
- ARC
- Xiezhi

</details>

<details open>
<summary><b>Medical Examinations</b></summary>

- CMB

</details>
      </td>
    </tr>
</td>
    </tr>
  </tbody>
  <tbody>
    <tr align="center" valign="bottom">
      <td>
        <b>Understanding</b>
      </td>
      <td>
        <b>Long Context</b>
      </td>
      <td>
        <b>Safety</b>
      </td>
      <td>
        <b>Code</b>
      </td>
    </tr>
    <tr valign="top">
      <td>
<details open>
<summary><b>Reading Comprehension</b></summary>

- C3
- CMRC
- DRCD
- MultiRC
- RACE
- DROP
- OpenBookQA
- SQuAD2.0

</details>

<details open>
<summary><b>Content Summary</b></summary>

- CSL
- LCSTS
- XSum
- SummScreen

</details>

<details open>
<summary><b>Content Analysis</b></summary>

- EPRSTMT
- LAMBADA
- TNEWS

</details>
      </td>
      <td>
<details open>
<summary><b>Long Context Understanding</b></summary>

- LEval
- LongBench
- GovReports
- NarrativeQA
- Qasper

</details>
      </td>
      <td>
<details open>
<summary><b>Safety</b></summary>

- CivilComments
- CrowsPairs
- CValues
- JigsawMultilingual
- TruthfulQA

</details>
<details open>
<summary><b>Robustness</b></summary>

- AdvGLUE

</details>
      </td>
      <td>
<details open>
<summary><b>Code</b></summary>

- HumanEval
- HumanEvalX
- MBPP
- APPs
- DS1000

</details>
      </td>
    </tr>
</td>
    </tr>
  </tbody>
</table>

## 📖 Model Support

<table align="center">
  <tbody>
    <tr align="center" valign="bottom">
      <td>
        <b>Open-source Models</b>
      </td>
      <td>
        <b>API Models</b>
      </td>
      <!-- <td>
        <b>Custom Models</b>
      </td> -->
    </tr>
    <tr valign="top">
      <td>

- [InternLM](https://github.com/InternLM/InternLM)
- [LLaMA](https://github.com/facebookresearch/llama)
- [LLaMA3](https://github.com/meta-llama/llama3)
- [Vicuna](https://github.com/lm-sys/FastChat)
- [Alpaca](https://github.com/tatsu-lab/stanford_alpaca)
- [Baichuan](https://github.com/baichuan-inc)
- [WizardLM](https://github.com/nlpxucan/WizardLM)
- [ChatGLM2](https://github.com/THUDM/ChatGLM2-6B)
- [ChatGLM3](https://github.com/THUDM/ChatGLM3-6B)
- [TigerBot](https://github.com/TigerResearch/TigerBot)
- [Qwen](https://github.com/QwenLM/Qwen)
- [BlueLM](https://github.com/vivo-ai-lab/BlueLM)
- [Gemma](https://huggingface.co/google/gemma-7b)
- ...

</td>
<td>

- OpenAI
- Gemini
- Claude
- ZhipuAI(ChatGLM)
- Baichuan
- ByteDance(YunQue)
- Huawei(PanGu)
- 360
- Baidu(ERNIEBot)
- MiniMax(ABAB-Chat)
- SenseTime(nova)
- Xunfei(Spark)
- ……

</td>

</tr>
  </tbody>
</table>

<p align="right"><a href="#top">🔝Back to top</a></p>

## 🔜 Roadmap

- [x] Subjective Evaluation
  - [ ] Release CompassAreana
  - [x] Subjective evaluation.
- [x] Long-context
  - [x] Long-context evaluation with extensive datasets.
  - [ ] Long-context leaderboard.
- [x] Coding
  - [ ] Coding evaluation leaderboard.
  - [x] Non-python language evaluation service.
- [x] Agent
  - [ ] Support various agenet framework.
  - [x] Evaluation of tool use of the LLMs.
- [x] Robustness
  - [x] Support various attack method

## 👷‍♂️ Contributing

We appreciate all contributions to improving OpenCompass. Please refer to the [contributing guideline](https://opencompass.readthedocs.io/en/latest/notes/contribution_guide.html) for the best practice.

<!-- Copy-paste in your Readme.md file -->

<!-- Made with [OSS Insight](https://ossinsight.io/) -->

<a href="https://github.com/open-compass/opencompass/graphs/contributors" target="_blank">
  <table>
    <tr>
      <th colspan="2">
        <br><img src="https://contrib.rocks/image?repo=open-compass/opencompass"><br><br>
      </th>
    </tr>
  </table>
</a>

## 🤝 Acknowledgements

Some code in this project is cited and modified from [OpenICL](https://github.com/Shark-NLP/OpenICL).

Some datasets and prompt implementations are modified from [chain-of-thought-hub](https://github.com/FranxYao/chain-of-thought-hub) and [instruct-eval](https://github.com/declare-lab/instruct-eval).

## 🖊️ Citation

```bibtex
@misc{2023opencompass,
    title={OpenCompass: A Universal Evaluation Platform for Foundation Models},
    author={OpenCompass Contributors},
    howpublished = {\url{https://github.com/open-compass/opencompass}},
    year={2023}
}
```

<p align="right"><a href="#top">🔝Back to top</a></p>

[github-contributors-link]: https://github.com/open-compass/opencompass/graphs/contributors
[github-contributors-shield]: https://img.shields.io/github/contributors/open-compass/opencompass?color=c4f042&labelColor=black&style=flat-square
[github-forks-link]: https://github.com/open-compass/opencompass/network/members
[github-forks-shield]: https://img.shields.io/github/forks/open-compass/opencompass?color=8ae8ff&labelColor=black&style=flat-square
[github-issues-link]: https://github.com/open-compass/opencompass/issues
[github-issues-shield]: https://img.shields.io/github/issues/open-compass/opencompass?color=ff80eb&labelColor=black&style=flat-square
[github-license-link]: https://github.com/open-compass/opencompass/blob/main/LICENSE
[github-license-shield]: https://img.shields.io/github/license/open-compass/opencompass?color=white&labelColor=black&style=flat-square
[github-release-link]: https://github.com/open-compass/opencompass/releases
[github-release-shield]: https://img.shields.io/github/v/release/open-compass/opencompass?color=369eff&labelColor=black&logo=github&style=flat-square
[github-releasedate-link]: https://github.com/open-compass/opencompass/releases
[github-releasedate-shield]: https://img.shields.io/github/release-date/open-compass/opencompass?labelColor=black&style=flat-square
[github-stars-link]: https://github.com/open-compass/opencompass/stargazers
[github-stars-shield]: https://img.shields.io/github/stars/open-compass/opencompass?color=ffcb47&labelColor=black&style=flat-square
[github-trending-shield]: https://trendshift.io/api/badge/repositories/6630
[github-trending-url]: https://trendshift.io/repositories/6630



            

Raw data

            {
    "_id": null,
    "home_page": "https://github.com/open-compass/opencompass",
    "name": "opencompass",
    "maintainer": "OpenCompass Authors",
    "docs_url": null,
    "requires_python": ">=3.8.0",
    "maintainer_email": null,
    "keywords": "AI, NLP, in-context learning, large language model, evaluation, benchmark, llm",
    "author": "OpenCompass Contributors",
    "author_email": null,
    "download_url": "https://files.pythonhosted.org/packages/4c/c0/047d8e98e8128ea54571506d55fab017c234798ba2ade115e41ad0104bbb/opencompass-0.2.4.tar.gz",
    "platform": null,
    "description": "<div align=\"center\">\n  <img src=\"docs/en/_static/image/logo.svg\" width=\"500px\"/>\n  <br />\n  <br />\n\n[![][github-release-shield]][github-release-link]\n[![][github-releasedate-shield]][github-releasedate-link]\n[![][github-contributors-shield]][github-contributors-link]<br>\n[![][github-forks-shield]][github-forks-link]\n[![][github-stars-shield]][github-stars-link]\n[![][github-issues-shield]][github-issues-link]\n[![][github-license-shield]][github-license-link]\n\n<!-- [![PyPI](https://badge.fury.io/py/opencompass.svg)](https://pypi.org/project/opencompass/) -->\n\n[\ud83c\udf10Website](https://opencompass.org.cn/) |\n[\ud83d\udcd6CompassHub](https://hub.opencompass.org.cn/home) |\n[\ud83d\udccaCompassRank](https://rank.opencompass.org.cn/home) |\n[\ud83d\udcd8Documentation](https://opencompass.readthedocs.io/en/latest/) |\n[\ud83d\udee0\ufe0fInstallation](https://opencompass.readthedocs.io/en/latest/get_started/installation.html) |\n[\ud83e\udd14Reporting Issues](https://github.com/open-compass/opencompass/issues/new/choose)\n\nEnglish | [\u7b80\u4f53\u4e2d\u6587](README_zh-CN.md)\n\n[![][github-trending-shield]][github-trending-url]\n\n</div>\n\n<p align=\"center\">\n    \ud83d\udc4b join us on <a href=\"https://discord.gg/KKwfEbFj7U\" target=\"_blank\">Discord</a> and <a href=\"https://r.vansin.top/?r=opencompass\" target=\"_blank\">WeChat</a>\n</p>\n\n> \\[!IMPORTANT\\]\n>\n> **Star Us**, You will receive all release notifications from GitHub without any delay ~ \u2b50\ufe0f\n\n## \ud83d\udce3 OpenCompass 2.0\n\nWe are thrilled to introduce OpenCompass 2.0, an advanced suite featuring three key components: [CompassKit](https://github.com/open-compass), [CompassHub](https://hub.opencompass.org.cn/home), and [CompassRank](https://rank.opencompass.org.cn/home).\n![oc20](https://github.com/tonysy/opencompass/assets/7881589/90dbe1c0-c323-470a-991e-2b37ab5350b2)\n\n**CompassRank** has been significantly enhanced into the leaderboards that now incorporates both open-source benchmarks and proprietary benchmarks. This upgrade allows for a more comprehensive evaluation of models across the industry.\n\n**CompassHub** presents a pioneering benchmark browser interface, designed to simplify and expedite the exploration and utilization of an extensive array of benchmarks for researchers and practitioners alike. To enhance the visibility of your own benchmark within the community, we warmly invite you to contribute it to CompassHub. You may initiate the submission process by clicking [here](https://hub.opencompass.org.cn/dataset-submit).\n\n**CompassKit** is a powerful collection of evaluation toolkits specifically tailored for Large Language Models and Large Vision-language Models. It provides an extensive set of tools to assess and measure the performance of these complex models effectively. Welcome to try our toolkits for in your research and products.\n\n<details>\n  <summary><kbd>Star History</kbd></summary>\n  <picture>\n    <source media=\"(prefers-color-scheme: dark)\" srcset=\"https://api.star-history.com/svg?repos=open-compass%2Fopencompass&theme=dark&type=Date\">\n    <img width=\"100%\" src=\"https://api.star-history.com/svg?repos=open-compass%2Fopencompass&type=Date\">\n  </picture>\n</details>\n\n## \ud83e\udded\tWelcome\n\nto **OpenCompass**!\n\nJust like a compass guides us on our journey, OpenCompass will guide you through the complex landscape of evaluating large language models. With its powerful algorithms and intuitive interface, OpenCompass makes it easy to assess the quality and effectiveness of your NLP models.\n\n\ud83d\udea9\ud83d\udea9\ud83d\udea9 Explore opportunities at OpenCompass! We're currently **hiring full-time researchers/engineers and interns**. If you're passionate about LLM and OpenCompass, don't hesitate to reach out to us via [email](mailto:zhangsongyang@pjlab.org.cn). We'd love to hear from you!\n\n\ud83d\udd25\ud83d\udd25\ud83d\udd25 We are delighted to announce that **the OpenCompass has been recommended by the Meta AI**, click [Get Started](https://ai.meta.com/llama/get-started/#validation) of Llama for more information.\n\n> **Attention**<br />\n> We launch the OpenCompass Collaboration project, welcome to support diverse evaluation benchmarks into OpenCompass!\n> Clike [Issue](https://github.com/open-compass/opencompass/issues/248) for more information.\n> Let's work together to build a more powerful OpenCompass toolkit!\n\n## \ud83d\ude80 What's New <a><img width=\"35\" height=\"20\" src=\"https://user-images.githubusercontent.com/12782558/212848161-5e783dd6-11e8-4fe0-bbba-39ffb77730be.png\"></a>\n\n- **\\[2024.04.22\\]** We supported the evaluation of [LLaMA3](configs/models/hf_llama/hf_llama3_8b.py) \u548c [LLaMA3-Instruct](configs/models/hf_llama/hf_llama3_8b_instruct.py), welcome to try! \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.02.29\\]** We supported the MT-Bench, AlpacalEval and AlignBench, more information can be found [here](https://opencompass.readthedocs.io/en/latest/advanced_guides/subjective_evaluation.html)\n- **\\[2024.01.30\\]** We release OpenCompass 2.0. Click  [CompassKit](https://github.com/open-compass), [CompassHub](https://hub.opencompass.org.cn/home), and [CompassRank](https://rank.opencompass.org.cn/home) for more information !\n\n> [More](docs/en/notes/news.md)\n\n## \u2728 Introduction\n\n![image](https://github.com/open-compass/opencompass/assets/22607038/f45fe125-4aed-4f8c-8fe8-df4efb41a8ea)\n\nOpenCompass is a one-stop platform for large model evaluation, aiming to provide a fair, open, and reproducible benchmark for large model evaluation. Its main features include:\n\n- **Comprehensive support for models and datasets**: Pre-support for 20+ HuggingFace and API models, a model evaluation scheme of 70+ datasets with about 400,000 questions, comprehensively evaluating the capabilities of the models in five dimensions.\n\n- **Efficient distributed evaluation**: One line command to implement task division and distributed evaluation, completing the full evaluation of billion-scale models in just a few hours.\n\n- **Diversified evaluation paradigms**: Support for zero-shot, few-shot, and chain-of-thought evaluations, combined with standard or dialogue-type prompt templates, to easily stimulate the maximum performance of various models.\n\n- **Modular design with high extensibility**: Want to add new models or datasets, customize an advanced task division strategy, or even support a new cluster management system? Everything about OpenCompass can be easily expanded!\n\n- **Experiment management and reporting mechanism**: Use config files to fully record each experiment, and support real-time reporting of results.\n\n## \ud83d\udcca Leaderboard\n\nWe provide [OpenCompass Leaderboard](https://rank.opencompass.org.cn/home) for the community to rank all public models and API models. If you would like to join the evaluation, please provide the model repository URL or a standard API interface to the email address `opencompass@pjlab.org.cn`.\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n## \ud83d\udee0\ufe0f Installation\n\nBelow are the steps for quick installation and datasets preparation.\n\n### \ud83d\udcbb Environment Setup\n\n#### Open-source Models with GPU\n\n```bash\nconda create --name opencompass python=3.10 pytorch torchvision pytorch-cuda -c nvidia -c pytorch -y\nconda activate opencompass\ngit clone https://github.com/open-compass/opencompass opencompass\ncd opencompass\npip install -e .\n```\n\n#### API Models with CPU-only\n\n```bash\nconda create -n opencompass python=3.10 pytorch torchvision torchaudio cpuonly -c pytorch -y\nconda activate opencompass\ngit clone https://github.com/open-compass/opencompass opencompass\ncd opencompass\npip install -e .\n# also please install requiresments packages via `pip install -r requirements/api.txt` for API models if needed.\n```\n\n### \ud83d\udcc2 Data Preparation\n\n```bash\n# Download dataset to data/ folder\nwget https://github.com/open-compass/opencompass/releases/download/0.2.2.rc1/OpenCompassData-core-20240207.zip\nunzip OpenCompassData-core-20240207.zip\n```\n\nSome third-party features, like Humaneval and Llama, may require additional steps to work properly, for detailed steps please refer to the [Installation Guide](https://opencompass.readthedocs.io/en/latest/get_started/installation.html).\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n## \ud83c\udfd7\ufe0f \ufe0fEvaluation\n\nAfter ensuring that OpenCompass is installed correctly according to the above steps and the datasets are prepared, you can evaluate the performance of the LLaMA-7b model on the MMLU and C-Eval datasets using the following command:\n\n```bash\npython run.py --models hf_llama_7b --datasets mmlu_ppl ceval_ppl\n```\n\nOpenCompass has predefined configurations for many models and datasets. You can list all available model and dataset configurations using the [tools](./docs/en/tools.md#list-configs).\n\n```bash\n# List all configurations\npython tools/list_configs.py\n# List all configurations related to llama and mmlu\npython tools/list_configs.py llama mmlu\n```\n\nYou can also evaluate other HuggingFace models via command line. Taking LLaMA-7b as an example:\n\n```bash\npython run.py --datasets ceval_ppl mmlu_ppl \\\n--hf-path huggyllama/llama-7b \\  # HuggingFace model path\n--model-kwargs device_map='auto' \\  # Arguments for model construction\n--tokenizer-kwargs padding_side='left' truncation='left' use_fast=False \\  # Arguments for tokenizer construction\n--max-out-len 100 \\  # Maximum number of tokens generated\n--max-seq-len 2048 \\  # Maximum sequence length the model can accept\n--batch-size 8 \\  # Batch size\n--no-batch-padding \\  # Don't enable batch padding, infer through for loop to avoid performance loss\n--num-gpus 1  # Number of minimum required GPUs\n```\n\n> **Note**<br />\n> To run the command above, you will need to remove the comments starting from `# ` first.\n\nThrough the command line or configuration files, OpenCompass also supports evaluating APIs or custom models, as well as more diversified evaluation strategies. Please read the [Quick Start](https://opencompass.readthedocs.io/en/latest/get_started/quick_start.html) to learn how to run an evaluation task.\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n## \ud83d\udcd6 Dataset Support\n\n<table align=\"center\">\n  <tbody>\n    <tr align=\"center\" valign=\"bottom\">\n      <td>\n        <b>Language</b>\n      </td>\n      <td>\n        <b>Knowledge</b>\n      </td>\n      <td>\n        <b>Reasoning</b>\n      </td>\n      <td>\n        <b>Examination</b>\n      </td>\n    </tr>\n    <tr valign=\"top\">\n      <td>\n<details open>\n<summary><b>Word Definition</b></summary>\n\n- WiC\n- SummEdits\n\n</details>\n\n<details open>\n<summary><b>Idiom Learning</b></summary>\n\n- CHID\n\n</details>\n\n<details open>\n<summary><b>Semantic Similarity</b></summary>\n\n- AFQMC\n- BUSTM\n\n</details>\n\n<details open>\n<summary><b>Coreference Resolution</b></summary>\n\n- CLUEWSC\n- WSC\n- WinoGrande\n\n</details>\n\n<details open>\n<summary><b>Translation</b></summary>\n\n- Flores\n- IWSLT2017\n\n</details>\n\n<details open>\n<summary><b>Multi-language Question Answering</b></summary>\n\n- TyDi-QA\n- XCOPA\n\n</details>\n\n<details open>\n<summary><b>Multi-language Summary</b></summary>\n\n- XLSum\n\n</details>\n      </td>\n      <td>\n<details open>\n<summary><b>Knowledge Question Answering</b></summary>\n\n- BoolQ\n- CommonSenseQA\n- NaturalQuestions\n- TriviaQA\n\n</details>\n      </td>\n      <td>\n<details open>\n<summary><b>Textual Entailment</b></summary>\n\n- CMNLI\n- OCNLI\n- OCNLI_FC\n- AX-b\n- AX-g\n- CB\n- RTE\n- ANLI\n\n</details>\n\n<details open>\n<summary><b>Commonsense Reasoning</b></summary>\n\n- StoryCloze\n- COPA\n- ReCoRD\n- HellaSwag\n- PIQA\n- SIQA\n\n</details>\n\n<details open>\n<summary><b>Mathematical Reasoning</b></summary>\n\n- MATH\n- GSM8K\n\n</details>\n\n<details open>\n<summary><b>Theorem Application</b></summary>\n\n- TheoremQA\n- StrategyQA\n- SciBench\n\n</details>\n\n<details open>\n<summary><b>Comprehensive Reasoning</b></summary>\n\n- BBH\n\n</details>\n      </td>\n      <td>\n<details open>\n<summary><b>Junior High, High School, University, Professional Examinations</b></summary>\n\n- C-Eval\n- AGIEval\n- MMLU\n- GAOKAO-Bench\n- CMMLU\n- ARC\n- Xiezhi\n\n</details>\n\n<details open>\n<summary><b>Medical Examinations</b></summary>\n\n- CMB\n\n</details>\n      </td>\n    </tr>\n</td>\n    </tr>\n  </tbody>\n  <tbody>\n    <tr align=\"center\" valign=\"bottom\">\n      <td>\n        <b>Understanding</b>\n      </td>\n      <td>\n        <b>Long Context</b>\n      </td>\n      <td>\n        <b>Safety</b>\n      </td>\n      <td>\n        <b>Code</b>\n      </td>\n    </tr>\n    <tr valign=\"top\">\n      <td>\n<details open>\n<summary><b>Reading Comprehension</b></summary>\n\n- C3\n- CMRC\n- DRCD\n- MultiRC\n- RACE\n- DROP\n- OpenBookQA\n- SQuAD2.0\n\n</details>\n\n<details open>\n<summary><b>Content Summary</b></summary>\n\n- CSL\n- LCSTS\n- XSum\n- SummScreen\n\n</details>\n\n<details open>\n<summary><b>Content Analysis</b></summary>\n\n- EPRSTMT\n- LAMBADA\n- TNEWS\n\n</details>\n      </td>\n      <td>\n<details open>\n<summary><b>Long Context Understanding</b></summary>\n\n- LEval\n- LongBench\n- GovReports\n- NarrativeQA\n- Qasper\n\n</details>\n      </td>\n      <td>\n<details open>\n<summary><b>Safety</b></summary>\n\n- CivilComments\n- CrowsPairs\n- CValues\n- JigsawMultilingual\n- TruthfulQA\n\n</details>\n<details open>\n<summary><b>Robustness</b></summary>\n\n- AdvGLUE\n\n</details>\n      </td>\n      <td>\n<details open>\n<summary><b>Code</b></summary>\n\n- HumanEval\n- HumanEvalX\n- MBPP\n- APPs\n- DS1000\n\n</details>\n      </td>\n    </tr>\n</td>\n    </tr>\n  </tbody>\n</table>\n\n## \ud83d\udcd6 Model Support\n\n<table align=\"center\">\n  <tbody>\n    <tr align=\"center\" valign=\"bottom\">\n      <td>\n        <b>Open-source Models</b>\n      </td>\n      <td>\n        <b>API Models</b>\n      </td>\n      <!-- <td>\n        <b>Custom Models</b>\n      </td> -->\n    </tr>\n    <tr valign=\"top\">\n      <td>\n\n- [InternLM](https://github.com/InternLM/InternLM)\n- [LLaMA](https://github.com/facebookresearch/llama)\n- [LLaMA3](https://github.com/meta-llama/llama3)\n- [Vicuna](https://github.com/lm-sys/FastChat)\n- [Alpaca](https://github.com/tatsu-lab/stanford_alpaca)\n- [Baichuan](https://github.com/baichuan-inc)\n- [WizardLM](https://github.com/nlpxucan/WizardLM)\n- [ChatGLM2](https://github.com/THUDM/ChatGLM2-6B)\n- [ChatGLM3](https://github.com/THUDM/ChatGLM3-6B)\n- [TigerBot](https://github.com/TigerResearch/TigerBot)\n- [Qwen](https://github.com/QwenLM/Qwen)\n- [BlueLM](https://github.com/vivo-ai-lab/BlueLM)\n- [Gemma](https://huggingface.co/google/gemma-7b)\n- ...\n\n</td>\n<td>\n\n- OpenAI\n- Gemini\n- Claude\n- ZhipuAI(ChatGLM)\n- Baichuan\n- ByteDance(YunQue)\n- Huawei(PanGu)\n- 360\n- Baidu(ERNIEBot)\n- MiniMax(ABAB-Chat)\n- SenseTime(nova)\n- Xunfei(Spark)\n- \u2026\u2026\n\n</td>\n\n</tr>\n  </tbody>\n</table>\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n## \ud83d\udd1c Roadmap\n\n- [x] Subjective Evaluation\n  - [ ] Release CompassAreana\n  - [x] Subjective evaluation.\n- [x] Long-context\n  - [x] Long-context evaluation with extensive datasets.\n  - [ ] Long-context leaderboard.\n- [x] Coding\n  - [ ] Coding evaluation leaderboard.\n  - [x] Non-python language evaluation service.\n- [x] Agent\n  - [ ] Support various agenet framework.\n  - [x] Evaluation of tool use of the LLMs.\n- [x] Robustness\n  - [x] Support various attack method\n\n## \ud83d\udc77\u200d\u2642\ufe0f Contributing\n\nWe appreciate all contributions to improving OpenCompass. Please refer to the [contributing guideline](https://opencompass.readthedocs.io/en/latest/notes/contribution_guide.html) for the best practice.\n\n<!-- Copy-paste in your Readme.md file -->\n\n<!-- Made with [OSS Insight](https://ossinsight.io/) -->\n\n<a href=\"https://github.com/open-compass/opencompass/graphs/contributors\" target=\"_blank\">\n  <table>\n    <tr>\n      <th colspan=\"2\">\n        <br><img src=\"https://contrib.rocks/image?repo=open-compass/opencompass\"><br><br>\n      </th>\n    </tr>\n  </table>\n</a>\n\n## \ud83e\udd1d Acknowledgements\n\nSome code in this project is cited and modified from [OpenICL](https://github.com/Shark-NLP/OpenICL).\n\nSome datasets and prompt implementations are modified from [chain-of-thought-hub](https://github.com/FranxYao/chain-of-thought-hub) and [instruct-eval](https://github.com/declare-lab/instruct-eval).\n\n## \ud83d\udd8a\ufe0f Citation\n\n```bibtex\n@misc{2023opencompass,\n    title={OpenCompass: A Universal Evaluation Platform for Foundation Models},\n    author={OpenCompass Contributors},\n    howpublished = {\\url{https://github.com/open-compass/opencompass}},\n    year={2023}\n}\n```\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n[github-contributors-link]: https://github.com/open-compass/opencompass/graphs/contributors\n[github-contributors-shield]: https://img.shields.io/github/contributors/open-compass/opencompass?color=c4f042&labelColor=black&style=flat-square\n[github-forks-link]: https://github.com/open-compass/opencompass/network/members\n[github-forks-shield]: https://img.shields.io/github/forks/open-compass/opencompass?color=8ae8ff&labelColor=black&style=flat-square\n[github-issues-link]: https://github.com/open-compass/opencompass/issues\n[github-issues-shield]: https://img.shields.io/github/issues/open-compass/opencompass?color=ff80eb&labelColor=black&style=flat-square\n[github-license-link]: https://github.com/open-compass/opencompass/blob/main/LICENSE\n[github-license-shield]: https://img.shields.io/github/license/open-compass/opencompass?color=white&labelColor=black&style=flat-square\n[github-release-link]: https://github.com/open-compass/opencompass/releases\n[github-release-shield]: https://img.shields.io/github/v/release/open-compass/opencompass?color=369eff&labelColor=black&logo=github&style=flat-square\n[github-releasedate-link]: https://github.com/open-compass/opencompass/releases\n[github-releasedate-shield]: https://img.shields.io/github/release-date/open-compass/opencompass?labelColor=black&style=flat-square\n[github-stars-link]: https://github.com/open-compass/opencompass/stargazers\n[github-stars-shield]: https://img.shields.io/github/stars/open-compass/opencompass?color=ffcb47&labelColor=black&style=flat-square\n[github-trending-shield]: https://trendshift.io/api/badge/repositories/6630\n[github-trending-url]: https://trendshift.io/repositories/6630\n\n\n",
    "bugtrack_url": null,
    "license": "Apache License 2.0",
    "summary": "A comprehensive toolkit for large model evaluation",
    "version": "0.2.4",
    "project_urls": {
        "Homepage": "https://github.com/open-compass/opencompass"
    },
    "split_keywords": [
        "ai",
        " nlp",
        " in-context learning",
        " large language model",
        " evaluation",
        " benchmark",
        " llm"
    ],
    "urls": [
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "671a3305f4147d81ab24869f7c3fe78c4c69fa75e848be7d91c51c7a66d75909",
                "md5": "a71da222d8b2fd32cee6dfb0028683f5",
                "sha256": "23052f05f890abc5afa20e4d088381ac7f43306a12db3d620df85b845b043c5f"
            },
            "downloads": -1,
            "filename": "opencompass-0.2.4-py3-none-any.whl",
            "has_sig": false,
            "md5_digest": "a71da222d8b2fd32cee6dfb0028683f5",
            "packagetype": "bdist_wheel",
            "python_version": "py3",
            "requires_python": ">=3.8.0",
            "size": 559877,
            "upload_time": "2024-04-23T09:22:10",
            "upload_time_iso_8601": "2024-04-23T09:22:10.677417Z",
            "url": "https://files.pythonhosted.org/packages/67/1a/3305f4147d81ab24869f7c3fe78c4c69fa75e848be7d91c51c7a66d75909/opencompass-0.2.4-py3-none-any.whl",
            "yanked": false,
            "yanked_reason": null
        },
        {
            "comment_text": "",
            "digests": {
                "blake2b_256": "4cc0047d8e98e8128ea54571506d55fab017c234798ba2ade115e41ad0104bbb",
                "md5": "b7b642b227eaf7638ac2c790ea3faf5f",
                "sha256": "09088816e0b6653cfc64b0427492037adda3a52a8032fc2c9ee99fef88a07faa"
            },
            "downloads": -1,
            "filename": "opencompass-0.2.4.tar.gz",
            "has_sig": false,
            "md5_digest": "b7b642b227eaf7638ac2c790ea3faf5f",
            "packagetype": "sdist",
            "python_version": "source",
            "requires_python": ">=3.8.0",
            "size": 331480,
            "upload_time": "2024-04-23T09:22:12",
            "upload_time_iso_8601": "2024-04-23T09:22:12.630861Z",
            "url": "https://files.pythonhosted.org/packages/4c/c0/047d8e98e8128ea54571506d55fab017c234798ba2ade115e41ad0104bbb/opencompass-0.2.4.tar.gz",
            "yanked": false,
            "yanked_reason": null
        }
    ],
    "upload_time": "2024-04-23 09:22:12",
    "github": true,
    "gitlab": false,
    "bitbucket": false,
    "codeberg": false,
    "github_user": "open-compass",
    "github_project": "opencompass",
    "travis_ci": false,
    "coveralls": false,
    "github_actions": true,
    "requirements": [],
    "lcname": "opencompass"
}
        
Elapsed time: 0.23489s