<div align="center">
<img src="docs/en/_static/image/logo.svg" width="500px"/>
<br />
<br />
[![][github-release-shield]][github-release-link]
[![][github-releasedate-shield]][github-releasedate-link]
[![][github-contributors-shield]][github-contributors-link]<br>
[![][github-forks-shield]][github-forks-link]
[![][github-stars-shield]][github-stars-link]
[![][github-issues-shield]][github-issues-link]
[![][github-license-shield]][github-license-link]
<!-- [![PyPI](https://badge.fury.io/py/opencompass.svg)](https://pypi.org/project/opencompass/) -->
[🌐Website](https://opencompass.org.cn/) |
[📖CompassHub](https://hub.opencompass.org.cn/home) |
[📊CompassRank](https://rank.opencompass.org.cn/home) |
[📘Documentation](https://opencompass.readthedocs.io/en/latest/) |
[🛠️Installation](https://opencompass.readthedocs.io/en/latest/get_started/installation.html) |
[🤔Reporting Issues](https://github.com/open-compass/opencompass/issues/new/choose)
English | [简体中文](README_zh-CN.md)
[![][github-trending-shield]][github-trending-url]
</div>
<p align="center">
👋 join us on <a href="https://discord.gg/KKwfEbFj7U" target="_blank">Discord</a> and <a href="https://r.vansin.top/?r=opencompass" target="_blank">WeChat</a>
</p>
> \[!IMPORTANT\]
>
> **Star Us**, You will receive all release notifications from GitHub without any delay ~ ⭐️
<details>
<summary><kbd>Star History</kbd></summary>
<picture>
<source media="(prefers-color-scheme: dark)" srcset="https://api.star-history.com/svg?repos=open-compass%2Fopencompass&theme=dark&type=Date">
<img width="100%" src="https://api.star-history.com/svg?repos=open-compass%2Fopencompass&type=Date">
</picture>
</details>
## 🧭 Welcome
to **OpenCompass**!
Just like a compass guides us on our journey, OpenCompass will guide you through the complex landscape of evaluating large language models. With its powerful algorithms and intuitive interface, OpenCompass makes it easy to assess the quality and effectiveness of your NLP models.
🚩🚩🚩 Explore opportunities at OpenCompass! We're currently **hiring full-time researchers/engineers and interns**. If you're passionate about LLM and OpenCompass, don't hesitate to reach out to us via [email](mailto:zhangsongyang@pjlab.org.cn). We'd love to hear from you!
🔥🔥🔥 We are delighted to announce that **the OpenCompass has been recommended by the Meta AI**, click [Get Started](https://ai.meta.com/llama/get-started/#validation) of Llama for more information.
> **Attention**<br />
> We launch the OpenCompass Collaboration project, welcome to support diverse evaluation benchmarks into OpenCompass!
> Clike [Issue](https://github.com/open-compass/opencompass/issues/248) for more information.
> Let's work together to build a more powerful OpenCompass toolkit!
## 🚀 What's New <a><img width="35" height="20" src="https://user-images.githubusercontent.com/12782558/212848161-5e783dd6-11e8-4fe0-bbba-39ffb77730be.png"></a>
- **\[2024.09.19\]** We now support [Qwen2.5](https://huggingface.co/Qwen)(0.5B to 72B) with multiple backend(huggingface/vllm/lmdeploy). Feel free to give them a try! 🔥🔥🔥
- **\[2024.09.17\]** We now support OpenAI o1(`o1-mini-2024-09-12` and `o1-preview-2024-09-12`). Feel free to give them a try! 🔥🔥🔥
- **\[2024.09.05\]** We now support answer extraction through model post-processing to provide a more accurate representation of the model's capabilities. As part of this update, we have integrated [XFinder](https://github.com/IAAR-Shanghai/xFinder) as our first post-processing model. For more detailed information, please refer to the [documentation](opencompass/utils/postprocessors/xfinder/README.md), and give it a try! 🔥🔥🔥
- **\[2024.08.20\]** OpenCompass now supports the [SciCode](https://github.com/scicode-bench/SciCode): A Research Coding Benchmark Curated by Scientists. 🔥🔥🔥
- **\[2024.08.16\]** OpenCompass now supports the brand new long-context language model evaluation benchmark — [RULER](https://arxiv.org/pdf/2404.06654). RULER provides an evaluation of long-context including retrieval, multi-hop tracing, aggregation, and question answering through flexible configurations. Check out the [RULER](configs/datasets/ruler/README.md) evaluation config now! 🔥🔥🔥
- **\[2024.08.09\]** We have released the example data and configuration for the CompassBench-202408, welcome to [CompassBench](https://opencompass.readthedocs.io/zh-cn/latest/advanced_guides/compassbench_intro.html) for more details. 🔥🔥🔥
- **\[2024.08.01\]** We supported the [Gemma2](https://huggingface.co/collections/google/gemma-2-release-667d6600fd5220e7b967f315) models. Welcome to try! 🔥🔥🔥
- **\[2024.07.23\]** We supported the [ModelScope](www.modelscope.cn) datasets, you can load them on demand without downloading all the data to your local disk. Welcome to try! 🔥🔥🔥
- **\[2024.07.17\]** We are excited to announce the release of NeedleBench's [technical report](http://arxiv.org/abs/2407.11963). We invite you to visit our [support documentation](https://opencompass.readthedocs.io/en/latest/advanced_guides/needleinahaystack_eval.html) for detailed evaluation guidelines. 🔥🔥🔥
- **\[2024.07.04\]** OpenCompass now supports InternLM2.5, which has **outstanding reasoning capability**, **1M Context window and** and **stronger tool use**, you can try the models in [OpenCompass Config](https://github.com/open-compass/opencompass/tree/main/configs/models/hf_internlm) and [InternLM](https://github.com/InternLM/InternLM) .🔥🔥🔥.
- **\[2024.06.20\]** OpenCompass now supports one-click switching between inference acceleration backends, enhancing the efficiency of the evaluation process. In addition to the default HuggingFace inference backend, it now also supports popular backends [LMDeploy](https://github.com/InternLM/lmdeploy) and [vLLM](https://github.com/vllm-project/vllm). This feature is available via a simple command-line switch and through deployment APIs. For detailed usage, see the [documentation](docs/en/advanced_guides/accelerator_intro.md).🔥🔥🔥.
> [More](docs/en/notes/news.md)
## 📊 Leaderboard
We provide [OpenCompass Leaderboard](https://rank.opencompass.org.cn/home) for the community to rank all public models and API models. If you would like to join the evaluation, please provide the model repository URL or a standard API interface to the email address `opencompass@pjlab.org.cn`.
<p align="right"><a href="#top">🔝Back to top</a></p>
## 🛠️ Installation
Below are the steps for quick installation and datasets preparation.
### 💻 Environment Setup
We highly recommend using conda to manage your python environment.
- #### Create your virtual environment
```bash
conda create --name opencompass python=3.10 -y
conda activate opencompass
```
- #### Install OpenCompass via pip
```bash
pip install -U opencompass
## Full installation (with support for more datasets)
# pip install "opencompass[full]"
## Environment with model acceleration frameworks
## Manage different acceleration frameworks using virtual environments
## since they usually have dependency conflicts with each other.
# pip install "opencompass[lmdeploy]"
# pip install "opencompass[vllm]"
## API evaluation (i.e. Openai, Qwen)
# pip install "opencompass[api]"
```
- #### Install OpenCompass from source
If you want to use opencompass's latest features, or develop new features, you can also build it from source
```bash
git clone https://github.com/open-compass/opencompass opencompass
cd opencompass
pip install -e .
# pip install -e ".[full]"
# pip install -e ".[vllm]"
```
### 📂 Data Preparation
You can choose one for the following method to prepare datasets.
#### Offline Preparation
You can download and extract the datasets with the following commands:
```bash
# Download dataset to data/ folder
wget https://github.com/open-compass/opencompass/releases/download/0.2.2.rc1/OpenCompassData-core-20240207.zip
unzip OpenCompassData-core-20240207.zip
```
#### Automatic Download from OpenCompass
We have supported download datasets automatic from the OpenCompass storage server. You can run the evaluation with extra `--dry-run` to download these datasets.
Currently, the supported datasets are listed in [here](https://github.com/open-compass/opencompass/blob/main/opencompass/utils/datasets_info.py#L259). More datasets will be uploaded recently.
#### (Optional) Automatic Download with ModelScope
Also you can use the [ModelScope](www.modelscope.cn) to load the datasets on demand.
Installation:
```bash
pip install modelscope[framework]
export DATASET_SOURCE=ModelScope
```
Then submit the evaluation task without downloading all the data to your local disk. Available datasets include:
```bash
humaneval, triviaqa, commonsenseqa, tydiqa, strategyqa, cmmlu, lambada, piqa, ceval, math, LCSTS, Xsum, winogrande, openbookqa, AGIEval, gsm8k, nq, race, siqa, mbpp, mmlu, hellaswag, ARC, BBH, xstory_cloze, summedits, GAOKAO-BENCH, OCNLI, cmnli
```
Some third-party features, like Humaneval and Llama, may require additional steps to work properly, for detailed steps please refer to the [Installation Guide](https://opencompass.readthedocs.io/en/latest/get_started/installation.html).
<p align="right"><a href="#top">🔝Back to top</a></p>
## 🏗️ ️Evaluation
After ensuring that OpenCompass is installed correctly according to the above steps and the datasets are prepared. Now you can start your first evaluation using OpenCompass!
- Your first evaluation with OpenCompass!
OpenCompass support setting your configs via CLI or a python script. For simple evaluation settings we recommend using CLI, for more complex evaluation, it is suggested using the script way. You can find more example scripts under the configs folder.
```bash
# CLI
opencompass --models hf_internlm2_5_1_8b_chat --datasets demo_gsm8k_chat_gen
# Python scripts
opencompass ./configs/eval_chat_demo.py
```
You can find more script examples under [configs](./configs) folder.
- API evaluation
OpenCompass, by its design, does not really discriminate between open-source models and API models. You can evaluate both model types in the same way or even in one settings.
```bash
export OPENAI_API_KEY="YOUR_OPEN_API_KEY"
# CLI
opencompass --models gpt_4o_2024_05_13 --datasets demo_gsm8k_chat_gen
# Python scripts
opencompass ./configs/eval_api_demo.py
# You can use o1_mini_2024_09_12/o1_preview_2024_09_12 for o1 models, we set max_completion_tokens=8192 as default.
```
- Accelerated Evaluation
Additionally, if you want to use an inference backend other than HuggingFace for accelerated evaluation, such as LMDeploy or vLLM, you can do so with the command below. Please ensure that you have installed the necessary packages for the chosen backend and that your model supports accelerated inference with it. For more information, see the documentation on inference acceleration backends [here](docs/en/advanced_guides/accelerator_intro.md). Below is an example using LMDeploy:
```bash
# CLI
opencompass --models hf_internlm2_5_1_8b_chat --datasets demo_gsm8k_chat_gen -a lmdeploy
# Python scripts
opencompass ./configs/eval_lmdeploy_demo.py
```
- Supported Models
OpenCompass has predefined configurations for many models and datasets. You can list all available model and dataset configurations using the [tools](./docs/en/tools.md#list-configs).
```bash
# List all configurations
python tools/list_configs.py
# List all configurations related to llama and mmlu
python tools/list_configs.py llama mmlu
```
If the model is not on the list but supported by Huggingface AutoModel class, you can also evaluate it with OpenCompass. You are welcome to contribute to the maintenance of the OpenCompass supported model and dataset lists.
```bash
opencompass --datasets demo_gsm8k_chat_gen --hf-type chat --hf-path internlm/internlm2_5-1_8b-chat
```
If you want to use multiple GPUs to evaluate the model in data parallel, you can use `--max-num-worker`.
```bash
CUDA_VISIBLE_DEVICES=0,1 opencompass --datasets demo_gsm8k_chat_gen --hf-type chat --hf-path internlm/internlm2_5-1_8b-chat --max-num-worker 2
```
> \[!TIP\]
>
> `--hf-num-gpus` is used for model parallel(huggingface format), `--max-num-worker` is used for data parallel.
> \[!TIP\]
>
> configuration with `_ppl` is designed for base model typically.
> configuration with `_gen` can be used for both base model and chat model.
Through the command line or configuration files, OpenCompass also supports evaluating APIs or custom models, as well as more diversified evaluation strategies. Please read the [Quick Start](https://opencompass.readthedocs.io/en/latest/get_started/quick_start.html) to learn how to run an evaluation task.
<p align="right"><a href="#top">🔝Back to top</a></p>
## 📣 OpenCompass 2.0
We are thrilled to introduce OpenCompass 2.0, an advanced suite featuring three key components: [CompassKit](https://github.com/open-compass), [CompassHub](https://hub.opencompass.org.cn/home), and [CompassRank](https://rank.opencompass.org.cn/home).
![oc20](https://github.com/tonysy/opencompass/assets/7881589/90dbe1c0-c323-470a-991e-2b37ab5350b2)
**CompassRank** has been significantly enhanced into the leaderboards that now incorporates both open-source benchmarks and proprietary benchmarks. This upgrade allows for a more comprehensive evaluation of models across the industry.
**CompassHub** presents a pioneering benchmark browser interface, designed to simplify and expedite the exploration and utilization of an extensive array of benchmarks for researchers and practitioners alike. To enhance the visibility of your own benchmark within the community, we warmly invite you to contribute it to CompassHub. You may initiate the submission process by clicking [here](https://hub.opencompass.org.cn/dataset-submit).
**CompassKit** is a powerful collection of evaluation toolkits specifically tailored for Large Language Models and Large Vision-language Models. It provides an extensive set of tools to assess and measure the performance of these complex models effectively. Welcome to try our toolkits for in your research and products.
## ✨ Introduction
![image](https://github.com/open-compass/opencompass/assets/22607038/f45fe125-4aed-4f8c-8fe8-df4efb41a8ea)
OpenCompass is a one-stop platform for large model evaluation, aiming to provide a fair, open, and reproducible benchmark for large model evaluation. Its main features include:
- **Comprehensive support for models and datasets**: Pre-support for 20+ HuggingFace and API models, a model evaluation scheme of 70+ datasets with about 400,000 questions, comprehensively evaluating the capabilities of the models in five dimensions.
- **Efficient distributed evaluation**: One line command to implement task division and distributed evaluation, completing the full evaluation of billion-scale models in just a few hours.
- **Diversified evaluation paradigms**: Support for zero-shot, few-shot, and chain-of-thought evaluations, combined with standard or dialogue-type prompt templates, to easily stimulate the maximum performance of various models.
- **Modular design with high extensibility**: Want to add new models or datasets, customize an advanced task division strategy, or even support a new cluster management system? Everything about OpenCompass can be easily expanded!
- **Experiment management and reporting mechanism**: Use config files to fully record each experiment, and support real-time reporting of results.
## 📖 Dataset Support
<table align="center">
<tbody>
<tr align="center" valign="bottom">
<td>
<b>Language</b>
</td>
<td>
<b>Knowledge</b>
</td>
<td>
<b>Reasoning</b>
</td>
<td>
<b>Examination</b>
</td>
</tr>
<tr valign="top">
<td>
<details open>
<summary><b>Word Definition</b></summary>
- WiC
- SummEdits
</details>
<details open>
<summary><b>Idiom Learning</b></summary>
- CHID
</details>
<details open>
<summary><b>Semantic Similarity</b></summary>
- AFQMC
- BUSTM
</details>
<details open>
<summary><b>Coreference Resolution</b></summary>
- CLUEWSC
- WSC
- WinoGrande
</details>
<details open>
<summary><b>Translation</b></summary>
- Flores
- IWSLT2017
</details>
<details open>
<summary><b>Multi-language Question Answering</b></summary>
- TyDi-QA
- XCOPA
</details>
<details open>
<summary><b>Multi-language Summary</b></summary>
- XLSum
</details>
</td>
<td>
<details open>
<summary><b>Knowledge Question Answering</b></summary>
- BoolQ
- CommonSenseQA
- NaturalQuestions
- TriviaQA
</details>
</td>
<td>
<details open>
<summary><b>Textual Entailment</b></summary>
- CMNLI
- OCNLI
- OCNLI_FC
- AX-b
- AX-g
- CB
- RTE
- ANLI
</details>
<details open>
<summary><b>Commonsense Reasoning</b></summary>
- StoryCloze
- COPA
- ReCoRD
- HellaSwag
- PIQA
- SIQA
</details>
<details open>
<summary><b>Mathematical Reasoning</b></summary>
- MATH
- GSM8K
</details>
<details open>
<summary><b>Theorem Application</b></summary>
- TheoremQA
- StrategyQA
- SciBench
</details>
<details open>
<summary><b>Comprehensive Reasoning</b></summary>
- BBH
</details>
</td>
<td>
<details open>
<summary><b>Junior High, High School, University, Professional Examinations</b></summary>
- C-Eval
- AGIEval
- MMLU
- GAOKAO-Bench
- CMMLU
- ARC
- Xiezhi
</details>
<details open>
<summary><b>Medical Examinations</b></summary>
- CMB
</details>
</td>
</tr>
</td>
</tr>
</tbody>
<tbody>
<tr align="center" valign="bottom">
<td>
<b>Understanding</b>
</td>
<td>
<b>Long Context</b>
</td>
<td>
<b>Safety</b>
</td>
<td>
<b>Code</b>
</td>
</tr>
<tr valign="top">
<td>
<details open>
<summary><b>Reading Comprehension</b></summary>
- C3
- CMRC
- DRCD
- MultiRC
- RACE
- DROP
- OpenBookQA
- SQuAD2.0
</details>
<details open>
<summary><b>Content Summary</b></summary>
- CSL
- LCSTS
- XSum
- SummScreen
</details>
<details open>
<summary><b>Content Analysis</b></summary>
- EPRSTMT
- LAMBADA
- TNEWS
</details>
</td>
<td>
<details open>
<summary><b>Long Context Understanding</b></summary>
- LEval
- LongBench
- GovReports
- NarrativeQA
- Qasper
</details>
</td>
<td>
<details open>
<summary><b>Safety</b></summary>
- CivilComments
- CrowsPairs
- CValues
- JigsawMultilingual
- TruthfulQA
</details>
<details open>
<summary><b>Robustness</b></summary>
- AdvGLUE
</details>
</td>
<td>
<details open>
<summary><b>Code</b></summary>
- HumanEval
- HumanEvalX
- MBPP
- APPs
- DS1000
</details>
</td>
</tr>
</td>
</tr>
</tbody>
</table>
## 📖 Model Support
<table align="center">
<tbody>
<tr align="center" valign="bottom">
<td>
<b>Open-source Models</b>
</td>
<td>
<b>API Models</b>
</td>
<!-- <td>
<b>Custom Models</b>
</td> -->
</tr>
<tr valign="top">
<td>
- [Alpaca](https://github.com/tatsu-lab/stanford_alpaca)
- [Baichuan](https://github.com/baichuan-inc)
- [BlueLM](https://github.com/vivo-ai-lab/BlueLM)
- [ChatGLM2](https://github.com/THUDM/ChatGLM2-6B)
- [ChatGLM3](https://github.com/THUDM/ChatGLM3-6B)
- [Gemma](https://huggingface.co/google/gemma-7b)
- [InternLM](https://github.com/InternLM/InternLM)
- [LLaMA](https://github.com/facebookresearch/llama)
- [LLaMA3](https://github.com/meta-llama/llama3)
- [Qwen](https://github.com/QwenLM/Qwen)
- [TigerBot](https://github.com/TigerResearch/TigerBot)
- [Vicuna](https://github.com/lm-sys/FastChat)
- [WizardLM](https://github.com/nlpxucan/WizardLM)
- [Yi](https://github.com/01-ai/Yi)
- ……
</td>
<td>
- OpenAI
- Gemini
- Claude
- ZhipuAI(ChatGLM)
- Baichuan
- ByteDance(YunQue)
- Huawei(PanGu)
- 360
- Baidu(ERNIEBot)
- MiniMax(ABAB-Chat)
- SenseTime(nova)
- Xunfei(Spark)
- ……
</td>
</tr>
</tbody>
</table>
<p align="right"><a href="#top">🔝Back to top</a></p>
## 🔜 Roadmap
- [x] Subjective Evaluation
- [x] Release CompassAreana.
- [x] Subjective evaluation.
- [x] Long-context
- [x] Long-context evaluation with extensive datasets.
- [ ] Long-context leaderboard.
- [x] Coding
- [ ] Coding evaluation leaderboard.
- [x] Non-python language evaluation service.
- [x] Agent
- [ ] Support various agent frameworks.
- [x] Evaluation of tool use of the LLMs.
- [x] Robustness
- [x] Support various attack methods.
## 👷♂️ Contributing
We appreciate all contributions to improving OpenCompass. Please refer to the [contributing guideline](https://opencompass.readthedocs.io/en/latest/notes/contribution_guide.html) for the best practice.
<!-- Copy-paste in your Readme.md file -->
<!-- Made with [OSS Insight](https://ossinsight.io/) -->
<a href="https://github.com/open-compass/opencompass/graphs/contributors" target="_blank">
<table>
<tr>
<th colspan="2">
<br><img src="https://contrib.rocks/image?repo=open-compass/opencompass"><br><br>
</th>
</tr>
</table>
</a>
## 🤝 Acknowledgements
Some code in this project is cited and modified from [OpenICL](https://github.com/Shark-NLP/OpenICL).
Some datasets and prompt implementations are modified from [chain-of-thought-hub](https://github.com/FranxYao/chain-of-thought-hub) and [instruct-eval](https://github.com/declare-lab/instruct-eval).
## 🖊️ Citation
```bibtex
@misc{2023opencompass,
title={OpenCompass: A Universal Evaluation Platform for Foundation Models},
author={OpenCompass Contributors},
howpublished = {\url{https://github.com/open-compass/opencompass}},
year={2023}
}
```
<p align="right"><a href="#top">🔝Back to top</a></p>
[github-contributors-link]: https://github.com/open-compass/opencompass/graphs/contributors
[github-contributors-shield]: https://img.shields.io/github/contributors/open-compass/opencompass?color=c4f042&labelColor=black&style=flat-square
[github-forks-link]: https://github.com/open-compass/opencompass/network/members
[github-forks-shield]: https://img.shields.io/github/forks/open-compass/opencompass?color=8ae8ff&labelColor=black&style=flat-square
[github-issues-link]: https://github.com/open-compass/opencompass/issues
[github-issues-shield]: https://img.shields.io/github/issues/open-compass/opencompass?color=ff80eb&labelColor=black&style=flat-square
[github-license-link]: https://github.com/open-compass/opencompass/blob/main/LICENSE
[github-license-shield]: https://img.shields.io/github/license/open-compass/opencompass?color=white&labelColor=black&style=flat-square
[github-release-link]: https://github.com/open-compass/opencompass/releases
[github-release-shield]: https://img.shields.io/github/v/release/open-compass/opencompass?color=369eff&labelColor=black&logo=github&style=flat-square
[github-releasedate-link]: https://github.com/open-compass/opencompass/releases
[github-releasedate-shield]: https://img.shields.io/github/release-date/open-compass/opencompass?labelColor=black&style=flat-square
[github-stars-link]: https://github.com/open-compass/opencompass/stargazers
[github-stars-shield]: https://img.shields.io/github/stars/open-compass/opencompass?color=ffcb47&labelColor=black&style=flat-square
[github-trending-shield]: https://trendshift.io/api/badge/repositories/6630
[github-trending-url]: https://trendshift.io/repositories/6630
Raw data
{
"_id": null,
"home_page": "https://github.com/open-compass/opencompass",
"name": "opencompass",
"maintainer": "OpenCompass Authors",
"docs_url": null,
"requires_python": ">=3.8.0",
"maintainer_email": null,
"keywords": "AI, NLP, in-context learning, large language model, evaluation, benchmark, llm",
"author": "OpenCompass Contributors",
"author_email": null,
"download_url": "https://files.pythonhosted.org/packages/46/59/2c49396bf5148a0a0b964546494aa0724591e8aec99755b9a1c52c97371c/opencompass-0.3.3.tar.gz",
"platform": null,
"description": "<div align=\"center\">\n <img src=\"docs/en/_static/image/logo.svg\" width=\"500px\"/>\n <br />\n <br />\n\n[![][github-release-shield]][github-release-link]\n[![][github-releasedate-shield]][github-releasedate-link]\n[![][github-contributors-shield]][github-contributors-link]<br>\n[![][github-forks-shield]][github-forks-link]\n[![][github-stars-shield]][github-stars-link]\n[![][github-issues-shield]][github-issues-link]\n[![][github-license-shield]][github-license-link]\n\n<!-- [![PyPI](https://badge.fury.io/py/opencompass.svg)](https://pypi.org/project/opencompass/) -->\n\n[\ud83c\udf10Website](https://opencompass.org.cn/) |\n[\ud83d\udcd6CompassHub](https://hub.opencompass.org.cn/home) |\n[\ud83d\udccaCompassRank](https://rank.opencompass.org.cn/home) |\n[\ud83d\udcd8Documentation](https://opencompass.readthedocs.io/en/latest/) |\n[\ud83d\udee0\ufe0fInstallation](https://opencompass.readthedocs.io/en/latest/get_started/installation.html) |\n[\ud83e\udd14Reporting Issues](https://github.com/open-compass/opencompass/issues/new/choose)\n\nEnglish | [\u7b80\u4f53\u4e2d\u6587](README_zh-CN.md)\n\n[![][github-trending-shield]][github-trending-url]\n\n</div>\n\n<p align=\"center\">\n \ud83d\udc4b join us on <a href=\"https://discord.gg/KKwfEbFj7U\" target=\"_blank\">Discord</a> and <a href=\"https://r.vansin.top/?r=opencompass\" target=\"_blank\">WeChat</a>\n</p>\n\n> \\[!IMPORTANT\\]\n>\n> **Star Us**, You will receive all release notifications from GitHub without any delay ~ \u2b50\ufe0f\n\n<details>\n <summary><kbd>Star History</kbd></summary>\n <picture>\n <source media=\"(prefers-color-scheme: dark)\" srcset=\"https://api.star-history.com/svg?repos=open-compass%2Fopencompass&theme=dark&type=Date\">\n <img width=\"100%\" src=\"https://api.star-history.com/svg?repos=open-compass%2Fopencompass&type=Date\">\n </picture>\n</details>\n\n## \ud83e\udded\tWelcome\n\nto **OpenCompass**!\n\nJust like a compass guides us on our journey, OpenCompass will guide you through the complex landscape of evaluating large language models. With its powerful algorithms and intuitive interface, OpenCompass makes it easy to assess the quality and effectiveness of your NLP models.\n\n\ud83d\udea9\ud83d\udea9\ud83d\udea9 Explore opportunities at OpenCompass! We're currently **hiring full-time researchers/engineers and interns**. If you're passionate about LLM and OpenCompass, don't hesitate to reach out to us via [email](mailto:zhangsongyang@pjlab.org.cn). We'd love to hear from you!\n\n\ud83d\udd25\ud83d\udd25\ud83d\udd25 We are delighted to announce that **the OpenCompass has been recommended by the Meta AI**, click [Get Started](https://ai.meta.com/llama/get-started/#validation) of Llama for more information.\n\n> **Attention**<br />\n> We launch the OpenCompass Collaboration project, welcome to support diverse evaluation benchmarks into OpenCompass!\n> Clike [Issue](https://github.com/open-compass/opencompass/issues/248) for more information.\n> Let's work together to build a more powerful OpenCompass toolkit!\n\n## \ud83d\ude80 What's New <a><img width=\"35\" height=\"20\" src=\"https://user-images.githubusercontent.com/12782558/212848161-5e783dd6-11e8-4fe0-bbba-39ffb77730be.png\"></a>\n\n- **\\[2024.09.19\\]** We now support [Qwen2.5](https://huggingface.co/Qwen)(0.5B to 72B) with multiple backend(huggingface/vllm/lmdeploy). Feel free to give them a try! \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.09.17\\]** We now support OpenAI o1(`o1-mini-2024-09-12` and `o1-preview-2024-09-12`). Feel free to give them a try! \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.09.05\\]** We now support answer extraction through model post-processing to provide a more accurate representation of the model's capabilities. As part of this update, we have integrated [XFinder](https://github.com/IAAR-Shanghai/xFinder) as our first post-processing model. For more detailed information, please refer to the [documentation](opencompass/utils/postprocessors/xfinder/README.md), and give it a try! \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.08.20\\]** OpenCompass now supports the [SciCode](https://github.com/scicode-bench/SciCode): A Research Coding Benchmark Curated by Scientists. \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.08.16\\]** OpenCompass now supports the brand new long-context language model evaluation benchmark \u2014 [RULER](https://arxiv.org/pdf/2404.06654). RULER provides an evaluation of long-context including retrieval, multi-hop tracing, aggregation, and question answering through flexible configurations. Check out the [RULER](configs/datasets/ruler/README.md) evaluation config now! \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.08.09\\]** We have released the example data and configuration for the CompassBench-202408, welcome to [CompassBench](https://opencompass.readthedocs.io/zh-cn/latest/advanced_guides/compassbench_intro.html) for more details. \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.08.01\\]** We supported the [Gemma2](https://huggingface.co/collections/google/gemma-2-release-667d6600fd5220e7b967f315) models. Welcome to try! \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.07.23\\]** We supported the [ModelScope](www.modelscope.cn) datasets, you can load them on demand without downloading all the data to your local disk. Welcome to try! \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.07.17\\]** We are excited to announce the release of NeedleBench's [technical report](http://arxiv.org/abs/2407.11963). We invite you to visit our [support documentation](https://opencompass.readthedocs.io/en/latest/advanced_guides/needleinahaystack_eval.html) for detailed evaluation guidelines. \ud83d\udd25\ud83d\udd25\ud83d\udd25\n- **\\[2024.07.04\\]** OpenCompass now supports InternLM2.5, which has **outstanding reasoning capability**, **1M Context window and** and **stronger tool use**, you can try the models in [OpenCompass Config](https://github.com/open-compass/opencompass/tree/main/configs/models/hf_internlm) and [InternLM](https://github.com/InternLM/InternLM) .\ud83d\udd25\ud83d\udd25\ud83d\udd25.\n- **\\[2024.06.20\\]** OpenCompass now supports one-click switching between inference acceleration backends, enhancing the efficiency of the evaluation process. In addition to the default HuggingFace inference backend, it now also supports popular backends [LMDeploy](https://github.com/InternLM/lmdeploy) and [vLLM](https://github.com/vllm-project/vllm). This feature is available via a simple command-line switch and through deployment APIs. For detailed usage, see the [documentation](docs/en/advanced_guides/accelerator_intro.md).\ud83d\udd25\ud83d\udd25\ud83d\udd25.\n\n> [More](docs/en/notes/news.md)\n\n## \ud83d\udcca Leaderboard\n\nWe provide [OpenCompass Leaderboard](https://rank.opencompass.org.cn/home) for the community to rank all public models and API models. If you would like to join the evaluation, please provide the model repository URL or a standard API interface to the email address `opencompass@pjlab.org.cn`.\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n## \ud83d\udee0\ufe0f Installation\n\nBelow are the steps for quick installation and datasets preparation.\n\n### \ud83d\udcbb Environment Setup\n\nWe highly recommend using conda to manage your python environment.\n\n- #### Create your virtual environment\n\n ```bash\n conda create --name opencompass python=3.10 -y\n conda activate opencompass\n ```\n\n- #### Install OpenCompass via pip\n\n ```bash\n pip install -U opencompass\n\n ## Full installation (with support for more datasets)\n # pip install \"opencompass[full]\"\n\n ## Environment with model acceleration frameworks\n ## Manage different acceleration frameworks using virtual environments\n ## since they usually have dependency conflicts with each other.\n # pip install \"opencompass[lmdeploy]\"\n # pip install \"opencompass[vllm]\"\n\n ## API evaluation (i.e. Openai, Qwen)\n # pip install \"opencompass[api]\"\n ```\n\n- #### Install OpenCompass from source\n\n If you want to use opencompass's latest features, or develop new features, you can also build it from source\n\n ```bash\n git clone https://github.com/open-compass/opencompass opencompass\n cd opencompass\n pip install -e .\n # pip install -e \".[full]\"\n # pip install -e \".[vllm]\"\n ```\n\n### \ud83d\udcc2 Data Preparation\n\nYou can choose one for the following method to prepare datasets.\n\n#### Offline Preparation\n\nYou can download and extract the datasets with the following commands:\n\n```bash\n# Download dataset to data/ folder\nwget https://github.com/open-compass/opencompass/releases/download/0.2.2.rc1/OpenCompassData-core-20240207.zip\nunzip OpenCompassData-core-20240207.zip\n```\n\n#### Automatic Download from OpenCompass\n\nWe have supported download datasets automatic from the OpenCompass storage server. You can run the evaluation with extra `--dry-run` to download these datasets.\nCurrently, the supported datasets are listed in [here](https://github.com/open-compass/opencompass/blob/main/opencompass/utils/datasets_info.py#L259). More datasets will be uploaded recently.\n\n#### (Optional) Automatic Download with ModelScope\n\nAlso you can use the [ModelScope](www.modelscope.cn) to load the datasets on demand.\n\nInstallation:\n\n```bash\npip install modelscope[framework]\nexport DATASET_SOURCE=ModelScope\n```\n\nThen submit the evaluation task without downloading all the data to your local disk. Available datasets include:\n\n```bash\nhumaneval, triviaqa, commonsenseqa, tydiqa, strategyqa, cmmlu, lambada, piqa, ceval, math, LCSTS, Xsum, winogrande, openbookqa, AGIEval, gsm8k, nq, race, siqa, mbpp, mmlu, hellaswag, ARC, BBH, xstory_cloze, summedits, GAOKAO-BENCH, OCNLI, cmnli\n```\n\nSome third-party features, like Humaneval and Llama, may require additional steps to work properly, for detailed steps please refer to the [Installation Guide](https://opencompass.readthedocs.io/en/latest/get_started/installation.html).\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n## \ud83c\udfd7\ufe0f \ufe0fEvaluation\n\nAfter ensuring that OpenCompass is installed correctly according to the above steps and the datasets are prepared. Now you can start your first evaluation using OpenCompass!\n\n- Your first evaluation with OpenCompass!\n\n OpenCompass support setting your configs via CLI or a python script. For simple evaluation settings we recommend using CLI, for more complex evaluation, it is suggested using the script way. You can find more example scripts under the configs folder.\n\n ```bash\n # CLI\n opencompass --models hf_internlm2_5_1_8b_chat --datasets demo_gsm8k_chat_gen\n\n # Python scripts\n opencompass ./configs/eval_chat_demo.py\n ```\n\n You can find more script examples under [configs](./configs) folder.\n\n- API evaluation\n\n OpenCompass, by its design, does not really discriminate between open-source models and API models. You can evaluate both model types in the same way or even in one settings.\n\n ```bash\n export OPENAI_API_KEY=\"YOUR_OPEN_API_KEY\"\n # CLI\n opencompass --models gpt_4o_2024_05_13 --datasets demo_gsm8k_chat_gen\n\n # Python scripts\n opencompass ./configs/eval_api_demo.py\n\n # You can use o1_mini_2024_09_12/o1_preview_2024_09_12 for o1 models, we set max_completion_tokens=8192 as default.\n ```\n\n- Accelerated Evaluation\n\n Additionally, if you want to use an inference backend other than HuggingFace for accelerated evaluation, such as LMDeploy or vLLM, you can do so with the command below. Please ensure that you have installed the necessary packages for the chosen backend and that your model supports accelerated inference with it. For more information, see the documentation on inference acceleration backends [here](docs/en/advanced_guides/accelerator_intro.md). Below is an example using LMDeploy:\n\n ```bash\n # CLI\n opencompass --models hf_internlm2_5_1_8b_chat --datasets demo_gsm8k_chat_gen -a lmdeploy\n\n # Python scripts\n opencompass ./configs/eval_lmdeploy_demo.py\n ```\n\n- Supported Models\n\n OpenCompass has predefined configurations for many models and datasets. You can list all available model and dataset configurations using the [tools](./docs/en/tools.md#list-configs).\n\n ```bash\n # List all configurations\n python tools/list_configs.py\n # List all configurations related to llama and mmlu\n python tools/list_configs.py llama mmlu\n ```\n\n If the model is not on the list but supported by Huggingface AutoModel class, you can also evaluate it with OpenCompass. You are welcome to contribute to the maintenance of the OpenCompass supported model and dataset lists.\n\n ```bash\n opencompass --datasets demo_gsm8k_chat_gen --hf-type chat --hf-path internlm/internlm2_5-1_8b-chat\n ```\n\n If you want to use multiple GPUs to evaluate the model in data parallel, you can use `--max-num-worker`.\n\n ```bash\n CUDA_VISIBLE_DEVICES=0,1 opencompass --datasets demo_gsm8k_chat_gen --hf-type chat --hf-path internlm/internlm2_5-1_8b-chat --max-num-worker 2\n ```\n\n> \\[!TIP\\]\n>\n> `--hf-num-gpus` is used for model parallel(huggingface format), `--max-num-worker` is used for data parallel.\n\n> \\[!TIP\\]\n>\n> configuration with `_ppl` is designed for base model typically.\n> configuration with `_gen` can be used for both base model and chat model.\n\nThrough the command line or configuration files, OpenCompass also supports evaluating APIs or custom models, as well as more diversified evaluation strategies. Please read the [Quick Start](https://opencompass.readthedocs.io/en/latest/get_started/quick_start.html) to learn how to run an evaluation task.\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n## \ud83d\udce3 OpenCompass 2.0\n\nWe are thrilled to introduce OpenCompass 2.0, an advanced suite featuring three key components: [CompassKit](https://github.com/open-compass), [CompassHub](https://hub.opencompass.org.cn/home), and [CompassRank](https://rank.opencompass.org.cn/home).\n![oc20](https://github.com/tonysy/opencompass/assets/7881589/90dbe1c0-c323-470a-991e-2b37ab5350b2)\n\n**CompassRank** has been significantly enhanced into the leaderboards that now incorporates both open-source benchmarks and proprietary benchmarks. This upgrade allows for a more comprehensive evaluation of models across the industry.\n\n**CompassHub** presents a pioneering benchmark browser interface, designed to simplify and expedite the exploration and utilization of an extensive array of benchmarks for researchers and practitioners alike. To enhance the visibility of your own benchmark within the community, we warmly invite you to contribute it to CompassHub. You may initiate the submission process by clicking [here](https://hub.opencompass.org.cn/dataset-submit).\n\n**CompassKit** is a powerful collection of evaluation toolkits specifically tailored for Large Language Models and Large Vision-language Models. It provides an extensive set of tools to assess and measure the performance of these complex models effectively. Welcome to try our toolkits for in your research and products.\n\n## \u2728 Introduction\n\n![image](https://github.com/open-compass/opencompass/assets/22607038/f45fe125-4aed-4f8c-8fe8-df4efb41a8ea)\n\nOpenCompass is a one-stop platform for large model evaluation, aiming to provide a fair, open, and reproducible benchmark for large model evaluation. Its main features include:\n\n- **Comprehensive support for models and datasets**: Pre-support for 20+ HuggingFace and API models, a model evaluation scheme of 70+ datasets with about 400,000 questions, comprehensively evaluating the capabilities of the models in five dimensions.\n\n- **Efficient distributed evaluation**: One line command to implement task division and distributed evaluation, completing the full evaluation of billion-scale models in just a few hours.\n\n- **Diversified evaluation paradigms**: Support for zero-shot, few-shot, and chain-of-thought evaluations, combined with standard or dialogue-type prompt templates, to easily stimulate the maximum performance of various models.\n\n- **Modular design with high extensibility**: Want to add new models or datasets, customize an advanced task division strategy, or even support a new cluster management system? Everything about OpenCompass can be easily expanded!\n\n- **Experiment management and reporting mechanism**: Use config files to fully record each experiment, and support real-time reporting of results.\n\n## \ud83d\udcd6 Dataset Support\n\n<table align=\"center\">\n <tbody>\n <tr align=\"center\" valign=\"bottom\">\n <td>\n <b>Language</b>\n </td>\n <td>\n <b>Knowledge</b>\n </td>\n <td>\n <b>Reasoning</b>\n </td>\n <td>\n <b>Examination</b>\n </td>\n </tr>\n <tr valign=\"top\">\n <td>\n<details open>\n<summary><b>Word Definition</b></summary>\n\n- WiC\n- SummEdits\n\n</details>\n\n<details open>\n<summary><b>Idiom Learning</b></summary>\n\n- CHID\n\n</details>\n\n<details open>\n<summary><b>Semantic Similarity</b></summary>\n\n- AFQMC\n- BUSTM\n\n</details>\n\n<details open>\n<summary><b>Coreference Resolution</b></summary>\n\n- CLUEWSC\n- WSC\n- WinoGrande\n\n</details>\n\n<details open>\n<summary><b>Translation</b></summary>\n\n- Flores\n- IWSLT2017\n\n</details>\n\n<details open>\n<summary><b>Multi-language Question Answering</b></summary>\n\n- TyDi-QA\n- XCOPA\n\n</details>\n\n<details open>\n<summary><b>Multi-language Summary</b></summary>\n\n- XLSum\n\n</details>\n </td>\n <td>\n<details open>\n<summary><b>Knowledge Question Answering</b></summary>\n\n- BoolQ\n- CommonSenseQA\n- NaturalQuestions\n- TriviaQA\n\n</details>\n </td>\n <td>\n<details open>\n<summary><b>Textual Entailment</b></summary>\n\n- CMNLI\n- OCNLI\n- OCNLI_FC\n- AX-b\n- AX-g\n- CB\n- RTE\n- ANLI\n\n</details>\n\n<details open>\n<summary><b>Commonsense Reasoning</b></summary>\n\n- StoryCloze\n- COPA\n- ReCoRD\n- HellaSwag\n- PIQA\n- SIQA\n\n</details>\n\n<details open>\n<summary><b>Mathematical Reasoning</b></summary>\n\n- MATH\n- GSM8K\n\n</details>\n\n<details open>\n<summary><b>Theorem Application</b></summary>\n\n- TheoremQA\n- StrategyQA\n- SciBench\n\n</details>\n\n<details open>\n<summary><b>Comprehensive Reasoning</b></summary>\n\n- BBH\n\n</details>\n </td>\n <td>\n<details open>\n<summary><b>Junior High, High School, University, Professional Examinations</b></summary>\n\n- C-Eval\n- AGIEval\n- MMLU\n- GAOKAO-Bench\n- CMMLU\n- ARC\n- Xiezhi\n\n</details>\n\n<details open>\n<summary><b>Medical Examinations</b></summary>\n\n- CMB\n\n</details>\n </td>\n </tr>\n</td>\n </tr>\n </tbody>\n <tbody>\n <tr align=\"center\" valign=\"bottom\">\n <td>\n <b>Understanding</b>\n </td>\n <td>\n <b>Long Context</b>\n </td>\n <td>\n <b>Safety</b>\n </td>\n <td>\n <b>Code</b>\n </td>\n </tr>\n <tr valign=\"top\">\n <td>\n<details open>\n<summary><b>Reading Comprehension</b></summary>\n\n- C3\n- CMRC\n- DRCD\n- MultiRC\n- RACE\n- DROP\n- OpenBookQA\n- SQuAD2.0\n\n</details>\n\n<details open>\n<summary><b>Content Summary</b></summary>\n\n- CSL\n- LCSTS\n- XSum\n- SummScreen\n\n</details>\n\n<details open>\n<summary><b>Content Analysis</b></summary>\n\n- EPRSTMT\n- LAMBADA\n- TNEWS\n\n</details>\n </td>\n <td>\n<details open>\n<summary><b>Long Context Understanding</b></summary>\n\n- LEval\n- LongBench\n- GovReports\n- NarrativeQA\n- Qasper\n\n</details>\n </td>\n <td>\n<details open>\n<summary><b>Safety</b></summary>\n\n- CivilComments\n- CrowsPairs\n- CValues\n- JigsawMultilingual\n- TruthfulQA\n\n</details>\n<details open>\n<summary><b>Robustness</b></summary>\n\n- AdvGLUE\n\n</details>\n </td>\n <td>\n<details open>\n<summary><b>Code</b></summary>\n\n- HumanEval\n- HumanEvalX\n- MBPP\n- APPs\n- DS1000\n\n</details>\n </td>\n </tr>\n</td>\n </tr>\n </tbody>\n</table>\n\n## \ud83d\udcd6 Model Support\n\n<table align=\"center\">\n <tbody>\n <tr align=\"center\" valign=\"bottom\">\n <td>\n <b>Open-source Models</b>\n </td>\n <td>\n <b>API Models</b>\n </td>\n <!-- <td>\n <b>Custom Models</b>\n </td> -->\n </tr>\n <tr valign=\"top\">\n <td>\n\n- [Alpaca](https://github.com/tatsu-lab/stanford_alpaca)\n- [Baichuan](https://github.com/baichuan-inc)\n- [BlueLM](https://github.com/vivo-ai-lab/BlueLM)\n- [ChatGLM2](https://github.com/THUDM/ChatGLM2-6B)\n- [ChatGLM3](https://github.com/THUDM/ChatGLM3-6B)\n- [Gemma](https://huggingface.co/google/gemma-7b)\n- [InternLM](https://github.com/InternLM/InternLM)\n- [LLaMA](https://github.com/facebookresearch/llama)\n- [LLaMA3](https://github.com/meta-llama/llama3)\n- [Qwen](https://github.com/QwenLM/Qwen)\n- [TigerBot](https://github.com/TigerResearch/TigerBot)\n- [Vicuna](https://github.com/lm-sys/FastChat)\n- [WizardLM](https://github.com/nlpxucan/WizardLM)\n- [Yi](https://github.com/01-ai/Yi)\n- \u2026\u2026\n\n</td>\n<td>\n\n- OpenAI\n- Gemini\n- Claude\n- ZhipuAI(ChatGLM)\n- Baichuan\n- ByteDance(YunQue)\n- Huawei(PanGu)\n- 360\n- Baidu(ERNIEBot)\n- MiniMax(ABAB-Chat)\n- SenseTime(nova)\n- Xunfei(Spark)\n- \u2026\u2026\n\n</td>\n\n</tr>\n </tbody>\n</table>\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n## \ud83d\udd1c Roadmap\n\n- [x] Subjective Evaluation\n - [x] Release CompassAreana.\n - [x] Subjective evaluation.\n- [x] Long-context\n - [x] Long-context evaluation with extensive datasets.\n - [ ] Long-context leaderboard.\n- [x] Coding\n - [ ] Coding evaluation leaderboard.\n - [x] Non-python language evaluation service.\n- [x] Agent\n - [ ] Support various agent frameworks.\n - [x] Evaluation of tool use of the LLMs.\n- [x] Robustness\n - [x] Support various attack methods.\n\n## \ud83d\udc77\u200d\u2642\ufe0f Contributing\n\nWe appreciate all contributions to improving OpenCompass. Please refer to the [contributing guideline](https://opencompass.readthedocs.io/en/latest/notes/contribution_guide.html) for the best practice.\n\n<!-- Copy-paste in your Readme.md file -->\n\n<!-- Made with [OSS Insight](https://ossinsight.io/) -->\n\n<a href=\"https://github.com/open-compass/opencompass/graphs/contributors\" target=\"_blank\">\n <table>\n <tr>\n <th colspan=\"2\">\n <br><img src=\"https://contrib.rocks/image?repo=open-compass/opencompass\"><br><br>\n </th>\n </tr>\n </table>\n</a>\n\n## \ud83e\udd1d Acknowledgements\n\nSome code in this project is cited and modified from [OpenICL](https://github.com/Shark-NLP/OpenICL).\n\nSome datasets and prompt implementations are modified from [chain-of-thought-hub](https://github.com/FranxYao/chain-of-thought-hub) and [instruct-eval](https://github.com/declare-lab/instruct-eval).\n\n## \ud83d\udd8a\ufe0f Citation\n\n```bibtex\n@misc{2023opencompass,\n title={OpenCompass: A Universal Evaluation Platform for Foundation Models},\n author={OpenCompass Contributors},\n howpublished = {\\url{https://github.com/open-compass/opencompass}},\n year={2023}\n}\n```\n\n<p align=\"right\"><a href=\"#top\">\ud83d\udd1dBack to top</a></p>\n\n[github-contributors-link]: https://github.com/open-compass/opencompass/graphs/contributors\n[github-contributors-shield]: https://img.shields.io/github/contributors/open-compass/opencompass?color=c4f042&labelColor=black&style=flat-square\n[github-forks-link]: https://github.com/open-compass/opencompass/network/members\n[github-forks-shield]: https://img.shields.io/github/forks/open-compass/opencompass?color=8ae8ff&labelColor=black&style=flat-square\n[github-issues-link]: https://github.com/open-compass/opencompass/issues\n[github-issues-shield]: https://img.shields.io/github/issues/open-compass/opencompass?color=ff80eb&labelColor=black&style=flat-square\n[github-license-link]: https://github.com/open-compass/opencompass/blob/main/LICENSE\n[github-license-shield]: https://img.shields.io/github/license/open-compass/opencompass?color=white&labelColor=black&style=flat-square\n[github-release-link]: https://github.com/open-compass/opencompass/releases\n[github-release-shield]: https://img.shields.io/github/v/release/open-compass/opencompass?color=369eff&labelColor=black&logo=github&style=flat-square\n[github-releasedate-link]: https://github.com/open-compass/opencompass/releases\n[github-releasedate-shield]: https://img.shields.io/github/release-date/open-compass/opencompass?labelColor=black&style=flat-square\n[github-stars-link]: https://github.com/open-compass/opencompass/stargazers\n[github-stars-shield]: https://img.shields.io/github/stars/open-compass/opencompass?color=ffcb47&labelColor=black&style=flat-square\n[github-trending-shield]: https://trendshift.io/api/badge/repositories/6630\n[github-trending-url]: https://trendshift.io/repositories/6630\n\n\n",
"bugtrack_url": null,
"license": "Apache License 2.0",
"summary": "A comprehensive toolkit for large model evaluation",
"version": "0.3.3",
"project_urls": {
"Homepage": "https://github.com/open-compass/opencompass"
},
"split_keywords": [
"ai",
" nlp",
" in-context learning",
" large language model",
" evaluation",
" benchmark",
" llm"
],
"urls": [
{
"comment_text": "",
"digests": {
"blake2b_256": "e6b784c5330d6fabbce9ca28d4569b802264455ec9650796626d3da7a7d8ffb7",
"md5": "3bf43a3ab215ba01ed532250ba008839",
"sha256": "efb1d47b1169c9f74f969941c9de996db002fab8daebf691b598c72d5a31972a"
},
"downloads": -1,
"filename": "opencompass-0.3.3-py3-none-any.whl",
"has_sig": false,
"md5_digest": "3bf43a3ab215ba01ed532250ba008839",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": ">=3.8.0",
"size": 2175712,
"upload_time": "2024-09-30T08:58:38",
"upload_time_iso_8601": "2024-09-30T08:58:38.205707Z",
"url": "https://files.pythonhosted.org/packages/e6/b7/84c5330d6fabbce9ca28d4569b802264455ec9650796626d3da7a7d8ffb7/opencompass-0.3.3-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": "",
"digests": {
"blake2b_256": "46592c49396bf5148a0a0b964546494aa0724591e8aec99755b9a1c52c97371c",
"md5": "86f5f6f48bf6964bafffcab8d1479474",
"sha256": "56278122e6e98dc06364262a5fc1f96401235236463b92e00767970784894e69"
},
"downloads": -1,
"filename": "opencompass-0.3.3.tar.gz",
"has_sig": false,
"md5_digest": "86f5f6f48bf6964bafffcab8d1479474",
"packagetype": "sdist",
"python_version": "source",
"requires_python": ">=3.8.0",
"size": 967107,
"upload_time": "2024-09-30T08:58:40",
"upload_time_iso_8601": "2024-09-30T08:58:40.762855Z",
"url": "https://files.pythonhosted.org/packages/46/59/2c49396bf5148a0a0b964546494aa0724591e8aec99755b9a1c52c97371c/opencompass-0.3.3.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2024-09-30 08:58:40",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "open-compass",
"github_project": "opencompass",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"requirements": [],
"lcname": "opencompass"
}