File size: 2,204 Bytes
6d98f28
e0e8cbd
 
 
5f9068f
 
 
e0e8cbd
 
 
 
310d540
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
CAMEL-13B-Combined-Data is a chat large language model obtained by finetuning LLaMA-13B model on a total of 229K conversations collected through our [CAMEL](https://arxiv.org/abs/2303.17760) framework, 100K English public conversations from ShareGPT that can be found [here](https://github.com/lm-sys/FastChat/issues/90#issuecomment-1493250773), and 52K instructions from Alpaca dataset that can be found [here](https://github.com/tatsu-lab/stanford_alpaca/blob/761dc5bfbdeeffa89b8bff5d038781a4055f796a/alpaca_data.json). We evaluate our model offline using EleutherAI's language model evaluation harness used by Huggingface's Open LLM Benchmark. CAMEL<sup>*</sup>-13B scores an average of 58.9.

| Model       | size | ARC-C  (25 shots, acc_norm) | HellaSwag  (10 shots, acc_norm) | MMLU  (5 shots, acc_norm) | TruthfulQA  (0 shot, mc2) | Average | Delta |
|-------------|:----:|:---------------------------:|:-------------------------------:|:-------------------------:|:-------------------------:|:-------:|-------|
| LLaMA       |  13B |             56.3            |               80.9              |            46.7           |            39.9           |   56.0  |   -   |
| Vicuna      |  13B |             52.8            |               80.1              |            50.5           |            51.8           |   58.8  |  2.8  |
| CAMEL<sup>*</sup>  |  13B |             56.1            |               79.9              |            50.5           |            49.0           |   58.9  |  2.9  |

---
license: cc-by-nc-4.0
---

# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_camel-ai__CAMEL-13B-Combined-Data)

| Metric                | Value                     |
|-----------------------|---------------------------|
| Avg.                  | 46.07   |
| ARC (25-shot)         | 55.63          |
| HellaSwag (10-shot)   | 79.25    |
| MMLU (5-shot)         | 49.74         |
| TruthfulQA (0-shot)   | 47.42   |
| Winogrande (5-shot)   | 75.45   |
| GSM8K (5-shot)        | 7.13        |
| DROP (3-shot)         | 7.86         |