facat commited on
Commit
3e6cdea
β€’
1 Parent(s): 2db079a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +99 -1
README.md CHANGED
@@ -31,4 +31,102 @@ widget:
31
  pipeline_tag: text-generation
32
  ---
33
 
34
- Powered by SUSTech x IDEA-CCNL, based on `01-ai/Yi-34B`.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31
  pipeline_tag: text-generation
32
  ---
33
 
34
+ .
35
+
36
+ ## Introduction
37
+
38
+ **SUS-CHhat** is powered by SUSTech x IDEA-CCNL, based on `01-ai/Yi-34B`
39
+
40
+ ## News
41
+
42
+ <details open>
43
+ <summary>🎯 <b>2023/11/23</b>: The chat models are open to public.</summary>
44
+
45
+ This release contains two chat models based on previous released base models, two 8-bits models quantized by GPTQ, two 4-bits models quantized by AWQ.
46
+
47
+ - `Yi-34B-Chat`
48
+ - `Yi-34B-Chat-4bits`
49
+ - `Yi-34B-Chat-8bits`
50
+ - `Yi-6B-Chat`
51
+ - `Yi-6B-Chat-4bits`
52
+ - `Yi-6B-Chat-8bits`
53
+
54
+ You can try some of them interactively at:
55
+
56
+ - [HuggingFace](https://huggingface.co/spaces/01-ai/Yi-34B-Chat)
57
+ - [Replicate](https://replicate.com/01-ai)
58
+ </details>
59
+
60
+ <details open>
61
+ <summary>πŸ”” <b>2023/11/23</b>: The Yi Series Models Community License Agreement is updated to v2.1.</summary>
62
+ </details>
63
+
64
+ <details>
65
+ <summary>πŸ”₯ <b>2023/11/08</b>: Invited test of Yi-34B chat model.</summary>
66
+
67
+ Application form:
68
+
69
+ - [English](https://cn.mikecrm.com/l91ODJf)
70
+ - [Chinese](https://cn.mikecrm.com/gnEZjiQ)
71
+
72
+ </details>
73
+
74
+ <details>
75
+ <summary>🎯 <b>2023/11/05</b>: The base model of <code>Yi-6B-200K</code> and <code>Yi-34B-200K</code>.</summary>
76
+
77
+ This release contains two base models with the same parameter sizes of previous
78
+ release, except that the context window is extended to 200K.
79
+
80
+ </details>
81
+
82
+ <details>
83
+ <summary>🎯 <b>2023/11/02</b>: The base model of <code>Yi-6B</code> and <code>Yi-34B</code>.</summary>
84
+
85
+ The first public release contains two bilingual (English/Chinese) base models
86
+ with the parameter sizes of 6B and 34B. Both of them are trained with 4K
87
+ sequence length and can be extended to 32K during inference time.
88
+
89
+ </details>
90
+
91
+ ## Model Performance
92
+
93
+ ### Base Model Performance
94
+
95
+ | Model | MMLU | CMMLU | C-Eval | GAOKAO | BBH | Common-sense Reasoning | Reading Comprehension | Math & Code |
96
+ | :------------ | :------: | :------: | :------: | :------: | :------: | :--------------------: | :-------------------: | :---------: |
97
+ | | 5-shot | 5-shot | 5-shot | 0-shot | 3-shot@1 | - | - | - |
98
+ | LLaMA2-34B | 62.6 | - | - | - | 44.1 | 69.9 | 68.0 | 26.0 |
99
+ | LLaMA2-70B | 68.9 | 53.3 | - | 49.8 | 51.2 | 71.9 | 69.4 | 36.8 |
100
+ | Baichuan2-13B | 59.2 | 62.0 | 58.1 | 54.3 | 48.8 | 64.3 | 62.4 | 23.0 |
101
+ | Qwen-14B | 66.3 | 71.0 | 72.1 | 62.5 | 53.4 | 73.3 | 72.5 | **39.8** |
102
+ | Skywork-13B | 62.1 | 61.8 | 60.6 | 68.1 | 41.7 | 72.4 | 61.4 | 24.9 |
103
+ | InternLM-20B | 62.1 | 59.0 | 58.8 | 45.5 | 52.5 | 78.3 | - | 30.4 |
104
+ | Aquila-34B | 67.8 | 71.4 | 63.1 | - | - | - | - | - |
105
+ | Falcon-180B | 70.4 | 58.0 | 57.8 | 59.0 | 54.0 | 77.3 | 68.8 | 34.0 |
106
+ | Yi-6B | 63.2 | 75.5 | 72.0 | 72.2 | 42.8 | 72.3 | 68.7 | 19.8 |
107
+ | Yi-6B-200K | 64.0 | 75.3 | 73.5 | 73.9 | 42.0 | 72.0 | 69.1 | 19.0 |
108
+ | **Yi-34B** | **76.3** | **83.7** | 81.4 | 82.8 | **54.3** | **80.1** | 76.4 | 37.1 |
109
+ | Yi-34B-200K | 76.1 | 83.6 | **81.9** | **83.4** | 52.7 | 79.7 | **76.6** | 36.3 |
110
+
111
+ While benchmarking open-source models, we have observed a disparity between the
112
+ results generated by our pipeline and those reported in public sources (e.g.
113
+ OpenCompass). Upon conducting a more in-depth investigation of this difference,
114
+ we have discovered that various models may employ different prompts,
115
+ post-processing strategies, and sampling techniques, potentially resulting in
116
+ significant variations in the outcomes. Our prompt and post-processing strategy
117
+ remains consistent with the original benchmark, and greedy decoding is employed
118
+ during evaluation without any post-processing for the generated content. For
119
+ scores that were not reported by the original authors (including scores reported
120
+ with different settings), we try to get results with our pipeline.
121
+
122
+ To evaluate the model's capability extensively, we adopted the methodology
123
+ outlined in Llama2. Specifically, we included PIQA, SIQA, HellaSwag, WinoGrande,
124
+ ARC, OBQA, and CSQA to assess common sense reasoning. SquAD, QuAC, and BoolQ
125
+ were incorporated to evaluate reading comprehension. CSQA was exclusively tested
126
+ using a 7-shot setup, while all other tests were conducted with a 0-shot
127
+ configuration. Additionally, we introduced GSM8K (8-shot@1), MATH (4-shot@1),
128
+ HumanEval (0-shot@1), and MBPP (3-shot@1) under the category "Math & Code". Due
129
+ to technical constraints, we did not test Falcon-180 on QuAC and OBQA; the score
130
+ is derived by averaging the scores on the remaining tasks. Since the scores for
131
+ these two tasks are generally lower than the average, we believe that
132
+ Falcon-180B's performance was not underestimated.