Update README.md
Browse files
README.md
CHANGED
@@ -94,7 +94,7 @@ Model release and download links are provided in the table below:
|
|
94 |
| Baichuan 2-13B | 68.9 | 67.2 | 70.8 | 78.1 | 74.1 | 66.3 |
|
95 |
| QWEN-14B | 93.0 | 90.3 | **80.2** | 79.8 | 71.4 | 66.3 |
|
96 |
| InternLM-20B | 86.4 | 83.3 | 78.1 | **80.3** | 71.8 | 68.3 |
|
97 |
-
| **Orion-14B-Base** | **93.
|
98 |
|
99 |
### 3.1.3. LLM evaluation results of OpenCompass testsets
|
100 |
| Model | Average | Examination | Language | Knowledge | Understanding | Reasoning |
|
@@ -104,7 +104,7 @@ Model release and download links are provided in the table below:
|
|
104 |
| Baichuan 2-13B | 49.4 | 51.8 | 47.5 | 48.9 | 58.1 | 44.2 |
|
105 |
| QWEN-14B | 62.4 | 71.3 | 52.67 | 56.1 | 68.8 | 60.1 |
|
106 |
| InternLM-20B | 59.4 | 62.5 | 55.0 | **60.1** | 67.3 | 54.9 |
|
107 |
-
|**Orion-14B-Base**| **64.
|
108 |
|
109 |
### 3.1.4. Comparison of LLM performances on Japanese testsets
|
110 |
| Model |**Average**| JCQA | JNLI | MARC | JSQD | JQK | XLS | XWN | MGSM |
|
|
|
94 |
| Baichuan 2-13B | 68.9 | 67.2 | 70.8 | 78.1 | 74.1 | 66.3 |
|
95 |
| QWEN-14B | 93.0 | 90.3 | **80.2** | 79.8 | 71.4 | 66.3 |
|
96 |
| InternLM-20B | 86.4 | 83.3 | 78.1 | **80.3** | 71.8 | 68.3 |
|
97 |
+
| **Orion-14B-Base** | **93.2** | **91.3** | 78.5 | 79.5 | **78.8** | **70.2** |
|
98 |
|
99 |
### 3.1.3. LLM evaluation results of OpenCompass testsets
|
100 |
| Model | Average | Examination | Language | Knowledge | Understanding | Reasoning |
|
|
|
104 |
| Baichuan 2-13B | 49.4 | 51.8 | 47.5 | 48.9 | 58.1 | 44.2 |
|
105 |
| QWEN-14B | 62.4 | 71.3 | 52.67 | 56.1 | 68.8 | 60.1 |
|
106 |
| InternLM-20B | 59.4 | 62.5 | 55.0 | **60.1** | 67.3 | 54.9 |
|
107 |
+
|**Orion-14B-Base**| **64.3** | **71.4** | **55.0** | 60.0 | **71.9** | **61.6** |
|
108 |
|
109 |
### 3.1.4. Comparison of LLM performances on Japanese testsets
|
110 |
| Model |**Average**| JCQA | JNLI | MARC | JSQD | JQK | XLS | XWN | MGSM |
|