Update README.md
Browse files
README.md
CHANGED
@@ -84,16 +84,19 @@ and is comparable with Mistral-7B-Instruct-v0.1 on MMLU and MT-Bench in English.
|
|
84 |
|
85 |
\* Taiwan-LLM models responds to multi-turn questions (English) in Traditional Chinese.
|
86 |
|
87 |
-
| Category ACC of TMMLU+
|
88 |
|-----------------------------------------------------|--------------|----------------|------------|------------|
|
|
|
|
|
|
|
89 |
| Yi-6B-Chat | 26.28 | 33.48 | 29.48 | 27.62 |
|
90 |
| Breeze-7B-Instruct-v0.1 | 37.45 | 48.35 | 40.26 | 40.44 |
|
91 |
-
|
|
92 |
| Qwen-7B-Chat | 32.89 | 44.26 | 38.21 | 37.83 |
|
93 |
| Taiwan-LLM-13B-v2.0-chat | 29.68 | 37.13 | 30.31 | 30.55 |
|
94 |
-
|
|
95 |
-
|
96 |
-
|
97 |
|
98 |
## Examples
|
99 |
|
|
|
84 |
|
85 |
\* Taiwan-LLM models responds to multi-turn questions (English) in Traditional Chinese.
|
86 |
|
87 |
+
| Category ACC of TMMLU+ (0 shot) | STEM | Social Science | Humanities | Other |
|
88 |
|-----------------------------------------------------|--------------|----------------|------------|------------|
|
89 |
+
| gpt-3.5-turbo-1106 | | | | |
|
90 |
+
| Yi-34B-Chat | 46.36 | 65.02 | 52.84 | 52.21 |
|
91 |
+
| Qwen-14B-Chat | 46.51 | 58.20 | 51.12 | 49.38 |
|
92 |
| Yi-6B-Chat | 26.28 | 33.48 | 29.48 | 27.62 |
|
93 |
| Breeze-7B-Instruct-v0.1 | 37.45 | 48.35 | 40.26 | 40.44 |
|
94 |
+
| Breeze-7B-Instruct-64k-v0.1 | 37.45 | 48.35 | 40.26 | 40.44 |
|
95 |
| Qwen-7B-Chat | 32.89 | 44.26 | 38.21 | 37.83 |
|
96 |
| Taiwan-LLM-13B-v2.0-chat | 29.68 | 37.13 | 30.31 | 30.55 |
|
97 |
+
| Taiwan-LLM-7B-v2.1-chat | 26.53 | 29.47 | 26.11 | 26.90 |
|
98 |
+
|
99 |
+
|
100 |
|
101 |
## Examples
|
102 |
|