File size: 2,541 Bytes
01b95c0
b741f55
aae5291
 
 
 
 
01b95c0
aae5291
a9fddb3
 
 
 
aae5291
d35efa7
aae5291
d35efa7
 
 
aae5291
d35efa7
 
 
 
aae5291
d35efa7
 
aae5291
d35efa7
 
 
 
 
aae5291
 
 
 
 
 
 
 
 
 
 
5b6411f
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
---
license: apache-2.0
widget:
  - text: "你好! 你叫什么名字!"
    output:
      text: "你好,我的名字叫聚言,很高兴见到你。"
pipeline_tag: text-generation
---

[OrionStarAI/OrionStar-Yi-34B-Chat-Llama](https://huggingface.co/OrionStarAI/OrionStar-Yi-34B-Chat-Llama/tree/main) 

*This model is identical to [OrionStarAI/OrionStar-Yi-34B](https://huggingface.co/OrionStarAI/OrionStar-Yi-34B/tree/main) 
with the only difference being that the tensors have been renamed to follow the LLaMA format for automatic evaluation on the HF leaderboard.*

# Model Introduction

- OrionStar-Yi-34B-Chat from OrionStarAI is based on the open-source Yi-34B model, fine-tuned on a high-quality corpus
  of over 15 million sentences. OrionStar-Yi-34B-Chat aims to provide an excellent interactive experience for users in
  the large model community.

- The Yi series models, open-sourced by the 01-ai team, have shown impressive performance on various benchmarks in
  Chinese, English, and general domains. OrionStar-Yi-34B-Chat further explores the potential of Yi-34B. Through
  extensive fine-tuning on a large and high-quality corpus, OrionStar-Yi-34B-Chat performs exceptionally well on
  evaluation data. We strive to make it an outstanding open-source alternative in the ChatGPT domain!

- Our fine-tuned model is completely open for academic research, but please adhere to the [agreement](#license) and
  the [Yi License](https://github.com/01-ai/Yi/blob/main/MODEL_LICENSE_AGREEMENT.txt).

- Model Evaluation Results

We use [opencompass](https://opencompass.org.cn) to perform 5-shot on the following general domain datasets Testing.
The evaluation results of other models are taken
from [opencompass leaderboard](https://opencompass.org.cn/leaderboard-llm).

|                           | C-Eval    | MMLU   | CMMLU     |
|---------------------------|-----------|--------|-----------|
| **GPT-4**                 | 69.9      | **83** | 71        |
| **ChatGPT**               | 52.5      | 69.1   | 53.9      | 			
| **Claude-1**              | 52        | 65.7   | -         |
| **TigerBot-70B-Chat-V2**  | 57.7      | 65.9   | 59.9      |
| **WeMix-LLaMA2-70B**      | 55.2      | 71.3   | 56        |  			
| **LLaMA-2-70B-Chat**      | 44.3      | 63.8   | 43.3      |
| **Qwen-14B-Chat**         | 71.7      | 66.4   | 70        |
| **Baichuan2-13B-Chat**    | 56.7      | 57     | 58.4      |      	
| **OrionStar-Yi-34B-Chat** | **77.71** | 78.32  | **73.52** |  


**Discord Link:** https://discord.gg/zumjDWgdAs