renillhuang commited on
Commit
a3c2313
β€’
1 Parent(s): 14fb749

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -6
README.md CHANGED
@@ -40,15 +40,29 @@ pipeline_tag: text-generation
40
  - [πŸ”— Model Download](#model-download)
41
  - [πŸ”– Model Benchmark](#model-benchmark)
42
  - [πŸ“Š Model Inference](#model-inference)
43
- - [πŸ₯‡ Company Introduction](#company-introduction)
44
  - [πŸ“œ Declarations & License](#declarations-license)
 
45
 
46
  # 1. Model Introduction
47
 
48
- - Orion-14B-Chat-Int4 is quantized using awq from Orion-14B-Chat while reducing model size by 70% and improving inference speed by 30%, with performance loss less than 1%.
49
- <div align="center">
50
- <img src="./assets/imgs/model_cap_en.png" alt="model_cap_en" width="50%" />
51
- </div>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
 
53
  - Orion-14B series models including:
54
  - **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
@@ -321,7 +335,7 @@ Truly Useful Robots", OrionStar empowers more people through AI technology.
321
 
322
  **The core strengths of OrionStar lies in possessing end-to-end AI application capabilities,** including big data preprocessing, large model pretraining, fine-tuning, prompt engineering, agent, etc. With comprehensive end-to-end model training capabilities, including systematic data processing workflows and the parallel model training capability of hundreds of GPUs, it has been successfully applied in various industry scenarios such as government affairs, cloud services, international e-commerce, and fast-moving consumer goods.
323
 
324
- Companies with demands for deploying large-scale model applications are welcome to contact us.
325
  **Enquiry Hotline: 400-898-7779**<br>
326
  **E-mail: ai@orionstar.com**
327
 
 
40
  - [πŸ”— Model Download](#model-download)
41
  - [πŸ”– Model Benchmark](#model-benchmark)
42
  - [πŸ“Š Model Inference](#model-inference)
 
43
  - [πŸ“œ Declarations & License](#declarations-license)
44
+ - [πŸ₯‡ Company Introduction](#company-introduction)
45
 
46
  # 1. Model Introduction
47
 
48
+ - Orion-14B series models are open-source multilingual large language models trained from scratch by OrionStarAI. The base model is trained on 2.5T multilingual corpus, including Chinese, English, Japanese, Korean, etc, and it exhibits superior performance in these languages. For details, please refer to [tech report](https://github.com/OrionStarAI/Orion/blob/master/doc/Orion14B_v3.pdf).
49
+
50
+ - The Orion-14B series models exhibit the following features:
51
+ - Among models with 20B-parameter scale level, Orion-14B-Base model shows outstanding performance in comprehensive evaluations.
52
+ - Strong multilingual capabilities, significantly outperforming in Japanese and Korean testsets.
53
+ - The fine-tuned models demonstrate strong adaptability, excelling in human-annotated blind tests.
54
+ - The long-chat version supports extremely long texts, performing exceptionally well at a token length of 200k and can support up to a maximum of 320k.
55
+ - The quantized versions reduce model size by 70%, improve inference speed by 30%, with performance loss less than 1%.
56
+ <table style="border-collapse: collapse; width: 100%;">
57
+ <tr>
58
+ <td style="border: none; padding: 10px; box-sizing: border-box;">
59
+ <img src="./assets/imgs/opencompass_en.png" alt="opencompass" style="width: 100%; height: auto;">
60
+ </td>
61
+ <td style="border: none; padding: 10px; box-sizing: border-box;">
62
+ <img src="./assets/imgs/model_cap_en.png" alt="modelcap" style="width: 100%; height: auto;">
63
+ </td>
64
+ </tr>
65
+ </table>
66
 
67
  - Orion-14B series models including:
68
  - **Orion-14B-Base:** A multilingual large language foundational model with 14 billion parameters, pretrained on a diverse dataset of 2.5 trillion tokens.
 
335
 
336
  **The core strengths of OrionStar lies in possessing end-to-end AI application capabilities,** including big data preprocessing, large model pretraining, fine-tuning, prompt engineering, agent, etc. With comprehensive end-to-end model training capabilities, including systematic data processing workflows and the parallel model training capability of hundreds of GPUs, it has been successfully applied in various industry scenarios such as government affairs, cloud services, international e-commerce, and fast-moving consumer goods.
337
 
338
+ Companies with demands for deploying large-scale model applications are welcome to contact us.<br>
339
  **Enquiry Hotline: 400-898-7779**<br>
340
  **E-mail: ai@orionstar.com**
341