Update README.md
Browse files
README.md
CHANGED
@@ -6,9 +6,9 @@ language:
|
|
6 |
- en
|
7 |
---
|
8 |
|
9 |
-
# Model Card for Breeze-7B-Base-v0_1
|
10 |
|
11 |
-
Breeze-7B is a language model family that builds on top of [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1), specifically intended for Traditional Chinese use.
|
12 |
|
13 |
[Breeze-7B-Base](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0_1) is the base model for the Breeze-7B series.
|
14 |
It is suitable for use if you have substantial fine-tuning data to tune it for your specific use case.
|
@@ -20,7 +20,7 @@ Breeze-7B-Instruct to enable a 64k-token context length. Roughly speaking, that
|
|
20 |
|
21 |
*Update (Feb. 21st, 2024): Breeze-7B-Instruct-64k-v0_1 has been temporarily removed from Hugging Face due to its actual performance in long context tests not meeting expectations.*
|
22 |
|
23 |
-
The current release version of Breeze-7B is
|
24 |
|
25 |
Practicality-wise:
|
26 |
- Breeze-7B-Base expands the original vocabulary with additional 30,000 Traditional Chinese tokens. With the expanded vocabulary, everything else being equal, Breeze-7B operates at twice the inference speed for Traditional Chinese to Mistral-7B and Llama 7B. [See [Inference Performance](#inference-performance).]
|
|
|
6 |
- en
|
7 |
---
|
8 |
|
9 |
+
# Model Card for MediaTek Research Breeze-7B-Base-v0_1
|
10 |
|
11 |
+
MediaTek Research Breeze-7B (hereinafter referred to as Breeze-7B) is a language model family that builds on top of [Mistral-7B](https://huggingface.co/mistralai/Mistral-7B-v0.1), specifically intended for Traditional Chinese use.
|
12 |
|
13 |
[Breeze-7B-Base](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v0_1) is the base model for the Breeze-7B series.
|
14 |
It is suitable for use if you have substantial fine-tuning data to tune it for your specific use case.
|
|
|
20 |
|
21 |
*Update (Feb. 21st, 2024): Breeze-7B-Instruct-64k-v0_1 has been temporarily removed from Hugging Face due to its actual performance in long context tests not meeting expectations.*
|
22 |
|
23 |
+
*Update (Mar. 7th, 2024): The current release version of Breeze-7B is v1.0. See [Breeze-7B-Base-v1_0](https://huggingface.co/MediaTek-Research/Breeze-7B-Base-v1_0).*
|
24 |
|
25 |
Practicality-wise:
|
26 |
- Breeze-7B-Base expands the original vocabulary with additional 30,000 Traditional Chinese tokens. With the expanded vocabulary, everything else being equal, Breeze-7B operates at twice the inference speed for Traditional Chinese to Mistral-7B and Llama 7B. [See [Inference Performance](#inference-performance).]
|