yentinglin
commited on
Commit
•
3838e02
1
Parent(s):
a9ea6ae
Update README.md
Browse files
README.md
CHANGED
@@ -43,7 +43,7 @@ Llama-3-Taiwan-70B is a large language model finetuned for Traditional Mandarin
|
|
43 |
- Inference Framework: [NVIDIA TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM)
|
44 |
- Base model: [Llama-3 70B](https://llama.meta.com/llama3/)
|
45 |
- Hardware: [NVIDIA DGX H100](https://www.nvidia.com/zh-tw/data-center/dgx-h100/) on Taipei-1
|
46 |
-
- Context length: 8K tokens (
|
47 |
- Batch size: 2M tokens per step
|
48 |
|
49 |
# Evaluation
|
|
|
43 |
- Inference Framework: [NVIDIA TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM)
|
44 |
- Base model: [Llama-3 70B](https://llama.meta.com/llama3/)
|
45 |
- Hardware: [NVIDIA DGX H100](https://www.nvidia.com/zh-tw/data-center/dgx-h100/) on Taipei-1
|
46 |
+
- Context length: 8K tokens ([128k version](https://huggingface.co/yentinglin/Llama-3-Taiwan-70B-Instruct-128k))
|
47 |
- Batch size: 2M tokens per step
|
48 |
|
49 |
# Evaluation
|