yentinglin commited on
Commit
3838e02
1 Parent(s): a9ea6ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -43,7 +43,7 @@ Llama-3-Taiwan-70B is a large language model finetuned for Traditional Mandarin
43
  - Inference Framework: [NVIDIA TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM)
44
  - Base model: [Llama-3 70B](https://llama.meta.com/llama3/)
45
  - Hardware: [NVIDIA DGX H100](https://www.nvidia.com/zh-tw/data-center/dgx-h100/) on Taipei-1
46
- - Context length: 8K tokens (Large-context model coming soon)
47
  - Batch size: 2M tokens per step
48
 
49
  # Evaluation
 
43
  - Inference Framework: [NVIDIA TensorRT-LLM](https://github.com/NVIDIA/TensorRT-LLM)
44
  - Base model: [Llama-3 70B](https://llama.meta.com/llama3/)
45
  - Hardware: [NVIDIA DGX H100](https://www.nvidia.com/zh-tw/data-center/dgx-h100/) on Taipei-1
46
+ - Context length: 8K tokens ([128k version](https://huggingface.co/yentinglin/Llama-3-Taiwan-70B-Instruct-128k))
47
  - Batch size: 2M tokens per step
48
 
49
  # Evaluation