Update README.md
Browse files
README.md
CHANGED
@@ -7,6 +7,20 @@ This is an HF version of the [Vicuna 7B 1.1 model](https://huggingface.co/lmsys/
|
|
7 |
|
8 |
It was created by merging the deltas provided in the above repo with the original Llama 7B model, [using the code provided on their Github page](https://github.com/lm-sys/FastChat#vicuna-weights).
|
9 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10 |
# Vicuna Model Card
|
11 |
|
12 |
## Model details
|
|
|
7 |
|
8 |
It was created by merging the deltas provided in the above repo with the original Llama 7B model, [using the code provided on their Github page](https://github.com/lm-sys/FastChat#vicuna-weights).
|
9 |
|
10 |
+
## My Vicuna 1.1 model repositories
|
11 |
+
|
12 |
+
I have the following Vicuna 1.1 repositories available:
|
13 |
+
|
14 |
+
**13B models:**
|
15 |
+
* [Unquantized 13B 1.1 model for GPU - HF format](https://huggingface.co/TheBloke/vicuna-13B-1.1-HF)
|
16 |
+
* [GPTQ quantized 4bit 13B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g)
|
17 |
+
* [GPTQ quantized 4bit 13B 1.1 for CPU - GGML format for `llama.cpp`](https://huggingface.co/TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g-GGML)
|
18 |
+
|
19 |
+
**7B models:**
|
20 |
+
* [Unquantized 7B 1.1 model for GPU - HF format](https://huggingface.co/TheBloke/vicuna-7B-1.1-HF)
|
21 |
+
* [GPTQ quantized 4bit 7B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g)
|
22 |
+
* [GPTQ quantized 4bit 7B 1.1 for CPU - GGML format for `llama.cpp`](https://huggingface.co/TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g-GGML)
|
23 |
+
|
24 |
# Vicuna Model Card
|
25 |
|
26 |
## Model details
|