Upload of AutoGPTQ quantized model
Browse files- config.json +1 -1
- gptq_model-4bit-128g.safetensors +1 -1
config.json
CHANGED
@@ -48,7 +48,7 @@
|
|
48 |
"rope_theta": 500000.0,
|
49 |
"tie_word_embeddings": true,
|
50 |
"torch_dtype": "float16",
|
51 |
-
"transformers_version": "4.46.
|
52 |
"use_cache": true,
|
53 |
"vocab_size": 128256
|
54 |
}
|
|
|
48 |
"rope_theta": 500000.0,
|
49 |
"tie_word_embeddings": true,
|
50 |
"torch_dtype": "float16",
|
51 |
+
"transformers_version": "4.46.3",
|
52 |
"use_cache": true,
|
53 |
"vocab_size": 128256
|
54 |
}
|
gptq_model-4bit-128g.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1558482936
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e52135012a4d83869fe23bb644d7493f51c65734d9ec34e56b6c0f0d0b8bd742
|
3 |
size 1558482936
|