Transformers
GGUF
llama
text-generation-inference
TheBloke commited on
Commit
90bb9da
1 Parent(s): 94edf6a

Initial GGUF model commit

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -55,8 +55,8 @@ Here is a list of clients and libraries, along with their expected timeline for
55
  ## Repositories available
56
 
57
  * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/PuddleJumper-13B-GPTQ)
58
- * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference (deprecated)](https://huggingface.co/TheBloke/PuddleJumper-13B-GGML)
59
  * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/PuddleJumper-13B-GGUF)
 
60
  * [Kai Howard's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/totally-not-an-llm/PuddleJumper-13b)
61
  <!-- repositories-available end -->
62
 
 
55
  ## Repositories available
56
 
57
  * [GPTQ models for GPU inference, with multiple quantisation parameter options.](https://huggingface.co/TheBloke/PuddleJumper-13B-GPTQ)
 
58
  * [2, 3, 4, 5, 6 and 8-bit GGUF models for CPU+GPU inference](https://huggingface.co/TheBloke/PuddleJumper-13B-GGUF)
59
+ * [2, 3, 4, 5, 6 and 8-bit GGML models for CPU+GPU inference (deprecated)](https://huggingface.co/TheBloke/PuddleJumper-13B-GGML)
60
  * [Kai Howard's original unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/totally-not-an-llm/PuddleJumper-13b)
61
  <!-- repositories-available end -->
62