nisten's picture
Fixed 5am spelling mistakes
4ad5e94 verified
|
raw
history blame
3.49 kB
metadata
base_model:
  - deepseek-ai/DeepSeek-Coder-V2-Instruct

Custom quantizations of deepseek-coder-v2-instruct optimized for cpu inference.

This iq4xm one uses GGML TYPE IQ_4_XS 4bit in combination with q8_0 bit so it runs fast with minimal loss and takes advantage of int8 optimizations on most newer server cpus.

While it required custom code to make, it is compatible with standard llama.cpp from github or just search nisten in lmstudio.

The following 4bit version is the one I use myself, it gets 17tps on 64 arm cores.

You don't need to consolidate the files anymore, just point llama-cli to the first one and it'll handle the rest fine.

Then to run in commandline interactive mode (prompt.txt file is optional) just do:

./llama-cli --temp 0.4 -m deepseek_coder_v2_cpu_iq4xm.gguf-00001-of-00004.gguf -c 32000 -co -cnv -i -f prompt.txt
deepseek_coder_v2_cpu_iq4xm.gguf-00001-of-00004.gguf
deepseek_coder_v2_cpu_iq4xm.gguf-00002-of-00004.gguf
deepseek_coder_v2_cpu_iq4xm.gguf-00003-of-00004.gguf
deepseek_coder_v2_cpu_iq4xm.gguf-00004-of-00004.gguf

To download the models MUCH faster on linux apt install aria2, on mac: brew install aria2

sudo apt install -y aria2

aria2c -x 8 -o deepseek_coder_v2_cpu_iq4xm.gguf-00001-of-00004.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_iq4xm.gguf-00001-of-00004.gguf

aria2c -x 8 -o deepseek_coder_v2_cpu_iq4xm.gguf-00002-of-00004.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_iq4xm.gguf-00002-of-00004.gguf

aria2c -x 8 -o deepseek_coder_v2_cpu_iq4xm.gguf-00003-of-00004.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_iq4xm.gguf-00003-of-00004.gguf

aria2c -x 8 -o deepseek_coder_v2_cpu_iq4xm.gguf-00004-of-00004.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_iq4xm.gguf-00004-of-00004.gguf

And for downloading the Q8_0 version converted in the most lossless way possible from hf bf16 download these:


aria2c -x 8 -o deepseek_coder_v2_cpu_q8_0-00001-of-00006.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_q8_0-00001-of-00006.gguf

aria2c -x 8 -o deepseek_coder_v2_cpu_q8_0-00002-of-00006.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_q8_0-00002-of-00006.gguf

aria2c -x 8 -o deepseek_coder_v2_cpu_q8_0-00003-of-00006.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_q8_0-00003-of-00006.gguf

aria2c -x 8 -o deepseek_coder_v2_cpu_q8_0-00004-of-00006.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_q8_0-00004-of-00006.gguf

aria2c -x 8 -o deepseek_coder_v2_cpu_q8_0-00005-of-00006.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_q8_0-00005-of-00006.gguf

aria2c -x 8 -o deepseek_coder_v2_cpu_q8_0-00006-of-00006.gguf \
  https://huggingface.co/nisten/deepseek-coder-v2-inst-cpu-optimized-gguf/resolve/main/deepseek_coder_v2_cpu_q8_0-00006-of-00006.gguf

Enjoy and remember to accelerate!

-Nisten