Help with convert-hf-to-gguf-update.py

#31
by Zibri - opened

I created Meta-Llama-3-8B-Instruct-f32-Q8_0.gguf to test it... but the tokenizer is messed up by the previous version...
how do I use convert-hf-to-gguf-update.py to fix it?

Zibri changed discussion title from Help. to Help with convert-hf-to-gguf-update.py
AetherArchitectural org
β€’
edited May 18

You need to convert again from the beginning.

You need to use the ...update.py script as explained in the Warning box, in the card, to get the llama-bpe tokenizer config files, then replace the ones in your original HF-Model files folder.

HF model β†’ outfile BF16 GGUF β†’ Quant

I recommend just using the lossless script from this repo, during the model download, replace the config files and it should work.

AetherArchitectural org

Adding myself here so I can reply more easily.

AetherArchitectural org
β€’
edited Jun 24

Currently you should be good to use the default configs with the recent models, since they should all have them correctly by now.

FantasiaFoundry changed discussion status to closed

Sign up or log in to comment