hello,i wonder why i use Automodel.from_pretrained and have some logs like Some weights of the model checkpoint at Llama-2-7B-Chat-GPTQ were not used when initializing LlamaForCausalLM. But the GPU use is about 8G,my huggingface is 4.40.2
· Sign up or log in to comment