pygmalion-13b-4bit-128g / quantize_config.json
Pernekhan's picture
Add quantize_config.json file to make it work with engines like vLLM
36f6209
raw
history blame contribute delete
85 Bytes
{
"bits": 4,
"desc_act": false,
"group_size": 128,
"true_sequential": true
}