falcon-40b-instruct quantized with GPTQ using the script in https://github.com/huggingface/text-generation-inference/pull/438
- group size: 128
- act order: true
- nsamples: 128
- dataset: wikitext2
- Downloads last month
- 6
Inference API (serverless) does not yet support model repos that contain custom code.