mpt-7b-8k-chat-gptq / flash_attn_triton.py

Commit History

GPTQ quantized MPT model
081aec9

casperhansen commited on