24GB VRAM Optimal Quants Collection When asked what I use locally on a 24GB card, this is what I point to. I favor exl2s for long context, GGUF for very short context. • 12 items • Updated 28 days ago • 2
Downtown-Case/EVA-UNIT-01_EVA-Qwen2.5-32B-v0.1-exl2-4.1bpw Text Generation • Updated 28 days ago • 12