24GB Models
Collection
Models optimized for 24GB VRAM • 8 items • Updated
Prioritized ctx over bpw for this one since it's primarily an agentic model. Would recommend Gemma 4 31B instead for non-agentic.
5.10 bpw, a mixture of Q5_K and Q4_K
Fits ~90k F16, ~170k Q8 CTX on a 24GiB GPU using Vulkan
We're not able to determine the quantization variants.
Base model
Qwen/Qwen3.6-27B