CanItRun Logocanitrun.

Qwen 3.6 27B

Qwen 3.6 27B needs roughly 13.5GB VRAM at Q4 quantization (54.0GB at FP16). 55 GPUs we track can run it fully in VRAM at 8k context.

Alibaba27B params256k contextApache 2.0Commercial use ok

VRAM at each quantization

Assumes 8k context. KV cache grows linearly with context length.

QuantWeightsKV cacheTotal
FP1654.0 GB1.61 GB62.3 GB
Q827.0 GB1.61 GB32.0 GB
Q6_K20.3 GB1.61 GB24.5 GB
Q5_K_M16.9 GB1.61 GB20.7 GB
Q4_K_M13.5 GB1.61 GB16.9 GB
Q3_K_M10.8 GB1.61 GB13.9 GB
Q2_K8.1 GB1.61 GB10.9 GB

Benchmarks

GPUs that run Qwen 3.6 27B natively (55)

Plus 3 GPUs that run it with CPU offload (slower)

Notes

Stability and agentic coding workflows with thinking preservation.

Hugging Face ↗Ollama ↗Released 2026-04-01