CanItRun Logocanitrun.

Coding LLMs

7models · local AI VRAM requirements & GPU compatibility

Coding-focused open-weight models excel at code generation, debugging, refactoring, and technical documentation. These models are optimized for instruction-following on programming tasks, often scoring high on HumanEval and similar benchmarks. Because you need quick iteration, tokens/sec matters — pair each model with a GPU that fits it in VRAM for fastest generation.

Want to check your specific GPU? Use the homepage calculator to see which of these models fit your hardware with estimated tokens per second.