NVIDIA DGX Spark (128GB)
The NVIDIA DGX Spark (128GB) has 128GB VRAM and 273 GB/s memory bandwidth. It can run 49 of our 54 tracked models natively in VRAM at 8k context.
| Vendor | NVIDIA |
| Architecture | Grace Blackwell |
| VRAM | 128 GB (unified) |
| Memory type | LPDDR5X |
| Memory bandwidth | 273 GB/s |
| Compute backend | CUDA |
| Tier | Workstation |
| Released | 2025 |
| Models (native) | 49 / 54 |
| Models (offload) | 0 / 54 |
Software: ARM-based Grace CPU + Blackwell GPU share unified memory. llama.cpp CUDA backend works; NVIDIA NIM microservices also supported.
Models this GPU runs natively in VRAM (49)
- Qwen3 235B-A22B (MoE)235B · MMLU-Pro —Q3_K_M · ~34.1 t/s
- Mixtral 8x22B Instruct v0.1141B · MMLU-Pro 40.0Q6_K · ~10.3 t/s
- Qwen 3.5 122B-A10B (MoE)122B · MMLU-Pro —Q6_K · ~40 t/s
- Llama 4 Scout 109B109B · MMLU-Pro 70.0Q6_K · ~23.6 t/s
- Qwen 2.5 72B Instruct72B · MMLU-Pro 58.1Q8 · ~3.8 t/s
- Llama 3.3 70B Instruct70B · MMLU-Pro 68.9Q8 · ~3.9 t/s
- DeepSeek R1 Distill Llama 70B70B · MMLU-Pro 70.0Q8 · ~3.9 t/s
- Llama 3.1 70B Instruct70B · MMLU-Pro 66.4Q8 · ~3.9 t/s
- Mixtral 8x7B Instruct v0.146.7B · MMLU-Pro 29.7FP16 · ~11.6 t/s
- Command-R 35B35B · MMLU-Pro 33.0FP16 · ~3.9 t/s
- Qwen 3.5 35B-A3B (MoE)35B · MMLU-Pro —FP16 · ~50.1 t/s
- Qwen 3.6 35B35B · MMLU-Pro —FP16 · ~3.9 t/s
- Yi 1.5 34B Chat34.4B · MMLU-Pro 37.0FP16 · ~4 t/s
- Qwen3 32B32.8B · MMLU-Pro —FP16 · ~4.2 t/s
- Qwen 2.5 32B Instruct32.5B · MMLU-Pro 55.1FP16 · ~4.2 t/s
- Qwen 2.5 Coder 32B Instruct32.5B · MMLU-Pro 50.4FP16 · ~4.2 t/s
- DeepSeek R1 Distill Qwen 32B32.5B · MMLU-Pro 65.0FP16 · ~4.2 t/s
- Gemma 4 31B31B · MMLU-Pro —FP16 · ~4.4 t/s
- Qwen3 30B-A3B (MoE)30B · MMLU-Pro —FP16 · ~50.1 t/s
- Gemma 2 27B Instruct27.2B · MMLU-Pro 38.0FP16 · ~5 t/s
- Gemma 3 27B Instruct27B · MMLU-Pro —FP16 · ~5.1 t/s
- Qwen 3.6 27B27B · MMLU-Pro —FP16 · ~5.1 t/s
- Gemma 4 26B (MoE)26B · MMLU-Pro —FP16 · ~39.5 t/s
- Mistral Small 3.1 24B Instruct24B · MMLU-Pro —FP16 · ~5.7 t/s
- Mistral Small 22B22.2B · MMLU-Pro 49.2FP16 · ~6.1 t/s
- Qwen3 14B14.8B · MMLU-Pro —FP16 · ~9.2 t/s
- Qwen 2.5 14B Instruct14.7B · MMLU-Pro 51.2FP16 · ~9.3 t/s
- Phi-4 14B Instruct14B · MMLU-Pro 56.1FP16 · ~9.8 t/s
- Mistral Nemo 12B Instruct12.2B · MMLU-Pro 35.6FP16 · ~11.2 t/s
- Gemma 3 12B Instruct12.2B · MMLU-Pro —FP16 · ~11.2 t/s
- Gemma 2 9B Instruct9.2B · MMLU-Pro 32.0FP16 · ~14.8 t/s
- Llama 3.1 8B Instruct8B · MMLU-Pro 37.5FP16 · ~17.1 t/s
- DeepSeek R1 Distill Llama 8B8B · MMLU-Pro 41.0FP16 · ~17.1 t/s
- Qwen3 8B8B · MMLU-Pro —FP16 · ~17.1 t/s
- Qwen 2.5 7B Instruct7.6B · MMLU-Pro 36.5FP16 · ~18 t/s
- Mistral 7B Instruct v0.37.25B · MMLU-Pro 30.0FP16 · ~18.8 t/s
- Gemma 3 4B Instruct4B · MMLU-Pro —FP16 · ~34.1 t/s
- Gemma 4 E4B4B · MMLU-Pro —FP16 · ~34.1 t/s
- Phi-3.5 Mini Instruct3.8B · MMLU-Pro 35.6FP16 · ~35.9 t/s
- Llama 3.2 3B Instruct3.2B · MMLU-Pro 24.0FP16 · ~42.7 t/s
- Qwen 2.5 3B Instruct3.1B · MMLU-Pro 32.4FP16 · ~44 t/s
- Gemma 2 2B Instruct2.6B · MMLU-Pro 17.8FP16 · ~52.5 t/s
- Gemma 4 E2B2B · MMLU-Pro —FP16 · ~68.3 t/s
- SmolLM2 1.7B Instruct1.7B · MMLU-Pro 19.0FP16 · ~80.3 t/s
- Qwen 2.5 1.5B Instruct1.5B · MMLU-Pro 16.8FP16 · ~91 t/s
- Llama 3.2 1B Instruct1.24B · MMLU-Pro 12.5FP16 · ~110.1 t/s
- Gemma 3 1B Instruct1B · MMLU-Pro —FP16 · ~136.5 t/s
- Qwen 2.5 0.5B Instruct0.5B · MMLU-Pro 10.0FP16 · ~273 t/s
- SmolLM2 360M Instruct0.36B · MMLU-Pro 8.0FP16 · ~379.2 t/s