March 6, 2026 · 8 min read 2/9 First Steps: Running Models on Ollama (20B → 120B) The standard recipe works but wastes the hardware. Scaling from 20B to 120B on Ollama shows the potential — and the ceiling. #DGX Spark #Ollama #Open WebUI #AI #Local AI