Gemma 3: Google’s Multimodal Local LLM Explained
A practical guide to running Google’s Gemma 3 locally with Ollama: the 1B, 4B, 12B, and 27B variants and their VRAM requirements, native multimodal image analysis at every size above 1B, CLI and Python usage including image inputs, how Gemma 3 4B compares to Llama 3.2 8B on reasoning tasks, the 12B as a multimodal sweet spot, 27B for frontier-class local quality on Apple Silicon, configuring a 32K context Modelfile, strong multilingual support, and how to choose between Gemma 3 and other local model families.