The milestone of Gemma 4
I believe Google’s release of Gemma 4 is a quiet milestone, and it might be more consequential to the overall arc of “how we use LLMs” than the mammoth models now rumbling behind closed doors.
Google has somehow managed to extend Gemini’s visual acuity into these open-weights models. My application has to do with handwriting recognition, plus the calculation of bounding boxes for blobs of text, and the 31B version performs as well as Gemini 3 Flash … and nearly as well as Gemini 3.1 Pro?! (This isn’t just vibes, but quantitative scoring.) Yet Gemma 4 31B is a model I can run however and wherever I want … it runs (quantized) on my old 2017-era deep learning rig with its three 12GB GPUs. It runs in the secure enclaves on Tinfoil.
A big brilliant model is cool, but I do not find it exciting in the way I find Gemma 4.
To the blog home page