Model Selection ComparisonsComparison

Gemma 4 multilingual quality vs other small models

April 7, 2026r/LocalLLaMA

In r/LocalLLaMA, builders argue Gemma 4 is unusually strong for European languages for its size, while others flag real-world inference instability in common local runtimes, complicating adoption decisions.

Gemma 4 is a huge improvement in many European languages, including Danish, Dutch, French and Italian
The benchmarks look really impressive for such small models.
Gemma 4 31B is (of all tested models):
I mean, I think it's a very good model, but I'm still seeing inference bugs (random typos, not closing the think tag, getting stuck generating 15K tokens in an agentic task) in latest LM Studio beta
I'm using their official version of Gemma 4 26B A4B @ Q4_K_M, with Q8 KV quant.
r/LocalLLaMA
benchmarkssmall modelslocal inferencegemmallama cppsmall modelsdeepmind gemma

See what authorities are saying right now

This finding is one of many signals tracked across Artificial Intelligence. The live feed updates every few hours with new authority voices, debates, and emerging ideas.

← Back to Artificial Intelligence