Robert Scoble asks for the best local model to run anywhere, Ethan Mollick praises Gemma 4 E4B as strong on-device, and Dave Morin promotes liberating OpenClaw with open or local models via Hugging Face tools.
“What is the best local model to run on ANY computer or device?”
“Gemma 4 E4B is impressive for an on-device LLM.”
“GPT-4ish quality”
“Liberate your @openclaw with an open model or local model”
This finding is one of many signals tracked across Indiehacking. The live feed updates every few hours with new authority voices, debates, and emerging ideas.
← Back to Indiehacking