4GB "Gemini Nano" model GGUF anyone?
Chrome background downloads a 4GB on-device AI model for text summarization.
Deep Dive
A user on Reddit reported seeing an article that claims Chrome silently downloads a ~4GB AI model (likely "Gemini Nano") for text summarization. The user is now asking for its exact name/version and whether a GGUF file exists to run it locally with llama.cpp instead of letting Chrome run it in the background.
Key Points
- Google's Gemini Nano is ~4GB, downloaded silently to Chrome for on-device text summarization.
- Users are requesting a GGUF version to run the model locally with llama.cpp, bypassing Chrome.
- The model runs entirely offline, reducing latency and keeping data on the device.
Why It Matters
On-device AI in browsers raises new questions about user consent, local control, and the future of private AI assistants.