Chrome adds on-device AI for speed and privacy
Google Chrome has begun installing a local copy of the Gemini Nano model when certain AI features are enabled, bringing faster, lower-latency AI tools directly to users' desktops. These on-device capabilities power things like smart replies, writing assistance, autofill suggestions, and scam detection — delivering many of the benefits of advanced AI without relying entirely on cloud inference.
Local model file and storage trade-off. The model appears as a weights.bin file in the browser directory and can be roughly 4GB in size on some systems. That additional storage use has surfaced for users who noticed unexplained reductions in available disk space. The file is downloaded when specific Chrome AI features are enabled, so the cost is tied to opting into those tools.
Why this is a net win. Running Gemini Nano on-device reduces network round-trips, so responses arrive faster and can work when connectivity is limited. It also improves privacy by keeping more processing local rather than sending every request to the cloud. Those benefits are important for real-time features like autofill, inline suggestions, and scam detection that need quick, private decisions.
Managing the trade-offs. Users who prefer not to store the model locally can disable or opt out of specific Chrome AI features to prevent the download. Meanwhile, the move reflects a broader industry trend toward efficient, edge-capable AI; we can expect model compression and app-level controls to further shrink footprints and give users more choice over storage and privacy in future updates.