CEO's Column
Search
More
Foundation Models

Google’s New AI Model, Gemma 3n, Brings Multimodal AI to Smartphones

ByRishabh Srihari
2025-05-21.about 2 months ago
Google’s New AI Model, Gemma 3n, Brings Multimodal AI to Smartphones
Google’s Gemma 3n brings powerful multimodal AI to smartphones, enabling offline text, image, and voice processing on devices with as little as 2GB RAM.

Artificial intelligence is no longer confined to the cloud. With the launch of Gemma 3n, Google is putting serious AI power directly into the hands of mobile users.

Unveiled at Google I/O 2025, Gemma 3n is the latest in Google’s family of “open” AI models—and it’s built to run natively on devices like phones, tablets, and laptops. That means it can process text, images, audio, and video without needing an internet connection or access to cloud servers.

AI That Runs Where You Are

What sets Gemma 3n apart is its ability to operate on devices with less than 2GB of RAM. This unlocks a new wave of on-device intelligence, enabling AI-powered apps to work even in offline or low-connectivity environments.

Also read: Microsoft Expands AI Offerings with xAI, Mistral, and New Coding Agent

In an age where privacy and latency matter more than ever, on-device AI offers real advantages. There’s no need to upload data to a remote server, and no lag caused by network delays. Whether analyzing images, interpreting voice input, or processing video in real-time, Gemma 3n keeps everything local.

The result: a faster, safer, and more seamless user experience.

Health and Accessibility Get New Models Too

Alongside Gemma 3n, Google introduced two more models: MedGemma and SignGemma.

MedGemma is a specialized AI designed for healthcare. It can read and understand medical text and images, giving developers tools to build diagnostic, wellness, or educational applications powered by multimodal intelligence.

SignGemma, meanwhile, focuses on accessibility. Trained to interpret American Sign Language (ASL) and translate it into spoken-language text, it paves the way for inclusive apps tailored to deaf and hard-of-hearing users.

Both models are part of Google’s push to democratize AI, offering open tools that can be customized for real-world impact.

Questions Around “Open” Remain

Though branded as open models, Gemma releases have drawn criticism for restrictive licensing terms. Developers have raised concerns about commercial use limitations, which can complicate deployment in certain markets.

Despite this, the models are being downloaded and adopted at scale. Gemma’s growing popularity suggests that performance and local execution may outweigh licensing hesitation for many developers.

With Gemma 3n now capable of running on everyday smartphones, Google is turning powerful AI into a mobile-native technology. And in doing so, it's redefining where—and how—AI can work.

Related Topics

Large Language Models (LLMs)LLMs

Subscribe to NG.ai News for real-time AI insights, personalized updates, and expert analysis—delivered straight to your inbox.