Google has released its Gemma 3n AI model, positioned as an advancement for on-device AI and bringing multimodal capabilities and higher performance to edge devices.
Previewed in May, Gemma 3n is multimodal by design, with native support for image, audio, video, and text inputs and outputs, Google said. Optimized for edge devices such as phones, tablets, laptops, desktops, or single cloud accelerators, Gemma 3n models are available in two sizes based on “effective” parameters, E2B and E4B. Whereas the raw parameter counts for E2B and E4B are 5B and 8B, respectively, these models run with a memory footprint comparable to traditional 2B and 4B models, running with as little as 2GB and 3GB of memory, Google said.
Announced as a production release June 26, Gemma 3n models can be downloaded from Hugging Face and Kaggle. Developers also can try out Gemma 3n in Google AI Studio.