Posted by David Chou, Product Supervisor and Caren Chang, Developer Relations Engineer
At Google, we’re dedicated to bringing probably the most succesful AI fashions on to the Android gadgets in your pocket. At the moment, we’re thrilled to announce the discharge of our newest state-of-the-art open mannequin: Gemma 4.
These fashions are the muse for the following era of Gemini Nano, so code you write as we speak for Gemma 4 will robotically work on Gemini Nano 4-enabled gadgets that will probably be out there later this yr. With Gemini Nano 4, you’ll profit from our further efficiency optimizations so you’ll be able to ship to manufacturing throughout the Android ecosystem with probably the most environment friendly on-device inference.
You will get early entry to this mannequin as we speak by way of the AICore Developer Preview.
Choose the Gemini Nano 4 Quick mannequin within the Developer Preview UI
to see its blazing quick inference velocity in motion earlier than you write any code
As a result of Gemma 4 natively helps over 140 languages, you’ll be able to count on improved localized, multilingual experiences to your world viewers. Moreover, Gemma 4 affords industry-leading efficiency with multimodal understanding, permitting your apps to grasp and course of textual content, photographs, and audio. To provide the finest stability of efficiency and effectivity, Gemma 4 on Android is available in two sizes:
- E4B: Designed for larger reasoning energy and sophisticated duties.
- E2B: Optimized for optimum velocity (3x sooner than the E4B mannequin!) and decrease latency.
The brand new mannequin is as much as 4x sooner than earlier variations and makes use of as much as 60% much less battery. Beginning as we speak, you’ll be able to experiment with improved capabilities together with:
- Reasoning: Chain-of-thought instructions and conditional statements can now be anticipated to return larger high quality outcomes. For instance: “Decide if the next remark for a dialogue thread passes the group pointers. The remark doesn’t move the group guideline if it incorporates a number of of those reason_for_flag: profanity, derogatory language, hate speech”. If the overview passes the group pointers, return {true}. In any other case, return {false, reason_for_flag}.”
- Math: With higher math abilities, the mannequin can now extra precisely reply questions. For instance: “If I get 26 paychecks per yr, how a lot ought to I contribute every paycheck to succeed in my financial savings purpose of $10,000 over the course of a yr?”
- Time understanding: The mannequin is now extra succesful when reasoning about time, making it extra correct to be used circumstances that contain calendars, reminders, and alarms. For instance: “The occasion is at 6PM on August 18th, and a reminder needs to be despatched out 10 hours earlier than the occasion. Return the time and date the reminder needs to be despatched.”
- Picture understanding: Use circumstances that contain OCR (Optical Character Recognition) – reminiscent of chart understanding, visible knowledge extraction, and handwriting recognition – will now return extra correct outcomes.
Be a part of the Developer Preview as we speak to obtain these fashions in preview fashions and begin constructing next-generation options immediately.
Begin constructing with Gemma 4
Begin testing the mannequin
You possibly can check out the mannequin with out code by following the Developer Preview information. If you wish to leap straight into integrating these fashions together with your present workflow, we’ve made that seamless. Head over to Android Studio to refine your immediate and construct with the acquainted ML Equipment Immediate API. We’ve launched a brand new capacity to specify a mannequin, permitting you to focus on the E2B (quick) or E4B (full) variants for testing.
// Outline the configuration with a particular monitor and desire val previewFullConfig = generationConfig { modelConfig = ModelConfig { releaseTrack = ModelReleaseTrack.PREVIEW desire = ModelPreference.FULL } } // Initialize the GenerativeModel with the configuration val previewModel = GenerativeModel.getClient(previewFullConfig) // Confirm that the precise preview mannequin is accessible val previewModelStatus = previewModel.checkStatus() if (previewModelStatus == FeatureStatus.AVAILABLE) { // Proceed with inference val response = previewModel.generateContent("If I get 26 paychecks per yr, how a lot I ought to contribute every paycheck to succeed in my financial savings purpose of $10k over the course of a yr? Return solely the quantity.") } else { // Deal with the case the place the preview mannequin will not be out there // (e.g., print out log statements) }
What to anticipate through the Developer Preview
The purpose of this Developer Preview is to provide you a head begin on refining immediate accuracy and exploring new use circumstances to your particular apps.
We will probably be making a number of updates all through the preview interval, together with assist for instrument calling, structured output, system prompts, and considering mode in Immediate API, making it simpler to take full benefit of the brand new capabilities and vital efficiency optimizations in Gemma 4.
The preview fashions can be found for testing on AICore-enabled gadgets. These fashions will run on the most recent era of specialised AI accelerators from Google, MediaTek, and Qualcomm Applied sciences. On different gadgets, the fashions will initially run on a CPU implementation that’s not consultant of ultimate manufacturing efficiency. In case your gadget will not be AICore-enabled, you can even check these fashions through the AI Edge Gallery app. We’ll present assist for extra gadgets sooner or later.
Methods to get began
Able to see what Gemma 4 can do to your customers?
- Choose-in: Join the AICore Developer Preview.
- Obtain: As soon as opted in, you’ll be able to set off the obtain of the most recent Gemma 4 fashions on to your supported check gadget.
- Construct: Replace your ML Equipment implementation to focus on the brand new fashions and begin constructing in Android Studio.


