Automated Immediate Optimization (APO)
To additional assist carry your ML Package Immediate API use circumstances to manufacturing, we’re excited to announce Automated Immediate Optimization (APO) concentrating on On-System fashions on Vertex AI. Automated Immediate Optimization is a device that helps you routinely discover the optimum immediate to your use circumstances.
The period of On-System AI is now not a promise—it’s a manufacturing actuality. With the discharge of Gemini Nano v3, we’re putting unprecedented language understanding and multimodal capabilities straight into the palms of customers. Via the Gemini Nano household of fashions, we now have huge protection of supported units throughout the Android Ecosystem. However for builders constructing the subsequent era of clever apps, entry to a strong mannequin is simply the first step. The true problem lies in customization: How do you tailor a basis mannequin to expert-level efficiency to your particular use case with out breaking the constraints of cellular {hardware}?
Within the server-side world, the bigger LLMs are typically extremely succesful and require much less area adaptation. Even when wanted, extra superior choices equivalent to LoRA (Low-Rank Adaptation) fine-tuning may be possible choices. Nonetheless, the distinctive structure of Android AICore prioritizes a shared, memory-efficient system mannequin. Which means deploying customized LoRA adapters for each particular person app comes with challenges on these shared system providers.
However there may be an alternate path that may be equally impactful. By leveraging Automated Immediate Optimization (APO) on Vertex AI, builders can obtain high quality approaching fine-tuning, all whereas working seamlessly throughout the native Android execution atmosphere. By specializing in superior system instruction, APO allows builders to tailor mannequin conduct with larger robustness and scalability than conventional fine-tuning options.
Observe: Gemini Nano V3 is a top quality optimized model of the extremely acclaimed Gemma 3N mannequin. Any immediate optimizations which might be made on the open supply Gemma 3N mannequin will apply to Gemini Nano V3 as nicely. On supported units, ML Package GenAI APIs leverage the nano-v3 mannequin to maximise the standard for Android Builders

