Microsoft combines accelerated computing with cloud scale engineering to deliver superior AI capabilities to our clients. For years, we’ve labored with NVIDIA to combine {hardware}, software program and infrastructure to energy lots of at present’s most necessary AI breakthroughs.
What’s new at NVIDIA GTC
- Expanded Microsoft Foundry capabilities to construct, deploy and function production-ready AI brokers on NVIDIA accelerators and open NVIDIA Nemotron fashions
- New Azure AI infrastructure optimized for inference-heavy, reasoning-based workloads, together with the primary hyperscale cloud to energy on next-generation NVIDIA Vera Rubin NVL72 methods
- Deeper integration throughout Microsoft Foundry, Microsoft Cloth and NVIDIA Omniverse libraries and open frameworks to help Bodily AI methods from simulation to actual‑world operations
From Frontier fashions to production-ready brokers
On the basis of this method is Microsoft Foundry: serving because the working system for constructing, deploying and working AI at enterprise scale. Foundry builds on Azure to deliver collectively fashions, instruments, information and observability right into a single system designed for manufacturing brokers. Right now we’re increasing these capabilities throughout Foundry Agent Service and NVIDIA Nemotron fashions.
The following-generation Foundry Agent Service and Observability in Foundry Management Aircraft are actually typically obtainable, enabling organizations to construct and function AI brokers at manufacturing scale. Foundry Agent Service permits groups to rapidly develop brokers that motive, plan and act throughout instruments, information and workflows. As soon as created, Foundry Management Aircraft offers the developer end-to-end visibility into agent habits, unlocking each developer productiveness in addition to enterprise belief. Firms akin to Corvus Power are already utilizing Foundry to interchange guide inspection workflows with agent-driven operational intelligence throughout their world fleet.
We’re additional simplifying the trail from prototype to manufacturing with the supply of Voice Dwell API integration with Foundry Agent Service, in public preview, which allows builders to construct voice-first, multimodal, real-time agentic experiences. This pairs with the overall availability of a refreshed Microsoft Foundry portal and expanded integrations for Palo Alto Networks’ Prisma AIRS and Zenity, delivering deeper builder experiences and runtime safety throughout your complete agent lifecycle.
NVIDIA Nemotron fashions are additionally now obtainable via Microsoft Foundry, becoming a member of the widest number of fashions on any cloud, together with the newest reasoning, frontier and open fashions. This bolsters our latest partnership announcement bringing Fireworks AI to Microsoft Foundry, enabling clients to fine-tune open-weight fashions like NVIDIA Nemotron into low-latency property that may be distributed to the sting.
Scaling AI infrastructure for the world’s most demanding workloads
Inference AI workloads are reshaping price, efficiency and system design necessities. To operationalize agentic AI at scale, clients want purpose-built infrastructure for inference‑heavy, reasoning‑primarily based workloads that may be deployed and operated constantly throughout world and controlled environments.
Microsoft’s AI infrastructure method is engineered to seamlessly deliver next-generation NVIDIA methods into Azure datacenters which are designed for energy, cooling networking and speedy generational upgrades. This enables our clients to maneuver with velocity and agility and keep at the forefront from technology to technology.
In lower than a yr, we’ve deployed lots of of 1000’s of liquid-cooled Grace Blackwell GPUs throughout our world datacenter footprint, and now we’re excited to be the first hyperscale cloud to energy on NVIDIA’s latest Vera Rubin NVL72 in our labs. Over the following few months, Vera Rubin NVL72 can be rolled out into our trendy, liquid-cooled Azure datacenters.
Microsoft’s infrastructure innovation with NVIDIA additionally extends to sovereign and controlled environments to provide clients management of each the place AI runs and the way it evolves over time. Just lately, we introduced Foundry Native help for contemporary infrastructure and huge AI fashions, and at present we now have preliminary help for NVIDIA Vera Rubin platform on Azure Native, extending accelerated AI capabilities to customer-controlled environments. This method permits organizations to plan for next-generation AI workloads, together with reasoning-based and agentic methods, whereas sustaining Azure-consistent operations, governance and safety via our unified software program layer with Azure Arc and Foundry Native.
Bringing AI into the bodily world
As AI strikes past digital experiences, Microsoft and NVIDIA are collaborating to help the following wave of Bodily AI. At GTC, this work facilities on NVIDIA Bodily AI Information Manufacturing unit Blueprint, with Microsoft Foundry because the platform for internet hosting and working Bodily AI methods on Azure at cloud scale.
By integrating this blueprint with Azure companies as a part of a Bodily AI Toolchain, Microsoft allows builders to construct, practice and function bodily AI and robotics workflows that join bodily property, simulation and cloud coaching environments into repeatable, enterprise-grade pipelines. To help, we’re introducing a public Azure Bodily AI Toolchain GitHub repository built-in with the Nvidia Bodily AI Information Manufacturing unit and with core Azure companies.
To additional the affect of AI in actual‑world, bodily environments, at present Microsoft and NVIDIA are deepening the combination between Microsoft Cloth and NVIDIA Omniverse libraries, connecting stay operational information with bodily correct digital twins and simulation. This enables organizations to see what’s taking place throughout their bodily methods, perceive it in actual time and use AI to determine what to do subsequent. In apply, clients in manufacturing and operations and past are utilizing this method to maneuver past dashboards and alerts to coordinated, AI‑pushed motion throughout machines, services and workflows.
From innovation to affect
Microsoft is delivering dependable, manufacturing‑scale AI by bringing collectively its world AI infrastructure, platforms and actual‑world methods with the newest innovation from NVIDIA. For patrons, this implies the power to function intelligence repeatedly, operating inference-heavy, reasoning-based and bodily AI workloads with the efficiency, safety and governance required for actual companies and controlled industries.
Whether or not powering always-on brokers, scaling next-generation AI infrastructure or deploying clever methods in factories, vitality services and sovereign environments, Microsoft and Nvidia are serving to clients transfer sooner from perception to motion.
Yina Arenas leads product technique and execution for Microsoft Foundry, overseeing the top–to–finish AI product portfolio, infrastructure, developer experiences and basis mannequin integration throughout OpenAI, Anthropic, Mistral, DeepSeek and others. She delivers an enterprise prepared, manufacturing grade AI platform trusted by world clients for safe, dependable and scalable AI.
