Together, Microsoft and NVIDIA are accelerating a few of the most groundbreaking improvements in AI. We are excited to proceed innovating with a number of new bulletins from Microsoft and NVIDIA that additional improve our full stack collaboration.
Together, Microsoft and NVIDIA are accelerating a few of the most groundbreaking improvements in AI. This long-standing collaboration has been on the core of the AI revolution over the previous few years, from bringing industry-leading supercomputing efficiency within the cloud to supporting breakthrough frontier fashions and options like ChatGPT in Microsoft Azure OpenAI Service and Microsoft Copilot.
Today, there are a number of new bulletins from Microsoft and NVIDIA that additional improve the complete stack collaboration to assist form the way forward for AI. This consists of integrating the latest NVIDIA Blackwell platform with Azure AI companies infrastructure, incorporating NVIDIA NIM microservices into Azure AI Foundry, and empowering builders, startups, and organizations of all sizes like NBA, BMW, Dentsu, Harvey and OriGen, to speed up their improvements and clear up probably the most difficult issues throughout domains.
Empowering all builders and innovators with agentic AI
Microsoft and NVIDIA collaborate deeply throughout all the expertise stack, and with the rise of agentic AI, they’re thrilled to share a number of new choices which might be obtainable in Azure AI Foundry. First is that Azure AI Foundry now presents NVIDIA NIM microservices. NIM gives optimized containers for greater than two dozen standard basis fashions, permitting builders to deploy generative AI functions and brokers shortly. These new integrations can speed up inferencing workloads for fashions obtainable on Azure, offering vital efficiency enhancements, enormously supporting the rising use of AI brokers. Key options embrace optimized mannequin throughput for NVIDIA accelerated computing platforms, prebuilt microservices deployable anyplace, and enhanced accuracy for particular use circumstances. In addition, we are going to quickly be integrating the NVIDIA Llama Nemotron Reason open reasoning mannequin. NVIDIA Llama Nemotron Reason is a robust AI mannequin household designed for superior reasoning.
Epic, a number one digital well being file firm, is planning to benefit from the newest integration of NVIDIA NIM on Azure AI Foundry, bettering AI functions to ship higher healthcare and affected person outcomes.
The launch of NVIDIA NIM microservices in Azure AI Foundry presents a safe and environment friendly means for Epic to deploy open-source generative AI fashions that enhance affected person care, increase clinician and operational effectivity, and uncover new insights to drive medical innovation. In collaboration with UW Health and UC San Diego Health, we’re additionally researching strategies to guage scientific summaries with these superior fashions. Together, we’re utilizing the newest AI expertise in ways in which really enhance the lives of clinicians and sufferers.
Drew McCombs, VP Cloud and Analytics, Epic
Further, Microsoft can be working intently with NVIDIA to optimize inference efficiency for standard, open-source language fashions and guarantee they’re obtainable on Azure AI Foundry so clients can take full benefit of the efficiency and effectivity advantages from basis fashions. The latest addition of this collaboration is the efficiency optimization for Meta Llama fashions utilizing TensorRT-LLM. Developers can now use the optimized Llama fashions from the mannequin catalog in Azure AI Foundry to expertise enhancements in throughput with out extra steps.
“At Synopsys, we rely on cutting-edge AI models to drive innovation, and the optimized Meta Llama models on Azure AI Foundry have delivered exceptional performance. We’ve seen substantial improvements in both throughput and latency, allowing us to accelerate our workloads while optimizing costs. These advancements make Azure AI Foundry an ideal platform for scaling AI applications efficiently.”
Arun Venkatachar, VP Engineering, Synopsys Central Engineering
At the identical time, Microsoft is worked up to be increasing its mannequin catalog in Azure AI Foundry even additional with the addition of Mistral Small 3.1, which is coming quickly, an enhanced model of Mistral Small 3, that includes multimodal capabilities and an prolonged context size of as much as 128k.
Microsoft can be asserting the basic availability of Azure Container Apps serverless graphics processing models (GPUs) with assist for NVIDIA NIM. Serverless GPUs enable enterprises, startups, and software program improvement corporations to seamlessly run AI workloads on-demand with computerized scaling, optimized chilly begin, and per-second billing with scale all the way down to zero when not in use to scale back operational overhead. With the assist of NVIDIA NIM, improvement groups can simply construct and deploy generative AI functions alongside current functions inside the identical networking, safety, and isolation boundary.
Expanding Azure AI Infrastructure with NVIDIA
The evolution of reasoning fashions and agentic AI techniques is reworking the factitious intelligence panorama. Robust and purpose-built infrastructure is essential to their success. Today, Microsoft is worked up to announce the basic availability of Azure ND GB200 V6 digital machine (VM) sequence accelerated by NVIDIA GB200 NVL72 and NVIDIA Quantum InfiniBand networking. This addition to the Azure AI Infrastructure portfolio, alongside current digital machines that use NVIDIA H200 and NVIDIA H100 GPUs, spotlight Microsoft’s dedication to optimizing infrastructure for the following wave of advanced AI duties like planning, reasoning, and adapting in real-time.
As we push the boundaries of AI, our partnership with Azure and the introduction of the NVIDIA Blackwell platform signify a big leap ahead. The NVIDIA GB200 NVL72, with its unparalleled efficiency and connectivity, tackles probably the most advanced AI workloads, enabling companies to innovate sooner and extra securely. By integrating this expertise with Azure’s safe infrastructure, we’re unlocking the potential of reasoning AI.
Ian Buck, Vice President of Hyperscale and HPC, NVIDIA
The mixture of high-performance NVIDIA GPUs with low-latency NVIDIA InfiniBand networking and Azure’s scalable architectures are important to deal with the brand new large knowledge throughput and intensive processing calls for. Furthermore, complete integration of safety, governance, and monitoring instruments from Azure helps highly effective, reliable AI functions that adjust to regulatory requirements.
Built with Microsoft’s customized infrastructure system and the NVIDIA Blackwell platform, on the datacenter stage every blade options two NVIDIA GB200 Grace™ Blackwell Superchips and NVIDIA NVLink™ Switch scale-up networking, which helps as much as 72 NVIDIA Blackwell GPUs in a single NVLink area. Additionally, it incorporates the newest NVIDIA Quantum InfiniBand, permitting for scaling out to tens of hundreds of Blackwell GPUs on Azure, offering two occasions the AI supercomputing efficiency from earlier GPU generations primarily based on GEMM benchmark evaluation.
As Microsoft’s work with NVIDIA continues to develop and form the way forward for AI, the corporate additionally seems ahead to bringing the efficiency of NVIDIA Blackwell Ultra GPUs and the NVIDIA RTX PRO 6000 Blackwell Server Edition to Azure. Microsoft is ready to launch the NVIDIA Blackwell Ultra GPU-based VMs later in 2025. These VMs promise to ship distinctive efficiency and effectivity for the following wave of agentic and generative AI workloads.
Azure AI’s infrastructure, superior by NVIDIA accelerated computing, persistently delivers excessive efficiency at scale for AI workloads as evidenced by main {industry} benchmarks like Top500 supercomputing and MLPerf outcomes.1,2 Recently, Azure Virtual Machines utilizing NVIDIA’s H200 GPUs achieved distinctive efficiency within the MLPerf Training v4.1 benchmarks throughout varied AI duties. Azure demonstrated main cloud efficiency by scaling 512 H200 GPUs in a cluster, attaining a 28% speedup over H100 GPUs within the newest MLPerf coaching runs by MLCommons.3 This highlights Azure’s means to effectively scale massive GPU clusters. Microsoft is worked up that clients are using this efficiency on Azure to coach superior fashions and get effectivity for generative inferencing.
Empowering companies with Azure AI Infrastructure
Meter is coaching a big basis mannequin on Azure AI Infrastructure to automate networking end-to-end. The efficiency and energy of Azure will considerably scale Meter’s AI coaching and inference, aiding within the improvement of fashions with billions of parameters throughout text-based configurations, time-series telemetry, and structured networking knowledge. With assist from Microsoft, Meter’s fashions goal to enhance how networks are designed, configured, and managed—addressing a big problem for progress.
Black Forest Labs, a generative AI start-up with the mission to develop and advance state-of-the-art deep studying fashions for media, has prolonged its partnership with Azure. Azure AI companies infrastructure is already getting used to deploy its flagship FLUX fashions, the world’s hottest text-to-image media fashions, serving thousands and thousands of high-quality pictures on a regular basis with unprecedented velocity and artistic management. Building on this basis, Black Forest Labs will undertake the brand new ND GB200 v6 VMs to speed up the event and deployment of its next-gen AI fashions, pushing the boundaries of innovation in generative AI for media. Black Forest Labs has been a Microsoft associate since its inception, working collectively to safe probably the most superior, environment friendly, and scalable infrastructure for coaching and delivering its frontier fashions.
We are increasing our partnership with Microsoft Azure to mix BFL’s distinctive analysis experience in generative AI with Azure’s highly effective infrastructure. This collaboration permits us to construct and ship the very best picture and video fashions sooner and at higher scale, offering our clients with state-of-the-art visible AI capabilities for media manufacturing, promoting, product design, content material creation and past.
Robin Rombach, CEO, Black Forest Labs
Creating new potentialities for innovators throughout industries
Microsoft and NVIDIA have launched preconfigured NVIDIA Omniverse and NVIDIA Isaac Sim digital desktop workstations, and Omniverse Kit App Streaming, on the Azure market. Powered by Azure Virtual Machines utilizing NVIDIA GPUs, these choices present builders the whole lot they should get began creating and self-deploying digital twin and robotics simulation functions and companies for the period of bodily AI. Several Microsoft and NVIDIA ecosystem companions together with Bright Machines, Kinetic Vision, Sight Machine, and SoftServe are adopting these capabilities to construct options that may allow the following wave of digitalization for the world’s producers.
There are many progressive options constructed by AI startups on Azure. Opaque Systems helps clients safeguard their knowledge utilizing confidential computing; Faros AI gives software program engineering insights, permitting clients to optimize sources and improve decision-making, together with measuring the ROI of their AI coding assistants; Bria AI gives a visible generative AI platform that enables builders to make use of AI picture technology responsibly, offering cutting-edge fashions skilled solely on fully-licensed datasets; Pangaea Data is delivering higher affected person outcomes by enhancing screening and therapy on the level of care; and Basecamp Research is driving biodiversity discovery with AI and in depth genomic datasets.
Experience the newest improvements from Azure and NVIDIA
Today’s bulletins on the NVIDIA GTC AI Conference underscore Azure’s dedication to pushing the boundaries of AI improvements. With state-of-the-art merchandise, deep collaboration, and seamless integrations, we proceed to ship the expertise that helps and empowers builders and clients in designing, customizing, and deploying their AI options effectively. Learn extra at this yr’s occasion and discover the probabilities that NVIDIA and Azure maintain for the long run.
- Visit us at Booth 514 at NVIDIA GTC.
Sources:
2Benchmark Work | Benchmarks MLCommons
3Leading AI Scalability Benchmarks with Microsoft Azure – Signal65