At NVIDIA GTC, Microsoft and NVIDIA are saying new choices throughout a breadth of resolution areas from main AI infrastructure to new platform integrations, and business breakthroughs. Today’s information expands our long-standing collaboration, which has paved the way in which for revolutionary AI improvements that clients at the moment are bringing to fruition.
Microsoft and NVIDIA collaborate on Grace Blackwell 200 Superchip for next-generation AI fashions
Microsoft and NVIDIA are bringing the ability of the NVIDIA Grace Blackwell 200 (GB200) Superchip to Microsoft Azure. The GB200 is a brand new processor designed particularly for large-scale generative AI workloads, information processing, and excessive efficiency workloads, that includes up to an enormous 16 TB/s of reminiscence bandwidth and as much as an estimated 45 instances the inference on trillion parameter fashions relative to the earlier Hopper era of servers.
Microsoft has labored carefully with NVIDIA to make sure their GPUs, together with the GB200, can deal with the newest giant language fashions (LLMs) educated on Azure AI infrastructure. These fashions require monumental quantities of knowledge and compute to coach and run, and the GB200 will allow Microsoft to assist clients scale these assets to new ranges of efficiency and accuracy.
Microsoft may even deploy an end-to-end AI compute material with the not too long ago introduced NVIDIA Quantum-X800 InfiniBand networking platform. By profiting from its in-network computing capabilities with SHARPv4, and its added assist for FP8 for modern AI strategies, NVIDIA Quantum-X800 extends the GB200’s parallel computing duties into large GPU scale.
Azure might be one of many first cloud platforms to ship on GB200-based situations
Microsoft has dedicated to bringing GB200-based situations to Azure to assist clients and Microsoft’s AI companies. The new Azure instances-based on the newest GB200 and NVIDIA Quantum-X800 InfiniBand networking will assist speed up the era of frontier and foundational fashions for pure language processing, pc imaginative and prescient, speech recognition, and extra. Azure clients will be capable to use GB200 Superchip to create and deploy state-of-the-art AI options that may deal with large quantities of knowledge and complexity, whereas accelerating time to market.
Azure additionally affords a spread of companies to assist clients optimize their AI workloads, reminiscent of Microsoft Azure CycleCloud, Azure Machine Learning, Microsoft Azure AI Studio, Microsoft Azure Synapse Analytics, and Microsoft Azure Arc. These companies present clients with an end-to-end AI platform that may deal with information ingestion, processing, coaching, inference, and deployment throughout hybrid and multi-cloud environments.
Delivering on the promise of AI to clients worldwide
With a robust basis of Azure AI infrastructure that makes use of the newest NVIDIA GPUs, Microsoft is infusing AI throughout each layer of the know-how stack, serving to clients drive new advantages and productiveness positive factors. Now, with greater than 53,000 Azure AI clients, Microsoft offers entry to the most effective collection of basis and open-source fashions, together with each LLMs and small language fashions (SLMs), all built-in deeply with infrastructure information and instruments on Azure.
The not too long ago introduced partnership with Mistral AI can also be an important instance of how Microsoft is enabling main AI innovators with entry to Azure’s cutting-edge AI infrastructure, to speed up the event and deployment of next-generation LLMs. Azure’s rising AI mannequin catalogue affords, greater than 1,600 fashions, letting clients select from the newest LLMs and SLMs, together with OpenAI, Mistral AI, Meta, Hugging Face, Deci AI, NVIDIA, and Microsoft Research. Azure clients can select the most effective mannequin for his or her use case.
“We are thrilled to embark on this partnership with Microsoft. With Azure’s cutting-edge AI infrastructure, we are reaching a new milestone in our expansion propelling our innovative research and practical applications to new customers everywhere. Together, we are committed to driving impactful progress in the AI industry and delivering unparalleled value to our customers and partners globally.”
Arthur Mensch, Chief Executive Officer, Mistral AI
General availability of Azure NC H100 v5 VM sequence, optimized for generative inferencing and high-performance computing
Microsoft additionally introduced the overall availability of Azure NC H100 v5 VM sequence, designed for mid-range coaching, inferencing, and excessive efficiency compute (HPC) simulations; it affords excessive efficiency and effectivity.
As generative AI purposes develop at unimaginable velocity, the basic language fashions that empower them will develop additionally to incorporate each SLMs and LLMs. In addition, synthetic slim intelligence (ANI) fashions will proceed to evolve, targeted on extra exact predictions somewhat than creation of novel information to proceed to reinforce its use instances. Their purposes embrace duties reminiscent of picture classification, object detection, and broader pure language processing.
Using the sturdy capabilities and scalability of Azure, we provide computational instruments that empower organizations of all sizes, no matter their assets. Azure NC H100 v5 VMs is yet one more computational device made usually obtainable at present that can just do that.
The Azure NC H100 v5 VM sequence is predicated on the NVIDIA H100 NVL platform, which affords two lessons of VMs, starting from one to 2 NVIDIA H100 94GB PCIe Tensor Core GPUs related by NVLink with 600 GB/s of bandwidth. This VM sequence helps PCIe Gen5, which offers the best communication speeds (128GB/s bi-directional) between the host processor and the GPU. This reduces the latency and overhead of knowledge switch and permits sooner and extra scalable AI and HPC purposes.
The VM sequence additionally helps NVIDIA multi-instance GPU (MIG) know-how, enabling clients to partition every GPU into as much as seven situations, offering flexibility and scalability for various AI workloads. This VM sequence affords as much as 80 Gbps community bandwidth and as much as 8 TB of native NVMe storage on full node VM sizes.
These VMs are perfect for coaching fashions, working inferencing duties, and creating cutting-edge purposes. Learn extra in regards to the Azure NC H100 v5-series.
“Snorkel AI is proud to partner with Microsoft to help organizations rapidly and cost-effectively harness the power of data and AI. Azure AI infrastructure delivers the performance our most demanding ML workloads require plus simplified deployment and streamlined management features our researchers love. With the new Azure NC H100 v5 VM series powered by NVIDIA H100 NVL GPUs, we are excited to continue to can accelerate iterative data development for enterprises and OSS users alike.”
Paroma Varma, Co-Founder and Head of Research, Snorkel AI
Microsoft and NVIDIA ship breakthroughs for healthcare and life sciences
Microsoft is increasing its collaboration with NVIDIA to assist rework the healthcare and life sciences business by way of the combination of cloud, AI, and supercomputing.
By utilizing the worldwide scale, safety, and superior computing capabilities of Azure and Azure AI, together with NVIDIA’S DGX Cloud and NVIDIA Clara suite, healthcare suppliers, pharmaceutical and biotechnology firms, and medical machine builders can now quickly speed up innovation throughout your complete medical analysis to care supply worth chain for the good thing about sufferers worldwide. Learn extra.
New Omniverse APIs allow clients throughout industries to embed large graphics and visualization capabilities
Today, NVIDIA’s Omniverse platform for creating 3D purposes will now be obtainable as a set of APIs working on Microsoft Azure, enabling clients to embed superior graphics and visualization capabilities into present software program purposes from Microsoft and accomplice ISVs.
Built on OpenUSD, a common information interchange, NVIDIA Omniverse Cloud APIs on Azure do the combination work for patrons, giving them seamless bodily primarily based rendering capabilities on the entrance finish. Demonstrating the worth of those APIs, Microsoft and NVIDIA have been working with Rockwell Automation and Hexagon to indicate how the bodily and digital worlds will be mixed for elevated productiveness and effectivity. Learn extra.
Microsoft and NVIDIA envision deeper integration of NVIDIA DGX Cloud with Microsoft Fabric
The two firms are additionally collaborating to deliver NVIDIA DGX Cloud compute and Microsoft Fabric collectively to energy clients’ most demanding information workloads. This signifies that NVIDIA’s workload-specific optimized runtimes, LLMs, and machine studying will work seamlessly with Fabric.
NVIDIA DGX Cloud and Fabric integration embrace extending the capabilities of Fabric by bringing in NVIDIA DGX Cloud’s giant language mannequin customization to deal with data-intensive use instances like digital twins and climate forecasting with Fabric OneLake because the underlying information storage. The integration may even present DGX Cloud as an possibility for patrons to speed up their Fabric information science and information engineering workloads.
Accelerating innovation within the period of AI
For years, Microsoft and NVIDIA have collaborated from {hardware} to programs to VMs, to construct new and revolutionary AI-enabled options to deal with complicated challenges within the cloud. Microsoft will proceed to develop and improve its international infrastructure with essentially the most cutting-edge know-how in each layer of the stack, delivering improved efficiency and scalability for cloud and AI workloads and empowering clients to attain extra throughout industries and domains.
Join Microsoft at NVIDIA CTA AI Conference, March 18 by way of 21, at sales space #1108 and attend a session to study extra about options on Azure and NVIDIA.
Learn extra about Microsoft AI options