NVIDIA Blackwell Ultra AI Factory Platform Paves Way for Age of AI Reasoning
NVIDIA has unveiled the next evolution of its Blackwell AI factory platform, called Blackwell Ultra, designed to advance AI reasoning capabilities. The platform includes the GB300 NVL72 rack-scale solution and HGX B300 NVL16 system, delivering 1.5x more AI performance than its predecessor.
The GB300 NVL72 connects 72 Blackwell Ultra GPUs and 36 Arm Neoverse-based NVIDIA Grace CPUs in a rack-scale design. The HGX B300 NVL16 offers 11x faster inference on large language models, 7x more compute, and 4x larger memory compared to the Hopper generation.
Key features include:
- Integration with NVIDIA Spectrum-X Ethernet and Quantum-X800 InfiniBand platforms
- New open-source NVIDIA Dynamo inference framework for enhanced AI services
- Support for agentic AI and physical AI applications
Major tech companies and cloud service providers, including AWS, Google Cloud, and Microsoft Azure, will offer Blackwell Ultra-powered instances starting from the second half of 2025.
NVIDIA ha svelato la prossima evoluzione della sua piattaforma Blackwell AI factory, chiamata Blackwell Ultra, progettata per migliorare le capacità di ragionamento dell'IA. La piattaforma include la soluzione rack-scale GB300 NVL72 e il sistema HGX B300 NVL16, che offre 1,5 volte le prestazioni in IA rispetto al suo predecessore.
Il GB300 NVL72 collega 72 GPU Blackwell Ultra e 36 CPU NVIDIA Grace basate su Arm Neoverse in un design rack-scale. L'HGX B300 NVL16 offre un'inferenza 11 volte più veloce su modelli di linguaggio di grandi dimensioni, 7 volte più potenza di calcolo e 4 volte più memoria rispetto alla generazione Hopper.
Le caratteristiche principali includono:
- Integrazione con le piattaforme Ethernet NVIDIA Spectrum-X e InfiniBand Quantum-X800
- Nuovo framework di inferenza open-source NVIDIA Dynamo per servizi IA avanzati
- Supporto per applicazioni di IA agentica e fisica
Grandi aziende tecnologiche e fornitori di servizi cloud, tra cui AWS, Google Cloud e Microsoft Azure, offriranno istanze alimentate da Blackwell Ultra a partire dalla seconda metà del 2025.
NVIDIA ha presentado la próxima evolución de su plataforma Blackwell AI factory, llamada Blackwell Ultra, diseñada para avanzar en las capacidades de razonamiento de la IA. La plataforma incluye la solución de rack GB300 NVL72 y el sistema HGX B300 NVL16, que ofrece 1.5 veces más rendimiento en IA que su predecesor.
El GB300 NVL72 conecta 72 GPU Blackwell Ultra y 36 CPU NVIDIA Grace basadas en Arm Neoverse en un diseño de rack. El HGX B300 NVL16 ofrece una inferencia 11 veces más rápida en modelos de lenguaje grandes, 7 veces más potencia de cálculo y 4 veces más memoria en comparación con la generación Hopper.
Las características clave incluyen:
- Integración con las plataformas Ethernet NVIDIA Spectrum-X y Quantum-X800 InfiniBand
- Nuevo marco de inferencia open-source NVIDIA Dynamo para servicios de IA mejorados
- Soporte para aplicaciones de IA agentiva y física
Grandes empresas tecnológicas y proveedores de servicios en la nube, incluidos AWS, Google Cloud y Microsoft Azure, ofrecerán instancias impulsadas por Blackwell Ultra a partir de la segunda mitad de 2025.
NVIDIA는 AI 추론 능력을 발전시키기 위해 설계된 Blackwell AI factory 플랫폼의 다음 진화인 Blackwell Ultra를 공개했습니다. 이 플랫폼에는 이전 모델보다 1.5배 더 많은 AI 성능을 제공하는 GB300 NVL72 랙 스케일 솔루션과 HGX B300 NVL16 시스템이 포함되어 있습니다.
GB300 NVL72는 72개의 Blackwell Ultra GPU와 36개의 Arm Neoverse 기반 NVIDIA Grace CPU를 랙 스케일 디자인으로 연결합니다. HGX B300 NVL16은 대형 언어 모델에서 11배 더 빠른 추론, 7배 더 많은 컴퓨팅 파워, 그리고 Hopper 세대에 비해 4배 더 큰 메모리를 제공합니다.
주요 기능은 다음과 같습니다:
- NVIDIA Spectrum-X 이더넷 및 Quantum-X800 InfiniBand 플랫폼과의 통합
- 향상된 AI 서비스를 위한 새로운 오픈 소스 NVIDIA Dynamo 추론 프레임워크
- 행위 AI 및 물리적 AI 응용 프로그램 지원
AWS, Google Cloud 및 Microsoft Azure를 포함한 주요 기술 기업 및 클라우드 서비스 제공업체는 2025년 하반기부터 Blackwell Ultra 기반 인스턴스를 제공할 예정입니다.
NVIDIA a dévoilé la prochaine évolution de sa plateforme Blackwell AI factory, appelée Blackwell Ultra, conçue pour faire avancer les capacités de raisonnement de l'IA. La plateforme comprend la solution rack-scale GB300 NVL72 et le système HGX B300 NVL16, offrant 1,5 fois plus de performances en IA que son prédécesseur.
Le GB300 NVL72 connecte 72 GPU Blackwell Ultra et 36 CPU NVIDIA Grace basés sur Arm Neoverse dans un design rack-scale. L'HGX B300 NVL16 offre une inférence 11 fois plus rapide sur de grands modèles de langage, 7 fois plus de puissance de calcul et 4 fois plus de mémoire par rapport à la génération Hopper.
Les principales caractéristiques incluent :
- Intégration avec les plateformes Ethernet NVIDIA Spectrum-X et InfiniBand Quantum-X800
- Nouveau framework d'inférence open-source NVIDIA Dynamo pour des services IA améliorés
- Support pour des applications d'IA agentique et physique
De grandes entreprises technologiques et des fournisseurs de services cloud, notamment AWS, Google Cloud et Microsoft Azure, proposeront des instances alimentées par Blackwell Ultra à partir de la seconde moitié de 2025.
NVIDIA hat die nächste Evolution seiner Blackwell AI factory-Plattform vorgestellt, die Blackwell Ultra genannt wird und darauf abzielt, die KI-Argumentationsfähigkeiten zu verbessern. Die Plattform umfasst die Rack-Scale-Lösung GB300 NVL72 und das System HGX B300 NVL16, das 1,5-mal mehr KI-Leistung als sein Vorgänger bietet.
Der GB300 NVL72 verbindet 72 Blackwell Ultra GPUs und 36 Arm Neoverse-basierte NVIDIA Grace CPUs in einem Rack-Design. Das HGX B300 NVL16 bietet eine 11-mal schnellere Inferenz bei großen Sprachmodellen, 7-mal mehr Rechenleistung und 4-mal mehr Speicher im Vergleich zur Hopper-Generation.
Zu den Hauptmerkmalen gehören:
- Integration mit den NVIDIA Spectrum-X Ethernet- und Quantum-X800 InfiniBand-Plattformen
- Neues Open-Source NVIDIA Dynamo Inferenz-Framework für verbesserte KI-Dienste
- Unterstützung für agentische KI und physische KI-Anwendungen
Große Technologieunternehmen und Cloud-Dienstleister, darunter AWS, Google Cloud und Microsoft Azure, werden ab der zweiten Hälfte von 2025 Instanzen anbieten, die von Blackwell Ultra betrieben werden.
- 1.5x performance improvement over previous generation
- 11x faster inference on large language models
- 50x increase in revenue opportunity for AI factories compared to Hopper
- Significant technological advancement in AI processing capabilities
- Strong partner ecosystem with major cloud providers and manufacturers
- Extended time to market (availability starting second half 2025)
- High infrastructure requirements for implementation
Insights
NVIDIA's Blackwell Ultra announcement represents a significant leap forward in the company's AI infrastructure portfolio, strategically positioning NVIDIA to capitalize on the emerging AI reasoning market. The new platform delivers 1.5x higher AI performance than its predecessor while expanding Blackwell's revenue opportunity by an impressive 50x compared to Hopper-based AI factories.
The technical specifications are remarkable, with the HGX B300 NVL16 delivering 11x faster inference, 7x more compute, and 4x larger memory than the previous Hopper generation. These performance gains directly address the computational demands of next-generation AI applications including reasoning, agentic AI, and physical AI – all representing expanding market opportunities.
The announcement solidifies NVIDIA's ecosystem dominance through comprehensive partner adoption. All major cloud providers (AWS, Google Cloud, Microsoft Azure, Oracle) and numerous server manufacturers (Dell, HPE, Lenovo, etc.) will offer Blackwell Ultra solutions, creating a robust deployment pipeline scheduled for H2 2025.
The simultaneous introduction of the open-source NVIDIA Dynamo inference framework is particularly notable as it optimizes token revenue generation – a critical metric for AI factory economics – while reducing response times and model serving costs. This software innovation effectively maximizes the ROI for enterprises deploying NVIDIA's hardware.
NVIDIA continues executing its full-stack strategy by providing not just hardware but the comprehensive software ecosystem (NVIDIA AI Enterprise, NIM microservices, CUDA-X) that drives adoption and creates switching costs for customers.
The Blackwell Ultra announcement demonstrates NVIDIA's forward-thinking approach to meeting the next wave of enterprise AI demands. By focusing on reasoning capabilities, NVIDIA is addressing a important barrier to AI implementation – the ability to perform complex problem-solving rather than simple pattern matching.
The test-time scaling approach is particularly innovative, allowing models to leverage massive compute resources during inference to explore multiple solution paths. This directly enables more accurate responses for complex enterprise use cases where reliability is paramount.
For enterprises building AI factories, the 50x increase in revenue opportunity compared to Hopper installations represents a compelling business case for investment. This aligns with the broader enterprise trend toward building proprietary AI infrastructure rather than relying exclusively on API-based services.
The integration with NVIDIA's networking portfolio (Spectrum-X Ethernet and Quantum-X800 InfiniBand) addresses critical enterprise requirements for low latency and consistent performance at scale. The 800 Gb/s throughput per GPU eliminates data movement bottlenecks that currently limit many enterprise AI deployments.
Most importantly, NVIDIA has secured the entire AI supply chain with this announcement – from chips to systems, networking, and software. Enterprise customers gain a unified procurement path through established vendors (Cisco, Dell, HPE) while maintaining access to NVIDIA's innovation. The H2 2025 availability gives enterprises sufficient lead time to budget and plan for implementation while creating a compelling upgrade path from current deployments.
- Top Computer Makers, Cloud Service Providers and GPU Cloud Providers to Boost Training and Test-Time Scaling Inference, From Reasoning to Agentic and Physical AI
- New Open-Source NVIDIA Dynamo Inference Software to Scale Up Reasoning AI Services With Leaps in Throughput, Faster Response Time and Reduced Total Cost of Ownership
- NVIDIA Spectrum-X Enhanced 800G Ethernet Networking for AI Infrastructure Significantly Reduces Latency and Jitter
SAN JOSE, Calif., March 18, 2025 (GLOBE NEWSWIRE) -- NVIDIA today announced the next evolution of the NVIDIA Blackwell AI factory platform, NVIDIA Blackwell Ultra — paving the way for the age of AI reasoning.
NVIDIA Blackwell Ultra boosts training and test-time scaling inference — the art of applying more compute during inference to improve accuracy — to enable organizations everywhere to accelerate applications such as AI reasoning, agentic AI and physical AI.
Built on the groundbreaking Blackwell architecture introduced a year ago, Blackwell Ultra includes the NVIDIA GB300 NVL72 rack-scale solution and the NVIDIA HGX™ B300 NVL16 system. The GB300 NVL72 delivers 1.5x more AI performance than the NVIDIA GB200 NVL72, as well as increases Blackwell’s revenue opportunity by 50x for AI factories, compared with those built with NVIDIA Hopper™.
“AI has made a giant leap — reasoning and agentic AI demand orders of magnitude more computing performance,” said Jensen Huang, founder and CEO of NVIDIA. “We designed Blackwell Ultra for this moment — it’s a single versatile platform that can easily and efficiently do pretraining, post-training and reasoning AI inference.”
NVIDIA Blackwell Ultra Enables AI Reasoning
The NVIDIA GB300 NVL72 connects 72 Blackwell Ultra GPUs and 36 Arm
Neoverse-based NVIDIA Grace™ CPUs in a rack-scale design, acting as a single massive GPU built for test-time scaling. With the NVIDIA GB300 NVL72, AI models can access the platform’s increased compute capacity to explore different solutions to problems and break down complex requests into multiple steps, resulting in higher-quality responses.
GB300 NVL72 is also expected to be available on NVIDIA DGX™ Cloud, an end-to-end, fully managed AI platform on leading clouds that optimizes performance with software, services and AI expertise for evolving workloads. NVIDIA DGX SuperPOD™ with DGX GB300 systems uses the GB300 NVL72 rack design to provide customers with a turnkey AI factory.
The NVIDIA HGX B300 NVL16 features 11x faster inference on large language models, 7x more compute and 4x larger memory compared with the Hopper generation to deliver breakthrough performance for the most complex workloads, such as AI reasoning.
In addition, the Blackwell Ultra platform is ideal for applications including:
- Agentic AI, which uses sophisticated reasoning and iterative planning to autonomously solve complex, multistep problems. AI agent systems go beyond instruction-following. They can reason, plan and take actions to achieve specific goals.
- Physical AI, enabling companies to generate synthetic, photorealistic videos in real time for the training of applications such as robots and autonomous vehicles at scale.
NVIDIA Scale-Out Infrastructure for Optimal Performance
Advanced scale-out networking is a critical component of AI infrastructure that can deliver top performance while reducing latency and jitter.
Blackwell Ultra systems seamlessly integrate with the NVIDIA Spectrum-X™ Ethernet and NVIDIA Quantum-X800 InfiniBand platforms, with 800 Gb/s of data throughput available for each GPU in the system, through an NVIDIA ConnectX®-8 SuperNIC. This delivers best-in-class remote direct memory access capabilities to enable AI factories and cloud data centers to handle AI reasoning models without bottlenecks.
NVIDIA BlueField®-3 DPUs, also featured in Blackwell Ultra systems, enable multi-tenant networking, GPU compute elasticity, accelerated data access and real-time cybersecurity threat detection.
Global Technology Leaders Embrace Blackwell Ultra
Blackwell Ultra-based products are expected to be available from partners starting from the second half of 2025.
Cisco, Dell Technologies, Hewlett Packard Enterprise, Lenovo and Supermicro are expected to deliver a wide range of servers based on Blackwell Ultra products, in addition to Aivres, ASRock Rack, ASUS, Eviden, Foxconn, GIGABYTE, Inventec, Pegatron, Quanta Cloud Technology (QCT), Wistron and Wiwynn.
Cloud service providers Amazon Web Services, Google Cloud, Microsoft Azure and Oracle Cloud Infrastructure and GPU cloud providers CoreWeave, Crusoe, Lambda, Nebius, Nscale, Yotta and YTL will be among the first to offer Blackwell Ultra-powered instances.
NVIDIA Software Innovations Reduce AI Bottlenecks
The entire NVIDIA Blackwell product portfolio is supported by the full-stack NVIDIA AI platform. The NVIDIA Dynamo open-source inference framework — also announced today — scales up reasoning AI services, delivering leaps in throughput while reducing response times and model serving costs by providing the most efficient solution for scaling test-time compute.
NVIDIA Dynamo is new AI inference-serving software designed to maximize token revenue generation for AI factories deploying reasoning AI models. It orchestrates and accelerates inference communication across thousands of GPUs, and uses disaggregated serving to separate the processing and generation phases of large language models on different GPUs. This allows each phase to be optimized independently for its specific needs and ensures maximum GPU resource utilization.
Blackwell systems are ideal for running new NVIDIA Llama Nemotron Reason models and the NVIDIA AI-Q Blueprint, supported in the NVIDIA AI Enterprise software platform for production-grade AI. NVIDIA AI Enterprise includes NVIDIA NIM™ microservices, as well as AI frameworks, libraries and tools that enterprises can deploy on NVIDIA-accelerated clouds, data centers and workstations.
The Blackwell platform builds on NVIDIA’s ecosystem of powerful development tools, NVIDIA CUDA-X™ libraries, over 6 million developers and 4,000+ applications scaling performance across thousands of GPUs.
Learn more by watching the NVIDIA GTC keynote and register for sessions from NVIDIA and industry leaders at the show, which runs through March 21.
About NVIDIA
NVIDIA (NASDAQ: NVDA) is the world leader in accelerated computing.
For further information, contact:
Kristin Uchiyama
NVIDIA Corporation
+1-408-313-0448
kuchiyama@nvidia.com
Certain statements in this press release including, but not limited to, statements as to: the benefits, impact, availability, and performance of NVIDIA’s products, services, and technologies; third parties adopting or offering NVIDIA’s products and technologies; Blackwell Ultra being able to easily and efficiently do pretraining, post-training and reasoning AI inference; and advanced networking being a critical component of AI infrastructure that can deliver top performance while reducing latency and jitter are forward-looking statements that are subject to risks and uncertainties that could cause results to be materially different than expectations. Important factors that could cause actual results to differ materially include: global economic conditions; our reliance on third parties to manufacture, assemble, package and test our products; the impact of technological development and competition; development of new products and technologies or enhancements to our existing product and technologies; market acceptance of our products or our partners' products; design, manufacturing or software defects; changes in consumer preferences or demands; changes in industry standards and interfaces; unexpected loss of performance of our products or technologies when integrated into systems; as well as other factors detailed from time to time in the most recent reports NVIDIA files with the Securities and Exchange Commission, or SEC, including, but not limited to, its annual report on Form 10-K and quarterly reports on Form 10-Q. Copies of reports filed with the SEC are posted on the company's website and are available from NVIDIA without charge. These forward-looking statements are not guarantees of future performance and speak only as of the date hereof, and, except as required by law, NVIDIA disclaims any obligation to update these forward-looking statements to reflect future events or circumstances.
Many of the products and features described herein remain in various stages and will be offered on a when-and-if-available basis. The statements above are not intended to be, and should not be interpreted as a commitment, promise, or legal obligation, and the development, release, and timing of any features or functionalities described for our products is subject to change and remains at the sole discretion of NVIDIA. NVIDIA will have no liability for failure to deliver or delay in the delivery of any of the products, features or functions set forth herein
© 2025 NVIDIA Corporation. All rights reserved. NVIDIA, the NVIDIA logo, BlueField, Connect-X, CUDA-X, NVIDIA DGX, NVIDIA DGX SuperPOD, NVIDIA Grace, NVIDIA HGX, NVIDIA Hopper, NVIDIA NIM and NVIDIA Spectrum-X are trademarks and/or registered trademarks of NVIDIA Corporation in the U.S. and/or other countries. Other company and product names may be trademarks of the respective companies with which they are associated. Features, pricing, availability, and specifications are subject to change without notice.
A photo accompanying this announcement is available at https://www.globenewswire.com/NewsRoom/AttachmentNg/7bb5b0bf-daad-41dc-8d0f-d1706984d616
