NVIDIA has recently announced the launch of the HGX H200 GPU, an advancement over the previous HGX H100.
This new chip, designed primarily for training AI systems, integrates several performance enhancements.
Compatibility with the hardware and software of the H100 systems has been maintained, facilitating a smooth upgrade process.
The H200, developed on NVIDIA’s Hopper architecture, features the H200 Tensor Core GPU.
Its introduction of HBM3e memory is a standout, providing 141GB at a speed of 4.8 terabytes per second.
This large memory, almost double that of its predecessor, the A100, has been aimed at enhancing the acceleration of generative AI and large language models (LLMs).
The H200’s capabilities are further exemplified by its near doubling of inference speed on complex models such as Llama 2, in comparison to the H100.
Various configurations, including four– and eight-way GPU setups and integration into the NVIDIA GH200 Grace Hopper Superchip with HBM3e, will be offered.
Major cloud providers like Google Cloud, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure, along with companies such as CoreWeave, Lambda, and Vultr, are set to deploy H200-based instances by 2024.
Boasting over 32 petaflops of FP8 deep learning compute and 1.1TB of high-bandwidth memory, the H200’s eight-way configuration represents a significant step in NVIDIA’s chipset technology.
Its availability is anticipated to begin in the second quarter of 2024.
This release, more than a year after the debut of the H100, which was NVIDIA’s inaugural GPU based on the Hopper architecture, shows the company’s continuous evolution in AI and supercomputing technologies.
NVIDIA’s H200 is expected to significantly impact various application workloads, especially in training and inference for large language models exceeding 175 billion parameters.
The NVIDIA HGX H200 stands out with its Hopper architecture and H200 Tensor Core GPU. It introduces HBM3e memory, providing a substantial memory upgrade with 141GB at 4.8 terabytes per second, almost doubling the capacity of the A100. This makes it more efficient in handling AI and large language models.
The HGX H200 is expected to be available starting in the second quarter of 2024. Major cloud service providers like Google Cloud, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure, along with companies such as CoreWeave, Lambda, and Vultr, are set to deploy H200-based instances.
The NVIDIA HGX H200 offers a significant leap in performance, delivering over 32 petaflops of FP8 deep learning compute and 1.1TB of high-bandwidth memory. Its eight-way configuration is particularly powerful for AI training and large language model processing.
Highlights Apple has reportedly paused development of the foldable iPad Ultra because of several challenges.…
Highlights Bloomberg reports the iPhone 18 Pro series will feature major camera hardware improvements. It…
Highlights Apple might launch a new Siri Mode in the Camera app with improved Visual…
Highlights OnePlus and Realme have reportedly merged operations. Li Jie (OnePlus China President) will lead…
Highlights May 2026 India major debuts include Vivo X300 Ultra/FE, OnePlus Nord CE 6/Lite and…
Highlights The Tecno Spark 50 Pro is expected to feature a large 6,000 mAh battery…
This website uses cookies.