NVIDIA has recently announced the launch of the HGX H200 GPU, an advancement over the previous HGX H100.
This new chip, designed primarily for training AI systems, integrates several performance enhancements.
Compatibility with the hardware and software of the H100 systems has been maintained, facilitating a smooth upgrade process.
The H200, developed on NVIDIA’s Hopper architecture, features the H200 Tensor Core GPU.
Its introduction of HBM3e memory is a standout, providing 141GB at a speed of 4.8 terabytes per second.
This large memory, almost double that of its predecessor, the A100, has been aimed at enhancing the acceleration of generative AI and large language models (LLMs).
The H200’s capabilities are further exemplified by its near doubling of inference speed on complex models such as Llama 2, in comparison to the H100.
Various configurations, including four– and eight-way GPU setups and integration into the NVIDIA GH200 Grace Hopper Superchip with HBM3e, will be offered.
Major cloud providers like Google Cloud, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure, along with companies such as CoreWeave, Lambda, and Vultr, are set to deploy H200-based instances by 2024.
Boasting over 32 petaflops of FP8 deep learning compute and 1.1TB of high-bandwidth memory, the H200’s eight-way configuration represents a significant step in NVIDIA’s chipset technology.
Its availability is anticipated to begin in the second quarter of 2024.
This release, more than a year after the debut of the H100, which was NVIDIA’s inaugural GPU based on the Hopper architecture, shows the company’s continuous evolution in AI and supercomputing technologies.
NVIDIA’s H200 is expected to significantly impact various application workloads, especially in training and inference for large language models exceeding 175 billion parameters.
The NVIDIA HGX H200 stands out with its Hopper architecture and H200 Tensor Core GPU. It introduces HBM3e memory, providing a substantial memory upgrade with 141GB at 4.8 terabytes per second, almost doubling the capacity of the A100. This makes it more efficient in handling AI and large language models.
The HGX H200 is expected to be available starting in the second quarter of 2024. Major cloud service providers like Google Cloud, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure, along with companies such as CoreWeave, Lambda, and Vultr, are set to deploy H200-based instances.
The NVIDIA HGX H200 offers a significant leap in performance, delivering over 32 petaflops of FP8 deep learning compute and 1.1TB of high-bandwidth memory. Its eight-way configuration is particularly powerful for AI training and large language model processing.
Highlights Instagram will discontinue support for end-to-end encrypted direct messages after May 8, 2026. Impacted…
Highlights Apple’s crease-free foldable display for the iPhone Fold is expected to enter mass production…
Highlights Realme might soon introduce the Realme 16T 5G and Realme P4R 5G in India.…
Highlights Honor is tipped to launch the X80 GT with a massive 13,080mAh silicon-carbon battery.…
Highlights The Vivo X500 series is tipped to debut MediaTek’s new Dimensity 9600 and 9600…
Highlights Leaked renders show the Galaxy A37 and Galaxy A57 in all colour variants. Both…
This website uses cookies.