NVIDIA has recently announced the launch of the HGX H200 GPU, an advancement over the previous HGX H100.
This new chip, designed primarily for training AI systems, integrates several performance enhancements.
Compatibility with the hardware and software of the H100 systems has been maintained, facilitating a smooth upgrade process.
The H200, developed on NVIDIA’s Hopper architecture, features the H200 Tensor Core GPU.
Its introduction of HBM3e memory is a standout, providing 141GB at a speed of 4.8 terabytes per second.
This large memory, almost double that of its predecessor, the A100, has been aimed at enhancing the acceleration of generative AI and large language models (LLMs).
The H200’s capabilities are further exemplified by its near doubling of inference speed on complex models such as Llama 2, in comparison to the H100.
Various configurations, including four– and eight-way GPU setups and integration into the NVIDIA GH200 Grace Hopper Superchip with HBM3e, will be offered.
Major cloud providers like Google Cloud, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure, along with companies such as CoreWeave, Lambda, and Vultr, are set to deploy H200-based instances by 2024.
Boasting over 32 petaflops of FP8 deep learning compute and 1.1TB of high-bandwidth memory, the H200’s eight-way configuration represents a significant step in NVIDIA’s chipset technology.
Its availability is anticipated to begin in the second quarter of 2024.
This release, more than a year after the debut of the H100, which was NVIDIA’s inaugural GPU based on the Hopper architecture, shows the company’s continuous evolution in AI and supercomputing technologies.
NVIDIA’s H200 is expected to significantly impact various application workloads, especially in training and inference for large language models exceeding 175 billion parameters.
The NVIDIA HGX H200 stands out with its Hopper architecture and H200 Tensor Core GPU. It introduces HBM3e memory, providing a substantial memory upgrade with 141GB at 4.8 terabytes per second, almost doubling the capacity of the A100. This makes it more efficient in handling AI and large language models.
The HGX H200 is expected to be available starting in the second quarter of 2024. Major cloud service providers like Google Cloud, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure, along with companies such as CoreWeave, Lambda, and Vultr, are set to deploy H200-based instances.
The NVIDIA HGX H200 offers a significant leap in performance, delivering over 32 petaflops of FP8 deep learning compute and 1.1TB of high-bandwidth memory. Its eight-way configuration is particularly powerful for AI training and large language model processing.
Highlights Google Maps now locates nearby EV charging stations AI-powered summaries describe charging station locations…
Highlights Indian phone market grew 11.5% in Q1 2024 with 34 million units shipped Mass…
Highlights Apple trains international employees for Vision Pro launch outside the US. Training held in…
Highlights Dell set to launch XPS 13 9345 and Inspiron 14 7441 Plus with Snapdragon…
Highlights Samsung postpones launch of Galaxy Z Fold 6 Fan Edition due to competition. Focus…
Highlights Google to release Pixel 9, Pixel 9 Pro, and Pixel 9 Pro XL this…