NVIDIA has recently announced the launch of the HGX H200 GPU, an advancement over the previous HGX H100.
This new chip, designed primarily for training AI systems, integrates several performance enhancements.
Compatibility with the hardware and software of the H100 systems has been maintained, facilitating a smooth upgrade process.
The H200, developed on NVIDIA’s Hopper architecture, features the H200 Tensor Core GPU.
Its introduction of HBM3e memory is a standout, providing 141GB at a speed of 4.8 terabytes per second.
This large memory, almost double that of its predecessor, the A100, has been aimed at enhancing the acceleration of generative AI and large language models (LLMs).
The H200’s capabilities are further exemplified by its near doubling of inference speed on complex models such as Llama 2, in comparison to the H100.
Various configurations, including four– and eight-way GPU setups and integration into the NVIDIA GH200 Grace Hopper Superchip with HBM3e, will be offered.
Major cloud providers like Google Cloud, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure, along with companies such as CoreWeave, Lambda, and Vultr, are set to deploy H200-based instances by 2024.
Boasting over 32 petaflops of FP8 deep learning compute and 1.1TB of high-bandwidth memory, the H200’s eight-way configuration represents a significant step in NVIDIA’s chipset technology.
Its availability is anticipated to begin in the second quarter of 2024.
This release, more than a year after the debut of the H100, which was NVIDIA’s inaugural GPU based on the Hopper architecture, shows the company’s continuous evolution in AI and supercomputing technologies.
NVIDIA’s H200 is expected to significantly impact various application workloads, especially in training and inference for large language models exceeding 175 billion parameters.
The NVIDIA HGX H200 stands out with its Hopper architecture and H200 Tensor Core GPU. It introduces HBM3e memory, providing a substantial memory upgrade with 141GB at 4.8 terabytes per second, almost doubling the capacity of the A100. This makes it more efficient in handling AI and large language models.
The HGX H200 is expected to be available starting in the second quarter of 2024. Major cloud service providers like Google Cloud, Amazon Web Services, Microsoft Azure, and Oracle Cloud Infrastructure, along with companies such as CoreWeave, Lambda, and Vultr, are set to deploy H200-based instances.
The NVIDIA HGX H200 offers a significant leap in performance, delivering over 32 petaflops of FP8 deep learning compute and 1.1TB of high-bandwidth memory. Its eight-way configuration is particularly powerful for AI training and large language model processing.
Highlights Vivo Y31d debuts in Cambodia and Vietnam with Glow White and Starlight Grey colour…
Highlights Motorola introduced Moto G17 and Moto G17 Power in select EMEA markets. Both phones…
Highlights Moto G67 launched featuring a 6.78-inch FHD+ AMOLED display, MediaTek Dimensity 6300, up to…
Highlights Samsung Galaxy A07 5G launched in the Philippines in Black and Light Violet colours.…
Highlights Google Photos now lets users turn still images into short AI-generated videos with added…
Highlights Redmi Turbo 5 will feature a 6.59-inch 1.5K OLED display, Dimensity 8500-Ultra chipset, LPDDR5x…
This website uses cookies.