Friday, December 6, 2024

NVIDIA HGX H200-Powered GIGABYTE Computer Servers

- Advertisement -

Two new 8-GPU baseboard servers have been added to the GIGABYTE G593 series. Giga Computing, a GIGABYTE subsidiary and a leader in the industry for generative AI servers and cutting-edge cooling technologies, supports the NVIDIA HGX H200, a GPU platform that is perfect for large AI datasets, scientific simulations, and other memory-intensive workloads.

NVIDIA HGX H200

G593 Series for AI & HPC Scale-up Computing

The G593 series provides reliable, demanding performance in its small 5U chassis with high airflow for amazing compute density thanks to dedicated real estate for cooling GPUs.

- Advertisement -

The tried-and-true GIGABYTE G593 series server, designed specifically for an 8-GPU baseboard, is the ideal match for the NVIDIA H200 Tensor Core GPU, which uses the same amount of power as the air-cooled NVIDIA HGX H100-based systems. In comparison to the NVIDIA H100 Tensor Core GPU, the NVIDIA HGX H200 GPU delivers a significant boost in memory capacity and bandwidth to help reduce the memory bandwidth limits on AI, particularly AI inference. With a maximum capacity of 141GB of HBM3e memory and a memory bandwidth of 4.8TB/s, the NVIDIA HGX H200 GPU offers a 1.7X memory capacity increase and a 1.4X throughput improvement.

The new GIGABYTE servers have the same general design and optimised specifications, with the exception of the CPU platform, for all customer applications in accelerated computing. These servers will soon be listed together with more eligible GIGABYTE G593 series servers, such as four G593 series servers that have completed functional and performance tests and are NVIDIA-Certified Systems in the NVIDIA Qualified System Catalogue.

The high degree of networking and storage is included into the G593 series. Using NVIDIA GPUDirect Storage, one of the NVIDIA Magnum IO technologies, which provides a direct link from GPU memory to storage for increased bandwidth and lower latency, the series, which supports up to eight U.2 Gen5 NVMe SSDs, can process data quickly. There are twelve expansion slots available for NVIDIA InfiniBand or Ethernet networking via a NIC or DPU for connecting to other servers in the cluster or for data transport outside the data centre. There is a 4+2 3000W redundant power supply design to provide enough of power and continuous operation.

Combining for a Rack-scale Solution with GIGAPOD

It is impossible to exaggerate the significance of a well-designed computer cluster, particularly in light of the rapid advancements in AI and the demand for fresh perspectives in industry and academia. GIGA POD, a rack-scale solution that GIGABYTE introduced last year, has been a huge success when deployed with NVIDIA HGX systems. Thus far, NVIDIA HGX H100 has been integrated into all of these systems using the G593 series server. GIGA POD is expected to help provide even greater performance and efficiency for accelerated computing operations with the addition of support for the NVIDIA HGX H200 GPU.

- Advertisement -

Designed with AI and High-Performance Computing in mind Multiple GPUs with lightning-fast interconnections and a fully accelerated software stack are needed for AI, complicated simulations, and large datasets. To deliver the best application performance and the quickest time to insights, the NVIDIA HGX AI supercomputing platform combines the entire capability of NVIDIA GPUs, NVIDIA NVLink, NVIDIA networking, and fully optimised AI and high-performance computing (HPC) software stacks.

Superior Full-Stack Accelerated Computing Platform

NVIDIA Blackwell Tensor Core GPUs are integrated with high-speed interconnects in the NVIDIA HGX B200 and HGX B100, ushering in a new age of accelerated computation and generative artificial intelligence for data centres. Blackwell-based HGX systems are a premium accelerated scale-up platform that can handle workloads in generative AI, data analytics, and HPC that require up to 15X higher inference performance than the previous generation.

For the best AI performance, NVIDIA HGX offers sophisticated networking solutions at up to 400 gigabits per second (Gb/s) using NVIDIA Quantum-2 InfiniBand and Spectrum-X Ethernet. To support cloud networking, composable storage, zero-trust security, and GPU compute elasticity in hyperscale AI clouds, HGX also incorporates NVIDIA BlueField-3 data processing units (DPUs).

NVIDIA HGX AI Supercomputer

Designed with AI and high-performance computing in mind

Large datasets, intricate simulations, and artificial intelligence all demand several GPUs with lightning-fast connections and an entirely enhanced software stack. To deliver the best application performance and the quickest time to insights, the NVIDIA HGX AI supercomputing platform combines the entire capability of NVIDIA GPUs, NVIDIA NVLink, NVIDIA networking, and fully optimised AI and high-performance computing (HPC) software stacks.

Superior Full-Stack Accelerated Computing Platform

NVIDIA Blackwell Tensor Core GPUs are integrated with high-speed interconnects in the NVIDIA HGX B200 and HGX B100, ushering in a new age of accelerated computation and generative artificial intelligence for data centres. Blackwell-based HGX systems are a premium accelerated scale-up platform that can handle workloads in generative AI, data analytics, and HPC that require up to 15X higher inference performance than the previous generation.

For the best AI performance, NVIDIA HGX offers sophisticated networking solutions at up to 400 gigabits per second (Gb/s) using NVIDIA Quantum-2 InfiniBand and Spectrum-X Ethernet. To support cloud networking, composable storage, zero-trust security, and GPU compute elasticity in hyperscale AI clouds, HGX also incorporates NVIDIA BlueField-3 data processing units (DPUs).

Inference from Deep Learning: Effectiveness and Adaptability

Real-Time Deduction for the Upcoming Large Language Model Generation

For huge models like GPT-MoE-1.8T, the HGX B200 achieves up to 15X greater inference performance compared to the previous NVIDIA Hopper generation. Utilising improvements from the Nemo Framework, TensorRT-LLM, and unique Blackwell Tensor Core technology, the second-generation Transformer Engine speeds up inference for Mixture-of-Experts (MoE) and large language models (LLMs).

GPT-MoE-1.8T Performance and Versatility
Image credit to Nvidia

Deep Learning Instruction: Effectiveness and Expandability

Superior Training Results

With 8-bit floating point (FP8) and increased precisions, the second-generation Transformer Engine allows for an astounding 3X faster training time for big language models, such as GPT-MoE-1.8T. NVIDIA Magnum IO software, InfiniBand networking, and fifth-generation NVLink, which offers 1.8TB/s of GPU-to-GPU connectivity, round out this innovation. When combined, these guarantee effective scalability for businesses and large-scale GPU computing clusters.

GPT-MoE-1.8T model training speedup
Image cresit to Nvidia

Enhancing HGX Through NVIDIA Networking

The modern computing unit is the data centre, and networking is essential to extending the performance of applications throughout it. When combined with NVIDIA Quantum InfiniBand, HGX offers exceptional performance and efficiency, ensuring that all available computing resources are fully utilised.

HGX works best with the NVIDIA Spectrum-X networking platform, which provides the best AI performance over Ethernet, in AI cloud data centres that employ Ethernet. For optimal resource utilisation and performance isolation, it has BlueField-3 DPUs and Spectrum-X switches. This allows it to produce consistent, predictable results for thousands of simultaneous AI operations at any scale. Zero-trust security and powerful cloud multi-tenancy are made possible by Spectrum-X. Israel-1, a hyperscale generative AI supercomputer built using Dell PowerEdge XE9680 servers based on the NVIDIA HGX 8-GPU architecture, BlueField-3 DPUs, and Spectrum-4 switches, is a reference design created by NVIDIA.

Nvidia HGX H200 Price

  • Disclaimer: There are a number of factors that might cause the price of NVIDIA HGX H200 systems to fluctuate dramatically, including:
  • Number of GPUs: The price increases with the number of GPUs included.
  • Other parts: The cost can be affected by memory, storage, networking, and other hardware.
  • Vendor: Prices and configurations vary across different sellers.
  • Market conditions: Price can be affected by supply and demand.

Sample Costs

For a general idea, consider the following recent example:

Lenovo NVIDIA HGX H200: Approximately $289,000 was the price of a system with eight H200 GPUs.

NVIDIA HGX H200 Specs

NVIDIA HGX H200 is available in single baseboards with four or eight H200 or H100 GPUs or eight Blackwell GPUs. These powerful combinations of hardware and software lay the foundation for unprecedented AI supercomputing performance.

 HGX B200HGX B100
Form Factor8x NVIDIA B200 SXM8x NVIDIA B100 SXM
FP4 Tensor Core144 PFLOPS112 PFLOPS
FP8/FP6 Tensor Core72 PFLOPS56 PFLOPS
INT8 Tensor Core72 POPS56 POPS
FP16/BF16 Tensor Core36 PFLOPS28 PFLOPS
TF32 Tensor Core18 PFLOPS14 PFLOPS
FP32640 TFLOPS480 TFLOPS
FP64320 TFLOPS240 TFLOPS
FP64 Tensor Core320 TFLOPS240 TFLOPS
MemoryUp to 1.5TBUp to 1.5TB
NVLinkFifth generationFifth generation
NVIDIA NVSwitchFourth generationFourth generation
NVSwitch GPU-to-GPU Bandwidth1.8TB/s1.8TB/s
Total Aggregate Bandwidth14.4TB/s14.4TB/s
- Advertisement -
agarapuramesh
agarapurameshhttps://govindhtech.com
Agarapu Ramesh was founder of the Govindhtech and Computer Hardware enthusiast. He interested in writing Technews articles. Working as an Editor of Govindhtech for one Year and previously working as a Computer Assembling Technician in G Traders from 2018 in India. His Education Qualification MSc.
RELATED ARTICLES

Recent Posts

Popular Post

Govindhtech.com Would you like to receive notifications on latest updates? No Yes