NVIDIA H200 NVL

  • GPU architecture NVIDIA Hopper

  • GPU accelerator optimized for generative AI and HPC computing

  • 14592 NVIDIA CUDA cores for large-scale parallel computing

  • 456 NVIDIA Tensor cores for AI model training and inference

  • 141 GB of HBM3e memory with ECC - huge space for the largest AI models

  • Memory bandwidth up to 4,8 TB/s

  • Interface PCIe 5.0 x16 providing very high data throughput

  • Passive cooling designed for servers and GPU clusters

  • Maximum power consumption: 600 W

Product out of stock

Free shipping from €300

Promocja cenowa na model HDR-15-5

Product intended for professional use only
NVIDIA H200 NVL

NVIDIA H200 NVL

Description

NVIDIA H200 NVL GPU accelerator for generative AI and the largest LLM models

NVIDIA H200 Tensor Core GPU is one of the latest and most advanced compute accelerators designed for artificial intelligence infrastructure and high performance computing (HPC). Based on the NVIDIA Hopper architecture, the chip delivers breakthrough performance and massive memory bandwidth to support the most demanding AI workloads.

The H200 NVLGPU is designed to train and run very large artificial intelligence models, including generative AI systems used in enterprises and data centres. With a new generation of HBM3e memory, the accelerator offers significantly higher bandwidth and memory capacity than previous generations of GPUs.

NVIDIA H200 NVL is the first GPU accelerator to use HBM3e memory, enabling significant acceleration of generative AI, large language models and scientific computing.

NVIDIA Hopper architecture and HBM3e memory

The accelerator uses the NVIDIA Hopper architecture, which is designed to maximise the acceleration of AI computing. The GPU is equipped with 14592 CUDA cores and 456 Tensor cores to accelerate mathematical operations used in deep learning.

The H200's most significant innovation is the use of 141 GB of HBM3e memory with ECC error correction and a bandwidth of up to 4.8 TB/s. This enables the GPU to support the largest artificial intelligence models and huge data sets used in modern AI systems.

14592
CUDA cores
456
Tensor cores
141 GB
of HBM3e memory
4.8 TB/s
memory bandwidth

GPU for generative artificial intelligence

The NVIDIA H200 NVL is designed as an accelerator for the most advanced generative artificial intelligence systems. The GPU enables massive AI models to be trained and rapidly deployed in production environments.

Large Language Models
training and inference of LLM models
Multimodal AI
VLM and systems combining text, image and video
RAG Systems
AI using enterprise knowledge
AI in the cloud
AI-as-a-Service platforms
Data analytics
big data and AI analytics
HPC simulation
academic and research computing

Accelerator for next-generation data centres

The NVIDIA H200 NVL GPU is designed to run in modern server platforms and GPU clusters used in data centres. The PCIe 5.0 x16 interface provides ultra-high bandwidth communication between the GPU and the server system.

With passive cooling and a maximum power consumption of 600 W, the accelerator is designed for professional AI infrastructure and HPC systems serving the most demanding workloads.

GPUs for AI infrastructure and supercomputing

The increasing complexity of artificial intelligence models is driving the need for massive computing power and very high memory bandwidth. The NVIDIA H200 NVL addresses these needs, enabling scalable AI platforms used in enterprises, data centres and research institutes.

The accelerator is a key component of the infrastructure used to train the largest artificial intelligence models and conduct advanced scientific computing.

Technical Specification

Specification H200 SXM H200 NVL
FP64 34 TFLOPS 30 TFLOPS
FP64 Tensor Core 67 TFLOPS 60 TFLOPS
FP32 67 TFLOPS 60 TFLOPS
TF32 Tensor Core² 989 TFLOPS 835 TFLOPS
BFLOAT16 Tensor Core² 1,979 TFLOPS 1,671 TFLOPS
FP16 Tensor Core² 1,979 TFLOPS 1,671 TFLOPS
FP8 Tensor Core² 3,958 TFLOPS 3,341 TFLOPS
INT8 Tensor Core² 3,958 TFLOPS 3,341 TFLOPS
GPU Memory 141 GB 141 GB
GPU Memory Bandwidth 4.8 TB/s 4.8 TB/s
Decoders 7 NVDEC
7 JPEG
7 NVDEC
7 JPEG
Confidential Computing Supported Supported
Max Thermal Design Power (TDP) Up to 700W (configurable) Up to 600W (configurable)
Multi-Instance GPUs Up to 7 MIGs @ 18GB each Up to 7 MIGs @ 16.5GB each
Form Factor SXM PCIe
Dual-slot air-cooled
Interconnect NVIDIA NVLink: 900GB/s
PCIe Gen5: 128GB/s
2- or 4-way NVIDIA NVLink bridge: 900GB/s per GPU
PCIe Gen5: 128GB/s
Server Options NVIDIA HGX H200 partner and NVIDIA-Certified Systems with 4 or 8 GPUs NVIDIA MGX H200 NVL partner and NVIDIA-Certified Systems with up to 8 GPUs
NVIDIA AI Enterprise Add-on Included Included

¹ Preliminary specifications. May be subject to change.
² With sparsity.

Contact an Elmark specialist

Have questions? Need advice? Call or write to us!