NVIDIA H100 NVL

  • GPU architecture NVIDIA Hopper

  • GPU accelerator optimised for AI, LLM and data centre computing

  • 14592 NVIDIA CUDA cores for large-scale parallel computing

  • 456 NVIDIA Tensor cores to train and infer AI models

  • 94 GB of HBM2e memory with ECC to support very large artificial intelligence models

  • Memory bandwidth up to 3,9 TB/s

  • Interface PCIe 5.0 x16 for maximum data throughput

  • Passive cooling designed for servers and GPU clusters

  • Maximum power consumption: 400 W

Product out of stock

Free shipping from €300

Promocja cenowa na model HDR-15-5

Product intended for professional use only
NVIDIA H100 NVL

NVIDIA H100 NVL

Description

NVIDIA H100 NVL - GPU accelerator for large AI models

NVIDIA H100 NVL is an advanced GPU accelerator designed specifically to support the largest artificial intelligence models and AI generative infrastructure. Based on the NVIDIA Hopper architecture, the chip provides exceptional performance for training and inference of AI models used in modern data centres.

The H100 NVL GPU is designed to work with massive language models and artificial intelligence systems used in enterprises. Thanks to its very large memory capacity and extreme data throughput, it enables efficient processing of models with billions of parameters.

The NVIDIA H100 NVL is designed for generative AI and large language models - enabling you to train and run the most advanced artificial intelligence systems in data centre environments.

NVIDIA Hopper Architecture and Tensor Cores

The H100 NVLGPU uses the NVIDIA Hopper architecture, which is designed to accelerate AI computing and analyze massive data sets. The accelerator features 14592 CUDA cores and 456 Tensor cores to accelerate matrix operations used in deep learning.

The card also offers 94GB of HBM2e memory with ECC error correction and a bandwidth of up to 3.9 TB/s to support the largest artificial intelligence models on the market today.

14592
CUDA cores
456
Tensor cores
94 GB
of HBM2e memory
3.9 TB/s
memory bandwidth

GPU for generative artificial intelligence

The NVIDIA H100 NVL is used in the computing infrastructure for the most advanced AI systems. The accelerator enables both the training of huge models and their deployment in production environments.

Large Language Models
LLM training and inference
Multimodal AI
models combining text, image and video
RAG systems
AI with enterprise knowledge base
AI in the cloud
AI-as-a-Service platforms
Data analytics
big data and enterprise AI
HPC
academic computing and simulation

GPU infrastructure for data centers

The NVIDIA H100 NVL is designed to run in modern server systems and GPU clusters supporting the most demanding computing workloads. The PCIe 5.0 x16 interface provides very high bandwidth communication with the server system.

With passive cooling and a maximum power consumption of 400 W, the card is designed for installation in professional GPU platforms used in data centres and AI infrastructures.

GPUs for LLM and AI enterprise infrastructures

The increasing complexity of artificial intelligence models is driving the need for more and more processing power and memory bandwidth. The NVIDIA H100 NVL is designed with these requirements in mind and is one of the key solutions used to build AI infrastructure in enterprises and data centres.

.

Technical Specification

H100 SXM H100 NVL
FP64 34 teraFLOPS 30 teraFLOPS
FP64 Tensor Core 67 teraFLOPS 60 teraFLOPS
FP32 67 teraFLOPS 60 teraFLOPS
TF32 Tensor Core* 989 teraFLOPS 835 teraFLOPS
BFLOAT16 Tensor Core* 1,979 teraFLOPS 1,671 teraFLOPS
FP16 Tensor Core* 1,979 teraFLOPS 1,671 teraFLOPS
FP8 Tensor Core* 3,958 teraFLOPS 3,341 teraFLOPS
INT8 Tensor Core* 3,958 TOPS 3,341 TOPS
GPU Memory 80 GB 94 GB
GPU Memory Bandwidth 3.35 TB/s 3.9 TB/s
Decoders 7 NVDEC, 7 JPEG 7 NVDEC, 7 JPEG
Max Thermal Design Power (TDP) Up to 700 W (configurable) 350-400 W (configurable)
Multi-Instance GPUs Up to 7 MIGs @ 10GB each Up to 7 MIGs @ 12GB each
Form Factor SXM PCIe dual-slot air-cooled
Interconnect NVIDIA NVLink™: 900 GB/s, PCIe Gen5: 128 GB/s NVIDIA NVLink: 600 GB/s, PCIe Gen5: 128 GB/s
Server Options NVIDIA HGX H100, Partner and NVIDIA-Certified Systems™ with 4 or 8 GPUs NVIDIA DGX H100 with 8 GPUs, Partner and NVIDIA-Certified Systems with 1–8 GPUs
NVIDIA Enterprise Add-on Included Included
Note *With sparsity

Contact an Elmark specialist

Have questions? Need advice? Call or write to us!