NVIDIA H100

  • GPU architecture NVIDIA Hopper

  • GPU accelerator optimized for AI, HPC and data centres

  • 14592 NVIDIA CUDA cores for large-scale parallel computing

  • 456 next-generation Tensor cores for AI model training and inference

  • 80 GB of HBM2e memory with ECC to support very large AI models and datasets

  • Memory bandwidth up to 2 TB/s

  • Interface PCIe 5.0 x16 for maximum communication bandwidth with the system

  • Passive cooling designed for servers and GPU clusters

  • Maximum power consumption: 350 W

Product out of stock

Free shipping from €300

Promocja cenowa na model HDR-15-5

Product intended for professional use only
NVIDIA H100

NVIDIA H100

Description

NVIDIA H100 GPU Accelerator for Generative Artificial Intelligence

NVIDIA H100 Tensor Core GPU is one of the most advanced compute accelerators built for modern data centres and artificial intelligence infrastructure. Based on the NVIDIA Hopper architecture, the chip is specifically designed to support the most demanding workloads for training and running next-generation artificial intelligence models.

The H100GPU is the foundation for many of today's AI platforms used in enterprise, data centre and research environments. With its massive computing power and ultra-high memory bandwidth, it enables it to train models with billions of parameters and support complex data analytics systems.

NVIDIA H100 is designed as an accelerator for generative AI - enabling it to train and run large-scale language models, multimodal systems, and AI platforms used in enterprise and cloud computing environments.

NVIDIA Hopper architecture - next-generation AI acceleration

The NVIDIA Hopper architecture introduces a new generation of Tensor cores and significant improvements in deep learning and data analytics processing. The H100 accelerator features 14592 CUDA cores and 456 Tensor cores to accelerate the matrix operations that underpin modern artificial intelligence models.

With 80GB of HBM2e memory with ECC error correction and bandwidth reaching 2 TB/s, the GPU enables the processing of massive data sets and the training of AI models with a very high number of parameters.

14592
CUDA cores
456
Tensor cores
80 GB
of HBM2e memory
2 TB/s
memory bandwidth

GPU designed for generative AI models

The NVIDIA H100 is one of the most widely used accelerators in generative AI infrastructure. The GPU enables both the training of massive AI models and their rapid deployment in production environments.

The accelerator is used in systems such as:

Large Language Models (LLM)
training and language model inference
Multi-modal models
VLM combining text, image and video
RAG systems
AI using enterprise knowledge
Data analytics
big data and enterprise AI
AI in the cloud
AI-as-a-Service platforms
HPC and simulation
advanced scientific computing

GPU infrastructure for data centers

The NVIDIA H100 is designed for installation in modern server systems and GPU clusters with very high computational density. The PCIe 5.0 x16 interface provides very high bandwidth communication between the GPU and the server system.

With passive cooling and a maximum power consumption of 350 W, the card is suitable for professional GPU platforms used in data centres and AI infrastructures.

Why the NVIDIA H100 is crucial for AI development

Modern artificial intelligence models require massive computing power and very high memory bandwidth. The NVIDIA H100 was designed with these requirements in mind and is the foundation of AI infrastructures used by technology companies, research institutes and cloud platforms around the world.

With its Hopper architecture and advanced Tensor cores, the H100 GPU enables AI models to significantly reduce training time and increase the performance of AI systems.

.

Technical Specification

H100 SXM H100 NVL
FP64 34 teraFLOPS 30 teraFLOPS
FP64 Tensor Core 67 teraFLOPS 60 teraFLOPS
FP32 67 teraFLOPS 60 teraFLOPS
TF32 Tensor Core* 989 teraFLOPS 835 teraFLOPS
BFLOAT16 Tensor Core* 1,979 teraFLOPS 1,671 teraFLOPS
FP16 Tensor Core* 1,979 teraFLOPS 1,671 teraFLOPS
FP8 Tensor Core* 3,958 teraFLOPS 3,341 teraFLOPS
INT8 Tensor Core* 3,958 TOPS 3,341 TOPS
GPU Memory 80 GB 94 GB
GPU Memory Bandwidth 3.35 TB/s 3.9 TB/s
Decoders 7 NVDEC, 7 JPEG 7 NVDEC, 7 JPEG
Max Thermal Design Power (TDP) Up to 700 W (configurable) 350-400 W (configurable)
Multi-Instance GPUs Up to 7 MIGs @ 10GB each Up to 7 MIGs @ 12GB each
Form Factor SXM PCIe dual-slot air-cooled
Interconnect NVIDIA NVLink™: 900 GB/s, PCIe Gen5: 128 GB/s NVIDIA NVLink: 600 GB/s, PCIe Gen5: 128 GB/s
Server Options NVIDIA HGX H100, Partner and NVIDIA-Certified Systems™ with 4 or 8 GPUs NVIDIA DGX H100 with 8 GPUs, Partner and NVIDIA-Certified Systems with 1–8 GPUs
NVIDIA Enterprise Add-on Included Included
Note *With sparsity

Contact an Elmark specialist

Have questions? Need advice? Call or write to us!