NVIDIA H100 NVL

The NVIDIA H100 NVL, built on the NVIDIA Hopper™ architecture, delivers industry-leading performance, scalability, and security for AI and high-performance computing (HPC) workloads. With innovations such as the Transformer Engine and fourth-generation Tensor Cores, it accelerates large language models (LLMs), securely handles enterprise workloads, and supercharges AI training with up to 4X performance improvement over previous models.

Calculator Icon

Pricing Calculator

GPU Cost in Seconds

Loading...
-
/ hourly cost

Estimate GPU Cloud Costs Instantly

Calculate your GPU cloud computing costs with our interactive pricing tool.

Billing Type

Product Type

GPU Type

Hardware Configuration

Contract Options

No contract discount applied
best value

GPU Plan Estimate

Hourly Cost
Loading...
-
Total cost per month
Loading...
-
Total cost
Loading...
-

Prices shown include all applicable discounts

Total cost per month
Loading...
-
Total cost
Loading...
-

Basic Product Information

NVIDIA H100 NVL
Tensor Core GPU

Product Name

NVIDIA H100 NVL Tensor Core GPU

Architecture

NVIDIA Hopper™

Memory

94GB HBM3

Compute Power

Up to 3,341 TFLOPS FP8 performance

Use Cases

LLM inference, AI training, large language models, high-performance computing (HPC), scientific computing

Price

From $2.29/hr

Key Advantages

Up to 4X Faster Training

Enhanced AI model training with FP8 precision.

Transformer Engine

Optimized for large language models like GPT-3.

60 TFLOPS FP64 Performance

Accelerates scientific and high-performance computing tasks.

Up to 30X Faster AI Inference

Delivers up to 30X performance increase on large models like Megatron 530B.

7X Performance for HPC Applications

Ideal for tasks like genome sequencing and 3D FFT.

Specifications

Performance Specifications

LLM Inference Performance

Up to 5X over NVIDIA A100 systems for LLMs up to 70 billion parameters

FP64

30 teraFLOPS

FP64 Tensor Core

60 teraFLOPS

FP32

60 teraFLOPS

TF32 Tensor Core

835 teraFLOPS (with sparsity)

BFLOAT16 Tensor Core

1,671 teraFLOPS (with sparsity)

FP16 Tensor Core

1,671 teraFLOPS (with sparsity)

FP8 Tensor Core

3,341 teraFLOPS (with sparsity)

INT8 Tensor Core

3,341 TOPS (with sparsity)

Decoders

7 NVDEC, 7 JPEG

Confidential Computing

Supported

Multi-Instance GPUs

Up to 7 MIGs @12GB each

Memory and Bandwidth

GPU Memory

94GB HBM3

Memory Bandwidth

3.9TB/s

Memory Clock

2619 MHz

Thermal and Power

Power

Configurable between 350-400W

Server Options

Compatible with partner and NVIDIA-Certified Systems that can accommodate 1 to 8 GPUs

Board Specifications

Form Factor

PCIe dual-slot air-cooled

Interconnect

NVIDIA NVLink: Offers 600GB/s bidirectional bandwidth
PCIe Gen5: Supports up to 128GB/s
NVLink Bridge: Can be connected with another H100 NVL using 2- or 4-way NVLink bridges for increased bandwidth

Supported Technologies

Multi-Instance GPU (MIG)

Up to 7 MIGs @ 12GB each

Confidential Computing

Provides hardware-based security for data in use.

AI Enterprise Software

NVIDIA AI Enterprise included for secure and scalable AI deployment.

Server Compatibility

NVL

NVL: Compatible with Partner and NVIDIA-Certified Systems with 1-8 GPUs
NVIDIA AI Enterprise: Included subscription

Additional Features

01

Transformer Engine

Uses mixed FP8 and FP16 precision to dramatically accelerate training and inference for large AI models.

02

NVLink Switch System

Scales multi-GPU communication up to 900GB/s, over 7X faster than PCIe Gen5.

03

Dynamic Programming (DPX) Instructions

Accelerates tasks like disease diagnosis and routing optimization by 7X compared to previous generations.

04

NVIDIA AI Enterprise Add-on

Included with the H100 NVL, providing access to a suite of AI tools and frameworks.

05

Single Root I/O Virtualization (SR-IOV)

Supported with up to 32 virtual functions.

06

Secure Boot (CEC)

Ensures secure boot and firmware updates.

07

Programmable Power

Allows configuration of the power cap using nvidia-smi or SMBPBI tools.

08

Bidirectional Airflow Support

Heat sink accepts airflow in either left-to-right or right-to-left directions

More Products

01NVIDIA H200

$2.95/GPU/hr

02NVIDIA L40s

From $0.99/hr

03NVIDIA A40

$0.50 /GPU/hr

04AMD MI300X

$2.50 /GPU/hr

Want to learn more?

×
By clicking the "submit" button, you agree to and accept our Terms & Conditions and Privacy Policy .
×
By clicking the "submit" button, you agree to and accept our Terms & Conditions and Privacy Policy .
×
By clicking the "submit" button, you agree to and accept our Terms & Conditions and Privacy Policy .
×
By clicking the "submit" button, you agree to and accept our Terms & Conditions and Privacy Policy .