The NVIDIA H200 Tensor Core GPU, based on the NVIDIA Hopper™ architecture, is engineered to deliver unprecedented acceleration for generative AI and high-performance computing (HPC) workloads. Featuring 141GB of HBM3e memory and optimized for maximum performance deployments, it delivers up to 2X faster large language model inference and up to 110X faster HPC performance compared to CPUs, making it the ultimate choice for AI factories, supercomputing centers, and demanding enterprise AI deployments.
Pricing Calculator
Calculate your GPU cloud computing costs with our interactive pricing tool.
| Hourly Cost |
|
| Total cost per month |
|
| Total cost |
|
Prices shown include all applicable discounts
NVIDIA H200 Tensor Core GPU
NVIDIA Hopper™
141GB HBM3e
Up to 4 PetaFLOPS of FP8 performance
AI inference, large language models (LLMs), scientific computing, HPC workloads
$2.95/GPU/hr
Offers larger and faster memory for high-performance tasks.
World's highest memory bandwidth for data-intensive workloads.
Industry-leading FP8 performance for AI acceleration.
Perfect for large language models like Llama2.
Greater performance at the same power profile as the H100.