20 people are viewing this right now
Estimated Shipping Time: 30-40 Days
Hurry Up! Only Few Left in Stock!
In Stock
NVIDIA H200
nvidia
NVIDIA H200
20 people are viewing this right now
Hurry Up! Only Few Left in Stock!
In Stock
NVIDIA H200
nvidia
NVIDIA H200
The NVIDIA H200 GPU supercharges generative AI and high-performance computing (HPC) workloads with game-changing performance and memory capabilities. As the first GPU with HBM3E, the H200’s larger and faster memory fuels the acceleration of generative AI and large language models (LLMs) while advancing scientific computing for HPC workloads.
In the ever-evolving landscape of AI, businesses rely on LLMs to address a diverse range of inference needs. An AI inference accelerator must deliver the highest throughput at the lowest TCO when deployed at scale for a massive user base.
The H200 boosts inference speed by up to 2X compared to H100 GPUs when handling LLMs like Llama2.
Memory bandwidth is crucial for HPC applications as it enables faster data transfer, reducing complex processing bottlenecks. For memory-intensive HPC applications like simulations, scientific research, and artificial intelligence, the H200’s higher memory bandwidth ensures that data can be accessed and manipulated efficiently, leading up to 110X faster time to results compared to CPUs.
With the introduction of the H200, energy efficiency and TCO reach new levels. This cutting-edge technology offers unparalleled performance, all within the same power profile as the H100. AI factories and supercomputing systems that are not only faster but also more eco-friendly, deliver an economic edge that propels the AI and scientific community forward.

NVIDIA H200 NVL is ideal for lower-power, air-cooled enterprise rack designs that require flexible configurations, delivering acceleration for every AI and HPC workload regardless of size. With up to four GPUs connected by NVIDIA NVLink™ and a 1.5x memory increase, large language model (LLM) inference can be accelerated up to 1.7x, and HPC applications achieve up to 1.3x more performance over the H100 NVL.
Specifications:
| H200 NVL | |
| FP64 | 30 TFLOPS |
| FP64 Tensor Core | 60 TFLOPS |
| FP32 | 60 TFLOPS |
| TF32 Tensor Core | 835 TFLOPS |
| BFLOAT16 Tensor Core | 1,671 TFLOPS |
| FP16 Tensor Core² | 1,671 TFLOPS |
| FP8 Tensor Core | 3,341 TFLOPS |
| INT8 Tensor Core | 3,341 TFLOPS |
| GPU Memory | 141GB |
| GPU Memory Bandwidth | 4.8TB/s |
| Decoders | 7 NVDEC 7 JPEG |
| Confidential Computing | Supported |
| Max Thermal Design Power (TDP) | Up to 600W (configurable) |
| Multi-Instance GPUs | Up to 7 MIGs @16.5GB each |
| Form Factor | PCIe Dual-slot air-cooled |
| Interconnect | 2- or 4-way NVIDIA NVLink bridge: 900GB/s per GPU PCIe Gen5: 128GB/s |
| Server Options | NVIDIA MGX H200 NVL partner and NVIDIA-Certified Systems with up to 8 GPUs |
| NVIDIA AI Enterprise | Included |
All orders are processed within 1-3 business days. Orders are not shipped or delivered on weekends or holidays. If we are experiencing a high volume of orders, shipments may be delayed by a few days. Please allow additional days in transit for delivery. If there will be a significant delay in shipment of your order, we will contact you via email or telephone.
Shipping charges for your order will be calculated and displayed at checkout. Delivery delays can occasionally occur.
You will receive a Shipment Confirmation email once your order has shipped containing your tracking number(s). The tracking number will be active within 24 hours.
Vishal Peripherals is liable for any products damaged or lost during shipping. If you received your order damaged, please take the video of unboxing and share through mail at support1@vishalperipherals.com.
Wishlist is empty.