NVIDIA DGX Platform

Effortless Power for Your First Step in AI

The NVIDIA H100 Tensor Core GPU offers unprecedented speed, scalability, and versatility for advanced AI workloads. As part of the NVIDIA DGX platform, H100 is the foundation of enterprise-class AI infrastructure.

H100 GPUs for Every AI Need

We provide access to the widest range of NVIDIA GPUs in the industry, allowing you to scale according to the demands of your workloads.

Single GPU Instances

Ideal for research teams or startups looking to prototype AI models and fine-tune algorithms with access to industry-leading hardware.

Supercomputer Clusters

Designed for enterprises handling large-scale training and inference workloads. Use H100 clusters for tasks like GPT-3-level training and distributed neural networks.

Integrated with NVIDIA DGX

The NVIDIA H100 is part of the DGX platform, used in DGX SuperPODs to deliver AI at enterprise scale, with the flexibility to run on-premises or in managed data centers.

Up to 4X Higher AI
Training on GPT-3

H100 features fourth-generation Tensor Cores and a Transformer Engine with FP8 precision that provides up to 4X faster training over the prior generation for GPT-3 (175B) models. The combination of fourth-generation NVLink, which offers 900 gigabytes per second (GB/s) of GPU-to-GPU interconnect; NDR Quantum-2 InfiniBand networking, which accelerates communication by every GPU across nodes; PCIe Gen5; and NVIDIA Magnum IO™ software delivers efficient scalability from small enterprise systems to massive, unified GPU clusters.

Deploying H100 GPUs at data center scale delivers outstanding performance and brings the next generation of exascale high-performance computing (HPC) and trillion-parameter AI within the reach of all researchers.

Projected performance subject to change.  GPT-3 175B training A100 cluster: HDR IB network, H100 cluster: NDR IB network | Mixture of Experts (MoE) Training Transformer Switch-XXL variant with 395B parameters on 1T token dataset,  A100 cluster: HDR IB network, H100 cluster: NDR IB network with NVLink Switch System where indicated.
Deploying H100 GPUs at data center scale delivers outstanding performance and brings the next generation of exascale high-performance computing (HPC) and trillion-parameter AI within the reach of all researchers.

DEPLOY EFFORTLESSLY

Pre-configured environments and familiar tools

Enjoy the convenience of one-click Jupyter access and effortlessly connect to the GPU instances from your browser. Benefit from pre-installed popular ML frameworks like Ubuntu, TensorFlow, PyTorch, CUDA, and cuDNN, all readily available.

Exceptional performance cloud built for deep learning

We provide access to the widest range of NVIDIA GPUs in the industry, allowing you to scale according to the demands of your workloads. Our infrastructure, native to Kubernetes, ensures rapid deployment, dynamic auto-scaling, and a modern networking architecture that grows with your needs.

Experience bare metal performance through Kubernetes

Eliminate hypervisors from your stack by deploying containerized workloads. Enjoy the advantages of bare-metal without the hassle of managing infrastructure.

Get Started
Deep expertise in high performance computing

Deep learning and high-performance compute is at the heart of what we do, and our infrastructure reflects this. Our cloud service is designed to minimize your setup time and enhance performance.

Get Started