No free plan. The cheapest way in is Enterprise Reserved Instance at ~$36,999/month per instance. Consider free alternatives in the cloud infrastructure category if budget is tight.
NVIDIA DGX Cloud pricing starts at approximately $36,999 per instance per month for an 8-GPU node with H100 or A100 GPUs, based on initial Microsoft Azure listings. Pricing is sold on reserved terms (typically monthly or annual) rather than hourly on-demand billing. All plans include NVIDIA AI Enterprise software, Base Command orchestration, and direct access to NVIDIA AI experts. Actual pricing varies by cloud partner (OCI, Azure, Google Cloud, AWS), GPU generation, and term length, and is negotiated through NVIDIA or the cloud provider's enterprise sales team.
DGX Cloud provides dedicated access to NVIDIA's flagship data center GPUs, including the H100 Tensor Core GPU (80GB HBM3) and A100 80GB. Each DGX Cloud node includes 8 GPUs connected by NVLink for 640GB of total GPU memory and multi-node configurations are connected by NVIDIA Quantum-2 InfiniBand at 400 Gb/s. NVIDIA has also announced Blackwell-based GB200 and GB300 NVL72 rack-scale systems coming to DGX Cloud, which will further accelerate trillion-parameter model training. Unlike shared cloud GPU offerings, DGX Cloud nodes are reserved, not preemptible.
DGX Cloud is infrastructure-first and optimized for training foundation models, while AWS SageMaker and Google Vertex AI are end-to-end ML platforms with broader tooling for deployment, feature stores, and AutoML. DGX Cloud delivers higher raw GPU performance per dollar for large-scale training because it uses NVIDIA reference architecture with dedicated InfiniBand fabric â not virtualized multi-tenant GPUs. Based on our analysis of 870+ AI tools, teams training models over 70B parameters typically choose DGX Cloud, while teams focused on managed ML pipelines and inference at variable scale choose SageMaker or Vertex. DGX Cloud also runs inside Azure, Google Cloud, OCI, and AWS, so customers can retain existing cloud billing relationships.
NVIDIA does not offer a self-service free trial for DGX Cloud in the traditional sense, but enterprise prospects can request a proof-of-concept engagement through NVIDIA's sales team. Developers who want to experiment with the same NVIDIA AI Enterprise software stack can use NVIDIA LaunchPad, which provides short-term free access to curated labs on DGX-class hardware. The NVIDIA NGC catalog also offers free access to pre-trained models and containers that run on DGX Cloud. For production workloads, expect a formal procurement process rather than a credit card checkout.
DGX Cloud is the core reserved-capacity service offering dedicated H100/A100 multi-node instances with NVIDIA AI Enterprise software. DGX Cloud Lepton, announced in 2025, is a GPU marketplace that aggregates compute capacity from a global network of NVIDIA cloud partners (GPU clouds like CoreWeave, Lambda, Nebius, and others), giving developers a unified API to access GPUs across providers. Lepton is designed for developers who want flexibility and broader GPU availability, while DGX Cloud proper is for enterprises committing to dedicated infrastructure. NVIDIA also offers DGX Cloud Serverless Inference for pay-per-call model deployment built on top of the same infrastructure.
See NVIDIA DGX Cloud plans and find the right tier for your needs.
See Pricing Plans âStill not sure? Read our full verdict â
Last verified March 2026