UKAI

Nvidia launches DGX Cloud Lepton to unify global GPU infrastructure for AI

Nvidia has unveiled DGX Cloud Lepton, a global platform designed to unify thousands of distributed GPUs into a single next-generation AI infrastructure. The launch follows Nvidia’s acquisition of Chinese startup Lepton AI earlier this year, co-founded by Yancing Jia, former Vice President of Technology at Alibaba, and Junjie Bai. Both have joined Nvidia to lead the project.

DGX Cloud Lepton offers AI developers a unified entry point into a global ecosystem of GPUs, removing geographical and provider limitations. It mirrors the early cloud revolution in storage by offering seamless and scalable access to compute power, regardless of underlying infrastructure.

The move comes as the cloud GPU market expands rapidly. Valued at $3.17 billion in 2023, it is expected to reach nearly $47 billion by 2032, with GPU as a Service forecast to grow from $5 billion in 2025 to $32 billion by 2034. Nvidia’s platform consolidates compute from major providers including AWS, CoreWeave, Lambda and Microsoft Azure, all accessible through a single software interface built on Nvidia’s AI stack.

This stack integrates key tools: NVIDIA Inference Microservices for rapid inference; NeMo for building and fine-tuning large models; Blueprints for assembling applications; and Cloud Functions for deploying modular AI tasks. Together, they offer a predictable and flexible environment for AI development.

DGX Cloud Lepton supports three workflows: Dev Pods for experimentation with tools like Jupyter Notebooks and VS Code; Batch Jobs for training and dataset generation across nodes with real-time monitoring; and Inference Endpoints for scalable API deployment with automatic resource scaling.

The platform also offers advanced features including observability tools, automated scaling, workspace customisation and regulatory compliance, making it suitable for startups, enterprises and government projects requiring secure and scalable infrastructure.

Nvidia is positioning DGX Cloud Lepton not as a rival to hyperscale providers, but as an integrative layer unifying their GPU resources into an open AI compute marketplace. Key partnerships with AWS, Microsoft Azure, CoreWeave and Hugging Face support this approach. Hugging Face provides training clusters for open model development, while venture-backed startups receive GPU credits and technical support to accelerate innovation.

DGX Cloud Lepton also responds to regional regulatory demands, offering localised GPU access to support data sovereignty and compliance, particularly in Europe and other regulated markets.

By consolidating global GPU capacity under a single interface, DGX Cloud Lepton could reshape AI infrastructure access. It supports ambitions to position the UK and other regions as leaders in responsible AI by enabling developers to scale efficiently without infrastructure barriers.

With demand for AI compute power growing, Nvidia’s platform signals a shift towards open, compliant and globally connected infrastructure—one that could accelerate the next wave of AI innovation.

Created by Amplify: AI-augmented, human-curated content.