TNS
VOXPOP
As a JavaScript developer, what non-React tools do you use most often?
Angular
0%
Astro
0%
Svelte
0%
Vue.js
0%
Other
0%
I only use React
0%
I don't use JavaScript
0%
NEW! Try Stackie AI
AI / AI Operations / Hardware

Developers Can Now ‘Uber’ GPUs With NVIDIA’s Lepton Platform

NVIDIA's new one-stop shop, called DGX Cloud Lepton, will allow developers to rent NVIDIA GPUs scattered worldwide across host providers.
May 20th, 2025 11:30am by
Featued image for: Developers Can Now ‘Uber’ GPUs With NVIDIA’s Lepton Platform

Do you desperately need an NVIDIA GPU? Don’t stress, you’ll soon be able to order one online.

NVIDIA has announced a one-stop shop called DGX Cloud Lepton, which will allow developers to rent NVIDIA GPUs scattered worldwide across host providers.

Developers can log into Lepton, specify their AI task, and rent a relevant GPU. Using the Uber analogy, GPUs will come to developers who don’t have time to go and find one.

“Similar to modern ride-sharing apps that connect riders to drivers, DGX Cloud Lepton provides a modern marketplace that connects developers to GPU compute and not just locally,” said Alexis Black Bjorlin, vice president of NVIDIA’s DGX Cloud, in a press briefing.

GPU availability isn’t a grave problem, but searching for the right service across dozens of cloud services or independent facilities can be a headache.

A host of NVIDIA GPUs will be available to order — including Blackwell, its latest GPU. Older GPUs, including Hopper (on which ChatGPT runs) and possibly Ampere, will also be available.

The GPUs will depend on the job. If users need large-scale training for models, it may be Blackwell. If users need inferencing, a Hopper or Ampere may be assigned. If users need low latency, then a GPU in a nearby region will be assigned.

The Problem: More GPU Suppliers

NVIDIA GPUs were in short supply up until last year. Microsoft, OpenAI, Amazon and Meta gobbled up most of the supplies. Even Apple was unable to acquire parts, which put them years behind in AI, according to a Bloomberg story this week.

Cloud providers initially charged a premium for access to NVIDIA GPUs. These were only available to large customers.

But as supply normalized, a bunch of independent GPU providers popped up to provide AI computing at cheaper prices. Most of these had previously mined cryptocurrency, but pivoted their hardware to AI as that market took off. Companies like Crusoe Energy, which sold its bitcoin mining business in March, acquired new GPUs and started renting them out as AI infrastructure. Another example is Voltage Park, which was founded by a crypto executive but now offers one-time H100 GPUs at $1.99 per hour.

NVIDIA’s Black Bjorlin acknowledged GPU capacity had grown, adding “finding and efficiently utilizing AI infrastructure in the right regions at high performance can be complex.”

NVIDIA’s Uber-Like Idea

Delivery of GPUs from Lepton will be from lesser known GPU providers — including CoreWeave, Crusoe, Lambda, Foxconn, and others. It’s not yet clear if users will be able to select the provider.

Lepton does not include the top-4 cloud providers, but users will be able “to bring their compute and the platforms that they currently use, so it’s certainly an option on this platform,” Bjorlin said.

Reading between the lines, NVIDIA may be creating its own cloud service and will compete with the likes of Amazon, Google and Microsoft, which also rent out GPUs.

Bjorlin also indicated that Lepton will be hybrid cloud-friendly, meaning that developers may be able to connect their data or AI workloads to rental GPUs at smaller providers.

Cheaper Alternatives

Pricing for Lepton wasn’t available, but the service requires some NVIDIA software and tooling, which isn’t cheap.

To avoid the NVIDIA premium, it may be cheaper to rent GPUs directly from smaller service providers that are already a part of Lepton.

A fully loaded NVIDIA H200 GPU from Crusoe Energy, for a single instance over a six-month term, will set users back about $120,000. A slower A100 will be about half the price.

The price of the H200 from CoreWeave is about $50 per hour.

Lepton’s strength is in its software backend, which takes care of the NVIDIA software and hardware development stack. Developers can take their programs directly to GPUs and cut off the development and microservices complexity in the middle.

Alternatively, it may be better to get hardware that runs locally — since the processing requirements for AI are also shrinking.

NVIDIA will soon start shipping DGX Spark — which delivers one petaflop of performance and can run inference on desktops. The hardware was described as “your own AI cloud sitting right next to you and it’s always on, always waiting for you,” by NVIDIA CEO Jensen Huang during a keynote.

How Lepton Will Work

Developers first need to sign up as an NVIDIA developer and create a cloud server. After that, developers can apply for access to the service.

My application for the service is currently under consideration. NVIDIA seems choosy in selecting users, limiting approvals to users who can properly exploit its GPUs and with money to spend. My applications for beta AI services have never been approved — nevertheless, developers should give it a shot.

If you get access, you can adopt LLMs or AI applications from NVIDIA’s Build website and deploy them to Lepton. It is possible to be granular and fine-tune the task for deployment to computing nodes.

Users can select the type of job and the GPU, establish the container with the development environment, and dispatch the AI job to GPUs. A training job will go to high-end GPUs, while inference will be sent to lower-end GPUs.

Developers can configure the GPU and manage multiple machines within the Lepton interface. That includes selecting a container image, selecting the instance type (CPU or GPU), and establishing the variables. NVIDIA provides templates for MPI and Torchrun to get started.

Users can also run Jupyter notebooks within containers. The pod can be accessed via SSH, browser, or through other tools. Lepton organizes and manages the nodes depending on the job and type of GPU.

Developers can bring their own hardware to integrate into Lepton for more self-managed nodes. But the hefty list of requirements includes 640GB of storage per GPU (recommended: a 20TB NVMe SSD), high-end x86 server CPUs, Ubuntu 22.04 LTS, and CUDA 12.4.1 or beyond.

Created with Sketch.
TNS DAILY NEWSLETTER Receive a free roundup of the most recent TNS articles in your inbox each day.