Skip to main content

GPU Product

Rent NVIDIA V100 32GB GPUs

32 GB of HBM2 at 900 GB/s — the cost-leader for memory-bound AI fine-tuning, HPC, and batch inference. From $0.29/hour on sustainable, second-life Volta silicon.

Technical Specifications

ArchitectureNVIDIA Volta
Memory Size32 GB HBM2
Memory Bandwidth900 GB/s
CUDA Cores5 120
Tensor Cores640 (1st gen)
NVIDIA V100 32GB SXM datacenter GPU module
Rental Options

V100 Rental Options

Rent NVIDIA V100 32GB GPUs from $0.29/hour. Hourly billing with no minimum commitment, plus 1-month and 3-month reservations for committed workloads. Capacity sourced from sustainable second-life datacenter hardware.

Comparison

V100 32GB vs L40

Both are datacenter GPUs in the CloudRift fleet, positioned for different workloads. The V100 wins on price (roughly half the hourly rate), HBM2 memory bandwidth, FP64 throughput, and NVLink for multi-GPU scale-out. The L40 wins on raw FP32, modern Ada Tensor Cores, and total VRAM — choose it when you need maximum inference throughput per card.

V100 32GBL40% Diff
CloudRift Price$0.29 / hr$0.63 / hr−54%
ArchitectureVoltaAda LovelaceN/A
Memory TypeHBM2 ECCGDDR6 ECCN/A
VRAM32 GB48 GB−33.3%
Bus Width4 096-bit384-bit+966%
Memory Bandwidth900 GB/s864 GB/s+4.2%
FP64 Performance~7.8 TFLOPS~1.4 TFLOPS+457%
FP32 Performance~15.7 TFLOPS~90.5 TFLOPS−82.7%
Tensor Performance~125 TFLOPS~362 TFLOPS−65.5%
CUDA Cores5 12018 176−71.8%
Tensor Cores640 (1st gen)568 (4th gen)+12.7%
Multi-GPU InterconnectNVLink 2.0PCIe 4.0 onlyN/A
Form FactorSXM3Dual-slot PCIeN/A
TDP350 W300 W+16.7%

Performance

Key performance metrics

HBM2 Memory Bandwidth

900 GB/s of HBM2 bandwidth on a 4 096-bit bus — the cost-leader for memory-bound workloads like scientific compute, mid-sized LLM fine-tuning, and large-batch inference.

Tensor Core Acceleration

640 first-generation Tensor Cores deliver up to ~125 TFLOPS of mixed-precision throughput — proven silicon for CUDA, cuDNN, and mature ML toolchains.

A Fraction of Hyperscaler Pricing

$0.29/hr on CloudRift vs ~$3.90/hr per GPU on AWS p3dn.24xlarge and ~$2.75/hr per GPU on Azure NDv2 — up to 13× cheaper than hyperscalers for the same V100 32GB silicon.

NVIDIA V100 FAQ

Common Questions About the V100

V100 32GB rentals on CloudRift start at $0.29/hour for a single GPU. Pay-as-you-go, no minimum commitment.
The V100 is a Volta-architecture datacenter GPU used for AI fine-tuning, batch inference, and HPC workloads like CFD, molecular dynamics, and computational chemistry. Its 32 GB of HBM2 memory and strong FP64 performance make it cost-efficient for memory-bound and double-precision workloads.
The V100 32GB has 32 GB of HBM2 memory on a 4 096-bit bus with 900 GB/s of bandwidth — meaningfully higher than GDDR-class cards in the same price tier.
Yes — for the right workload. The V100 has 640 first-generation Tensor Cores and 32 GB of HBM2 memory, which makes it a strong fit for fine-tuning 7B–13B parameter models, batch inference, and memory-bandwidth-bound HPC. For training the largest frontier models you want H100/H200 or MI350X, but the V100 remains the cost-leader for many production workloads.
A V100 32GB on CloudRift is $0.29/hr. The same silicon on AWS p3dn.24xlarge runs ~$3.90/hr per GPU, and on Azure NDv2 ~$2.75/hr per GPU — CloudRift is roughly 10–13× cheaper. Capacity comes from sustainable second-life datacenter hardware, redeployed instead of replaced.
The V100 32GB is $0.29/hr — roughly half the L40’s $0.63/hr. It has HBM2 memory (900 vs 864 GB/s), a much wider 4 096-bit bus, NVLink 2.0 for multi-GPU, and about 5× the FP64 throughput (~7.8 vs ~1.4 TFLOPS). The L40 has 50% more VRAM (48 GB), newer Ada Tensor Cores, and roughly 6× the FP32 throughput, so it is the better pick when you need maximum inference throughput per card or modern Ada-only features. Pick the V100 for memory-bandwidth-bound workloads, FP64-heavy HPC, multi-GPU training, or the lowest hourly rate.
Yes. Open the Console, pick a V100 instance, and launch — most instances are running in under a minute. Hourly billing, no long-term commitment.
Get in touch

Ready to get started?

Get in touch with our team to discuss your requirements and find the right solution for your infrastructure.