Log in Rent RTX 5090
RTX 5090 · live marketplace — from $0.39 / hr

Rent an RTX 5090
by the minute.
Not the month.

Rent an RTX 5090 for the only consumer card with 32 GB and native FP4. 32 GB GDDR7 at 1.79 TB/s, 21,760 Blackwell cores — fits Llama-3 13B FP16 single-card with 64-request KV cache, runs Flux batch-4 at ~5.8 it/s, generates native 720p Hunyuan video without offload latency. Spun up in under 90 seconds, billed per-minute, paid in BTC, USDT/USDC or CLORE. The consumer ceiling.

Per-minute billing SSH + Docker + Jupyter Spot & on-demand 21 regions
SHELL ~/projects/finetune
# install the official Python SDK + CLI $ pip install clore-ai $ clore search --gpu "RTX 5090" --max-price 3.0 47 servers · cheapest spot $0.39/hr · cheapest on-demand $0.62/hr $ clore deploy 64210 --image cloreai/ubuntu22.04-cuda12 --type on-demand --currency bitcoin order #82144 created · waiting for boot… running · ssh ready $ clore ssh 82144
GPU
RTX 5090 ×1
VRAM
32 GB
Rate
$0.62/hr
Status
Running
$0.39/hr
Starting on-demand price
32GB
GDDR7 VRAM per card
21
Regions with 5090 hosts
<90s
Cold-start to ready
workloads

Built for the workloads
that broke your last GPU.

5090s eat 70B-parameter inference and 4K-tile diffusion for breakfast. Pick the template, rent for the hours you need, walk away with a checkpoint.

32 GB on a consumer card, FP4 native

Only consumer GPU that fits 13B FP16 single-card with proper KV cache headroom. 1.79 TB/s GDDR7 and Blackwell FP4 paths give roughly 1.4× a 4090 on Flux production. The card that lets indie devs ship workloads they used to need an A6000 for.

Llama-3.3 70B QLoRA ~3.4× vs 4090

Diffusion & rendering

SDXL, Flux, and HunyuanVideo at native resolution. Blender Cycles with OptiX turning out 4K frames at 22 s/frame on a single card.

SDXL 1024² batch-8 ~2.7× vs 4090

Inference at scale

vLLM and TensorRT-LLM containers ship pre-tuned for Blackwell. 32 GB VRAM means 70B models on a single GPU — no tensor-parallel headache.

Llama-3.3 70B INT4 140 tok/s/user
why 5090

A generational jump,
at a per-minute rate.

Specs vs. what you've probably been renting. All numbers from Nvidia's reference spec sheets; pricing is the lowest on-demand rate live in the marketplace right now.

RTX 5090 RTX 4090 A100 80GB H100 80GB
Architecture Blackwell Ada Lovelace Ampere Hopper
CUDA cores 21,760 16,384 6,912 14,592
VRAM 32 GB GDDR7 24 GB GDDR6X 80 GB HBM2e 80 GB HBM3
Memory bandwidth 1,792 GB/s 1,008 GB/s 1,935 GB/s 3,350 GB/s
FP16 / BF16 (dense) ~210 TFLOPS ~165 TFLOPS 312 TFLOPS 756 TFLOPS
From / hr (on-demand) $0.39 $0.31 $1.20 $2.40

// prices are spot-market lows · refreshed every 60 s

pricing

Two ways to rent.
Pay only for the minutes you use.

Every server is priced by its host. These are the live floors across the marketplace — you'll see hundreds of variants once you're in.

Spot

$0.39 / hr
≈ 0.0000035 BTC · 260 CLORE
  • Lowest possible rate
  • Per-minute billing
  • Can be interrupted by on-demand renter
  • Best for batch training, rendering
Browse spot 5090s
MOST RENTED

On-demand

$0.62 / hr
≈ 0.0000056 BTC · 413 CLORE
  • Guaranteed availability
  • No preemption, ever
  • Per-minute billing
  • Best for inference, dev work, demos
Rent on-demand
Pay with
Bitcoin on-chain
CLORE native token
USDT / USDC ERC-20 · BEP-20
workflow

Four steps to a running 5090.

No sales call. No quota request. No three-week procurement. The first four commands are all you need.

01 / FILTER

Pick your card

Filter the marketplace by RTX 5090, country, GPU count, reliability score, network speed.

02 / RENT

Click rent

Choose a Docker image — PyTorch, vLLM, ComfyUI, Blender — or paste your own.

$ clore rent --gpu "RTX 5090"
03 / CONNECT

SSH or Jupyter

You get a public endpoint, an SSH key, and Jupyter on port 8888 in under 90 s.

04 / STOP

Stop anytime

Per-minute billing rounds to the second. Stop the instance and the meter stops with it.

faq

Questions hosts and renters ask.

Is the 5090 worth the premium over a 4090?

If you need >24 GB on a consumer card, yes — the 5090's 32 GB GDDR7 fits Llama-3 13B FP16 in single-card memory and runs ~1.4× transformer training throughput vs 4090. For 24 GB-or-less workloads, the 4090 is still the better $/throughput pick.

What can I actually run on a consumer GPU on CLORE?

Consumer cards on CLORE.AI cover most hobby and indie workflows: Stable Diffusion 1.5 and SDXL, ComfyUI/Automatic1111, Flux.1, LoRA and QLoRA fine-tuning of 7B-13B LLMs, Whisper transcription, video transcoding, Blender Cycles, and game-server hosting. Anything that fits in 8-32 GB VRAM and runs in Docker runs here. You get full root SSH plus a Jupyter template if you want one.

How fast does a rented server actually boot?

Cold-start lands in roughly 60-90 seconds for a typical Docker image: server allocation, container pull, GPU passthrough, SSH up. Pre-cached templates (PyTorch, ComfyUI, vLLM, Ollama) are faster because the image is already on the host. Once running you pay per minute, so a 10-minute experiment costs ten minutes of rental, not an hour.

Spot vs on-demand - what's the difference?

On-demand is a fixed per-hour price the host sets; the rental cannot be revoked while you have funds. Spot is auction-style: you bid, the highest bidder runs, and a higher bidder can preempt you. Spot is typically 30-50% cheaper. CLORE.AI charges 2.5% on spot and 10% on on-demand, split 50/50 with the host.

Is CLORE.AI cheaper than RunPod or Vast.ai?

Spot prices on CLORE.AI usually beat RunPod community pricing because there is no centralized markup; you rent directly from the host with a 2.5% spot fee. Vast.ai is the closest comparison, and on consumer cards CLORE.AI is generally within a few cents per hour. Hold CLORE in your wallet for Proof of Holding and you stack up to 50% off the marketplace fee.

Can I bring my own Docker image and SSH key?

Yes. Point at any registry - Docker Hub, GHCR, Quay, your private registry - then set env vars, port forwards, and your SSH public key in the rent dialog. Templates on the platform are just preset configs; nothing is locked down. You get full root inside the container with GPU passthrough.

workload spotlight

Real numbers on the RTX 5090.

32 GB GDDR7 at 1.79 TB/s — the only consumer card that fits Llama-3 13B FP16 single-card with native FP4 throughput.

Llama-3 13B FP16 single-card
vLLM + Blackwell FP4 KV cache
~3,200 tok/s aggregated, 64 concurrent

32 GB fits 13B FP16 weights + 64-request KV cache; FP4 KV is a Blackwell-only optimization.

Read the guide →
Flux.1 dev batch-4 FP4
ComfyUI + Blackwell FP4
~5.8 it/s @ 1024² batch 4

~1.4× a 4090 on Flux production thanks to GDDR7 bandwidth and FP4 tensor paths.

Read the guide →
Hunyuan-Video native 720p
ComfyUI + sequence parallel + fp8
~5 min per 5 s @ 720p, no offload

32 GB removes offload latency — single-card generative video at production cadence.

Read the guide →
consumer comparison

Consumer-tier comparison.

Side-by-side specs across the consumer tier. Click any row to see that GPU.

GPU
VRAM
CUDA cores
FP16 TFLOPS (tensor, dense)
Mem BW (GB/s)
Spot $/hr
SDXL 1024² it/s
Llama-3 8B tok/s
RTX 3070
8 GB GDDR6
5,888
~80
448
$0.10
~1.4
~50
RTX 3080
10 GB GDDR6X
8,704
~119
760
$0.14
~2.0
~85
RTX 3090
24 GB GDDR6X
10,496
~142
936
$0.18
~3.0
~110
RTX 4070
12 GB GDDR6X
5,888
~117
504
$0.16
~2.5
~60
RTX 4080
16 GB GDDR6X
9,728
~195
716
$0.27
~4.5
~95
RTX 4090 tier focus
24 GB GDDR6X
16,384
~165
1,008
$0.31
~7.5
~125
RTX 5080
16 GB GDDR7
10,752
~225
960
$0.28
~5.5
~115
RTX 5090 / this page
32 GB GDDR7
21,760
~419
1,792
$0.39
~10.0
~180
RTX 4070 Ti
12 GB GDDR6X
7,680
~160
504
$0.20
~3.2
~75
workload guides

Run these on your rented RTX 5090.

Step-by-step guides verified on CLORE.AI hardware. Pick a workload, copy the docker image, ship in minutes.

Image Generation
Flux.1 on CLORE.AI
Run Black Forest Labs' Flux for state-of-the-art image gen.
Language Models
vLLM serving
High-throughput LLM serving with PagedAttention.
Training
LLM fine-tuning
LoRA / QLoRA fine-tuning workflow.
Video Generation
Wan Video
Alibaba's Wan-2.1 text/image-to-video.
Video Generation
Hunyuan Video
Tencent's open video generation model.
Training
Kohya SS LoRA training
The standard SDXL LoRA training pipeline.
Comparisons
llm serving comparison
See all guides →
other gpus

Compare with similar cards.

RTX 4090
24 GB · from $0.31/hr
Rent →
RTX 6000 Ada
48 GB · from $0.55/hr
Rent →

Your training run
is 90 seconds away.

Hosts around the world are accepting workloads right now. Sign up, top up your wallet, and the next hour is yours.