Tokenized GPU cloud

Live fleet934 GPUs · 6 accelerator classes

The cloud for AI
infrastructure, made liquid.

Buy and trade compute tokens backed by AMD Instinct accelerators. Redeem for GPU sessions, API credits, or batch jobs — with the clarity of a modern AI platform.

AMD

Super Micro Computers

Micron Technology

Region

Lubbock, Texas

Low-latency GPU fabric

How it works

From tokens to production GPUs

Four steps — the same clarity you expect from a modern AI cloud.

01

Acquire Tokens

Purchase compute tokens backed by specific GPU hardware. Each token represents one GPU-hour of compute on that accelerator.

Buy Tokens
02

Redeem or Trade

Redeem tokens for live GPU sessions, API credits, or batch jobs. Or trade unused tokens on the open marketplace at market price.

Start Trading
03

Lock in Vaults

Lock tokens for 3, 6, or 12 months to earn yield, discounted compute rates, and priority queue access.

Explore Vaults
04

Deploy at Scale

Lease enterprise clusters with dedicated hardware, InfiniBand networking, and white-glove support for production workloads.

View Clusters

Infrastructure

We build AI-optimized, power-aware data centers

Lubbock Cloud couples West Texas grid dynamics with dense GPU footprints: liquid cooling paths, high-uptime networking, and interruptible workloads when markets demand it.

Tokens map to real racks — not synthetic credits. When you redeem, you land on hardware we operate, instrument, and support end to end.

Modern data center and network infrastructure

Case study

Tested with GenAI workloads

Teams run inference, fine-tuning, and batch pipelines on the same tokenized footprint — with metrics that match what you see in the marketplace.

Lubbock Cloud×ROCm inference

LLM inference where tokens meet telemetry

Goal: Give builders a single surface to buy capacity, deploy models, and trace cost per token in real time.

Approach: ROCm-native images, managed endpoints, and a job fabric that reports queue depth back to the dashboard you already use for balances.

  • Inference
  • ROCm
  • Tokens
  • Observability

Sub-100ms

target TTFT on tuned vLLM routes for 70B-class models on MI300X.

1 token

= 1 GPU-hour on the printed accelerator — no hidden conversion tables.

Full stack

API, Terraform-style provisioning hooks, and live job telemetry in one place.

We provide every essential resource for your AI journey

Hardware-backed tokens, a trading surface, and managed services — composed the way modern AI teams expect.

Data center server infrastructure

Latest AMD Instinct accelerators

MI210 through MI355X — matched 1:1 to liquid compute tokens.

Console

Cloud-native experience

$ lubbock tokens list --region=lub-1
LUB-MI300X 2.84 USD · 934 GPUs live
job submit --gpu MI325X --hours 8

Ready-to-run stack

Ship training and inference without assembling drivers by hand.

PyTorchvLLMROCmKubernetesSlurmRay

Fully managed services

Data and control planes you do not babysit.

  • PostgreSQL & Redis
  • Metrics & tracing
  • Secrets & API keys
  • Backup & snapshots
Engineers collaborating

Architects & expert support

Multi-node training, InfiniBand, and ROCm — covered by engineers who run the metal.

Enterprise

Dedicated clusters for inference & training

Pre-built topologies with reserved capacity, InfiniBand, and white-glove onboarding.

Request access →

Fleet

Accelerator availability

Live view of capacity across our Lubbock, Texas region.

AcceleratorArchitectureHBMFP16 TFLOPSTDPAvailability
AMDMI300X
CDNA 3192 GB HBM31,307750W
342/512
AMDMI325X
CDNA 3256 GB HBM3E1,307750W
128/256
AMDMI355X
CDNA 4288 GB HBM3E2,300800W
64/128
AMDMI250X
CDNA 2128 GB HBM2e383500W
198/320
AMDMI210
CDNA 264 GB HBM2e181300W
112/192
AMDMI300A
CDNA 3128 GB HBM31,307760W
88/160
AMDROCm-native

Managed ROCm services

Purpose-built AI services running natively on AMD Instinct hardware. No ROCm complexity — just APIs.

LLM Inference Endpoint

Deploy any open-weight LLM with vLLM on MI300X. Sub-100ms TTFT, automatic batching, OpenAI-compatible API.

MI300X$2.84/hr

Embeddings API

High-throughput text embeddings with BGE, E5, or custom models. Batch processing up to 10K docs/min.

MI300X$1.20/hr

Fine-Tuning Pipeline

Managed fine-tuning with LoRA, QLoRA, or full-parameter training. Weights stored in your vault.

MI325X$3.56/hr

Vector Database

GPU-accelerated vector search powered by FAISS on AMD hardware. Millisecond queries at billion-scale.

MI300X$0.95/hr

Image Generation

SDXL and Flux inference on ROCm. API and batch modes. Custom model uploads supported.

MI300X$2.10/hr

Video Generation

Coming Soon

Text-to-video and image-to-video pipelines on MI355X. Up to 4K resolution, 30fps output.

MI355X$5.12/hr

Get started

Own, trade, and deploy
GPU infrastructure

Stop renting by the hour. Build on tokenized compute with a product experience modeled for AI teams.