GPU


Simple and Affordable Pricing

All instances are Dedicated GPUs and backed with NVMe storage.

Pricing Table

Caution

To gain access to our GPUs, please reach out to us at [email protected] to request an increase in your quota.

Did you know?

Each GPU Instance has 55GB of disk per vCPU

eg: 16 * 55GiB = 880 GiB

L40s class

These instances are powered by Nvidia L40s. These instances are good for running LLMs using GPU inference like llama3.2-vision:11b-instruct-fp16, nemotron:70b-instruct-q4_K_M, qwen2.5-coder:32b-base-q4_K_M, etc.

Did you know?

It is better to use L40s for models which fit within 48GiB. Doing multi GPU training and Inference is not recommended. It is better to use MI300X or H100-SXM for multi GPU training or inference.

InstancevCPUMemoryvRAMReservedOn-Demand
l40s.1x16 vCPU128 GiB48 GiB$0.80/hr$1.55/hr
l40s.2x32 vCPU256 GiB96 GiB$1.60/hr$3.10/hr
l40s.4x64 vCPU512 GiB192 GiB$3.20/hr$6.20/hr

MI300X class

These instances are powered by AMD MI300X. These instances are good for running LLMs using Multi GPU inference like llama3.2-vision:11b-instruct-fp16, nemotron:70b-instruct-q4_K_M, qwen2.5-coder:32b-base-q4_K_M, etc.

InstancevCPUMemoryvRAMReservedOn-Demand
mi300x.1x16 vCPU128 GiB192 GiB$2.90/hr$5.62/hr
mi300x.8x128 vCPU1024 GiB1536 GiB$23.00/hr$29.90/hr

H100 class

These instances are powered by Nvidia H100 SXM. These instances are good for running LLMs using Multi GPU inference like llama3.2-vision:11b-instruct-fp16, nemotron:70b-instruct-q4_K_M, qwen2.5-coder:32b-base-q4_K_M, etc.

InstancevCPUMemoryvRAMReservedOn-Demand
h100.1x16 vCPU128 GiB80 GiB$2.00/hr$3.88/hr
h100.8x128 vCPU1024 GiB640 GiB$16/hr$20.8/hr