Note that this list is aimed at cloud GPUs where more expensive GPUs are comparatively cheap vs buying the whole GPU outright.| gpus.llm-utils.org
GPU Inference speed relative to 2x H100s (est) Speed / $ (relative est) Cost at Runpod Cost at FluidStack Cost at Lambda Labs 2x H100s 100% Not available Not available Not available in an on-demand 2x instance Not available in an on-demand 2x instance 2x 6000 Ada 48% 0.| gpus.llm-utils.org