GPU Inference speed relative to 2x H100s (est) Speed / $ (relative est) Cost at Runpod Cost at FluidStack Cost at Lambda Labs 2x H100s 100% Not available Not available Not available in an on-demand 2x instance Not available in an on-demand 2x instance 2x 6000 Ada 48% 0.| gpus.llm-utils.org