We’ve collaborated with AMD to deliver select ONNX-optimized versions of the Stable Diffusion model family, engineered to run faster and more efficiently on AMD Radeon™ GPUs and Ryzen™ AI APUs. This joint engineering effort focused on maximizing inference performance without compromising model output quality or our open licensing.
The result is a set of accelerated models that integrate into any ONNX Runtime-supported environment, making it easy to drop them into your existing workflows right out of the box. Whether you’re deploying Stable Diffusion 3.5 (SD3.5) variants, our most advanced image model, or Stable Diffusion XL Turbo (SDXL Turbo), these models are ready to power faster creative applications on AMD hardware.
As generative visual media adoption accelerates, it’s essential our models are optimized for leading hardware. This collaboration ensures builders and businesses can integrate Stable Diffusion into their production pipelines, making workflows faster, more efficient, and ready to scale.
Available models
AMD has optimized four models across SD3.5 and SDXL for improved performance.
SD3.5 Version:
AMD-optimized SD3.5 models deliver up to 2.6x faster inference when compared to the base PyTorch models.
SDXL Version:
With AMD optimization, SDXL 1.0 and SDXL Turbo achieve up to 3.8x faster inference, when compared to the base PyTorch models.