Stable Diffusion Models by Olive for OnnxRuntime CUDA
Collection
Stable Diffusion ONNX Models optimized by Olive for OnnxRuntime CUDA execution provider
•
7 items
•
Updated
This repository hosts the optimized versions of DreamShaper XL Turbo to accelerate inference with ONNX Runtime CUDA execution provider.
The models are generated by Olive with command like the following:
python stable_diffusion_xl.py --provider cuda --optimize --model_id Lykon/dreamshaper-xl-turbo
Base model
Lykon/dreamshaper-xl-turbo