Text-to-Image
GGUF
stable-diffusion-xl
quantization
unet
vae
clip

SDXL GGUF Quantized Model

This repository contains a quantized version of Stable Diffusion XL in the GGUF format. The model has been converted to different quantization levels, including Q4_K_S, Q5_K_S, and Q8, allowing for flexible deployment based on hardware capabilities. The UNet, VAE, and CLIP components are provided separately for better optimization and compatibility.

Quantization Details

Component Available Quantization
UNet Q4_K_S, Q5_K_S, Q8
VAE FP16
CLIP FP16

Files & Structure

  • sdxl-unet-q4_ks.gguf
  • sdxl-unet-q5_ks.gguf
  • sdxl-unet-q8.gguf
  • sdxl-vae-fp16.safetensors
  • sdxl-clip-fp16.safetensors

Each quantization level offers a trade-off between speed and quality. Q4_K_S provides the highest speed but lower quality, while Q8 retains more details with higher VRAM usage.

Usage

This model can be used with any GGUF-compatible inference engine, such as ComfyUI, Kohya's SDXL GGUF loader, or custom scripts supporting GGUF-based SDXL inference.

Hardware Requirements

  • Q4_K_S: Suitable for low-VRAM environments (2GB+)
  • Q5_K_S: Balanced performance and quality (3GB+ VRAM recommended)
  • Q8: Best quality, requires higher VRAM (4GB+ recommended)

Acknowledgments

This model is based on Stable Diffusion XL by Stability AI and has been quantized for improved accessibility across various hardware configurations.

For support and discussions, feel free to open an issue or reach out on HF中国镜像站 forums!


Downloads last month
258
GGUF
Model size
2.57B params
Architecture
sdxl

4-bit

5-bit

8-bit

Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for HyperX-Sentience/SDXL-GGUF

Quantized
(13)
this model