Back to List
GGUF-Org Z-Image GGUF Models
Official GGUF quantized versions of Z-Image Turbo models for low-VRAM deployment (6GB+), optimized for ComfyUI with multiple quantization levels
GGUF
Quantization
Low VRAM
ComfyUI
6GB Compatible
Consumer GPU
Overview
GGUF-Org provides officially converted GGUF quantized versions of Z-Image Turbo, enabling deployment on consumer-grade GPUs with as low as 6GB VRAM. Supports multiple quantization levels including Q3_K_S, IQ4_NL, and IQ4_XS for different VRAM/quality trade-offs.
Features
- Multiple quantization levels: Q3_K_S (6GB), IQ4_NL/IQ4_XS (8GB)
- Compatible with ComfyUI via GGUF-Connector or ComfyUI-GGUF extension
- Supports various Qwen3 text encoder variants
- Memory-mapped file format for efficient loading
- Maintains image quality while reducing VRAM requirements by up to 70%
- Includes VAE and text encoder models
Installation
Download GGUF files and place in ComfyUI directories: - Z-Image model → ./ComfyUI/models/diffusion_models - Qwen3-4B encoder → ./ComfyUI/models/text_encoders - VAE → ./ComfyUI/models/vae
Usage
Use with ComfyUI GGUF extension or gguf-connector CLI tool. Run 'ggc z1' to select and interact with GGUF models.
Requirements
- 6GB+ VRAM (RTX 2060 or equivalent)
- 16GB+ system RAM
- ComfyUI with GGUF extension
- SSD storage (not HDD)