Install this extension via the ComfyUI Manager by searching
for ComfyUI-GGUF
1. Click the Manager button in the main menu
2. Select Custom Nodes Manager button
3. Enter ComfyUI-GGUF in the search bar
After installation, click the Restart button to
restart ComfyUI. Then, manually
refresh your browser to clear the cache and access
the updated list of nodes.
Visit
ComfyUI Online
for ready-to-use ComfyUI environment
ComfyUI-GGUF adds GGUF quantization support for native ComfyUI models, enabling efficient performance on low-end GPUs by using lower bits per weight variable bitrate quants, particularly for transformer/DiT models.
ComfyUI-GGUF Introduction
ComfyUI-GGUF is an extension designed to support GGUF quantization for native ComfyUI models. GGUF, a format popularized by llama.cpp, allows for efficient model storage and execution, particularly on lower-end GPUs. This extension is particularly useful for AI artists who want to run complex models on less powerful hardware without sacrificing too much performance.
Key Benefits:
Efficient Model Execution: Run transformer/DiT models with lower bits per weight variable bitrate quants, making it feasible to use on low-end GPUs.
Enhanced Compatibility: Supports models stored in the GGUF format, which is optimized for performance and storage efficiency.
Ease of Use: Integrates seamlessly with ComfyUI, allowing you to replace the standard "Load Diffusion Model" with the "Unet Loader (GGUF)" node in your workflows.
How ComfyUI-GGUF Works
ComfyUI-GGUF leverages the GGUF format to store and execute models more efficiently. GGUF quantization reduces the number of bits used to represent each weight in the model, which in turn reduces the model's size and the computational resources required to run it. This is particularly beneficial for transformer/DiT models, which are less affected by quantization compared to regular UNET models.
Simplified Explanation:
Quantization: Think of quantization as compressing a high-resolution image into a smaller file size without losing much detail. Similarly, GGUF quantization compresses the model weights, making them smaller and faster to process.
Model Loading: Instead of loading a large, uncompressed model, ComfyUI-GGUF loads a quantized model, which is quicker and requires less memory.
ComfyUI-GGUF Features
Unet Loader (GGUF)
Function: Loads models stored in the GGUF format.
Customization: You can place your .gguf model files in the ComfyUI/models/unet folder and use them directly in your workflows.
Example: Replace the standard "Load Diffusion Model" node with the "Unet Loader (GGUF)" node in your existing workflows to take advantage of GGUF quantization.
Usage: These models are pre-quantized and ready to use, saving you the effort of quantizing them yourself.
Important Note
CLIP Device Setting: The "Force/Set CLIP Device" is not part of this node pack. Avoid installing it if you only have one GPU, and do not set it to cuda:0 to prevent out-of-memory (OOM) errors.
ComfyUI-GGUF Models
ComfyUI-GGUF supports models stored in the GGUF format. Here are the available pre-quantized models:
flux1-dev GGUF: Ideal for development and experimentation.
flux1-schnell GGUF: Optimized for faster performance.
When to Use Each Model:
flux1-dev GGUF: Use this model when you are developing new features or experimenting with different settings.
flux1-schnell GGUF: Choose this model for production environments where speed is crucial.
Troubleshooting ComfyUI-GGUF
Common Issues and Solutions
Out-of-Memory (OOM) Errors:
Solution: Ensure you are not using the "Force/Set CLIP Device" if you have only one GPU. Avoid setting it to cuda:0.
Model Not Loading:
Solution: Verify that your .gguf model files are placed in the ComfyUI/models/unet folder. Ensure you are using the "Unet Loader (GGUF)" node in your workflow.
Frequently Asked Questions
Q: Can I use LoRA or Controlnet with GGUF models?
A: Currently, LoRA and Controlnet are not supported due to the weights being quantized.
Learn More about ComfyUI-GGUF
Additional Resources
llama.cpp: Learn more about the GGUF format and its benefits.
Community Forums: Join discussions and get support from other AI artists and developers.
By understanding and utilizing ComfyUI-GGUF, you can significantly enhance your AI art projects, making them more efficient and accessible even on lower-end hardware.