ComfyUI > Nodes > ComfyUI-GGUF

ComfyUI Extension: ComfyUI-GGUF

Repo Name

ComfyUI-GGUF

Author
city96 (Account age: 552 days)
Nodes
View all nodes(1)
Latest Updated
2024-08-18
Github Stars
0.25K

How to Install ComfyUI-GGUF

Install this extension via the ComfyUI Manager by searching for ComfyUI-GGUF
  • 1. Click the Manager button in the main menu
  • 2. Select Custom Nodes Manager button
  • 3. Enter ComfyUI-GGUF in the search bar
After installation, click the Restart button to restart ComfyUI. Then, manually refresh your browser to clear the cache and access the updated list of nodes.

Visit ComfyUI Online for ready-to-use ComfyUI environment

  • Free trial available
  • High-speed GPU machines
  • 200+ preloaded models/nodes
  • Freedom to upload custom models/nodes
  • 50+ ready-to-run workflows
  • 100% private workspace with up to 200GB storage
  • Dedicated Support

Run ComfyUI Online

ComfyUI-GGUF Description

ComfyUI-GGUF adds GGUF quantization support for native ComfyUI models, enabling efficient performance on low-end GPUs by using lower bits per weight variable bitrate quants, particularly for transformer/DiT models.

ComfyUI-GGUF Introduction

ComfyUI-GGUF is an extension designed to support GGUF quantization for native ComfyUI models. GGUF, a format popularized by llama.cpp, allows for efficient model storage and execution, particularly on lower-end GPUs. This extension is particularly useful for AI artists who want to run complex models on less powerful hardware without sacrificing too much performance.

Key Benefits:

  • Efficient Model Execution: Run transformer/DiT models with lower bits per weight variable bitrate quants, making it feasible to use on low-end GPUs.
  • Enhanced Compatibility: Supports models stored in the GGUF format, which is optimized for performance and storage efficiency.
  • Ease of Use: Integrates seamlessly with ComfyUI, allowing you to replace the standard "Load Diffusion Model" with the "Unet Loader (GGUF)" node in your workflows.

How ComfyUI-GGUF Works

ComfyUI-GGUF leverages the GGUF format to store and execute models more efficiently. GGUF quantization reduces the number of bits used to represent each weight in the model, which in turn reduces the model's size and the computational resources required to run it. This is particularly beneficial for transformer/DiT models, which are less affected by quantization compared to regular UNET models.

Simplified Explanation:

  • Quantization: Think of quantization as compressing a high-resolution image into a smaller file size without losing much detail. Similarly, GGUF quantization compresses the model weights, making them smaller and faster to process.
  • Model Loading: Instead of loading a large, uncompressed model, ComfyUI-GGUF loads a quantized model, which is quicker and requires less memory.

ComfyUI-GGUF Features

Unet Loader (GGUF)

  • Function: Loads models stored in the GGUF format.
  • Customization: You can place your .gguf model files in the ComfyUI/models/unet folder and use them directly in your workflows.
  • Example: Replace the standard "Load Diffusion Model" node with the "Unet Loader (GGUF)" node in your existing workflows to take advantage of GGUF quantization.

Pre-Quantized Models

  • Available Models:
  • flux1-dev GGUF
  • flux1-schnell GGUF
  • Usage: These models are pre-quantized and ready to use, saving you the effort of quantizing them yourself.

Important Note

  • CLIP Device Setting: The "Force/Set CLIP Device" is not part of this node pack. Avoid installing it if you only have one GPU, and do not set it to cuda:0 to prevent out-of-memory (OOM) errors.

ComfyUI-GGUF Models

ComfyUI-GGUF supports models stored in the GGUF format. Here are the available pre-quantized models:

  • flux1-dev GGUF: Ideal for development and experimentation.
  • flux1-schnell GGUF: Optimized for faster performance.

When to Use Each Model:

  • flux1-dev GGUF: Use this model when you are developing new features or experimenting with different settings.
  • flux1-schnell GGUF: Choose this model for production environments where speed is crucial.

Troubleshooting ComfyUI-GGUF

Common Issues and Solutions

  1. Out-of-Memory (OOM) Errors:
  • Solution: Ensure you are not using the "Force/Set CLIP Device" if you have only one GPU. Avoid setting it to cuda:0.
  1. Model Not Loading:
  • Solution: Verify that your .gguf model files are placed in the ComfyUI/models/unet folder. Ensure you are using the "Unet Loader (GGUF)" node in your workflow.

Frequently Asked Questions

  • Q: Can I use LoRA or Controlnet with GGUF models?
  • A: Currently, LoRA and Controlnet are not supported due to the weights being quantized.

Learn More about ComfyUI-GGUF

Additional Resources

  • llama.cpp: Learn more about the GGUF format and its benefits.
  • Hugging Face Models: Access pre-quantized models for immediate use.
  • Community Forums: Join discussions and get support from other AI artists and developers. By understanding and utilizing ComfyUI-GGUF, you can significantly enhance your AI art projects, making them more efficient and accessible even on lower-end hardware.

ComfyUI-GGUF Related Nodes

RunComfy

© Copyright 2024 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals.