Install this extension via the ComfyUI Manager by searching
for ComfyUI-ELLA
1. Click the Manager button in the main menu
2. Select Custom Nodes Manager button
3. Enter ComfyUI-ELLA in the search bar
After installation, click the Restart button to
restart ComfyUI. Then, manually
refresh your browser to clear the cache and access
the updated list of nodes.
Visit
ComfyUI Online
for ready-to-use ComfyUI environment
ComfyUI-ELLA integrates ComfyUI with ELLA, a tool developed by TencentQQGYLab, enhancing user interface capabilities for streamlined and efficient operations.
ComfyUI-ELLA Introduction
ComfyUI-ELLA is an extension for ComfyUI that integrates the ELLA (Equip Diffusion Models with LLM for Enhanced Semantic Alignment) framework. This extension enhances the capabilities of text-to-image diffusion models by leveraging large language models (LLMs) to improve semantic alignment and text comprehension. ELLA addresses the limitations of traditional text encoders like CLIP, enabling the generation of images from complex and detailed prompts. This is particularly useful for AI artists who want to create intricate and multi-faceted artworks based on textual descriptions.
How ComfyUI-ELLA Works
ComfyUI-ELLA works by incorporating a novel module called the Timestep-Aware Semantic Connector (TSC). This module dynamically extracts semantic features from the LLM at different stages of the denoising process in diffusion models. Think of it as a translator that helps the model understand and interpret complex prompts over time, ensuring that the generated images accurately reflect the detailed descriptions provided. By adapting semantic features at various timesteps, ELLA enhances the model's ability to follow dense prompts, resulting in more accurate and detailed image generation.
ComfyUI-ELLA Features
ELLA Text Encode Node
Function: Automatically concatenates ELLA and CLIP conditions.
Customization: You can input your text prompt and adjust the sigma value to match the KSampler settings.
Example: Using a detailed prompt like "A serene landscape with mountains, a river, and a sunset" will result in a more nuanced and accurate image.
ELLA Apply Method
Function: Integrates ELLA with the ComfyUI ecosystem, ensuring compatibility with various nodes and workflows.
Customization: You can choose to apply ELLA only in positive conditions, which might yield different artistic effects.
Example: Applying ELLA only to positive conditions can highlight specific elements in your artwork, such as making the sunset more vibrant.
CLIP Concat Support
Function: Supports concatenation of CLIP conditioning to make LoRA trigger words effective.
Customization: Link ELLA output conditioning to the conditioning_to of the Conditioning (Concat) node.
Example: Using trigger words like "vintage" or "futuristic" in your prompt will influence the style of the generated image.
ComfyUI-ELLA Models
ComfyUI-ELLA uses specific models that need to be placed in designated directories:
ELLA Models: Place in ComfyUI/models/ella/. Example: ella-sd1.5-tsc-t5xl.safetensors.
FLAN-T5 XL Text Encoder: Place in ComfyUI/models/ella_encoder/. This should include files like config.json, model.safetensors, and tokenizer files.
These models enhance the text-to-image generation process by providing the necessary data and structure for the ELLA framework to function effectively.
What's New with ComfyUI-ELLA
Changelog
2024.4.30: Added ELLA Text Encode node for automatic concatenation of ELLA and CLIP conditions.
2024.4.24: Upgraded ELLA Apply method for better compatibility with the ComfyUI ecosystem.
2024.4.22: Fixed unstable image quality in multi-batch processing and added CLIP concat support.
2024.4.19: Initial repository setup and documentation of nodes.
These updates improve the functionality and compatibility of ComfyUI-ELLA, making it easier for AI artists to create detailed and accurate images from complex prompts.