Visit ComfyUI Online for ready-to-use ComfyUI environment
Create dynamic animated portraits from static images using advanced AI techniques for lifelike expressions and movements.
The LivePortraitProcess
node is designed to create dynamic and realistic animated portraits from static images. This node leverages advanced AI techniques to animate facial features based on driving images or videos, allowing you to bring still portraits to life. The primary goal of this node is to provide a seamless and intuitive way to generate animated portraits that can mimic expressions and movements from a reference video or image sequence. This is particularly useful for AI artists looking to add a layer of interactivity and realism to their digital art projects. The node handles various complex processes such as image cropping, keypoint extraction, feature transformation, and retargeting of facial expressions, making it a powerful tool for creating lifelike animations with minimal manual intervention.
The source_image
parameter is the static portrait image that you want to animate. This image serves as the base for all subsequent animations. The quality and resolution of this image can significantly impact the final output, so it is recommended to use high-resolution images for best results.
The driving_images
parameter consists of a sequence of images or a video that provides the motion and expressions to be transferred to the source image. These driving images dictate how the facial features of the source image will move and animate. The more expressive and varied the driving images, the more dynamic the resulting animation will be.
The dsize
parameter specifies the desired size of the output image. This parameter allows you to control the resolution of the animated portrait, ensuring it fits your specific requirements. The value should be provided in the format (width, height).
The scale
parameter adjusts the scaling factor for the image processing pipeline. This can be used to fine-tune the size of the facial features and overall image to better match the driving images. The default value is typically 1.0, meaning no scaling.
The vx_ratio
parameter controls the horizontal scaling ratio for the image. This can be useful for adjusting the width of the facial features to better align with the driving images. The value is usually between 0.0 and 1.0.
The vy_ratio
parameter controls the vertical scaling ratio for the image. Similar to vx_ratio
, this parameter helps in adjusting the height of the facial features. The value is usually between 0.0 and 1.0.
The pipeline
parameter is an instance of the LivePortraitPipeline
class, which orchestrates the entire animation process. This includes tasks like cropping, keypoint extraction, and feature transformation. This parameter is essential for the node to function correctly.
The lip_zero
parameter is a boolean flag that, when set to true, initializes the lip-open scalar to zero. This can be useful for ensuring that the mouth remains closed at the start of the animation. The default value is false.
The eye_retargeting
parameter is a boolean flag that enables or disables eye retargeting. When enabled, the node will adjust the eye movements to better match the driving images. The default value is false.
The lip_retargeting
parameter is a boolean flag that enables or disables lip retargeting. When enabled, the node will adjust the lip movements to better match the driving images. The default value is false.
The stitching
parameter is a boolean flag that enables or disables the stitching of the animated portrait back onto the original image. This can be useful for blending the animated features seamlessly with the original portrait. The default value is false.
The relative
parameter is a boolean flag that, when set to true, makes the animation relative to the initial pose of the source image. This can help in maintaining a more natural look. The default value is false.
The eyes_retargeting_multiplier
parameter is a scalar value that multiplies the eye retargeting adjustments. This allows for fine-tuning the intensity of the eye movements. The default value is typically 1.0.
The lip_retargeting_multiplier
parameter is a scalar value that multiplies the lip retargeting adjustments. This allows for fine-tuning the intensity of the lip movements. The default value is typically 1.0.
The onnx_device
parameter specifies the device to be used for ONNX model inference. Common options include 'CPU' and 'CUDA' for GPU acceleration. The default value is 'CUDA'.
The cropped_out_list
parameter is a list of cropped images that have been processed through the pipeline. These images are intermediate results that show the source image after various transformations like cropping and scaling.
The full_out_list
parameter is a list of fully processed images that have been animated and possibly stitched back onto the original portrait. These images represent the final output of the node, showcasing the animated portrait in its entirety.
dsize
, scale
, vx_ratio
, and vy_ratio
parameters to better match the source image with the driving images.eye_retargeting
and lip_retargeting
for more accurate facial animations, especially if the driving images have significant eye and lip movements.stitching
parameter to blend the animated features seamlessly with the original portrait, creating a more cohesive final output.dsize
, vx_ratio
, and vy_ratio
parameters to ensure that the images are correctly resized and aligned.© Copyright 2024 RunComfy. All Rights Reserved.