Visit ComfyUI Online for ready-to-use ComfyUI environment
Extract and process human poses from images or video frames using advanced pose detection models for realistic motion integration in creative projects.
MimicMotionGetPoses is a powerful node designed to extract and process human poses from images or video frames. This node leverages advanced pose detection models to identify and rescale keypoints of human bodies, faces, and hands, making it an essential tool for AI artists who want to incorporate realistic human motion into their projects. By using this node, you can seamlessly integrate human poses into your animations or visual effects, ensuring that the movements are natural and accurately scaled. The node processes input images, detects poses, and resizes them according to reference images, providing a consistent and high-quality output that can be used in various creative applications.
This parameter represents the pipeline used for processing the mimic motion. It is essential for defining the sequence of operations that will be applied to the input images to extract and rescale the poses. The pipeline ensures that the processing steps are executed in the correct order, leading to accurate and consistent results.
The reference image is a crucial input that provides a baseline for rescaling the detected poses. This image should contain a clear and well-defined human pose that will be used to adjust the scale of the poses detected in the input images. The reference image helps maintain consistency in the size and proportions of the detected poses across different frames.
This parameter consists of the images or video frames from which the human poses will be detected. The input images should be in a format that the node can process, typically as numpy arrays. The quality and resolution of these images can impact the accuracy of the pose detection, so it is recommended to use high-quality images for the best results.
This parameter defines the minimum configuration settings for the pose detection process. It helps in fine-tuning the detection algorithm to ensure that it captures the essential keypoints without being too sensitive to noise or irrelevant details. Adjusting this parameter can improve the accuracy of the detected poses.
Similar to cfg_min, this parameter sets the maximum configuration settings for the pose detection process. It helps in controlling the upper limits of the detection algorithm, ensuring that it does not miss any critical keypoints. Properly configuring this parameter can enhance the robustness of the pose detection.
The number of steps determines the granularity of the pose detection process. More steps can lead to more detailed and accurate pose detection, but it may also increase the processing time. Balancing the number of steps with the desired level of detail and processing efficiency is key to optimizing the node's performance.
The seed parameter is used for initializing the random number generator, ensuring that the pose detection process is reproducible. By setting a specific seed value, you can achieve consistent results across different runs of the node, which is particularly useful for debugging and fine-tuning the detection process.
This parameter controls the strength of noise augmentation applied to the input images. Noise augmentation can help improve the robustness of the pose detection algorithm by simulating various real-world conditions. Adjusting this parameter allows you to find the right balance between noise robustness and detection accuracy.
Frames per second (fps) is a parameter that defines the frame rate for processing video input. It determines how many frames will be processed per second, impacting the smoothness and continuity of the detected poses in video sequences. Setting an appropriate fps value is crucial for achieving realistic motion in animations.
This boolean parameter indicates whether the pose detection model should remain loaded in memory after processing. Keeping the model loaded can speed up subsequent processing tasks, but it may also consume more memory. This parameter helps manage the trade-off between processing speed and memory usage.
Context size defines the amount of surrounding context considered during the pose detection process. A larger context size can improve the accuracy of the detected poses by providing more information about the surrounding environment, but it may also increase the processing time.
This parameter specifies the overlap between consecutive context windows during the pose detection process. Overlapping context windows can help capture continuous motion more accurately, ensuring smooth transitions between frames. Adjusting this parameter can enhance the quality of the detected poses in video sequences.
The optional scheduler parameter allows you to specify a custom scheduler for the pose detection process. A scheduler can help manage the execution of different processing tasks, optimizing the overall performance of the node. This parameter is optional and can be left unset if not needed.
Pose strength controls the influence of the detected poses on the final output. A higher pose strength value will make the detected poses more prominent, while a lower value will blend them more subtly with the input images. Adjusting this parameter allows you to achieve the desired level of emphasis on the detected poses.
This parameter determines the strength of the image embedding applied to the detected poses. Image embedding helps integrate the detected poses more naturally into the input images, ensuring a seamless and realistic appearance. Adjusting this parameter can enhance the visual quality of the final output.
Pose start percent defines the starting point of the pose detection process as a percentage of the total input sequence. This parameter allows you to focus the detection on a specific portion of the input, which can be useful for processing only the relevant sections of a video or image sequence.
Similar to pose start percent, this parameter defines the ending point of the pose detection process as a percentage of the total input sequence. It helps in limiting the detection to a specific portion of the input, ensuring that only the relevant sections are processed.
The output tensor is a multi-dimensional array that contains the processed poses in a format suitable for further use in animations or visual effects. This tensor includes the rescaled keypoints of human bodies, faces, and hands, ensuring that the detected poses are accurately represented and ready for integration into your projects.
This parameter represents the subsequent frames or images in the output tensor, excluding the reference pose. It provides the processed poses for each input frame, allowing you to create smooth and continuous animations or visual effects based on the detected human poses.
© Copyright 2024 RunComfy. All Rights Reserved.