ComfyUI  >  Workflows  >  Hunyuan Image to Video | Breathtaking Motion Creator

Hunyuan Image to Video | Breathtaking Motion Creator

Hunyuan Image to Video is Tencent's latest image-to-video model, built on the HunyuanVideo foundation. It transforms static images into high-quality videos with smooth motion. Using image latent concatenation and a pre-trained Multimodal Large Language Model, it merges images with text prompts smoothly. HunyuanVideo I2V supports up to 720p resolution at 24fps for videos up to 5 seconds. It also enables customizable effects via LoRA training, allowing unique transformations like hair growth or emotional embraces. The Hunyuan Image to Video workflow includes wrapper nodes and native compatibility, with memory optimization through FP8 weights for efficiency.

ComfyUI Hunyuan image-to-video Workflow

HunyuanVideo-I2V Workflow in ComfyUI | Premium Image-to-Video Generation
Want to run this workflow?
  • Fully operational workflows
  • No missing nodes or models
  • No manual setups required
  • Features stunning visuals

ComfyUI Hunyuan image-to-video Examples

ComfyUI Hunyuan image-to-video Description

ComfyUI Hunyuan Image-to-Video Workflow Description

1. What is Hunyuan Image-to-Video Workflow?

The Hunyuan Image-to-Video workflow is a powerful pipeline designed to transform still images into high-quality videos with natural motion. Developed by Tencent, this cutting-edge technology enables users to create cinematic animations with smooth 24fps playback at resolutions up to 720p. By leveraging latent image concatenation and a Multimodal Large Language Model, Hunyuan Image-to-Video interprets image content and applies consistent motion patterns based on text prompts.

2. Benefits of Hunyuan Image-to-Video:

  • High-Resolution Output - Generates videos up to 720p at 24fps
  • Natural Motion Generation - Creates fluid, realistic animations from static images
  • Text-Guided Animation - Uses text prompts to guide motion and visual effects
  • Cinematic Quality - Produces professional-grade videos with high fidelity
  • Customizable Effects - Supports LoRA-trained effects like hair growth, facial expressions, and style adjustments
  • Optimized Memory Usage - Utilizes FP8 weights for better resource management

3. How to Use Hunyuan Image-to-Video Workflow

3.1 Generation Methods with Hunyuan Image-to-Video

Example Workflow:

Hunyuan Image-to-Video

  1. Prepare Inputs
    • In Load Image: Upload your source image
  2. Enter Motion Description
    • In HyVideo I2V Encode: Input a descriptive text prompt for the desired motion
  3. Refinement (Optional)
    • In HunyuanVideo Sampler: Adjust frames to control video length (default: 129 frames ≈ 5 seconds)
    • In HunyuanVideo TeaCache: Modify cache_factor for optimized memory usage
    • In HunyuanVideo Enhance A Video: Enable for temporal consistency and flickering reduction
  4. Output
    • In Video Combine: Check the preview and find saved result in ComfyUI > Output folder

3.2 Parameter Reference for Hunyuan Image-to-Video

Hunyuan Image-to-Video

  • HunyuanVideo Model Loader
    • model_name: hunyuan_video_I2V_fp8_e4m3fn.safetensors - Core model for image-to-video conversion
    • weight_precision: bf16 - Defines precision level for model weights
    • scale_weights: fp8_e4m3fn - Optimizes memory use
    • attention_implementation: flash_attn_varlen - Controls attention processing efficiency
  • HunyuanVideo Sampler
    • frames: 129 - Number of frames (5.4 seconds at 24fps)
    • steps: 20 - Sampling steps (higher values improve quality)
    • cfg: 6 - Controls prompt adherence strength
    • seed: varies - Ensures generation consistency
  • HyVideo I2V Encode
    • prompt: [text field] - Descriptive prompt for motion and style
    • add_prepend: true - Enables automatic text formatting

3.3 Advanced Optimization with Hunyuan Image-to-Video

  • Memory Optimization
    • HunyuanVideo BlockSwap: CPU offloading for VRAM efficiency
    • HunyuanVideo TeaCache: Controls cache behavior to balance memory vs. speed
    • scale_weights: FP8 weights (e4m3fn format) for memory reduction
  • Speed Optimization
    • HunyuanVideo Torch Compile Settings: Enables Torch compile for faster processing
    • attention_implementation: Selects efficient attention mechanisms for performance boost
    • offload_device: Configures GPU/CPU memory management

More Information

For more details on the Hunyuan Image-to-Video workflow, visit .

Acknowledgements

This workflow is powered by Hunyuan Image-to-Video, developed by Tencent. The ComfyUI integration includes wrapper nodes created by Kijai, enabling advanced features such as context windowing and direct image embedding support. Full credit goes to the original creators for their contributions to Hunyuan Image-to-Video workflow!

Want More ComfyUI Workflows?

RunComfy
Copyright 2025 RunComfy. All Rights Reserved.

RunComfy is the premier ComfyUI platform, offering ComfyUI online environment and services, along with ComfyUI workflows featuring stunning visuals. RunComfy also provides AI Playground, enabling artists to harness the latest AI tools to create incredible art.