The nodes and its associated workflow are fully developed by Kijai. We give all due credit to Kijai for this innovative work. On the RunComfy platform, we are simply presenting Kijai’s contributions to the community. It is important to note that there is currently no formal connection or partnership between RunComfy and Kijai. We deeply appreciate Kijai’s work!
Tora introduces a novel framework for generating high-quality videos by leveraging trajectory-based guidance in a diffusion transformer model. By focusing on motion trajectories, Tora achieves more realistic and temporally coherent video synthesis. This approach bridges the gap between spatial-temporal modeling and generative diffusion frameworks.
Please note that this version of Tora is based on the CogVideoX-5B model and is intended for academic research purposes only. For licensing details, please refer .
This is the CogVideoX Tora workflow, Left Side nodes are inputs, Middle are processing tora nodes, and right are the outputs node.
[!CAUTION] Only Horizontal Format Images of Dimensions - 720*480 will work. Other Dimensions will give error.
Positive
: Enter the actions taking place with the Subject based of Trajectory defined in the trajectory node (moving, flowing....etc).Negative
: Enter what you don't want to happen (Distorted hands, blurry...etc)Here you set the Trajectory path of the motion of the subject in the uploaded photo.
points_to_sample
: This Set the Number of frames for Rendering, or Duration of your video in frames.mask_width
: Default is 720. DO NOT CHANGE!mask_height
: Default is 480. DO NOT CHANGE!Node Guide:
Toggle handles visibility
Sampling_method:
These are the model downloader nodes, it will automatically download models in your comfyui in 2-3 mins.
Steps
: This value decides the quality of your render. Keep in between 25 - 35 for best and effcient value.cfg
: Default value is 6.0 for CogVideo Sampling.denoising strength
and Scheduler
: Do not change this.This node will set the strength of your motion trajectory.
strength
: High value will give distorted figure or flying dot. Use between 0.5 - 0.9.start_percent
: Use this value to ease in the effect of strength motion.end_percent
: - High value will give distorted figure or flying dot. Use between 0.3 - 0.7These nodes will give you 3 outputs.
"CogVideoX Tora: Trajectory-oriented Diffusion Transformer for Video Generation" presents an innovative approach to video generation by introducing trajectory-based guidance within a diffusion transformer framework. Unlike traditional video synthesis models that struggle to maintain temporal consistency and realistic motion, CogVideoX Tora explicitly focuses on modeling motion trajectories. This enables the system to generate coherent and visually convincing videos by understanding how objects and elements evolve over time. By combining the power of diffusion models, known for high-quality image generation, with the temporal reasoning capabilities of transformers, CogVideoX Tora bridges the gap between spatial and temporal modeling.
CogVideoX Tora's trajectory-oriented mechanism provides fine-grained control over object movements and dynamic interactions, making it particularly suitable for applications requiring precise motion guidance, such as video editing, animation, and special effects generation. The model's ability to maintain temporal consistency and realistic transitions enhances its applicability in creating smooth and coherent video content. By integrating trajectory priors, CogVideoX Tora not only improves motion dynamics but also reduces artifacts often seen in frame-based generation. This breakthrough sets a new benchmark for video synthesis, offering a powerful tool for creators and developers in fields like filmmaking, virtual reality, and video-based AI.
© Copyright 2024 RunComfy. All Rights Reserved.