Generative video modeling has made significant strides, yet ensuring structural and temporal consistency over long sequences remains a challenge. Current methods predominantly rely on RGB signals, leading to accumulated errors in object structure and motion over extended durations. To address these issues, we introduce WorldWeaver, a robust framework for long video generation that jointly models RGB frames and perceptual conditions within a unified long-horizon modeling scheme. Our training framework offers three key advantages. First, by jointly predicting perceptual conditions and color information from a unified representation, it significantly enhances temporal consistency and motion dynamics. Second, by leveraging depth cues, which we observe to be more resistant to drift than RGB, we construct a memory bank that preserves clearer contextual information, improving quality in long-horizon video generation. Third, we employ segmented noise scheduling for training prediction groups, which further mitigates drift and reduces computational cost. Extensive experiments on both diffusion- and rectified flow-based models demonstrate the effectiveness of WorldWeaver in reducing temporal drift and improving the fidelity of generated videos.
Given an input video, RGB, depth, and optical flow signals are encoded into a joint latent representation via a 3D VAE. The latents are split into a memory bank and prediction groups for the Diffusion Transformer. The memory bank stores historical frames and is excluded from loss computation; short-term memory retains a few fully denoised frames for fine details, while long-term memory keeps depth cues noise-free and adds low-level noise to RGB information. During training, prediction groups are assigned different noise levels according to the noise scheduler curve, aligned with the noise scheduling used during inference.
Below, we present qualitative results generated by our method, including human activity scenes and robotic arm manipulation tasks.
@article{liu2025worldweaver,
title={WorldWeaver: Generating Long-Horizon Video Worlds via Rich Perception
},
author={Liu, Zhiheng and Deng, Xueqing and Chen, Shoufa and Wang, Angtian and Guo, Qiushan and Han, Mingfei and Xue, Zeyue and Chen, Mengzhao and Luo, Ping and Yang, Linjie},
journal={arXiv preprint arXiv:2508.15720},
year={2025}
}