# historyguided_video_diffusion__f1dd6cb5.pdf History-Guided Video Diffusion Kiwhan Song* 1 Boyuan Chen* 1 Max Simchowitz 2 Yilun Du 3 Russ Tedrake 1 Vincent Sitzmann 1 Classifier-free guidance (CFG) is a key technique for improving conditional generation in diffusion models, enabling more accurate control while enhancing sample quality. It is natural to extend this technique to video diffusion, which generates video conditioned on a variable number of context frames, collectively referred to as history. However, we find two key challenges to guiding with variable-length history: architectures that only support fixed-size conditioning, and the empirical observation that CFG-style history dropout performs poorly. To address this, we propose the Diffusion Forcing Transformer (DFo T), a video diffusion architecture and theoretically grounded training objective that jointly enable conditioning on a flexible number of history frames. We then introduce History Guidance, a family of guidance methods uniquely enabled by DFo T. We show that its simplest form, vanilla history guidance, already significantly improves video generation quality and temporal consistency. A more advanced method, history guidance across time and frequency further enhances motion dynamics, enables compositional generalization to out-ofdistribution history, and can stably roll out extremely long videos. Project website: https: //boyuan.space/history-guidance 1 Introduction Diffusion models are effective generative models in domains such as image, sound, and video. Critical to their success is classifier-free guidance (CFG) (Ho & Salimans, 2022), which trades off between sample quality and diversity by jointly training a conditional and an unconditional diffusion model and combining their score estimates when sampling. In the realm of video generative models, CFG commonly relies on either text or image prompts as conditioning vari- *Equal contribution 1MIT 2Carnegie Mellon University 3Harvard University. Correspondence to: Kiwhan Song , Boyuan Chen . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). ables. Yet, another conditioning variable, namely the entire collection of previous video frames, or history, deserves further exploration. In this paper, we investigate the following question: Can we use different portions of history - variable lengths, subsets of frames, and even different image-domain frequencies - as a form of guidance for video generation? Importantly, CFG with flexible history is incompatible with existing diffusion model architectures and the most obvious fix significantly degrades sample quality (see Section 3). To address these limitations, we propose the Diffusion Forcing Transformer (DFo T), a video diffusion framework that enables flexible conditioning on any portion of the input history. Extending the noising-as-masking paradigm in Diffusion Forcing (Chen et al., 2024) to non-causal transformers, DFo T trains video diffusion models by applying independent noise levels to each frame. During sampling, portions of the history can be selectively masked with noise, enabling flexible conditioning and guidance. For instance, in CFG, the unconditional score corresponds to our model with the entire history masked out. Notably, DFo T is compatible with existing architectures such as Di T (Peebles & Xie, 2023) and U-Vi T (Hoogeboom et al., 2023; 2024) and can be efficiently implemented through fine-tuning of pre-trained video diffusion models. At sampling time, the DFo T facilitates a family of historyconditioned guidance methods, collectively referred to as History Guidance (HG). The simplest of these, Vanilla History Guidance (HG-v), uses an arbitrary length of history as the conditioning variable for CFG. Notably, even this simple method significantly enhances video quality. We further introduce two advanced methods enabled by the DFo T: Temporal History Guidance (HG-t) and Fractional History Guidance (HG-f) . These extend history guidance beyond a special case of CFG. Temporal History Guidance combines scores from different history windows. Fractional History Guidance conditions on history windows corrupted by varying levels of noise, effectively acting as a low-pass filter on historical frames. With minor modifications, it can also target specific frequency bandwidths to enhance the dynamic degree of generated videos (hence the frequency-based terminology). Together, we compose HG-t and HG-f to create a comprehensive history guidance paradigm, which we term history guidance across time and frequency (HG-tf). The Diffusion Forcing Transformer and associated History History-Guided Video Diffusion : single image : stably generate more than 800 frames with consistent transitions t=8 t=16 t=24 t=32 t=40 t=48 t=56 t=808 t=816 t=824 t=832 t=840 t=848 t=856 450 frames later, exploring outside: consistent transition indoor to outdoor (t=460,465,470,500) t=308 t=316 t=324 t=332 t=340 t=348 t=356 250 frames later, exploring a bedroom: smoothly exit the kitchen door (t=85,110,135,140) Figure 1. Diffusion Forcing Transformer with history guidance enables stable rollout of extremely long videos. We visualize 21 frames from an 862-frame long navigation video generated by our DFo T model from a single image in a test set video that the model has never seen before. Best viewed as videos on our project website. Guidance methods dramatically improve the quality and consistency of video generation, enabling the creation of exceptionally long videos through autoregressive extension, outperforming the de facto standard Di T diffusion and performing on par with industry models trained with an order of magnitude more compute. In Fig. 1, we showcase our method by using history guidance across time and frequency with DFo T to generate an 862-frame navigation video from a single image many times longer than prior results and the maximum video length in the training set. Our contributions can be summarized as follows: 1. We propose the Diffusion Forcing Transformer (DFo T), a competitive video diffusion framework that enables sampling-time conditioning using any portion of history, a capability that is difficult to achieve with existing models. 2. We introduce History Guidance (HG), a family of history-conditioned guidance methods enabled by DFo T that significantly enhance sample consistency, motion dynamics, and visual quality in video diffusion. 3. We empirically demonstrate the state-of-the-art performance and new capabilities enabled by our method, especially in long video generation. Additionally, we provide a theoretical justification of the training objective through a variational lower bound. 2 Preliminaries and Related Work Diffusion Models. Diffusion models (Sohl-Dickstein et al., 2015; Ho et al., 2020; Song et al., 2021) define a forward process that transforms a data distribution into white noise via a stochastic process over increasing noise levels k [0, 1]: xk = αkx0+σkϵ, where ϵ N(0, I). The goal of the model is to reverse this process by learning to estimate the score function sθ(xk, k) log pk(xk) (Vincent, 2011), which enables iterative denoising of a data point, gradually transforming it from white noise back to a sample from the original distribution. In practice, the score function is often parameterized as an affine function of alternative objectives such as the noise prediction ϵθ(xk, k) ϵ. Video Diffusion Models (VDMs). VDMs have enabled the generation of realistic, high-resolution videos (Brooks et al., 2024; Yang et al., 2024; Zheng et al., 2024; Kong et al., 2024). Their success is largely attributed to advancements such as transferring successful image diffusion models (Singer et al., 2022; Guo et al., 2023), scaling data and model (Blattmann et al., 2023a), improving transformerbased architectures (Peebles & Xie, 2023; Gupta et al., 2024; Jin et al., 2024), and enhancing computational efficiency through multi-stage approaches like latent VDMs (He et al., 2022; Blattmann et al., 2023b; Ma et al., 2024; Yin et al., 2024). Many of these models (Blattmann et al., 2023a; Yang et al., 2024) focus on generating videos from a single first image. In contrast, our model is designed to condition on arbitrary length histories, a crucial capability for autoregressively extending newly generated videos. Conditional Diffusion Sampling with Guidance. Classifier-free guidance (CFG) (Ho & Salimans, 2022) is a crucial technique for improving sample quality in diffusion models. CFG jointly trains conditional and unconditional models sθ(x, c, k) log pk(xk|c) and sθ(x, , k) log pk(xk) by randomly dropping out the conditioning c. During sampling, the true conditional score log pk(xk|c) is replaced with the weighted score log pk(xk) + ω log pk(xk|c) log pk(xk) , (1) where ω 1 is the guidance scale that pushes the sample towards the conditioning. In VDMs, CFG is predominantly History-Guided Video Diffusion (a) Conventional Video Diffusion (b) Diffusion Forcing Transformer Figure 2. Comparison of the conventional conditional video diffusion models and Diffusion Forcing Transformer. At training time, conventional (a) approaches treat history as part of the conditioning input, first encoded by an separate encoder and then injected to the Di T via Adaptive Layer Norm and scaling. The Diffusion Forcing Transformer (b) instead does not distinguish between history and generation target frames. It trains a Di T to denoise all frames of a sequence, where frames have independently varying noise levels. used for text guidance (Ho et al., 2022b; Wang et al., 2023). For frame conditioning, first frame guidance is commonplace in image-to-video models (Blattmann et al., 2023a; Yang et al., 2024), or fixed set of few frames (Blattmann et al., 2023b; Gupta et al., 2024; Watson et al., 2025), likewise in multi-view diffusion models (Gao et al., 2024). Our work generalizes CFG by enabling guidance with a variable number of conditioning frames and later extends beyond the conventional approach of subtracting an unconditioned score - similar to prior works in compositional generative models (Du & Kaelbling, 2024; Liu et al., 2022; Du et al., 2023), we compose score from multiple conditioning to combine their behaviors. Additionally, we eliminate the reliance on binary-dropout training, the default mechanism for enabling CFG, which we empirically show performs sub-optimally when extended to history guidance. Diffusion Forcing. Traditionally, diffusion models are trained using uniform noise levels across all tokens. Diffusion Forcing (DF) (Chen et al., 2024) proposes training sequence diffusion models with independently varied noise levels per frame. Although DF provides theoretical and empirical support for this approach, their work focuses on causal, state-space models. Caus Vid (Yin et al., 2024) builds on DF by scaling it to a causal transformer, creating an autoregressive video foundation model. Our work extends the flexibility of DF by developing both the theory and architecture for non-causal, state-free models, enabling new, unexplored capabilities in video generation. 3 Challenges when Guiding with History Video diffusion models are conditional diffusion models p(x|c), where x denotes frames to be generated, and c represents the conditioning (e.g. text prompt, or a few observed prior frames). For simplicity, we refer to the latter as history, even when the observed images could be e.g. a subset of keyframes that are spaced across time. Our discussion of c will focus exclusively on history conditioning and exclude text or other forms of conditioning in notation. Formally, let x T denote a T-frame video clips with indices T = {1, 2, . . . , T}. Define H T as the indices of history frames used for conditioning, and G = T \ H as the indices of the frames to be generated. Our objective is to model the conditional distribution p(x G|x H) with a diffusion model. We aim to extend classifier-free guidance (CFG) to this setting. Since the history x H serves as conditioning, sampling can be performed by estimating the following score: log pk(xk G)+ω log pk(xk G|x H) log pk(xk G) . (2) This approach differs from conventional CFG in two ways: 1) The generation x G and conditioning history x H belong to the same signal x T , differing only in their indices G, H T ; thus, the generated x G can be reused as conditioning x H for generating subsequent frames. 2) The history x H can be any subset of T , allowing its length to vary. Guiding with history, therefore, requires a model that can estimate both conditional and unconditional scores given arbitrary subsets of video frames. Below, we analyze how these differences present challenges for implementation within the current paradigm of video diffusion models (VDMs). Architectures with fixed-length conditioning. As shown in Figure 2a, Di T (Peebles & Xie, 2023) or U-Net-based diffusion models (Bao et al., 2023; Rombach et al., 2022) typically inject conditioning using Ada LN (Peebles & Xie, 2023; Perez et al., 2018) layers or by concatenating the conditioning with noisy input frames along the channel dimension. This design constrains conditioning to a fixedsize vector. While some models adopt sequence encoders for variable-length conditioning (e.g., for text inputs), these encoders are often pre-trained (Yang et al., 2024) and cannot share parameters with the diffusion model to encode history frames. Consequently, guidance has been limited to fixedlength and generally short history (Blattmann et al., 2023a; Xing et al., 2023; Yang et al., 2024; Watson et al., 2025). Framewise Binary Dropout performs poorly. Classifierfree guidance is typically implemented using a single network that jointly represents the conditional and uncondi- History-Guided Video Diffusion Figure 3. Sampling with DFo T and History Guidance. A DFo T can be used to estimate scores conditioned on differently masked histories using noise as masking. This includes clean (full history), fully masked (unconditional), subset masked (shorter history), or partially masked (low-frequency history). These scores can be composed when sampling to obtain a family of History Guidance methods. tional models. These are trained via binary dropout, where the conditioning variable c is randomly masked during training with a certain probability. History guidance can, in principle, be achieved by randomly dropping out subsets of history frames during training. However, our ablations (Sec. 6.2) reveal that this approach performs poorly. We hypothesize that this is due to inefficient token utilization: although the model processes all |T | frames via attention, only a random subset of |G| frames contribute to the loss. This becomes more pronounced as videos grow longer, making framewise binary dropout a suboptimal choice. 4 The Diffusion Forcing Transformer In this section, we introduce the Diffusion Forcing Transformer (DFo T), a simple yet powerful video diffusion framework designed to model score functions associated with different portions of history. This includes variable-length histories, arbitrary subsets of frames, and even history processed at different image-domain frequencies. DFo T improves video generation performance as a base model even without guidance. By addressing the challenges outlined in Section 3, DFo T further enables guidance with flexible history and a more advanced family of history guidance methods described in Section 5. Noise as Masking. The forward diffusion process turns the t-th frame xt of a video sequence into a noisy frame xkt t at noise levels kt [0, 1]. One can interpret this as progressively masking xt with noise (Chen et al., 2024) - x0 t is clean and hence unmasked, x1 t is fully masked and contains no information about the original xt. Intermediate noise levels (0 < kt < 1) yield a partially masked frame xkt t , retaining a noisy snapshot of the original frame s information. History as noise-free frames. Denoising generated frames xk G conditioned on history x H can be unified under the noiseas-masking framework. Specifically, this involves denoising the entire sequence of frames x H xk G with noise levels k T = [k1, k2, , k T ] defined as: ( 0 if t H k if t G. (3) This formulation treats history and generated frames as parts of the same input to the transformer, rather than separating history as a distinct conditioning input (see Figure 2 and Section 3). This unification allows any full-sequence transformer to be fine-tuned into a history-conditional model with variable-length history, simply by varying the noise levels within each sequence. Training: Per-frame Independent Noise Levels. As illustrated in Figure 2b, instead of setting noise levels to zero for all history frames, we adopt per-frame independent noise levels introduced in Diffusion Forcing (Chen et al., 2024). Each frame xt x T is assigned an independent noise level kt [0, 1], resulting in random sequences of noise levels k T in contrast with Equation 3. The DFo T model is then trained to minimize the following noise prediction loss, where ϵT denotes noise added to all frames: E k T ,x T ,ϵT h ϵT ϵθ(xk T T , k T ) 2i , (4) Crucially, noise levels are selected independently for all frames without distinguishing the past and the future. This enables parallel training while also allowing non-causal conditioning on partially masked future frames. In Appendix A.5, we further discuss a simplified objective when max(|H|) T and a causal adaptation of our model. In Appendix A.1, we justify this training objective as optimizing a (reweighted) valid Evidence Lower Bound (ELBO) on the expected log-likelihoods: Theorem 4.1 (Informal). The DFo T training objective (Equation (4)) optimizes a reweighting of an Evidence Lower Bound (ELBO) on the expected log-likelihoods. Compared to conventional video diffusion methods, where a single noise level k [0, 1] is uniformly applied to all generation frames x G, our approach provides two key benefits: (1) token utilization is improved by computing a loss History-Guided Video Diffusion Table 1. Comparison with generic diffusion models on Kinetics600. , , and indicate whether a model can condition on a single predefined, arbitrary under approximation, or arbitrary history. DFo T, both trained from scratch and fine-tuned, outperforms all generic diffusion baselines under the same architecture and is on par with industry models trained with more compute resources (see Appendix C.4). Flexible? Method FVD Industry size and compute MAGVIT-v2 (Yu et al., 2023b) 4.3 0.1 W.A.L.T (Gupta et al., 2024) 3.3 0.1 Rolling Diffusion (Ruhe et al., 2024) 5.2 Video Diffusion (Ho et al., 2022b) 16.2 0.3 MAGVIT (Yu et al., 2023a) 9.9 0.3 Same Architecture FS 95.5 0.4 BD 6.4 0.1 DFo T (scratch) 4.3 0.1 DFo T (fine-tuned from FS) 4.7 0.0 conditioned on all frames x T instead of a smaller subset; second, (2) this objective places variable history lengths in-distribution of the training objective, leading to more flexible use of history lengths as detailed below. Sampling: Conditioning on Arbitrary History. Unlike standard VDMs that require fixed-length history during sampling, DFo T allows conditioning on arbitrary history. To generate x G conditioned on x H at each sampling step with noise level k, we estimate the conditional score log pk(xk G|x H) by feeding the model noisy xk G and clean history frames x0 H. Sampling is then performed using standard score-based sampling schemes such as DDPM (Ho et al., 2020) or DDIM (Song et al., 2020). This flexibility in conditioning enables history guidance and its more advanced variants, as described in the next section. 5 History Guidance Leveraging the flexibility of Diffusion Forcing Transformer (DFo T), we introduce History Guidance (HG), a family of techniques for history-conditioned video generation. These methods enhance generation quality, improve motion dynamics, enable robustness to out-of-distribution (OOD) histories, and unlock novel capabilities such as compositional video generation. Please refer to Figure 3 for an overview. Simplest HG: Vanilla History Guidance. The simplest form of HG, referred to as Vanilla History Guidance (HG-v), directly performs classifier-free guidance (CFG) with a chosen history length, following Equation 2. The conditional score for any history H can be computed as described in the previous section. To perform CFG, we need to estimate the unconditional score log pk(xk G). Notably, the unconditional score is a special case of the conditional score with H = and can be estimated by masking history frames x H with complete noise. Even this simple form of HG Figure 4. Qualitative comparison on Kinetics-600. DFo T (both scratch and fine-tuned) generates higher-quality samples consistent with the history than baselines. FS omitted for poor quality. We show 6 of 16 frames; see Figure 14 for more comparisons. significantly improves generation quality and consistency. History Guidance Across Time and Frequency. While history guidance has been presented as a special case of CFG so far, its full potential extends far beyond CFG. Consider the following generalization of Equation 2: log pk(xk G) + P i ωi log pk(xk G|x k Hi Hi ) log pk(xk G) , (5) where the total score is a weighted sum of conditional scores, each conditioned on possibly different segments of history {Hi}, and each masked with a possibly different noise level k Hi. This formulation enables better generalization than a single score function conditioned on a full long history. By composing scores, each individual score component operates on a restricted conditional context, reducing the likelihood of being out-of-distribution (Du & Kaelbling, 2024). Appendix A.3 provides informal mathematical intuition on why summing conditional scores is permissible. Equation 5 effectively allows us to compose the scores conditioned on 1) different history subsequences, and 2) history frames that are partially noisy. We refer to these two principal axes as time and frequency, which together form a 2D plane of options that we refer to as History Guidance across Time and Frequency. For simplicity, we introduce composition along these two axes separately. Time Axis: Temporal History Guidance. Due to the curse of dimensionality, the amount of data that we require to guarantee constant data support grows exponentially with the length of history we wish to condition on. As a result, history conditioned models are particularly prone to outof-distribution (OOD) history without an inductive bias of sparse dependency. Common symptoms include blowing up or overfitting to irrelevant features. To address this, we propose Temporal History Guidance (HG-t), which composes History-Guided Video Diffusion 1 2 3 4 Guidance scale ω Better 170.4 1 2 3 4 Guidance scale ω Frame-wise Quality 1 2 3 4 Guidance scale ω Consistency 1 2 3 4 Guidance scale ω Dynamic Degree DFo T DFo T + HG-v DFo T + HG-f SD GT Figure 5. Various metrics as a function of guidance scale ω for vanilla and fractional history guidance on Kinetics-600, comparing against ω = 1 ( , w/o HG), SD, and ground truth (GT). FS is omitted for poor performance (FVD = 1040). Vanilla history guidance trades off dynamics diversity for quality consistency. Fractional history guidance better balances these trade-offs, achieving the best FVD. (a) Vanilla history guidance significantly improves frame quality and consistency with an increasing guidance scale. We sample with varying guidance scales ω = 1 (top, without history guidance), 1.5 (middle), and 3 (bottom). (b) Fractional history guidance resolves the issue of static videos, improving dynamics by guiding with lower frequencies. We sample with varying frequency scales, with k H = 0 (top, vanilla guidance leading to static videos), 0.3 (middle), and 0.6 (bottom). Figure 6. Qualitative results for vanilla fractional history guidance on Kinetics-600. Best viewed zoomed in. Red box = history frames. scores conditioned on different subsequences of history by setting k Hi = 0 in Equation 5. This composition can be performed with either: 1) long and short history {Hlong, Hshort}, aiming to trade-off between the two imperfect predictive models, reducing the likelihood of OOD while preserving both long and short-term dependencies, or 2) multiple short, overlapping in-distribution histories {Hshort1, Hshort2, }, to simulate the conditional distribution of the full history. Frequency Axis: Fractional History Guidance. We observe that a major failure mode of HG-v under high guidance scales is the generation of overly static videos with minimal motion. This occurs because HG-v encourages consistency with history, leading to a trivial solution of simply copying the most recent history frame. To address this, we propose Fractional History Guidance (HG-f), which guides the sampling process using fractionally masked history. Frac- tionally masking history retains only low-frequency information (Dieleman (2024), Appendix A.2), allowing highfrequency details (e.g., fine textures and fast motions) to remain unconstrained by guidance. This approach makes videos more dynamic while maintaining consistency, which is mainly associated with low-frequency details. Specifically, the HG-f score is given by: log pk(xk G|x H) + ω log pk xk G|xk H H log pk(xk G) , (6) where k H (0, 1) controls the degree of masking to focus on lower-frequency details, and ω is the guidance scale for the partially masked history xk H H . In principle, different history frames could contribute information at different frequency bands, such as high-frequency details from recent frames and low-frequency motion from earlier frames. While a detailed exploration of sophisticated sampling strategies is left to future work, our experiments show that even History-Guided Video Diffusion 0 25 50 75 100 125 150 Camera Rotation Angle ( ) In-Distribution Slightly OOD OOD Training Data FS SD DFo T DFo T + HG-v DFo T + HG-t Figure 7. Robust performance of temporal history guidance given OOD history unseen in the training data. Left: Baselines sharply lose performance transitioning from in-distribution, slightly OOD, to OOD tasks, while DFo T with HG-t shows minimal drop. Right: Baselines produce blurry, inconsistent frames with artifacts on slightly OOD history and unrecognizable frames on OOD history, whereas DFo T with HG-t generates high-quality, accurate samples. Each frame shown is one of four generated; see Figure 10 for full results. simple implementations of HG-f significantly improve motion dynamics without sacrificing consistency. 6 Experiments We empirically evaluate the performance of the Diffusion Forcing Transformer and history guidance. We first validate the DFo T as a generic video model without history guidance (Sec. 6.2), demonstrating the effectiveness of the modified training objective. Next, we examine the effectiveness and additional capabilities of history guidance (Secs. 6.3 and 6.4). Finally, we showcase very long videos generated by DFo T with history guidance (Sec. 6.5). 6.1 Experimental Setup Datasets. Throughout our experiments, we train and evaluate a separate DFo T model for each dataset as follows: Kinetics-600 (Kay et al. (2017), 128 128), a standard video prediction benchmark, Real Estate10K or RE10K (Zhou et al. (2018), 256 256), a dataset of real-world indoor scenes with camera pose annotations, and Minecraft (Yan et al. (2023), 256 256), a dataset of long-context Minecraft navigation videos with discrete actions. We employ Fruit Swapping, an imitation learning task adapted from Diffusion Forcing (Chen et al., 2024) to test the combined ability to handle long-term memory and reactive behavior with a physical robot. Details are in Appendix C.1. We use Kinetics-600 for benchmarking and quantitative comparisons, and the other three for studying new applications. Baselines. 1) Standard Diffusion (SD): A single-task model trained for specific test history lengths following the standard conditional diffusion setup (Gupta et al., 2024; Watson et al., 2025). 2) Binary-Dropout Diffusion (BD): An ablative baseline trained with framewise binary dropout for history guidance instead of independent per-frame noise levels. Note that BD requires DFo T s architecture as opposed to conditioning via adaptive Layer Norm to support flexible history lengths, effectively making it an ablation. 3) Full-Sequence Diffusion with Reconstruction Guidance (FS): An unconditional video diffusion model trained with maximum sequence length. Flexible-length conditioning is achieved during sampling via history replacement and reconstruction guidance (Ho et al., 2022b). Evaluation. To evaluate the overall video generation performance encompassing quality and diversity, we use Fr echet Video Distance (FVD, Unterthiner et al. (2018)). For a more detailed analysis of video quality, we use VBench (Huang et al., 2024), which provides separate scores for different aspects such as frame quality, consistency, and dynamics. For highly deterministic tasks, we evaluate according to Learned Perceptual Image Patch Similarity (LPIPS, Zhang et al. (2018)), computed frame-wise against the ground truth. Additional experimental details are provided in Appendix C. 6.2 Evaluating the Diffusion Forcing Transformer We validate DFo T as a competitive video generative model without history guidance by answering the questions: Q1: How does DFo T compare to the conventional video diffusion approach in standard video benchmarks? Q2: Does binary dropout diffusion (BD) perform competitively as an alternative training approach that also supports flexible history? Q3: Is DFo T empirically flexible enough to handle arbitrary sets of history frames? Q4: Can we fine-tine an existing model into DFo T? We summarize quantitative and qualitative results in Table 1 and Figure 4 respectively. Competitive Performance of DFo T (Q1) without Guidance. DFo T outperforms all baselines, including singletask standard diffusion (SD), despite SD being optimized for the eval s specific history length. This demonstrates DFo T s flexibility without sacrificing task-specific performance, aligning with observations from (Chen et al., 2024). Limited Performance of Binary Dropout (Q2). While BD enables flexible history conditioning, it suffers a significant performance drop compared to SD. Notably, BD produces artifacts and inconsistent generations (Figure 4), highlighting its inefficiency as an alternative to DFo T s History-Guided Video Diffusion training objective. Flexibility of DFo T (Q3). We demonstrate DFo T s flexibility by tasking it with various video generation tasks on RE10K, such as future prediction, frame interpolation, and mixed history setups. As shown in Figure 11, DFo T generates consistent, high-quality samples across all tasks. Fine-tune existing models into DFo T (Q4). As discussed in Sec. 4, an DFo T can be obtained by fine-tuning an existing video diffusion model. We fine-tune the full-sequence model on Kinetics-600 into a DFo T using only 12.5% of the training cost. The fine-tuned model surpasses all baselines and performs comparably to the DFo T trained from scratch (see Appendix D.1 for detailed analysis). This confirms the feasibility of fine-tuning large foundation models into DFo T to support history guidance. 6.3 Improving Video Generation via History Guidance We examine the effect of history guidance on video quality in terms of frame-wise quality, frame-to-frame consistency and dynamic degree of generated video. We benchmark 64-frame video generation using sliding window rollout on Kinetics-600, a challenging setup that requires outstanding consistency to avoid blowing up. Note that this is a setup where conventional image-to-video models struggle since they can only condition on the final generated frame to extend the video. We present quantitative and quantitative results in Figures 5 and Figure 6 respectively. Vanilla History Guidance. We visualize samples generated with vanilla history guidance with increasing guidance scale in Figure 6a. Stronger history guidance consistently improves frame quality and consistency, which is also reflected in their corresponding VBench scores in Figures 5b and 5c. In Figure 5a, we obtain the best FVD result with a small guidance scale of ω=1.5. Beyond that, FVD increases sharply, indicating a loss of diversity with higher guidance scales, similar to the quality-diversity trade-off of CFG. Fractional History Guidance. Despite notable quality improvements, we observe that vanilla history guidance tends to generate static videos at high guidance scales (ω 3), as illustrated in the top rows of Figure 6b, with significantly less motion than ground truth in Figure 5d. Fractional history guidance resolves this in the side-by-side visualization. We find that guiding with lower frequencies (higher k H) consistently increases dynamics while maintaining quality, as shown in Figure 6b. This further lowers the best FVD of vanilla history guidance (181.6) to 170.4, surpassing FS (1040), SD (247.5), and DFo T without guidance (208.0). 6.4 New Abilities via Temporal History Guidance Temporal history guidance brings new capabilities to DFo T, allowing it to solve tasks impossible for previous models. We discuss three representative tasks. Task 1. Robust to Out-of-Distribution (OOD) History. We evaluate robustness to OOD histories on Real Estate10K by creating scenarios with extreme camera rotations between history frames and ask the model to interpolate. Baselines fail to generalize, producing incoherent generations. In contrast, DFo T with temporal history guidance splits OOD histories into shorter, in-distribution subsequences, composing their scores to maintain both local and global dependencies. This enables DFo T to handle OOD histories effectively, as shown in Figure 7. Task 2. Long Context Generation. Minecraft is a video dataset that requires long context to achieve good FVD scores. We found generating coherent videos with long contexts often leads to OOD histories. Baselines prioritize consistency with the context at the expense of quality. Our hypothesis is that temporal guidance blends scores from long-context and short-context models, balancing memory retention with robustness to OOD. This strategy improves FVD scores from 97.63 to 79.19, achieving long-term coherent high-quality generations. See Appendix D.3 for details. Task 3. Long-horizon yet Reactive Imitation Learning. We test on a robotic manipulation task requiring both longterm memory for object rearrangement and short-term reactivity for disturbances. Each data point in the dataset contains either of these two behaviors but never both. Baselines fail to integrate the two behaviors, while DFo T combines full-history scores (for memory) with single-frame scores (for reactivity) using temporal history guidance. This allows the robot to recover from disturbances and complete tasks, achieving a success rate of 83% while baselines fail to perform the task completely. See Appendix D.4 for details. 6.5 Ultra Long Video Generation In Figure 1, we present a showcase that utilizes all of this paper s contributions - we extend a single image to an 862frame video in RE10K. Even the most high-performing prior methods can only roll out for dozens of frames under the same setup. This is made possible by enhanced quality, consistency, and rollout stability through history guidance, plus DFo T s flexibility that enables this. See Appendix C.9 for more details and Appendix D.6 for more samples (Figures 8a to 8d), including notable failures of other models. 7 Conclusion Enabling flexible conditioning on different portions of history, the Diffusion Forcing Transformer not only establishes itself as a competitive video diffusion framework but also gives rise to History Guidance, a family of powerful historyconditioned guidance methods that significantly enhances video quality, consistency, and motion degree. Additionally, we demonstrate that DFo T can be efficiently fine-tuned from existing models, suggesting future potentials of integrating History Guidance with current foundation models. History-Guided Video Diffusion Acknowledgements This work was supported by the National Science Foundation under Grant No. 2211259, by the Singapore DSTA under DST00OECI20300823 (New Representations for Vision, 3D Self-Supervised Learning for Label-Efficient Vision), by the Intelligence Advanced Research Projects Activity (IARPA) via Department of Interior/ Interior Business Center (DOI/IBC) under 140D0423C0075, by the Amazon Science Hub, and by the MIT-Google Program for Computing Innovation. Impact Statement This paper aims to advance the field of video generative modeling. As a video generative model, our approach may enable the creation of longer, higher-quality videos, with potential applications in robotics and other fields. However, we acknowledge the potential risks associated with misuse, such as the generation of harmful or unethical content. We emphasize the importance of ethical considerations and responsible use of this work. Bao, F., Nie, S., Xue, K., Cao, Y., Li, C., Su, H., and Zhu, J. All are worth words: A vit backbone for diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 22669 22679, 2023. Bellec, P. C. Optimal exponential bounds for aggregation of density estimators. Bernoulli, 23(1):219 248, 2017. Blattmann, A., Dockhorn, T., Kulal, S., Mendelevitch, D., Kilian, M., Lorenz, D., Levi, Y., English, Z., Voleti, V., Letts, A., et al. Stable video diffusion: Scaling latent video diffusion models to large datasets. ar Xiv preprint ar Xiv:2311.15127, 2023a. Blattmann, A., Rombach, R., Ling, H., Dockhorn, T., Kim, S. W., Fidler, S., and Kreis, K. Align your latents: Highresolution video synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 22563 22575, 2023b. Brooks, T., Peebles, B., Holmes, C., De Pue, W., Guo, Y., Jing, L., Schnurr, D., Taylor, J., Luhman, T., Luhman, E., et al. Video generation models as world simulators. Open AI Blog, 1:8, 2024. Carreira, J. and Zisserman, A. Quo vadis, action recognition? a new model and the kinetics dataset. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6299 6308, 2017. Chan, S. et al. Tutorial on diffusion models for imaging and vision. Foundations and Trends in Computer Graphics and Vision, 16(4):322 471, 2024. Chen, B., Monso, D. M., Du, Y., Simchowitz, M., Tedrake, R., and Sitzmann, V. Diffusion forcing: Next-token prediction meets full-sequence diffusion. Advances in Neural Information Processing Systems, 2024. Chen, T. On the importance of noise scheduling for diffusion models. ar Xiv preprint ar Xiv:2301.10972, 2023. Chi, C., Xu, Z., Feng, S., Cousineau, E., Du, Y., Burchfiel, B., Tedrake, R., and Song, S. Diffusion policy: Visuomotor policy learning via action diffusion. The International Journal of Robotics Research, pp. 02783649241273668, 2023. Dhariwal, P. and Nichol, A. Diffusion models beat gans on image synthesis. Advances in neural information processing systems, 34:8780 8794, 2021. Dieleman, S. Diffusion is spectral autoregression, 2024. URL https://sander.ai/2024/09/02/ spectral-autoregression.html. Du, Y. and Kaelbling, L. Compositional generative modeling: A single model is not all you need. ar Xiv preprint ar Xiv:2402.01103, 2024. Du, Y., Durkan, C., Strudel, R., Tenenbaum, J. B., Dieleman, S., Fergus, R., Sohl-Dickstein, J., Doucet, A., and Grathwohl, W. S. Reduce, reuse, recycle: Compositional generation with energy-based diffusion models and mcmc. In International conference on machine learning, pp. 8489 8510. PMLR, 2023. Gao, R., Holynski, A., Henzler, P., Brussee, A., Martin Brualla, R., Srinivasan, P. P., Barron, J. T., and Poole, B. Cat3d: Create anything in 3d with multi-view diffusion models. Advances in Neural Information Processing Systems, 2024. Gervet, T., Xian, Z., Gkanatsios, N., and Fragkiadaki, K. Act3d: 3d feature field transformers for multi-task robotic manipulation. In Conference on Robot Learning, pp. 3949 3965. PMLR, 2023. Guo, Y., Yang, C., Rao, A., Liang, Z., Wang, Y., Qiao, Y., Agrawala, M., Lin, D., and Dai, B. Animatediff: Animate your personalized text-to-image diffusion models without specific tuning. ar Xiv preprint ar Xiv:2307.04725, 2023. Gupta, A., Yu, L., Sohn, K., Gu, X., Hahn, M., Li, F.- F., Essa, I., Jiang, L., and Lezama, J. Photorealistic video generation with diffusion models. In European Conference on Computer Vision, pp. 393 411. Springer, 2024. Hang, T., Gu, S., Li, C., Bao, J., Chen, D., Hu, H., Geng, X., and Guo, B. Efficient diffusion training via min-snr weighting strategy. In Proceedings of the IEEE/CVF History-Guided Video Diffusion international conference on computer vision, pp. 7441 7451, 2023. He, Y., Yang, T., Zhang, Y., Shan, Y., and Chen, Q. Latent video diffusion models for high-fidelity long video generation. ar Xiv preprint ar Xiv:2211.13221, 2022. Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., and Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. Advances in neural information processing systems, 30, 2017. Ho, J. and Salimans, T. Classifier-free diffusion guidance. ar Xiv preprint ar Xiv:2207.12598, 2022. Ho, J., Jain, A., and Abbeel, P. Denoising diffusion probabilistic models. Advances in neural information processing systems, 33:6840 6851, 2020. Ho, J., Chan, W., Saharia, C., Whang, J., Gao, R., Gritsenko, A., Kingma, D. P., Poole, B., Norouzi, M., Fleet, D. J., et al. Imagen video: High definition video generation with diffusion models. ar Xiv preprint ar Xiv:2210.02303, 2022a. Ho, J., Salimans, T., Gritsenko, A., Chan, W., Norouzi, M., and Fleet, D. J. Video diffusion models. Advances in Neural Information Processing Systems, 35:8633 8646, 2022b. Hoogeboom, E., Heek, J., and Salimans, T. simple diffusion: End-to-end diffusion for high resolution images. In International Conference on Machine Learning, pp. 13213 13232. PMLR, 2023. Hoogeboom, E., Mensink, T., Heek, J., Lamerigts, K., Gao, R., and Salimans, T. Simpler diffusion (sid2): 1.5 fid on imagenet512 with pixel-space diffusion. ar Xiv preprint ar Xiv:2410.19324, 2024. Huang, S., Wang, Z., Li, P., Jia, B., Liu, T., Zhu, Y., Liang, W., and Zhu, S.-C. Diffusion-based generation, optimization, and planning in 3d scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16750 16761, 2023. Huang, Z., He, Y., Yu, J., Zhang, F., Si, C., Jiang, Y., Zhang, Y., Wu, T., Jin, Q., Chanpaisit, N., et al. Vbench: Comprehensive benchmark suite for video generative models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 21807 21818, 2024. Jin, Y., Sun, Z., Li, N., Xu, K., Jiang, H., Zhuang, N., Huang, Q., Song, Y., Mu, Y., and Lin, Z. Pyramidal flow matching for efficient video generative modeling. ar Xiv preprint ar Xiv:2410.05954, 2024. Karras, T., Aittala, M., Lehtinen, J., Hellsten, J., Aila, T., and Laine, S. Analyzing and improving the training dynamics of diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 24174 24184, 2024. Kay, W., Carreira, J., Simonyan, K., Zhang, B., Hillier, C., Vijayanarasimhan, S., Viola, F., Green, T., Back, T., Natsev, P., et al. The kinetics human action video dataset. ar Xiv preprint ar Xiv:1705.06950, 2017. Kingma, D. P. Auto-encoding variational bayes. ar Xiv preprint ar Xiv:1312.6114, 2013. Kingma, D. P. and Gao, R. Understanding the diffusion objective as a weighted integral of elbos. Advances in Neural Information Processing Systems, 2023. Kong, W., Tian, Q., Zhang, Z., Min, R., Dai, Z., Zhou, J., Xiong, J., Li, X., Wu, B., Zhang, J., et al. Hunyuanvideo: A systematic framework for large video generative models. ar Xiv preprint ar Xiv:2412.03603, 2024. Lin, B., Ge, Y., Cheng, X., Li, Z., Zhu, B., Wang, S., He, X., Ye, Y., Yuan, S., Chen, L., et al. Open-sora plan: Open-source large video generation model. ar Xiv preprint ar Xiv:2412.00131, 2024a. Lin, S., Liu, B., Li, J., and Yang, X. Common diffusion noise schedules and sample steps are flawed. In Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp. 5404 5411, 2024b. Liu, N., Li, S., Du, Y., Torralba, A., and Tenenbaum, J. B. Compositional visual generation with composable diffusion models. In European Conference on Computer Vision, pp. 423 439. Springer, 2022. Loshchilov, I. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. Ma, X., Wang, Y., Jia, G., Chen, X., Liu, Z., Li, Y.-F., Chen, C., and Qiao, Y. Latte: Latent diffusion transformer for video generation. ar Xiv preprint ar Xiv:2401.03048, 2024. Mildenhall, B., Srinivasan, P. P., Tancik, M., Barron, J. T., Ramamoorthi, R., and Ng, R. Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM, 65(1):99 106, 2021. Nichol, A. Q. and Dhariwal, P. Improved denoising diffusion probabilistic models. In International conference on machine learning, pp. 8162 8171. PMLR, 2021. Peebles, W. and Xie, S. Scalable diffusion models with transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 4195 4205, 2023. History-Guided Video Diffusion Perez, E., Strub, F., De Vries, H., Dumoulin, V., and Courville, A. Film: Visual reasoning with a general conditioning layer. In Proceedings of the AAAI conference on artificial intelligence, volume 32, 2018. Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748 8763. PMLR, 2021. Rigollet, P. and Tsybakov, A. B. Linear and convex aggregation of density estimators. Mathematical Methods of Statistics, 16:260 280, 2007. Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 10684 10695, 2022. Ruhe, D., Heek, J., Salimans, T., and Hoogeboom, E. Rolling diffusion models. In International Conference on Machine Learning, pp. 42818 42835. PMLR, 2024. Salimans, T. and Ho, J. Progressive distillation for fast sampling of diffusion models. ar Xiv preprint ar Xiv:2202.00512, 2022. Shoemake, K. Animating rotation with quaternion curves. In Proceedings of the 12th annual conference on Computer graphics and interactive techniques, pp. 245 254, 1985. Singer, U., Polyak, A., Hayes, T., Yin, X., An, J., Zhang, S., Hu, Q., Yang, H., Ashual, O., Gafni, O., et al. Make-avideo: Text-to-video generation without text-video data. ar Xiv preprint ar Xiv:2209.14792, 2022. Sohl-Dickstein, J., Weiss, E., Maheswaranathan, N., and Ganguli, S. Deep unsupervised learning using nonequilibrium thermodynamics. In Proceedings of the International Conference on Machine Learning (ICML), 2015. Song, J., Meng, C., and Ermon, S. Denoising diffusion implicit models. ar Xiv preprint ar Xiv:2010.02502, 2020. Song, Y., Sohl-Dickstein, J., Kingma, D. P., Kumar, A., Ermon, S., and Poole, B. Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021. Su, J., Lu, Y., Pan, S., Murtadha, A., Wen, B., and Liu, Y. Roformer: Enhanced transformer with rotary position embedding, 2023. Unterthiner, T., Van Steenkiste, S., Kurach, K., Marinier, R., Michalski, M., and Gelly, S. Towards accurate generative models of video: A new metric & challenges. ar Xiv preprint ar Xiv:1812.01717, 2018. Vincent, P. A connection between score matching and denoising autoencoders. Neural computation, 23(7):1661 1674, 2011. Wang, J., Yuan, H., Chen, D., Zhang, Y., Wang, X., and Zhang, S. Modelscope text-to-video technical report. ar Xiv preprint ar Xiv:2308.06571, 2023. Watson, D., Chan, W., Martin-Brualla, R., Ho, J., Tagliasacchi, A., and Norouzi, M. Novel view synthesis with diffusion models. International Conference on Learning Representations, 2023. Watson, D., Saxena, S., Li, L., Tagliasacchi, A., and Fleet, D. J. Controlling space and time with diffusion models. International Conference on Learning Representations, 2025. Xiao, G., Tian, Y., Chen, B., Han, S., and Lewis, M. Efficient streaming language models with attention sinks. International Conference on Learning Representations, 2024. Xing, J., Xia, M., Zhang, Y., Chen, H., Yu, W., Liu, H., Wang, X., Wong, T.-T., and Shan, Y. Dynamicrafter: Animating open-domain images with video diffusion priors. ar Xiv preprint ar Xiv:2310.12190, 2023. Yan, W., Hafner, D., James, S., and Abbeel, P. Temporally consistent transformers for video generation. In International Conference on Machine Learning, pp. 39062 39098. PMLR, 2023. Yang, Z., Teng, J., Zheng, W., Ding, M., Huang, S., Xu, J., Yang, Y., Hong, W., Zhang, X., Feng, G., et al. Cogvideox: Text-to-video diffusion models with an expert transformer. ar Xiv preprint ar Xiv:2408.06072, 2024. Yin, T., Zhang, Q., Zhang, R., Freeman, W. T., Durand, F., Shechtman, E., and Huang, X. From slow bidirectional to fast causal video generators. ar Xiv preprint ar Xiv:2412.07772, 2024. Yu, L., Cheng, Y., Sohn, K., Lezama, J., Zhang, H., Chang, H., Hauptmann, A. G., Yang, M.-H., Hao, Y., Essa, I., et al. Magvit: Masked generative video transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10459 10469, 2023a. Yu, L., Lezama, J., Gundavarapu, N. B., Versari, L., Sohn, K., Minnen, D., Cheng, Y., Birodkar, V., Gupta, A., Gu, X., et al. Language model beats diffusion tokenizer is key to visual generation. ar Xiv preprint ar Xiv:2310.05737, 2023b. Zhang, R., Isola, P., Efros, A. A., Shechtman, E., and Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE conference History-Guided Video Diffusion on computer vision and pattern recognition, pp. 586 595, 2018. Zheng, Z., Peng, X., Yang, T., Shen, C., Li, S., Liu, H., Zhou, Y., Li, T., and You, Y. Open-sora: Democratizing efficient video production for all. ar Xiv preprint ar Xiv:2412.20404, 2024. Zhou, T., Tucker, R., Flynn, J., Fyffe, G., and Snavely, N. Stereo magnification: Learning view synthesis using multiplane images. ar Xiv preprint ar Xiv:1805.09817, 2018. History-Guided Video Diffusion A Proofs, Explanations, and Extensions A.1 Derivation of an ELBO This section includes a derivation of an ELBO corresponding to the DFo T training objective. By taking a sequence modeling perspective, the derivation below streamlines that of the Diffusion Forcing ELBO in (Chen et al., 2024). Let T denote the index set associated with a sequence x, so that x T = (xt)t T is the whole sequence. We use the notation k = (kt)t T for the sequence of noise levels. A path ρ is a sequence of noising steps that transition from an unnoised sequence to a noised one. Specifically, Definition A.1 (Path). We define a path ρ as a sequence (kj)0 j N that begins at zero noise k0 = (0, 0, . . . , 0), and terminates at full noise k N = (K, K, . . . , K). Given a path ρ, we let xρ = xk0:N denote the sequence with (xkt t )t T . Note that there is nothing intrinsically causal or temporal about the indices t; indeed, we can define noising paths on other objects like trees or graphs. Examples of paths include: Autoregressive diffusion, where kj t is equal to K if t j/K , equal to 0 if t > j/K + 1, and equal to j K j/K otherwise. This path looks like (0, . . . , 0), (1, 0, . . . , 0),. . . , (K, 0, . . . , 0), (K, 1, 0, . . . , 0), increasing lexicographically. Full-sequence diffusion, where kj t = j and N = K; i.e. all points are denoised together. We can accomodate skips in noiseless, e.g. DDIM, or paths with linearly increasing noise, such as those considered in (Chen et al., 2024). Typically, we assume that kj t is non-decreasing in j (the noise level is monotonic up to k N = (K, . . . , K)). The essential property that we require is that our learned model and forward process factor nicely along such paths. It is straightforward to check that this is indeed the case for Diffusion Forcing Transformer with these monotonic paths: Definition A.2 (Factoring Property). We say that a model pθ and forward process q factor along a path ρ if for any path, ρ = (k1, . . . , k N) be a path, q(xk1:N | xk0) factors as q(xk1:N | xk0) = Qn j=1 q(xkj | xkj 1), and pθ factors as pθ(xk0:N ) = QN j=1 pθ(xkj 1 | xkj)pθ(xk N ), with pθ(xk N ) not depending on θ. When the model factors along paths, a general ELBO holds. We first state the general form, then specialize to Diffusion via Gaussian forward processes, and conclude with the proof of the general result. Theorem A.3. Suppose that (pθ, q) factor along a path ρ = (k1, . . . , k N). Then, for some constant C not depending on θ, we have ln p(xk0) C + Exk1:N q(xk1:N |xk0) ln pθ(xk0 | xk1) + j=1 DKL(pθ(xkj | xkj+1) q(xkj | xkj+1, xk0)) Consequently, if Ek1:N Dp denotes an expectation over paths ρ = (k1, . . . , k K) along which (pθ, q) factor, then ln p(xk0) C + Ek1:N Dp Exk1:N q(xk1:N |xk0) ln pθ(xk0 | xk1) + j=1 DKL(pθ(xkj | xkj+1) q(xkj | xkj+1, xk0)) We now specialize Theorem A.3 to Gaussian diffusion. For now, we focus on the x-prediction formulation of diffusion. The ϵ-prediction , used throughout the main body of the text and the v-prediction formalism, which is the one used in our implementation, can be derived similarly (see Section 2 of (Chan et al., 2024) for a clean exposition). The following theorem is derived directly by applying standard likelihood and KL-divergence computations for the DDPM (Ho et al., 2020; Chan et al., 2024) to Theorem A.3. For simplicity, we focus on paths with a single increment (e.g. DDPM), but extending to jumps (e.g. DDIM) is straightforward (albeit more notationally burdensome). History-Guided Video Diffusion Corollary A.4. Consider only paths ρ for which kj kj 1 entrywise, and for any t and j for which kj t > kj 1 t , kj t = kj 1 t + 1 increments by one. q(xkj+1 | xkj t ) = Y t:kj t 1: {(Hi, k Hi, ωi)}I i=1 = {(H, 0, ω)} Temporal History Guidance with I subsequences {Hi}I i=1 and guidance scales {ωi}I i=1: {(Hi, k Hi, ωi)}I i=1 = {(Hi, 0, ωi)}I i=1 Fractional History Guidance with a guidance scale ω and fractional masking level k H: {(Hi, k Hi, ωi)}I i=1 = {(H, 0, 1), (H, k H, ω 1)} History-Guided Video Diffusion A.5 Simplifying Training Objective Diffusion Forcing (Chen et al., 2024) proposes to train the entire sequence with independent noise per frame. A natural question to ask is whether this mixed objective includes too many tasks compared to what one actually needs. Here we provide some insights from our experiments throughout the project: When the number of frames is small e.g. 10 latent frames, there is no noticeable decrease in training efficiency - Diffusion Forcing seems to converge as fast as standard diffusion from both training and validation curves. However, when we grow the number of latent frames to 50, we start to witness decreased performance at sampling time. While we firmly believe that binary dropout is not the ideal way to achieve objective reduction from our experiments, we believe that one can easily reduce our training objective by only applying independent noise up to the maximum training length one wants to support. In particular, if one wants to generate the next 10 frames from previous 1 10 frames, it doesn t seem necessary for frame 11 to be independently masked as noise from time to time, since we will never need to mask it out for flexible conditioning. In addition, one may want to consider treating the number of history frames as a random variable at training time, sampling a length first and then applying uniform levels of masking to the history, though independent from the noise level of the generation target. We didn t investigate these simplifications in detail because we simply find Diffusion Forcing s training objective very versatile for many of the tasks we want to do, e.g. interpolation, and varying noise level sampling. However, we do believe that these schemes could worth more exploration if one is to scale up our method to a much bigger number of context frames. A.6 Causal Variant In principle, one can implement DFo T and History Guidance with a causal transformer as well. For example, Caus Vid (Yin et al., 2024) has proved the effectiveness of Diffusion Forcing on fast causal video synthesis and doesn t conflict with History Guidance. However, we d like to highlight that one can also use our non-causal DFo T to achieve causal sampling. Different from traditional transformer-based models, DFo T doesn t need to enforce an attention mask to achieve causality. Instead, at generation time, one can mask out the future with white noise to prevent any information from the future from leaking into the neural network. In fact, there might be use cases when one may want some low-frequency information from the future, and then one can fractionally mask out the future via noise as masking to achieve so. On the other hand, the motivation behind causal video diffusion models is often speed and real-time generation using KV caching. In that case, one either needs to train a causal DFo T directly or consult advanced techniques like attention sink (Xiao et al., 2024) to perform windowed attention effectively. A.7 Incorporating Other Conditioning Throughout our discussions in the main paper, conditioning is history exclusively. What if one wants to integrate the Diffusion Forcing Transformer into a text-conditioned diffusion model? One claim of the DFo T is that it doesn t require architectural changes so one can fine-tune an existing model into a DFo T model. This is still the case here: if one already has a text-conditioned video diffusion model, presumably built to accept such conditioning via an adaptive layer norm, one simply take DFo T as an add on to their existing architecture to obtain a DFo T model that accepts both text and history as conditioning. DFo T s Figure 2 does not assert that one cannot use an external Ada LN layer with DFo T, but is rather saying no architectural changes is needed. A.8 Extended Temporal History Guidance Temporal history guidance addresses the challenge of out-of-distribution (OOD) history by composing scores conditioned on different, shorter history subsequences, which are closer to being in-distribution. However, since the model receives the entire video sequence as input during sampling including both the history and the noisy frames being generated the OOD problem can arise throughout the entire video sequence, not just in the history portion. To mitigate this, we propose further decomposing the generation G into generation subsequences G1, G2, . . . , GJ G. In line with the original temporal history guidance, the history H is already decomposed into history subsequences H1, H2, . . . , HI H. This allows us to compose scores conditioned on even shorter, and thus more in-distribution, subsequences in {Hi}I i=1 {Gj}J j=1. Specifically, the composed score is given by: LJ j=1 PI i=1 log pk(xk Gj|x Hi) (21) where L denotes a frame-wise averaging operation. We refer to this method as Extended Temporal History Guidance, as it extends the concept of temporal history guidance by composing both history and generation subsequences. Empirically, we find this method to be more effective than the original temporal history guidance when the video sequence is clearly OOD (e.g., Real Estate10K OOD history experiment), and thus requires shorter subsequences to be in-distribution. Ultra Long Video Generation on Real Estate10K, with HG-v and HG-f Supplementary Visuals Before delving into further details, we list extensive figures (Figures 8 to 14) that supplement the main paper s content. Detailed descriptions for these figures can be found in Appendix D. (a) Long navigation video generated by DFo T with HG. # frames = 862. Ultra Long Video Generation on Real Estate10K, with HG-v and HG-f (b) Long navigation video generated by DFo T with HG. # frames = 917. Ultra Long Video Generation on Real Estate10K, with HG-v and HG-f (c) Long navigation video generated by DFo T with HG. # frames = 442. (d) Long navigation video generated by DFo T with HG. # frames = 442. Figure 8. Long navigation videos generated by DFo T with HG-v and HG-f, from a single history frame on Real Estate10K. We subsample with a stride of 8 frames for visualization. The videos exhibit consistent transitions navigating while through diverse indoor and outdoor scenes, maintaining high stability over hundreds of frames. This is enabled by the improved sample quality and consistency from HG, along with DFo T s flexibility that allows both interpolation and extrapolation. DFo T vs. SD: Long Rollout Comparison on Real Estate10K, with HG-v and HG-f Figure 9. Qualitative comparison of DFo T with HG vs. SD on long video generation. Given a single history frame , we task both models to generate videos of moving straight ahead and visualize them with a stride of 8 frames. While SD quickly diverges after t 30 frames, DFo T with HG maintains high stability until t = 72 and can roll out further. Robustness to Out-of-Distribution History on Real Estate10K, with HG-t (a) Given slightly OOD history with rotation angles in [120 , 130 ], baselines and DFo T with HG-v generate inconsistent frames with artifacts. In contrast, DFo T with HG-t generates consistent videos that highly resemble the ground truth. This is the region where HG-t starts showing its generalization gap with other methods. Robustness to Out-of-Distribution History on Real Estate10K, with HG-t (b) Given OOD history, all baselines completely fail yet DFo T with HG-t still manages to generate high-quality, accurate videos. Figure 10. Qualitative results of testing robustness to out-of-distribution history on Real Estate10K. We provide wide-angle, 4-frame history and task the models to generate the next 4 frames that interpolate between the history frames. As the angle increases, the history becomes more out-of-distribution, and thus we split the results into slightly OOD and OOD depending on the angle range. DFo T s Flexible Sampling on Real Estate10K, without HG Figure 11. An illustration of the empirical flexibility of DFo T, showing ten samples from Real Estate10K, where a single DFo T model infills the missing frames given different history. DFo T successfully generates consistent samples across ten diverse tasks, each varying in the history length from 1 to 6 frames and at different timestamps. Improved Video Generation on Real Estate10K, with HG-v Figure 12. Improved video generation quality with vanilla history guidance on Real Estate10K, for both extrapolation and interpolation tasks. HG-v, with an increasing guidance scale, enhances fidelity and consistency while effectively removing artifacts. Videos are sampled conditioned on two history frames , with varying guidance scales ω = 1 (top, without HG-v), 2 (middle), and 3 (bottom). Zoom into the boxed regions to see notable differences. Long Context Generation on Minecraft, with HG-t Figure 13. Visualization of long context generation on Minecraft. We visualize the generation up to the maximum length of the training set. Given 25 initial frames (red), DFo T with temporal history guidance (upper) can roll out stably without blowing up even without CFG. In contrast, one can clearly see that without temporal history guidance (lower), conditional generation easily becomes blurry in later frames. This is likely because the shorter-context model is less likely to fall out of distribution, using its generation power to compensate for the unconfident, blurry prediction from the longer-context model. DFo T vs. Baselines on Kinetics-600, without HG Figure 14. Additional qualitative comparison on Kinetics-600. We uniformly subsample 6 frames {0, 3, 6, 9, 12, 15} from 16-frame videos, conditioned on 5-frame histories. Both DFo T variants, scratch and fine-tuned, consistently align with the history, generating high-quality samples that closely resemble the ground truth. In contrast, the baselines, typically ordered as SD > BD > FS, struggle to maintain consistency and often exhibit artifacts. History-Guided Video Diffusion B Extended Related Work B.1 History-conditioned Guidance In this section, we discuss how CFG is employed for guiding with history in video diffusion models. The most common case is in Image-to-Video Diffusion Models (Blattmann et al., 2023a; Xing et al., 2023; Yang et al., 2024), where the model uses the first frame for guidance. Typically, the conditioning frame is incorporated into the architecture by concatenating it channel-wise with each frame to be generated, and additionally, the CLIP (Radford et al., 2021) embedding of the conditioning frame is used for cross-attention. Few Conditional Video Diffusion Models have pushed the boundary by guiding with fixed set of few frames. Specifically, Video LDM (Blattmann et al., 2023b) uses the first {1, 2} frames for guidance, W.A.L.T. (Gupta et al., 2024) guides with the first 2 latent tokens, i.e. {5} frames, and 4Di M (Watson et al., 2025) guides with the first {1, 2, 8} frames. Similarly, in Multi-view Diffusion Models, which is similar to video diffusion models but do not differentiate frame order, CAT3D (Gao et al., 2024) guides with the first {1, 3} frames. Architecturally, these models incorporate history frames in various ways. Video LDM concatenates a binary mask, indicating whether each history frame is masked, along with all masked history frames, feeding them to every temporal layer using a learnable downsampling encoder. W.A.L.T. simplifies this by directly concatenating the history frames and binary mask to the noisy generation input, omitting the encoder. 4Di M and CAT3D process the entire sequence both history and generation frames as a single sequence, with a binary mask concatenated along the channel dimension to indicate whether each frame is masked. In summary, guiding with history in video models has been explored to a limited extent. While these models differ in how they incorporate history frames into the architecture, they all process history frames separately from generated frames, except for 4Di M and CAT3D, leading to inflexibility of guidance. Additionally, these models are trained using CFG-style random dropout of history frames, which categorizes them as special cases of Binary-Dropout Diffusion, shown to be suboptimal. These limitations are highlighted in Section 3. In contrast, our work enables guiding with arbitrary, variable-length history frames without the need for binary-dropout training, facilitated by our modified training objective and architecture design. C Experimental Details Below, we provide additional details on datasets, architectures, training, evaluation metrics, and protocols for our experiments. C.1 Datasets Kinetics-600 (Kay et al., 2017) is a widely used benchmark dataset for video generation, featuring 600 classes of approximately 400K action videos. In addition to its role as a standard benchmark, the task is history-conditioned video generation, making it ideal for evaluating our methods. Following prior works, we use a resolution of 128 128 pixels. Despite the large volume of videos and their low resolution, generating high-quality samples from the Kinetics-600 dataset is challenging even with large models due to the diversity and complexity of the content, and thus qualifies as our primary benchmark. Real Estate10K (Zhou et al., 2018) is a dataset of home walkthrough videos, accompanied by camera pose annotations. While the dataset is predominantly used in novel view synthesis tasks, we utilize it for several reasons: 1) The camera poses allow for a more controlled evaluation of video models; for instance, we can easily switch between highly stochastic and deterministic tasks by altering the camera poses, 2) The dataset s nature enables the examination of the consistency of generated videos at a 3D level, and 3) The dataset s relatively smaller size compared to other text-conditioned video datasets makes it more computationally feasible to train our models, while still providing high-resolution videos. We use a resolution of 256 256 pixels. Minecraft (Yan et al., 2023) is a dataset of Minecraft gameplay videos, where the player randomly navigates using 3 actions: forward, left, and right. The dataset consists of 200K videos, each with a length of 300 frames, each frame has a corresponding action label. The dataset is designed in a way that good FVD can only be achieved with a long context under action conditioned setting. Specifically, the dataset contains many trajectories where the player turns around and visits areas that it had visited before. While the original dataset is 128 128 pixels, we train and evaluate on an upsampled version of 256 256 pixels, to generate higher-quality samples. Fruit Swapping is an imitation learning dataset associated with a fruit rearrangement task adopted from Diffusion History-Guided Video Diffusion Forcing (Chen et al., 2024). The task involves a tabletop setup where an apple and an orange are randomly put in two of the three empty clots. A single-arm robot is tasked with swapping the two fruits slots using the third, empty slot as shown in Figure 17. The task requires long-horizon memory since one must remember the initial configuration of the slots to determine the final, target configuration. While the three slots provide a discrete state, each slot has a diameter of 15 centimeters and the fruit can be anywhere in the slot as soon as half of its column resides inside the slot. The task is made even harder when an adversarial human deliberately perturbs the fruit within its slot during the task execution - if there are 10 possible locations within each slot, there would already be 103 combinations of waypoints. This requires a robot policy to be reactive to the fruit locations rather than memorizing all possible combinations. The dataset contains 300 expert demonstrations of the entire swapping task collected by a model-based planner, during which no disturbance happens. The robot may move an apple from slot 1 to the center of slot 2, move the orange from slot 3 to the center of slot 1, and then move the apple from slot 2 to slot 3. Notably, it had never seen a situation where the apple changed its location from center to edge during the middle of the manipulation due to adversarial humans. In addition, the dataset features 300 additional demonstrations of re-grasping, which is a very short recovery behavior when it narrowly misses the fruit. In these re-grasping demonstrations, the robot arm only repositions to grab the missed object without moving it to another slot. Therefore, the dataset contains 300 demonstrations that involve moving fruits but no regrasping, and 300 demonstrations of regrasping but no moving fruit. The former has an average length of 540 frames and the later has an average length of around 50 frames. C.2 Implementation Details We provide a summary of our implementation details in Table 2 and discuss them below. Pixel vs. Latent Diffusion. In this work, we validate DFo T and HG using both pixel and latent diffusion models. For Kinetics-600 and Minecraft, we train a latent diffusion model to enhance computational efficiency. Specifically, for Minecraft, we train an Image VAE (Kingma, 2013) from scratch, which compresses 256 256 images into 32 32 latents, following the approach of Stable Diffusion (Rombach et al., 2022). For Kinetics-600, we train a chunk-wise Video VAE that compresses {1, 4} 128 128 video chunks into 16 16 latents, to more aggressively reduce computational costs. This approach resembles Causal Video VAE, commonly used in prior works (Yu et al., 2023b; Gupta et al., 2024), which compresses an entire 17 128 128 video into 5 16 16 latents via causal convolutions. However, we choose to compress every 4 frames separately to preserve DFo T s flexibility. Moreover, this ensures that consistency is influenced solely by the performance of the diffusion model, not the VAE. We implement the Video VAE and training procedure following Open-Sora-Plan (Lin et al., 2024a). Lastly, for Real Estate10K, we train directly in pixel space, based on the observation that latent diffusion models struggle to correctly follow camera pose conditioning, leading to poor performance on this dataset. Architectures and training details differ significantly between pixel and latent diffusion models, as we discuss in the following sections. Architecture. We employ the Di T (Peebles & Xie, 2023) and U-Vi T (Hoogeboom et al., 2023; 2024) backbones for the latent and pixel diffusion models, respectively. Both are transformer-based architectures; however, the key difference is that Di T s transformer blocks operate at a single resolution, whereas U-Vi T incorporates multiple resolutions, with transformer blocks residing at each resolution. Due to this difference, we observe that the U-Vi T backbone scales better in the pixel space. For improved scalability and temporal consistency, instead of using factorized attention (Ho et al., 2022b), where attention is applied separately to spatial and temporal dimensions, we employ 3D attention that operates on all tokens simultaneously. In addition to this, we incorporate 3D Ro PE (Su et al., 2023; Gervet et al., 2023) as relative positional encodings for the T, H, W dimensions. All conditioning inputs, including noise levels, actions, and camera poses, are injected into the model using an Ada LN layer, following (Peebles & Xie, 2023). For noise levels, since each frame retains independent noise levels in DFo T, an Ada LN layer is applied separately to each token, using the noise level of the corresponding frame. Minecraft actions are converted into one-hot vectors, which are then transformed into embeddings through an MLP layer and added to the noise level embeddings. For camera pose conditioning in Real Estate10K, we compute the relative camera pose with respect to the first frame. Following the methodologies of 3Di M (Watson et al., 2023) and 4Di M (Watson et al., 2025), this relative pose is then converted into ray origins and directions, which are then transformed into 180-dimensional positional embeddings, similar to Nerf (Mildenhall et al., 2021). Across the resolutions of U-Vi T, the camera pose embeddings are spatially downsampled to match the resolution before being injected into the model. Diffusion. We use a cosine noise schedule (Nichol & Dhariwal, 2021) for all of our diffusion models. For the Real Estate10K and Minecraft models, we shift the noise schedule to be significantly noisier (Hoogeboom et al., 2023) by a factor of History-Guided Video Diffusion Table 2. Implementation details for DFo T and baseline models. Kinetics-600 Real Estate10K Minecraft Imitation Learning VAEs Input {1, 4} 128 128 Compression (ft, fs) {1, 4}, 8 1, 8 Latent channels 16 4 Training steps 600k 50k Optimizer Adam Adam Batch size 64 96 Learning rate 1e-4 4e-4 EMA 0.999 VDMs Input 17 128 128 8 256 256 50 256 256 21 32 32 Latent 5 16 16 50 32 32 Frame skip 1 10 Max 2 15 Backbone Di T U-Vi T Di T Attention UNet Patch size 1 2 2 1 Layer types Transformer [Res Net 2, Transformer 2] Transformer Attention, Conv Layers 28 [3, 3, 6, 20] 12 8 Hidden size 1152 [128, 256, 576, 1152] 768 128 Heads 16 9 12 4 Training steps 640k 500k 200k 100k Warmup steps 10k 10k 10k 10k Optimizer Adam W Adam W Adam W Adam W Batch size 192 96 96 64 Learning rate 2e-4 5e-5 1e-4 5e-4 Weight decay 0 1e-2 1e-3 1e-3 EMA 0.9999 0.9999 0.9999 Diffusion type Discrete Continuous Discrete Discrete Noise schedule Cosine Shifted Cosine Shifted Cosine Cosine Noise schedule shift 0.125 0.125 Parameterization v v v x0 Sampler DDIM DDIM DDIM DDIM Sampling steps 50 50 50 50 0.125, which we find markedly enhances sample quality, especially for Real Estate10K. This finding aligns with prior works (Chen, 2023; Hoogeboom et al., 2023) that highlight the importance of adding sufficient noise during training, especially when dealing with highly redundant images, such as those with high resolution. Another important design choice is the parameterization of diffusion models. We employ the v-parameterization (Salimans & Ho, 2022) for all models, which has been widely adopted in image and video diffusion models (Ho et al., 2022a; Lin et al., 2024b) due to its superior sample quality and quicker convergence, except for the robot model, where we use the x0-parameterization. Lastly, to expedite training, we use min-SNR loss reweighting (Hang et al., 2023) for Kinetics and robot learning, and sigmoid loss reweighting (Kingma & Gao, 2023; Hoogeboom et al., 2024) for Real Estate10K and Minecraft. Training. We train models for each dataset and for each model class (e.g., DFo T, SD, etc.), using the same pipeline within each dataset. We apply a frame skip, where training video clips are subsampled by a specific stride: a value of 1 for Kinetics-600, 2 for Minecraft, and 1 for Imitation Learning. For Real Estate10K, we use an increasing frame skip, starting from 10 and extending to the maximum frame skip possible within each video, to help the model learn various camera poses. Throughout all training, We employ the Adam W (Loshchilov, 2017) optimizer, with linear warmup and a constant learning rate. Additionally, we utilize fp16 precision for computational efficiency and clip gradients to a maximum norm of 1.0 to stabilize training. For robot imitation learning, we follow the setup in Diffusion Forcing (Chen et al., 2024) where we concatenate actions and the next observation together for diffusion, with the exception that we stack the next 15 actions History-Guided Video Diffusion together for every video frame. Sampling. For all experiments, we use the deterministic DDIM (Song et al., 2020) sampler with 50 steps. Sampling with history guidance, which requires multiple scores at every sampling step, is implemented by stacking the corresponding inputs across the batch dimension to compute the scores in parallel. These scores are then composed to obtain the final score for the DDIM update. Compute Resources. We utilize 12 H100 GPUs for training all of our video diffusion models, with each model requiring approximately 5 days to train under our chosen batch size. One exception is the Robot model, which is trained on 4 RTX4090 GPUs for 4 hours. We note that most of the video models converge in validation metrics with a fraction of our reported total training steps. However, we chose to train them longer because the industry baselines on these datasets (Yu et al., 2023a; Ruhe et al., 2024) are trained for a great number of epochs that are even unmatched by our final training steps. There was no noticeable overfitting throughout the process. C.3 Evaluation Metrics. Fr echet Video Distance (FVD, Unterthiner et al. (2018)). We employ FVD as the primary evaluation metric for video generation performance. Similar to FID (Heusel et al., 2017), FVD computes the Fr echet distance between the feature distributions of generated and real videos, with features extracted from a pre-trained I3D network (Carreira & Zisserman, 2017). Lower FVD scores indicate better video generation performance. Unlike image-wise metrics such as FID, FVD evaluates entire video sequences, capturing temporal consistency and dynamics in addition to quality and diversity, making it the most suitable metric for our video generation tasks. Moreover, FVD is computed for the entire video, including both history and generated frames, to assess the consistency between them. VBench (Huang et al., 2024). We use VBench, an evaluation suite designed to assess video generation models in a comprehensive manner, when separate evaluation for different aspects of video generation is needed. Among 16 sub-metrics, we focus on 5 metrics to assess three aspects: 1) Frame-wise Quality, calculated as the average of Aesthetic Quality and Imaging Quality, assesses the visual quality of individual frames; 2) (Temporal) Consistency, derived as the average of Subject Consistency and Background Consistency, evaluates the shortand long-term consistency of generated videos; and 3) Dynamic Degree assesses the degree of dynamics, i.e., the amount of motion in the generated videos. All metrics are better when higher, evaluate the generated videos independently without comparison to the ground truth, and are computed by averaging over all generated videos. Learned Perceptual Image Patch Similarity (LPIPS, Zhang et al. (2018)). We use LPIPS as an alternative metric for highly deterministic tasks, where video-wise metrics may not be as sensitive and accurate. LPIPS computes the perceptual similarity between the generated and corresponding ground truth frames, with lower scores indicating higher similarity. We compute LPIPS only for the generated frames, excluding the history frames, to evaluate whether the generated frames are visually similar to the ground truth frames. C.4 Details on Video Generation Benchmark (Section 6.2) Kinetics-600 Benchmark. We closely follow the experimental setup of prior works (Ho et al., 2022b; Yu et al., 2023a;b; Ruhe et al., 2024). On the test split of the dataset, we evaluate the models on a video prediction task, where the model is conditioned on the first 5 history frames and asked to predict the next 11 frames. Since our models, utilizing Video VAE, generate 3 future tokens corresponding to 12 frames, we drop the last frame to align with the prediction task. We report the FVD score computed on 50K generated 16-frame videos, using three different random seeds. Resource Comparison Against Industry-Level Literature Baselines. In Table 1, we show that DFo T not only outperforms generic diffusion baselines trained with the same pipeline but also holds its ground against strong literature baselines, including Video Diffusion (Ho et al., 2022b), MAGVIT (Yu et al., 2023a), MAGVIT-v2 (Yu et al., 2023b), W.A.L.T (Gupta et al., 2024), and Rolling Diffusion (Ruhe et al., 2024). We have selected only the highest-performing baselines from the literature for comparison, omitting others for brevity. A critical aspect of our evaluation is the comparison of computational resources. Our DFo T is trained with fewer resources compared to these industry-level baselines. Specifically, two primary factors affect the performance of diffusion models: network complexity and training batch size. Our DFo T model is a 673M parameter model with a Di T backbone, trained with a batch size of 196. (i) Network Complexity. As Video Diffusion and Rolling Diffusion have different backbones from ours, we compare the number of parameters; they are billion-parameter models, each with 1.1B and 1.2B, significantly larger than our model. History-Guided Video Diffusion For MAGVIT, MAGVIT-v2, and W.A.L.T, which are pure transformer models with similar backbones, we use Gflops as a measure of computational complexity, as suggested by (Peebles & Xie, 2023). Our model is of Di T/XL size, whereas the baselines are Di T/L size, making them slightly smaller. In terms of Gflops, our model has 1.5 times more Gflops compared to these baselines. (ii) Batch Size. Video Diffusion, MAGVIT, and MAGVIT-v2 are trained with a batch size of 256, while W.A.L.T and Rolling Diffusion are trained with a batch size of 512, which is significantly larger than ours. When considering both network complexity and training batch size, MAGVIT and MAGVIT-v2 use comparable resources to our model, whereas Video Diffusion, W.A.L.T, and Rolling Diffusion require significantly more resources. Despite this resource disadvantage, DFo T proves to be highly competitive with these strong baselines. It is only slightly behind W.A.L.T, comparable to MAGVIT-v2, and outperforms the rest. This highlights the superior performance of DFo T as a base video diffusion model. C.5 Details on History Guidance Experiment (Section 6.3) For the Kinetics-600 rollout experiment, the models generate the next 59 frames using sliding windows, given the first 5 history frames. The sliding windows are applied such that the model is always conditioned on the last 2 latent tokens and generates the next 3 latent tokens. As with the Kinetics-600 benchmark, we drop the last frame to align with the task. We assess the FVD and VBench scores on 1,024 generated 64-frame videos. History Guidance Scheme. To investigate the effect of HG-v and HG-f, we vary guidance scales using an equally spaced set of ω {1.0, 1.5, 2.0, 2.5, 3.0, 3.5, 4.0} for both methods. For HG-f, we use a fixed fractional masking degree of k H = 0.8, which we find to generate videos with sufficient dynamics. C.6 Details on OOD History Experiment (Section 6.4, Task 1) In Task 1 of Section 6.4, we have shown that video diffusion models easily fail to generalize when the conditioning history is OOD, and temporal history guidance resolves this challenge, through a systematic study on Real Estate10K. Below, we detail the experiment. What makes a history OOD? As shown in the training data distribution of Figure 7, we find that the rotation angle of the camera poses within a single training scene is typically small, rarely exceeding 100 . Hence, a history with a wider rotation angle, such as 150 , is considered OOD. Based on this observation, we assign the following tasks to the models: Given a 4-frame history, with varying rotation angles, generate 4 frames that interpolates between these frames. Evaluation Based on Rotation Angles. We categorize all scenes based on their rotation angles, into the bins of [0 , 10 ], [10 , 20 ], . . . , [170 , 180 ]. Based on the statistics of the training scenes, we conceptually classify the bins of [0 , 10 ], . . . , [90 , 100 ] as in-distribution, [100 , 110 ], . . . , [130 , 140 ] as slightly OOD (< 500 training scenes), and [140 , 150 ], . . . as OOD (< 100 training scenes). We then randomly select 32 test scenes (or less if the bin contains fewer scenes) from each bin. For each scene, we select 4 equally spaced frames from the beginning and end of it as the history, and designate the target frames as those in between. We evaluate by computing the LPIPS between the generated and target frames, and report the average LPIPS score for each bin, as shown in Figure 7. History Guidance Scheme. From a full history H = {0, 1, 2, 3}, we compose scores conditioned on the following two history subsequences: H1 = {0, 1, 2} and H2 = {1, 2, 3}, each with a guidance scale of ω1 = ω2 = 2. Additionally, we implement an extended version of temporal history guidance discussed in Appendix A.8, by also composing generation subsequences: G1 = {4, 5, 6} and G2 = {5, 6, 7} chosen from the full generation G = {4, 5, 6, 7}. For the baseline using vanilla history guidance, we apply a guidance scale of ω = 2 to the full history H. C.7 Details on Long Context Generation (Section 6.4, Task 2). We train a 50-frame DFo T model that can condition on history up to a length of 25 following the simplified objective Appendix A.5. Note that this is equivalent to 100 frames under the original video with a frameskip of 2, or one-third of the maximum video length. We sample an initial context of 25 from the dataset and use our trained model to auto-regressively diffuse the next 25 frames conditioned on the previous 25. We roll out 5 times, or 125 frames in total, converging the maximum video length in the dataset. History Guidance Scheme. During sampling, we compose the scores from one long-context model and one short-context model, with context lengths of 25 and 4 respectively. Subtracting the unconditioned score doesn t play a significant role on this dataset so we proceed to compose the above two scores only, with a simple weighting of 50% each. History-Guided Video Diffusion C.8 Details on Robot Imitation Learning (Section 6.4, Task 3). Baselines. We compare against other diffusion-based imitation learning methods using our same architecture and implementation. First, we compare against a typical Markovian model, which diffuses the next few actions only based on current observation. Then, we use a variant of this Markovian model, which can see the previous two frames as a short history but still no long-term memory. Notice that these two short history lengths represent the current mainstream approaches (Chi et al., 2023). In addition, we have a third baseline trained to condition on the entire history so far, representing a family of decision-making as sequence generation methods. For the convenience of notation, we will refer to these baselines as Markov model, 2-frame model, and full-history model. All baselines are trained to diffuse actions and next observations jointly. The Need to Compose Subtrajectories. As we mentioned in the dataset description, robot imitation learning is a sequence task that requires both long-term memory and local reactive behavior. While both are important to the final task s success, a short-context model will trivially fail most of the time since it won t remember which final state to proceed to. Therefore we focus on our experiment design on exploiting the failure mode of long-context models. One predominant failure mode is overfitting - since the imitation learning dataset is extremely small, a long-context model can attribute an action to any coincidental features. For example, all swapping trajectories in the dataset feature the behavior of putting the first fruit in the very center of the initially empty slot and coming back later to move it away from that center location. How should the model determine where it should pick up this fruit? There is little guarantee for it to determine correctly that it shall proceed to move its gripper right above that fruit versus just blindly going to the center. Whenever a human perturbs this fruit from the very center of the slot to the edge of the slot, an overfitted model will still move to the very center and proceed to grasp air, ignoring the actual location of that fruit. Therefore, theoretically, a full-history model would never be able to react to such perturbation, since it had never seen a trajectory with such perturbation and a successful trajectory would be out-of-distribution. Instead, it needs to mix in some behavior from a local reactive policy to perform the task, leveraging the fact that whenever a long history is out-of-distribution, you can always fall back to a shorter context model and imitate relevant sub-trajectories. Therefore, the only way to solve this task under the adversarial human is to stitch sub-trajectories together while keeping a long-term memory. History Guidance Scheme. To achieve the aforementioned stitched behavior, we compose three diffusion models with a context of 1 frames, 4 frames, and full history. We assign the full-history model with a small weight of 0.2, the 1 frame model, and the 4 frame model with a weight of 0.45 each. Like Minecraft, we didn t find subtracting unconditioned score super important in this task so we omitted it. The frames here refer to the bundle of the next 15 actions and the single future video frame after that as we mentioned earlier in implementation details. C.9 Details on Ultra Long Video Generation (Section 6.5). We provide additional details on generating long navigation videos on Real Estate10K, incorporating all advanced techniques associated with DFo T and history guidance. The generation of long navigation videos is divided into two phases: (i) a rollout phase, where the model generates a long video using a sliding window approach, and (ii) an interpolation phase, where the generated frames are further interpolated to create a smooth video. The process is detailed below. (i) Rollout Phase. During the rollout phase, starting with a single image randomly selected from the dataset, the model generates a long video using a sliding window, where it is conditioned on the last 4 frames to generate the next 4 frames. The first iteration is an exception, where the model is conditioned on the single image and generates the next 7 frames. Importantly, navigation cannot rely on the ground truth camera poses for two reasons: 1) videos in the dataset are relatively short (less than 300 frames), so we quickly exhaust available camera poses, and 2) the navigation task is highly stochastic, meaning the ground truth camera poses may not align with the generated frames (e.g., moving straight into a wall). To address this, we have developed a simple navigation UI, allowing a user to navigate freely in the scene by providing inputs after each sliding window iteration. Specifically, the user can specify the horizontal and vertical angles, relative to the current frame, for the desired navigation direction, as well as the movement distance. This input is converted into a sequence of camera poses, which are then used as conditioning input for the model to sample the next set of frames. This process is repeated until the desired video length is achieved. (ii) Interpolation Phase. Next, in the interpolation phase, leveraging DFo T s flexibility which supports interpolation, we interpolate between the generated frames by a factor of 7. Specifically, using every pair of consecutive generated frames as history, we interpolate 6 frames between them. Camera poses for the interpolated frames, which should be given as input to the model, are computed by linearly interpolating the camera poses of the frames at both ends. More specifically, rotation matrices are interpolated using SLERP (Shoemake, 1985), and translation vectors are linearly interpolated. History-Guided Video Diffusion 0 100k 200k 300k 400k 500k 600k Training Iterations Training Loss DFo T (scratch) DFo T (fine-tuned) (a) A comprehensive view of the training loss curves. DFo T (finetuned) achieves a low training loss early in the iterations and converges significantly faster than DFo T (scratch). 40k 80k 405k 640k Training Iterations Training Loss (80k) FVD: 4.7 (40k) FVD: 5.2 (640k) FVD: 4.3 DFo T (scratch) DFo T (fine-tuned) (b) A zoomed-in view of the training loss curves. Only after 80k iterations, DFo T (fine-tuned) displays a lower training loss than DFo T (scratch) trained for 640k iterations. Figure 15. Training loss curves for DFo T, trained from scratch and fine-tuned from the pre-trained FS model, on Kinetics-600. History Guidance Scheme. Finally, we discuss how history guidance is utilized throughout the navigation task. During the sliding window rollout, the default HG scheme is HG-f, which we find to be extremely stable during long rollouts. Specifically, we apply HG-f with a guidance scale of ω = 4 with a fractional masking degree of k H = 0.4, chosen to ensure optimal stability. Additionally, we switch to HG-v with a guidance scale of ω = 4 for more challenging situations, such as when the model needs to extrapolate to new areas. This is because HG-v performs better in such challenging scenarios, although it is less stable than HG-f, and thus is used sparingly. This switch is triggered when the model is asked to change the direction by more than 30 , or when the model is asked to move further than a certain distance. During the interpolation phase, we apply HG-v with a small guidance scale of ω = 1.5, to ensure the interpolated video is smooth and consistent. Stabilization. As an additional techinique, we also employ the stabilization technique proposed in Diffusion Forcing (Chen et al., 2024), where the previously generated frames are marked to be slightly noisy at a level of k = 0.02, to prevent error accumulation, thereby further stabilizing the long rollout. D Additional Experimental Results In this section, we present additional experimental results to (i) answer potential questions that may provide further insights into our proposed DFo T and HG, and (ii) further elaborate and provide additional samples for Section 6. D.1 Additional Results on Fine-tuning to DFo T Below we provide detailed results on fine-tuning a pre-trained full-sequence (FS) model to DFo T, both from training and sampling perspectives. Training Dynamics. We show the training loss curves of for two variants of DFo T, one trained from scratch for 640k iterations, and the other fine-tuned from the pre-trained FS model for 80k iterations, in Figure 15. We observe that the pre-trained model already provides a good initialization for DFo T, as the model starts with a low training loss and converges rapidly in the early iterations, in Figure 15a. Surprisingly, the fine-tuned model achieves a lower training loss than the model trained from scratch after only 80k iterations, as shown in Figure 15b. Moreover, after 40k iterations, the fine-tuned model exhibits a training loss comparable to the model trained from scratch for 405k iterations, which is 10x speedup. This highlights the superior efficiency and ease of training DFo T by fine-tuning from a pre-trained model. While this opens up the possibility of fine-tuning large foundational video diffusion models to DFo T with small computational cost, we leave this as future work. FVD Metric Evolution. In contrast to the training loss, Figure 15b (or Table 1) shows that the fine-tuned model achieves a slightly higher FVD score than the model trained from scratch, although being highly competitive even after 40k iterations. We attribute this discrepancy to the use of EMA, which is commonly employed in diffusion models to enhance sample quality (Ho et al., 2020; Dhariwal & Nichol, 2021). By default, we use an EMA decay of 0.9999, and thus the model weights used for sampling are affected by the last tens of thousands of training iterations. Therefore, the fine-tuned model s superior History-Guided Video Diffusion 1 2 3 4 Guidance scale ω Better 170.4 BD + HG-v DFo T + HG-v + HG-f (a) FVD as a function of guidance scale ω for DFo T and BD using HG. Both with HG-v, DFo T yields better FVD-ω curves than BD and thus achieves a lower best FVD score. Applying HG-f, which is specific to DFo T, enlarges the performance gap. (b) Qualitative comparison of DFo T and BD using HG-v with optimal guidance scales ω = 1.5. While DFo T generates consistent, high-quality samples, BD struggles to remain consistent with the history frames and produces artifacts. Red box = history frames. Figure 16. History Guidance works better with DFo T than with Binary-Dropout Diffusion (BD). training loss does not immediately translate to a lower FVD score, but we expect it to outperform the model trained from scratch after an additional short training period. While one may consider simply fine-tuning the model without EMA to speed up, EMA is crucial for sample quality; for example, at 80k iterations, FVD without EMA is 7.3, significantly higher than the 4.7 with EMA. This suggests that choosing a smaller EMA decay that still guarantees sample quality, through sophisticated strategies such as post hoc EMA tuning (Karras et al., 2024), may be a promising direction for future work. D.2 Ablation Study on Binary-Dropout Diffusion with Vanilla History Guidance While we have shown that Binary-Dropout Diffusion (BD) performs poorly as a base model (Q2 of Section 6.2), BD still can implement vanilla history guidance due to its binary dropout training. As such, a natural question is: How does BD perform with HG-v, compared to DFo T? To answer this question, we repeat the Kinetics-600 rollout experiment in Section 6.3 using BD with HG-v, comparing against DFo T with HG. See Figure 16 for the results. We observe that DFo T consistently outperforms BD across all guidance scales except for ω = 2.5, as shown in Figure 16a. Under their optimal guidance scales of ω = 1.5, DFo T achieves a lower FVD score of 181.6 compared to BD s 196.0, and qualitatively, generates more consistent, high-quality samples, as shown in Figure 16b. When using HG-f, which is only applicable to DFo T, DFo T further outperforms BD, achieving an FVD score of 170.4. These results highlight that DFo T is a better base model for implementing history guidance, both in performance and in a variety of guidance methods that can be applied. D.3 Detailed Results on Long Context Generation (Section 6.4, Task 2) We calculate the FVD on 1024 samples across all 125 generated frames. A simple conditional diffusion model with context full context achieves an FVD of 97.625 while our temporal guidance achieves an FVD of 79.19 (lower is better). We note that while traditionally FVD is a bad metric for videos with high intrinsic variance, it s well-suited for our benchmark since both action-conditioning and the dataset design constrain the possible variance. We visually observe that Diffusion Forcing Transformer s prediction aligns well with the ground truth semantically over the majority of the frames in a video, showing the variance is well-warranted. We visualize one randomly picked sample in Figure 13, showing that temporal guidance can maintain high-quality details far into the future even without CFG. In the meanwhile, the long-context model without temporal guidance can suffer from the high dimensional context, which makes it much more likely to see out-of-distribution frames in its history. D.4 Detailed Results on Long-horizon yet Reactive Imitation Learning (Section 6.4, Task 3) We examine the success rate of robot imitation learning quantitatively by randomizing the environment 100 times before testing the temporal guidance model as well as its baselines. We found that the Markov baseline fails to perform the task completely as expected since it has trouble sticking to a specific plan - it would move away from fruit and then move back halfway since it has no memory. The 4-frame model suffers from the same issue and cannot finish the task. It does react well to perturbations on the object and picks up the fruit from time to time, showing short context indeed prevents History-Guided Video Diffusion Figure 17. Visualization of the fruit-swapping task through a DFo T generated video. Two fruits are randomly put within two random slots. The robot is tasked with swapping its slots using the third slot and moving one fruit at a time. This task requires long-horizon memory because it needs to remember the initial location of the fruit for the task completion, but also react to different fruit locations within each slot, which is combinatorically impossible form the dataset. overfitting from temporal locality. We found that the full-history model, with the maximum possible memory, performs well whenever there is no human perturbation. However, as soon as the adversarial human perturbs the fruit during the task execution, this policy often blindly goes to the very center of the third slot while the object is already moved to the edge of the slot. The policy will then proceed to close its gripper, holding nothing, and then move to the next slot, thinking it has something in its hand. There are occasional cases when this doesn t happen and the model actually reacts to the adversarial perturbation, although infrequently and only happens to the case then perturbation from the slot center isn t too big. Overall this shows that using a full-context model naively can make the model suffer from overfitting and one may want to manually emphasize the temporal locality prior. Finally, we tested DFo T composed guidance and found it to achieve a much higher success rate of 83%, showing that it s actually stitching the subtrajectories to make decisions, or at least simultaneously borrowing the memory from the full-context model while staying locally reactive using the short-context model. In addition, we attempted a few stronger perturbations such that the adversarial human will deliberately knock off the fruit from the robot s gripper when it s closing. We found that temporal guidance can even react to this by regrasping and eventually finishing the whole swapping task. However, even temporal guidance achieves only 28% to this strong perturbation since it s way too out-of-distribution and may require more data. Qualitatively, we visualize a generated robot trajectory with an unseen configuration in Figure 17. D.5 Additional Qualitative Results We present additional qualitative results to supplement our main findings in Section 6. Please refer to Figures 8 to 12 and 14 for detailed visual comparisons, which are discussed below. DFo T vs. Baselines (Section 6.2, Q1). We present additional qualitative comparisons of DFo T against baselines in Figure 14, as an extension to the qualitative results shown in Figure 4. Consistent with the quantitative findings in Table 1, DFo T produces more consistent and higher-quality samples compared to all baselines. Empirical Flexibility of DFo T (Section 6.2, Q3). As evidence of the empirical flexibility of DFo T, we present additional qualitative results on Real Estate10K in Figure 11. Our DFo T model successfully generates consistent samples, given histories that vary both in length and timestamps. This highlights the effectiveness of our new training objective, which transforms DFo T into a flexible multi-task model, uniformly achieving high performance across diverse tasks. Improving Video Generation via History Guidance (Section 6.3). In addition to the results shown in Figure 6a for Kinetics-600, we present further qualitative results on Real Estate10K in Figure 12, highlighting the effectiveness of vanilla history guidance in improving video generation. With increasing guidance scales, the generated samples exhibit significantly higher frame quality and consistency, likewise to the results on Kinetics-600. This behavior is consistent across different tasks extrapolation and showcasing the broad applicability of history guidance in any history-conditioned video generation History-Guided Video Diffusion Robustness to Out-of-Distribution (OOD) History (Section 6.4, Task 1). We provide additional qualitative results for Task 1 from Section 6.4, as illustrated in Figure 10. These results demonstrate that HG-t enables DFo T to uniquely remain robust to OOD history. Failure cases clearly observed in baselines show that typically, video diffusion models only perform well when the history is in-distribution. By composing in-distribution short history windows, HG-t can effectively approximate strictly OOD histories that were unseen during training. D.6 Detailed results on Ultra Long Video Generation (Section 6.5). We present extended results from Section 6.5 below. DFo T vs. SD on Long Rollout. To begin with, we highlight the significant challenges of generating long navigation videos using the Real Estate10K dataset. Specifically, we investigate the performance of SD, the most conventional and competitive baseline. To mitigate the stochastic nature of navigation that complicates comparisons, we evaluate DFo T with HG and SD on a simple navigation task of moving straight, which is almost deterministic. We avoid using interpolation applicable only to DFo T to ensure a fair comparison. The results, shown in Figure 9, indicate that SD struggles to maintain consistency with the history frame, failing around frame 30. We attribute this to SD s inferior quality and consistency, along with its inability to recover from small errors during generation. In contrast, DFo T with HG succeeds to stably roll out beyond frame 72. Alongside the qualitative comparison, we note that 4Di M (Watson et al., 2025), an SD model that, to our knowledge, produces the longest and highest-quality videos on Real Estate10K among the methods in the literature, generates videos with a maximum length of 32 frames, which is significantly shorter than our long navigation videos. More Samples. We present four samples of long navigation videos generated by DFo T with HG in Figures 8a to 8d. These samples demonstrate the capability of DFo T with HG to stably generate extremely long videos. The generated videos are notably longer than those in the training dataset, which primarily cover a single room or small area, rather than multiple connected rooms or areas.