# videotetris_towards_compositional_texttovideo_generation__225fd7ab.pdf Video Tetris: Towards Compositional Text-to-Video Generation Ye Tian1 Ling Yang1 Haotian Yang2 Yuan Gao2 Yufan Deng1 Jingmin Chen1 Xintao Wang2 Zhaochen Yu1 Xin Tao2 Pengfei Wan2 Di Zhang2 Bin Cui1 1Peking University 2Kuaishou Technology Project: videotetris.github.io Code: https://github.com/Yang Ling0818/Video Tetris Diffusion models have demonstrated great success in text-to-video (T2V) generation. However, existing methods may face challenges when handling complex (long) video generation scenarios that involve multiple objects or dynamic changes in object numbers. To address these limitations, we propose Video Tetris, a novel framework that enables compositional T2V generation. Specifically, we propose spatio-temporal compositional diffusion to precisely follow complex textual semantics by manipulating and composing the attention maps of denoising networks spatially and temporally. Moreover, we propose an enhanced video data preprocessing to enhance the training data regarding motion dynamics and prompt understanding, equipped with a new reference frame attention mechanism to improve the consistency of auto-regressive video generation. Extensive experiments demonstrate that our Video Tetris achieves impressive qualitative and quantitative results in compositional T2V generation. 1 Introduction With the significant development of diffusion models [1 3] recently, advanced text-to-video models [4 7] have emerged and demonstrated impressive results. However, these models often struggle with generating complex scenes following compositional prompts, such as "A man on the left walking his dog on the right", which requires the model to compose various objects spatially and temporally. Moreover, with a growing interest in generating long videos, existing methods [8 10] try to explore multi-prompt long video generation, which is typically limited to simple single-object scene changes. These methods fail to manage scenarios where the number of objects changes dynamically, often resulting in bizarre transformations that do not accurately follow the input text. To overcome these challenges, we introduce Video Tetris, a novel and effective diffusion-based framework to enable compositional text-to-video generation. Firstly, we define compositional video generation as encompassing two primary tasks: (i) Video Generation with Compositional Prompts, which involves integrating objects with various attributes and relationships into a complex and coherent video; and (ii) Long Video Generation with Progressive Compositional Prompts, where progressive refers to the continuous changes in the position, quantity, and presence of objects with different attributes and relationships. Then, we introduce a novel Spatio-Temporal Compositional Diffusion, which manipulates the cross-attention value of denoising network temporally and spatially, synthesizing videos that faithfully follow complex or progressive instructions. Subsequently, to enhance the ability of long video generation models to grasp complex semantics and generate intricate scenes encompassing various attributes and relationships, we propose an Enhanced Video Contributed equally. Contact: yangling0818@163.com Corresponding Authors. 38th Conference on Neural Information Processing Systems (Neur IPS 2024). a) Video Genera,on with Composi,onal Prompts A heroic robot on the le, and a magical girl on the right are saving the day. A handsome young man is drinking coffee on a wooden table. ---------> (transi+ons to) A handsome young man and a beau:ful young lady on his le, are drinking coffee on a wooden table. Video Crafter2 Ours Animate Diff Pika Gen-2 A cute brown dog and a sleepy cat are napping in the sun. Streaming T2V A cute brown squirrel in Antarctica, on a pile of hazelnuts cinematic. ---------> (transitions to) A cute brown squirrel and a cute white squirrel in Antarctica, on a pile of hazelnuts cinematic Free Noise Ours Streaming T2V (b) Long Video Genera,on with Progressive Composi,onal Prompts Figure 1: (a): Comparison in Video Generation with Compositoinal Prompts. (b): Comparision in Long Video Generation with Progressive Compositional Prompts. Video Tetris demonstrates superior performance, exhibiting precise adherence to position information, diverse attributes, interaction, consistent scene transitions, and high motion dynamics in compositional video generation. Data Preprocessing pipeline, augmenting the training data with enhanced motion dynamics and prompt semantics, enabling the model to perform more effectively in long video generation with progressive compositional generation. Finally, we propose a consistency regularization method, namely Reference Frame Attention, that maintains content consistency in a coherent representation space with latent noise while being capable of accepting arbitrary image inputs, ensuring the consistency of multiple objects across different frames and positions. fig. 1(a) showcases our Video Tetris s superior performance in compositional short video generation. We accurately compose two distinct objects with their own attributes while maintaining their respective "left" and "right" positions and ensuring natural interaction between multiple objects. As for long video generation comparisons in fig. 1(b), Free Noise [10] either depicts characters appearing abruptly and inexplicably transforming a man into a woman or depicts a squirrel transforming from a hazelnut. Streaming T2V [11] fails to incorporate information about new characters altogether, ignoring quantity information and exhibits severe color distortion in later stages. In contrast, our Video Tetris excels in generating long videos with progressive compositional prompts, seamlessly integrating new characters into the video scenes while maintaining consistent and accurate positional and quantity information. Notably, in generating long videos of the same length, Free Noise produces only minor variations within the same scene, whereas Video Tetris demonstrates significantly higher motion dynamics, resulting in outputs that more closely resemble long narrative videos. Our contributions are summarized as follows: 1). We introduce a Spatio-Temporal Compositional Diffusion method for handling scenes with multiple objects and following progressive complex prompts. 2). We develop an Enhanced Video Data Preprocessing pipeline to enhance auto-regressive long video generation through motion dynamics and prompt semantics 3). We propose a consistency regularization method with Reference Frame Attention that maintains content coherence in compositional video generation. 4). Extensive experiments show that Video Tetris can generate state-of-the-art quality compositional videos, as well as produce high-quality long videos that align with progressive compositional prompts while maintaining the best consistency. 2 Related Work Text-to-Video Diffusion Models The field of text-to-video generation has seen significant advancements with the progress of diffusion models [1, 12, 13] and the development of large-scale video-text paired datasets [14, 15]. Early works such as LVDM [16] and Model Scope [7], adapted 2D image diffusion models by flattening the U-Net architecture to a 3D U-Net and training on extensive video datasets. Subsequently, methods like Animated Diff [5] have incorporated temporal attention modules into the existing 2D latent diffusion models, preserving the established efficacy of T2I models. More recently, several transformers-based diffusion methods [17, 18, 6] have enabled large-scale joint training of videos and images, leading to significant improvements in generation quality. Long Video Generation Most existing text-to-video diffusion models have been trained on fixedsize video datasets due to the increased computational complexity and resource constraints. Consequently, these models are often limited to generating a relatively small number of frames, leading to significant degradation in quality when tasked with generating longer videos. Several advancements [19, 8, 10, 20] have sought to overcome this limitation through various strategies. More recently, Vlogger [9] and Sparse Ctrl [21] employ a masked diffusion model for conditional frame input. Although these masked diffusion approaches facilitate longer video generation, they often encounter model inconsistencies and quality degradation due to domain shifts in input. Streaming T2V [11] proposes a new paradigm, utilizing a Control Net [22]-like conditioning scheme to enable auto-regressive video generation. However, due to the low quality of the training data, the final video outputs often exhibit inconsistent and low-quality artifacts. Compositional Video Generation While current video generation models can synthesize textguided videos, they often face challenges in generating videos featuring multiple objects or adhering to multiple complex instructions, which requires the model to compose objects with diverse temporal and spatial relationships. In the realm of text-to-video diffusion models, exploration of such scenarios remains incomplete. Several text-to-image methods like RPG [23] leverage additional layout or regional information to facilitate more intricate image generation [24 27, 23]. Within video diffusion techniques, approaches like LVD [28] and Video Director GPT [29] employ a layout-to-video generator to produce videos based on spatial configurations. However, these layout-based methods often offer Figure 2: The overall pipeline of Video Tetris. We introduce Spatio-Temporal Compositional module for compositional video generation and Reference Frame Attention for consistency regularization. For longer video generation, a Control Net [22]-like branch can be adopted for auto-regressive generation. only rudimentary and suboptimal spatial guidance, struggling particularly with overlapping objects, thereby resulting in videos with unnatural content. In contrast, our method adopts a compositional region diffusion approach. By explicitly modeling the spatial positions of objects with cross attention maps, our approach allows the objects to naturally integrate and blend during the denoising process, resulting in more realistic and coherent video output. Overview In this section, we introduce our method Video Tetris for compositional text-to-video generation. Our goal is to develop an efficient approach that enables text-to-video models to handle scenes with multiple objects and follow sequential complex instructions. We first introduce Spatio Temporal Compositional Region Diffusion in section 3.1, which allows different objects to naturally integrate and blend during the denoising process in a training-free manner. Furthermore, for the task of generating long videos with progressive complex prompts, we construct an auto-regressive model based on the Control Net [22] architecture and introduced a Enhanced Video Data Preprocessing pipeline in section 3.2 to collect a high-quality video-text pair dataset to train our auto-regressive model for enhanced motion dynamics and prompt understanding. Combined with Spatio-Temporal Compositional Region Diffusion, our auto-regressive model can generate long videos with seamless transitions between diverse target scenes. Finally, we propose a consistency regularization with Reference Frame Attention in section 3.3 for better object appearance preserving. 3.1 Spatio-Temporal Compositional Diffusion Motivation To achieve natural compositional generation, a straightforward approach is to use the layout as a condition to guide the generation process. However, this method presents several challenges: (i) Requiring large-scale training. Given the significant potential for improvement in layout-to-image models, training a layout-to-video model or training temporal convolution and attention layers for a layout-to-image model would require substantial computational resources and may struggle to keep pace with the latest advancements in text-to-video models. (ii) Layoutbased generation models impose significant constraints on object bounding boxes. For long video duration, the need to continuously adjust the positions and sizes of these boxes to maintain coherent video content introduces a complex planning process, which adds complexity to the overall method. Therefore, instead of training a layout-to-video model, we utilize cross-attention for precise generation [30 35] and propose a training-free approach that directly adjusts the cross-attention of different targets [23, 36 39], as is shown in fig. 3. This approach aims to overcome the limitations of layout-based methods and leverage the potential of more flexible and efficient generation techniques. Localizing Subobjects with Prompt Decomposition For a given prompt p, we first decompose it temporally into contents at different frames: p = {p1, p2, , pt}, where t denotes the total number of frames and pi denotes the given text prompt at i-th frame. Subsequently, for the i-th Text Prompt # 1 A little dolphin is exploring an old city under the sea Cross Attention Text Prompt # 2 A little dolphin is exploring an old city under the sea with a green cute turtle Cross Attention Text Prompt # 3 A little dolphin with her huge father and is exploring an old city under the sea with a green cute turtle Video Timeline Text Prompt: A little dolphin starts exploring an old city under the sea, she first found a green turtle at the bottom, then her huge father comes along to accompany her at the right side. Temporal Decomposing Spa>o-Temporal Composing Spatio-Temporal Composing Spa>o-Temporal Composing Cross A+en,on Cross Attention Figure 3: Illustration of Spatio-Temporal Compositional Diffusion. For a given story "A little dolphin starts exploring an old city under the sea, she first found a green turtle at the bottom, then her huge father comes along to accompany her at the right side.", we first decompose it temporally to Text Prompt #1, #2 and #3, then we decompose each of them spatially to compute each sub-region s cross attention maps. Finally, we compose them spatio-temporally to form a natural story. frame, we decompose the original pi spatially into different sub-objects: {pi 0, pi 1, , pi n} with their corresponding region masks M i = {M i 0, M i 1, , M i n}, where n denotes the number of different objects. In this way, we decompose a prompt list temporally and spatially to acquire each sub-object s corresponding region information in the video timeline. We then calculate the cross attention value for the j-th sub-object at i-th frame as follows: Cross Attni j = Softmax(Qi(Ki j)T d )V i j M i j, K = WK ϕ(pi j), V = WV ϕ(pi j) (1) where Qi represents the query for the latent frame features, WK, WV are linear projections, ϕ denotes the text encoder, and d is the latent projection dimension of the latent frame features. LLM-based Automatic Spatio-Temporal Decomposer (Optional) Alternatively, the spatiotemporal decomposition process can directly utilize a Large Language Model (LLM) to automate tasks, given the robust performance of LLMs in language comprehension, reasoning, summarization and region generation ablilities [23, 28, 27, 26]. We employ the in-context learning (ICL) capability of LLMs and guide the model to use chain-of-thought (Co T) [40] reasoning. Concretely, we first guide the LLM to decompose the story temporally, generating frame-wise prompts, and reception each one of them with LLM for better semantic richness. Then we use another LLM to decompose each prompt spatially into multiple prompts corresponding to different objects, assigning a region mask to each sub-prompt. The specific prompt templates that include task rules (instructions), in-context examples (demonstrations) are detailed in table 4, table 5 and table 6 of appendix A.1. Spatio-Temporal Subobjects Composition After we decompose the original prompt list temporally and spatially, we then compose them together from spatial to temporal. To this end, we first compute the cross-attention value of all sub-objects Cross Attni region at i-th frame with: Cross Attni region = j=0 Cross Attni j (2) Subsequently, to ensure a cohesive transition across the boundaries of distinct regions and a seamless integration between the background and the entities within each region, we employ the weighted sum of the Cross Attnregion and the Cross Attnoriginal for the original compositional prompt p with : Cross Attni original = Softmax(Qi(Ki)T d )V i, K = WK ϕ(pi), V = WV ϕ(pi) Cross Attni = α Cross Attni original + (1 α) Cross Attni region. (3) Here α parameter is utilized to adjust the balance between global information and individual characteristics, aiming to achieve video content more aligned with human aesthetic perception. Finally, we naturally concatenate all the cross-attention values computed along the temporal dimension: Cross Attn = Concat(Cross Attn1, Cross Attn2, , Cross Attnt) (4) In this way, either for a pre-trained text-to-video model such as Modelscope [7], Animatediff [5], Video Crafter2 [4] and Latte [6], or an auto-regressive model for longer video generation like Streaming T2V[11], this approach can be directly applied in a training-free manner to obtain compositional, consistent and aesthetically pleasing results. 3.2 Enhanced Video Data Preprocessing Enhancement of Motion Dynamics For auto-regressive video generation, we empirically find Streaming T2V [11] is the most effective in producing consistent content. However, there is a notable tendency for the occurrence of poor-quality cases and color degradation in the later stages of video generation. We attribute this issue to the suboptimal quality of the original training data. To enhance the motion consistency and stability of long video generation, it is imperative to filter the video data to retain high-quality content with consistent motion dynamics. Inspired by Stable Video Diffusion [41], we empirically observed a significant correlation between a video s optical flow [42] score its motion magnitude. Excessively low optical flow often corresponds to static video frames, while excessively high optical flow typically indicates frames with intense changes. To ensure the generation of smooth and suitable video data, we filter Panda-70M [15] by selecting videos with average optical flow scores computed by RAFT [43] falling within a specified range (s1 to s2). Enhancedment of Prompt Semantics While the Panda-70M s videos exhibit the best visual quality, the paired prompts tend to be relatively brief, which conflicts with our objective of generating videos that adhere to intricate, detailed, and compositional prompts. Directly using such data for training can result in a video generation model that inadequately comprehends complex compositional prompts. Inspired by recent text-to-image research [23, 44, 45], it has been demonstrated that high-quality prompts significantly enhance the output quality of visual content. Therefore, after filtering the initial set of videos, we perform a recaptioning process on the selected samples to ensure they are better aligned with our objectives. We employ three multimodal LLMs to generate spatio-temporally intricate and detailed descriptions of each video, followed by a local LLM to consolidate these descriptions, extract common elements, and add further information. More details on this process can be found in appendix A.2. 3.3 Consistency Regularization with Reference Frame Attention Given our approach involves the addition and removal of different objects in long videos, maintaining the consistency of each object throughout the video is crucial for final outputs. Most consistent ID control methods, such as IP-Adapter [46], Streaming T2V [11], Instant ID [47], and Vlogger [9], typically encode reference images using an image encoder, often CLIP [48], and then integrate the results into the cross-attention block. However, since CLIP is pre-trained on image-text pairs, its image embeddings are designed to align with text. Consistency control, on the other hand, focuses on ensuring that the feature information of the same object in different frames is similar, which does not involve text. We hypothesize that using CLIP for this purpose is an indirect approach and propose Reference Frame Attention to maintain the inter-frame consistency of object features. Formally, we first directly encode the reference images, which are usually the initial frames where the object appears, using the same autoencoder as the pre-trained T2V model. This ensures that the computational target during latent denoising is spatially consistent with the reference target within the hidden representation space. We then train a 2D convolutional layer and projection layer that are structurally identical to those in the original pipeline. This process can be represented as: xref = W(Conv(Auto Encoder(fk:k+l))), (5) Text prompt: A brave knight and a wise wizard are journeying through a forest. Text prompt: A talking sponge on the le> and a superhero baby on the right are having an adventure. Video Crafter2 Ours Animate Diff Pika Gen-2 Figure 4: Qualitative Results of Video Generation with Compositional Prompts in Comparision with SOTA Text-to-Video Models where W, Conv denote the projection layer and the 2D convolutional layer, fk:k+l denotes the l frames from index k that are chosen for refernce. After encoding, we insert a Reference Frame Attention block in each attention block that calculates the cross-attention between the current object and the reference object, supplementing the existing attention blocks: Ref Attn = Softmax(QKT d )V, K = WK xref, V = WV xref (6) It is noteworthy that to ensure the consistency of different objects across various regions, we need to separately multiply the corresponding object s region mask with Q, K, and V during this computation process, and in practical applications, when a new object emerges in the auto-regressive long video, we precompute its corresponding xref in the relevant regions for further process. 4 Experiments 4.1 Experimental Setups We conducted our experiments in two specific scenarios: Video Generation with Compositional Prompts and Long Video Generation for progressive Compositional Prompts. For the first scenario, we directly applied our Spatio-Temporal Compositional Diffusion on Video Crafter2 [4] to generate videos with F = 16 frames. For the second scenario, we employed the core Control Net [22]- like branch from Streaming T2V [11] as the backbone and processed the Panda-70M [15] dataset using the Enhanced Video Data Preprocessing methods in section 3.2 as the training set. For both scenarios, we used Chat GPT3 to generate 100 different prompts/prompt lists as input to the models, generated 6 videos for each prompt, and randomly selected one for comparison. Additional model hyperparameters and implementation details of Video Tetris are provided in appendix A.5. 3chat.openai.com Table 1: Quantitative Results of Video Generation with Compositional Prompts Method VBLIP-VQA VUnidet CLIP-SIM Animatediff [5] 0.3834 0.1921 0.8676 Video Crafter2 [4] 0.4510 0.1719 0.9249 Gen-2 [53] (Commercial) 0.4427 0.1503 0.9421 Pika [54] (Commercial) 0.4219 0.1782 0.9736 LVD [28] 0.4820 0.1934 0.8873 Video Tetris (Ours) 0.5563 0.2350 0.9312 4.2 Metrics To evaluate compositinal video generation, existing metrics, such as CLIPScore [48] and Fréchet Video Distance (FVD) [49], assess coarse text-video and video-video similarity but do not capture detailed correspondences in object-level attributes and spatial relationships. Instead, we extended the T2I-Comp Bench [50] to the video domain and introduced the following metrics for compositional text-to-video evaluation: VBLIP-VQA: the average BLIP [51]-VQA score averaged across all frames and VUnidet: the average Unidet [52] score averaged across all frames. In addition, we followed previous work [10] and used CLIP-SIM [48] to measure the content consistency of generated videos by calculating the CLIP [48] similarity among adjacent frames of generated videos. 4.3 Video Generation with Compositional Prompts Qualitative Results We compare our Video Tetris with several state-of-the-art text-to-video (T2V) models on their ability to generate videos based on complex compositional prompts. These models include open-source options like LVD [28], Video Crafter2 [4], and Animatediff [5], as well as commercial models Gen-2 [53] and Pika [54]. Using Video Crafter2 as a backbone, we directly evaluate our Spatio-Temporal Compositional Diffusion module s training-free performance. In fig. 4, we show text-to-video synthesis results. For the prompt, "A brave knight and a wise wizard are journeying through a forest," most models generate two similar characters, blending features and losing individual distinctions. This highlights challenges in semantic alignment and compositional modeling for open-source models. In contrast, our Video Tetris preserves the distinct characteristics of each object and integrates them seamlessly with the background without confining them to fixed regions. For the prompt, "A talking sponge on the left and a superhero baby on the right are having an adventure," models like Animate Diff split the image, while Runaway Gen-2, Pika, and Video Crafter2 produce misaligned characters. LVD produces entangled features, resulting in disordered representations. In contrast, our method accurately aligns objects to their specified positions while maintaining high video quality, outperforming other methods. Additional examples in fig. 8 demonstrate our model s capability to handle more complex prompts with multiple objects, maintaining high quality and adherence to compositional semantics. Quantitative Results We report our quantitative results in table 1. Our Video Tetris achieves the best VBLIP-VQA and VUnidet scores across all models, demonstrating our superiority for complex compositional generation. We also achieved a CLIP-SIM higher than the original backbone Video Cratfer 2[4] and comparable to commercial models thanks to accurate semantic understanding. This proves that better text-video alignment can benefit overall consistency. User Study For further evaluation, we conducted a user study comparing our method with other video generation models, reported in appendix A.4. Using GPT-4, we collected 100 compositional prompts and generated 100 video samples across diverse scenes, styles, and objects. Users compared model pairs by selecting their preferred video from three options: method 1, method 2, and comparable results. 4.4 Long Video Generation for Progressive Compositional Prompts Qualitative Results We compared our Video Tetris with state-of-the-art long video generation models Free Noise [10] and Streaming T2V [11]. Free Noise inherently supports multi-prompts, and we provide Streaming T2V with different prompts at various frame indexes for multi-prompt video generation. We present our qualitative experimental results in fig. 1 and fig. 5. For the multi- Free Noise Streaming T2V Ours Progressive Prompts : A brave young knight is journeying through a forest ---------> (transi+ons to) A brave young knight and a wise wizard are journeying through a forest Figure 5: Qualitative Results of Long Video Generation for Progressive Compositional Prompts. prompt sequence from "A brave young knight is journeying through a forest" to "A brave young knight and a wise wizard are journeying through a forest," Free Noise generates consistent content; however, the new character appears abruptly, and the two characters switch identities by the end. Additionally, Free Noise consistently produces near-static global motion, with neither the background nor character positions changing. Conversely, Streaming T2V produces bizarre videos in which the knight disappears for half the duration and a merged character appears. In contrast, our method successfully models stable and consistent changes in long videos. The new character appears naturally and integrates seamlessly with the existing background throughout the video. Moreover, our approach achieves significantly more dynamic motion compared to Free Noise. This further demonstrates our method s capability of generating long videos that fully adhere to the evolving semantics while maintaining overall consistency. Quantitative Results We report our quantitative results in table 2. We achieve the best VBLIPVQA and VUnidet scores across all models, demonstrating the robust generation capability of our model in compositional video generation. Free Noise archives a better CLIP-SIM score due to its unique noise scheduling method, but this empirically damages transitions in a natural story. Table 2: Quantitative Results of Long Video Generation for Progressive Compositional Prompts Method VBLIP-VQA VUnidet CLIP-SIM Free Noise [10] 0.4372 0.1823 0.9706 Streaming T2V [11] 0.2412 0.1367 0.6720 Video Tetris (Ours) 0.4839 0.2137 0.9521 4.5 Ablation Study Effect of Enhanced Video Data Preprocessing We conducted an ablation study about the Enhanced Video Data Preprocessing pipeline, and show the results in fig. 6 and table 3. We directly compare our auto-regressive generation results with the original Streaming T2V [11] using the original prompts and comparison methods. fig. 6 demonstrates the significant improvements we achieved. For the given prompt our model better captures the semantics of "early morning sunlight." In addition, we generate long videos with all test prompts in Streaming T2V, and report our MAWE [11], CLIP(image-text alignment), AE [55] and CLIP-SIM scores, which further proves our effectiveness. Table 3: Quantitative Comparison of Ablation Study. Method MAWE CLIP AE CLIP-SIM Free Noise [10] 49.53 32.14 4.79 0.91 Streaming T2V [11] 10.26 31.30 5.07 0.93 Video Tetris w/o Reference Frame Attention 10.21 33.50 7.21 0.92 Video Tetris (Ours) 9.98 34.80 8.07 0.96 Streaming T2V Ours w/o RFA Ours Text prompt: Close flyover over a large wheat field in the early morning sunlight. Figure 6: Ablation Study. Comparison Between the original Streaming T2V [11], Video Tetris w/o Reference Frame Attention and Video Tetris. Effect of Reference Frame Attention We also conducted an ablation study about the Reference Frame Attention, as demonstrated in fig. 6 and table 3. We observe from the result that our Reference Frame Attention achieves more consistent outputs, and the frequency of color artifacts significantly decreases, resulting in a more uniform overall color. This highlights the benefit of aligning reference and noise information semantically in the latent space. We provide more ablation studies about compositional approaches in appendix A.3. 5 Conclusion and Discussion Conclusion In this study, we addressed the limitations of current video generation models incapable of generating compositional video content and introduced a novel Video Tetris framework that enables high-quality compositional generation. We propose an efficient Spatio-Temporal Compositional module that decomposes and composes semantic information temporally and spatially in the crossattention space. Additionally, to further enhance consistency in auto-regressive long video generation, we introduced an Enhanced Video Data Preprocessing pipeline and designed a brand new Reference Frame Attention module. Extensive experiment results confirmed the superiority of our paradigm in extending the generative capabilities of video diffusion models. Limitations Our proposed method can generate both short and long compositional videos. For fixed text-to-video generation, we can directly enhance the compositional performance of existing models. However, for long videos, due to the current performance limitations of long video generation models, our method inevitably encounters some bad cases. Additionally, using Control Net [22] for auto-regressive long video generation results in huge computation cost and overly strong control information, leading to an excessive number of transition frames. Broader Impact Recent notable progress in text-to-video diffusion models has opened up new possibilities in creative design, autonomous media, and other fields. However, the dual-use nature of this technology raises concerns about its societal impact. There is a significant risk of misuse of video diffusion models, particularly in the impersonation of individuals. It is essential to emphasize that our algorithm is designed to enhance the quality of video generation, and we do not support or provide means for malicious uses. Acknowledgement This work is supported by National Natural Science Foundation of China (U23B2048, U22B2037), Beijing Municipal Science and Technology Project (Z231100010323002), research grant No. SH2024JK29 and High-performance Computing Platform of Peking University. [1] J. Ho, A. Jain, and P. Abbeel, Denoising diffusion probabilistic models, Advances in neural information processing systems, vol. 33, pp. 6840 6851, 2020. [2] Y. Song, J. Sohl-Dickstein, D. P. Kingma, A. Kumar, S. Ermon, and B. Poole, Score-based generative modeling through stochastic differential equations, ar Xiv preprint ar Xiv:2011.13456, 2020. [3] J. Sohl-Dickstein, E. Weiss, N. Maheswaranathan, and S. Ganguli, Deep unsupervised learning using nonequilibrium thermodynamics, in International conference on machine learning, pp. 2256 2265, PMLR, 2015. [4] H. Chen, Y. Zhang, X. Cun, M. Xia, X. Wang, C. Weng, and Y. Shan, Videocrafter2: Overcoming data limitations for high-quality video diffusion models, ar Xiv preprint ar Xiv:2401.09047, 2024. [5] Y. Guo, C. Yang, A. Rao, Z. Liang, Y. Wang, Y. Qiao, M. Agrawala, D. Lin, and B. Dai, Animatediff: Animate your personalized text-to-image diffusion models without specific tuning, International Conference on Learning Representations, 2024. [6] X. Ma, Y. Wang, G. Jia, X. Chen, Z. Liu, Y.-F. Li, C. Chen, and Y. Qiao, Latte: Latent diffusion transformer for video generation, ar Xiv preprint ar Xiv:2401.03048, 2024. [7] J. Wang, H. Yuan, D. Chen, Y. Zhang, X. Wang, and S. Zhang, Modelscope text-to-video technical report, ar Xiv preprint ar Xiv:2308.06571, 2023. [8] G. Oh, J. Jeong, S. Kim, W. Byeon, J. Kim, S. Kim, H. Kwon, and S. Kim, Mtvg: Multi-text video generation with text-to-video models, ar Xiv preprint ar Xiv:2312.04086, 2023. [9] S. Zhuang, K. Li, X. Chen, Y. Wang, Z. Liu, Y. Qiao, and Y. Wang, Vlogger: Make your dream a vlog, ar Xiv preprint ar Xiv:2401.09414, 2024. [10] H. Qiu, M. Xia, Y. Zhang, Y. He, X. Wang, Y. Shan, and Z. Liu, Freenoise: Tuning-free longer video diffusion via noise rescheduling, ar Xiv preprint ar Xiv:2310.15169, 2023. [11] R. Henschel, L. Khachatryan, D. Hayrapetyan, H. Poghosyan, V. Tadevosyan, Z. Wang, S. Navasardyan, and H. Shi, Streamingt2v: Consistent, dynamic, and extendable long video generation from text, ar Xiv preprint ar Xiv:2403.14773, 2024. [12] L. Yang, Z. Zhang, Y. Song, S. Hong, R. Xu, Y. Zhao, W. Zhang, B. Cui, and M.-H. Yang, Diffusion models: A comprehensive survey of methods and applications, ACM Computing Surveys, vol. 56, no. 4, pp. 1 39, 2023. [13] L. Yang, Z. Zhang, Z. Yu, J. Liu, M. Xu, S. Ermon, and B. CUI, Cross-modal contextualized diffusion models for text-guided visual generation and editing, in International Conference on Learning Representations, 2024. [14] Y. Wang, Y. He, Y. Li, K. Li, J. Yu, X. Ma, X. Li, G. Chen, X. Chen, Y. Wang, P. Luo, Z. Liu, Y. Wang, L. Wang, and Y. Qiao, Internvid: A large-scale video-text dataset for multimodal understanding and generation, in The Twelfth International Conference on Learning Representations, 2024. [15] T.-S. Chen, A. Siarohin, W. Menapace, E. Deyneka, H.-w. Chao, B. E. Jeon, Y. Fang, H.-Y. Lee, J. Ren, M.-H. Yang, et al., Panda-70m: Captioning 70m videos with multiple cross-modality teachers, ar Xiv preprint ar Xiv:2402.19479, 2024. [16] J. Ho, T. Salimans, A. Gritsenko, W. Chan, M. Norouzi, and D. J. Fleet, Video diffusion models, Advances in Neural Information Processing Systems, vol. 35, pp. 8633 8646, 2022. [17] O. Bar-Tal, H. Chefer, O. Tov, C. Herrmann, R. Paiss, S. Zada, A. Ephrat, J. Hur, Y. Li, T. Michaeli, et al., Lumiere: A space-time diffusion model for video generation, ar Xiv preprint ar Xiv:2401.12945, 2024. [18] A. Gupta, L. Yu, K. Sohn, X. Gu, M. Hahn, L. Fei-Fei, I. Essa, L. Jiang, and J. Lezama, Photorealistic video generation with diffusion models, ar Xiv preprint ar Xiv:2312.06662, 2023. [19] F.-Y. Wang, W. Chen, G. Song, H.-J. Ye, Y. Liu, and H. Li, Gen-l-video: Multi-text to long video generation via temporal co-denoising, ar Xiv preprint ar Xiv:2305.18264, 2023. [20] W. Weng, R. Feng, Y. Wang, Q. Dai, C. Wang, D. Yin, Z. Zhao, K. Qiu, J. Bao, Y. Yuan, C. Luo, Y. Zhang, and Z. Xiong, Art v: Auto-regressive text-to-video generation with diffusion models, 2023. [21] Y. Guo, C. Yang, A. Rao, M. Agrawala, D. Lin, and B. Dai, Sparsectrl: Adding sparse controls to text-to-video diffusion models, ar Xiv preprint ar Xiv:2311.16933, 2023. [22] L. Zhang, A. Rao, and M. Agrawala, Adding conditional control to text-to-image diffusion models, in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3836 3847, 2023. [23] L. Yang, Z. Yu, C. Meng, M. Xu, S. Ermon, and B. Cui, Mastering text-to-image diffusion: Recaptioning, planning, and generating with multimodal llms, in International Conference on Machine Learning, 2024. [24] Y. Li, H. Liu, Q. Wu, F. Mu, J. Yang, J. Gao, C. Li, and Y. J. Lee, Gligen: Open-set grounded text-to-image generation, in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 22511 22521, 2023. [25] J. Xie, Y. Li, Y. Huang, H. Liu, W. Zhang, Y. Zheng, and M. Z. Shou, Boxdiff: Text-to-image synthesis with training-free box-constrained diffusion, in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 7452 7461, 2023. [26] L. Lian, B. Li, A. Yala, and T. Darrell, Llm-grounded diffusion: Enhancing prompt understanding of text-to-image diffusion models with large language models, ar Xiv preprint ar Xiv:2305.13655, 2023. [27] L. Qu, S. Wu, H. Fei, L. Nie, and T.-S. Chua, Layoutllm-t2i: Eliciting layout guidance from llm for text-to-image generation, in Proceedings of the 31st ACM International Conference on Multimedia, pp. 643 654, 2023. [28] L. Lian, B. Shi, A. Yala, T. Darrell, and B. Li, Llm-grounded video diffusion models, in The Twelfth International Conference on Learning Representations, 2023. [29] H. Lin, A. Zala, J. Cho, and M. Bansal, Videodirectorgpt: Consistent multi-scene video generation via llm-guided planning, ar Xiv preprint ar Xiv:2309.15091, 2023. [30] Z. Tu, W. Pan, Y. Duan, J. Tang, and C. Li, Rgbt tracking via reliable feature configuration, Science China Information Sciences, vol. 65, no. 4, p. 142101, 2022. [31] C. Xu, Z. Cui, C. Wang, C. Zhou, and J. Yang, Learning cross-modal interaction for rgb-t tracking, Science China Information Sciences, vol. 38, p. 44, 2023. [32] C. Wang, S.-J. Cai, B.-X. Shi, and Z.-H. Chong, Visual topic semantic enhanced machine translation for multi-modal data efficiency, Journal of Computer Science and Technology, vol. 38, no. 6, pp. 1223 1236, 2023. [33] Y.-Y. Zhang, L. Niu, Y. Makihara, J.-F. Zhang, W.-J. Zhao, Y. Yagi, and L.-Q. Zhang, Natural image matting with attended global context, Journal of Computer Science and Technology, vol. 38, no. 3, pp. 659 673, 2023. [34] A. Wang, S. Wen, Y. Gao, Q. Li, K. Deng, and C. Pang, Toward enhancing room layout estimation by feature pyramid networks, Data Science and Engineering, vol. 7, no. 3, pp. 213 224, 2022. [35] Z. Lei, G. Zhang, L. Wu, K. Zhang, and R. Liang, A multi-level mesh mutual attention model for visual question answering, Data Science and Engineering, vol. 7, no. 4, pp. 339 353, 2022. [36] N. Liu, S. Li, Y. Du, A. Torralba, and J. B. Tenenbaum, Compositional visual generation with composable diffusion models, in European Conference on Computer Vision, pp. 423 439, Springer, 2022. [37] R. Wang, Z. Chen, C. Chen, J. Ma, H. Lu, and X. Lin, Compositional text-to-image synthesis with attention map control of diffusion models, in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38, pp. 5544 5552, 2024. [38] H. Chefer, Y. Alaluf, Y. Vinker, L. Wolf, and D. Cohen-Or, Attend-and-excite: Attention-based semantic guidance for text-to-image diffusion models, ACM Transactions on Graphics (TOG), vol. 42, no. 4, pp. 1 10, 2023. [39] W. Feng, X. He, T.-J. Fu, V. Jampani, A. Akula, P. Narayana, S. Basu, X. E. Wang, and W. Y. Wang, Training-free structured diffusion guidance for compositional text-to-image synthesis, ar Xiv preprint ar Xiv:2212.05032, 2022. [40] J. Wei, X. Wang, D. Schuurmans, M. Bosma, F. Xia, E. Chi, Q. V. Le, D. Zhou, et al., Chain-of-thought prompting elicits reasoning in large language models, Advances in neural information processing systems, vol. 35, pp. 24824 24837, 2022. [41] A. Blattmann, T. Dockhorn, S. Kulal, D. Mendelevitch, M. Kilian, D. Lorenz, Y. Levi, Z. English, V. Voleti, A. Letts, et al., Stable video diffusion: Scaling latent video diffusion models to large datasets, ar Xiv preprint ar Xiv:2311.15127, 2023. [42] G. Farnebäck, Two-frame motion estimation based on polynomial expansion, in Image Analysis: 13th Scandinavian Conference, SCIA 2003 Halmstad, Sweden, June 29 July 2, 2003 Proceedings 13, pp. 363 370, Springer, 2003. [43] Z. Teed and J. Deng, Raft: Recurrent all-pairs field transforms for optical flow, in Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part II 16, pp. 402 419, Springer, 2020. [44] E. Segalis, D. Valevski, D. Lumen, Y. Matias, and Y. Leviathan, A picture is worth a thousand words: Principled recaptioning improves image generation, ar Xiv preprint ar Xiv:2310.16656, 2023. [45] W. Zheng, J. Teng, Z. Yang, W. Wang, J. Chen, X. Gu, Y. Dong, M. Ding, and J. Tang, Cogview3: Finer and faster text-to-image generation via relay diffusion, ar Xiv preprint ar Xiv:2403.05121, 2024. [46] H. Ye, J. Zhang, S. Liu, X. Han, and W. Yang, Ip-adapter: Text compatible image prompt adapter for text-to-image diffusion models, ar Xiv preprint ar Xiv:2308.06721, 2023. [47] Q. Wang, X. Bai, H. Wang, Z. Qin, and A. Chen, Instantid: Zero-shot identity-preserving generation in seconds, ar Xiv preprint ar Xiv:2401.07519, 2024. [48] A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, et al., Learning transferable visual models from natural language supervision, in International conference on machine learning, pp. 8748 8763, PMLR, 2021. [49] T. Unterthiner, S. Van Steenkiste, K. Kurach, R. Marinier, M. Michalski, and S. Gelly, Towards accurate generative models of video: A new metric & challenges, ar Xiv preprint ar Xiv:1812.01717, 2018. [50] K. Huang, K. Sun, E. Xie, Z. Li, and X. Liu, T2i-compbench: A comprehensive benchmark for open-world compositional text-to-image generation, Advances in Neural Information Processing Systems, vol. 36, pp. 78723 78747, 2023. [51] J. Li, D. Li, C. Xiong, and S. Hoi, Blip: Bootstrapping language-image pre-training for unified visionlanguage understanding and generation, in International conference on machine learning, pp. 12888 12900, PMLR, 2022. [52] X. Zhou, V. Koltun, and P. Krähenbühl, Simple multi-dataset detection, in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 7571 7580, 2022. [53] Gen-2. https://research.runwayml.com/gen2,, 2024. [54] Pika labs. https://www.pika.art/,, 2023. [55] C. Schuhmann, R. Beaumont, R. Vencu, C. Gordon, R. Wightman, M. Cherti, T. Coombes, A. Katta, C. Mullis, M. Wortsman, et al., Laion-5b: An open large-scale dataset for training next generation image-text models, Advances in Neural Information Processing Systems, vol. 35, pp. 25278 25294, 2022. [56] J. Achiam, S. Adler, S. Agarwal, L. Ahmad, I. Akkaya, F. L. Aleman, D. Almeida, J. Altenschmidt, S. Altman, S. Anadkat, et al., Gpt-4 technical report, ar Xiv preprint ar Xiv:2303.08774, 2023. [57] H. Liu, W. Yan, M. Zaharia, and P. Abbeel, World model on million-length video and language with ringattention, ar Xiv preprint ar Xiv:2402.08268, 2024. [58] H. Zhang, X. Li, and L. Bing, Video-llama: An instruction-tuned audio-visual language model for video understanding, ar Xiv preprint ar Xiv:2306.02858, 2023. [59] M. Maaz, H. Rasheed, S. Khan, and F. S. Khan, Video-chatgpt: Towards detailed video understanding via large vision and language models, ar Xiv preprint ar Xiv:2306.05424, 2023. [60] B. Lin, B. Zhu, Y. Ye, M. Ning, P. Jin, and L. Yuan, Video-llava: Learning united visual representation by alignment before projection, ar Xiv preprint ar Xiv:2311.10122, 2023. [61] M. Chen, I. Laina, and A. Vedaldi, Training-free layout control with cross-attention guidance, 2023. A.1 LLM-based Automatic Spatio-Temporal Decomposer Large Language Models (LLMs) have showcased impressive language comprehension, reasoning, and summarization abilities. Using LLMs for specific region generation has been proven efficient and effective in previous works [23, 28, 27, 26]. In our work, we employ the in-context learning (ICL) capability of LLMs to generate reasonable and natural temporal information and spatial regions, eliminating the need for manual spatio-temporal prompt decomposing for each prompt. We construct prompt templates that include task rules (instructions), in-context examples (demonstrations), and the user s input prompt (test). The specific prompts used for decomposing prompts spatio-temporally and recaptioning sub-prompts are detailed in table 4, table 5 and table 6. In designing our LLM prompts, we focused on clearly defining the roles and guidelines for the LLM, building on insights from previous research [26, 23, 28]. Moreover, by guiding the model to use chain-of-thought(Co T) [40] reasoning, where it articulates its reasoning process during generation, we empirically achieved better outcomes. This Co T method produced more accurate suggestions compared to when the model s reasoning process was not explicitly detailed. In our experiment, we choose GPT-4 [56] as our LLM. # Your Role: Excellent Story Planner ## Objective: Analyze your input descriptions and plan a reasonable story by providing the frame-specific prompt. ## Process Steps 1. Read the user input story with his given total frames. 2. Analyze the story, and specify every object and its attribute. 3. Crafting a video timeline with a prompt and its corresponding frame index, Keep the frame index an integer multiple of 8. 4. Explain your understanding (reasoning) and then format your result as ## Examples - Example 1 User prompt: I would like to create a story about a man in a cafe. He first drinks coffee alone on a wooden table, and then a young lady with blonde hair comes to company. They started chatting joyfully at the end. Total frames: 80 Reasoning: This story contains three main objects: a man, a wooden table, and a young lady with blonde hair. We can split the 80 frames into 3 different parts to construct a story. Output: [ 0 : "A man is drinking coffee on a wooden table", "32": "A man and a young lady with blonde hair are drinking coffee on a wooden table", "64": "A man and a young lady with blonde hair are drinking coffee and chatting joyfully on a wooden table"] - Example 2: Your Current Task: Follow the steps closely and accurately identify frame index specific sub-prompts based on the given story and total frames. Ensure adherence to the above output format. User prompt: {the input user prompt} Reasoning: Table 4: Our full prompt for Decomposing Prompts Temporally. # Your Role: Excellent Prompt Recaptioner You are an excellent recaptioning bot. Your task is to recaption each given prompt with a more descriptive prompt while maintaining the original meaning with at least 40 words. You will be given with an caption of a video, this caption is very short and simple, only containing the main entity and perhaps the simplest description of the background. Please take the provided caption and expand to at least 40 words upon it by providing additional details. You can start this procedure by following these rules: ## Objective: Recaption each given prompt with a more descriptive prompt while maintaining the original meaning with at least 40 words. ## Process Steps 1. If the original prompt contains words about the camera view, such as "top view of" or "camera clockwise", remember to also contain them in the recaption. 2. Describe each entity appearing in this original caption with at least more than two adjectives, making every entity as detailed as possible. 3. Using your knowledge to fulfill the background or any other thing that should or should not appear in this frames. Adding as much details as you can to enrich the caption, but you shouldn t change the original meaning of the prompt or any main entity. 4. Your recaption should contain at least 40 words, and you should keep it within 60 words. 5. Your answer should strictly follow the form : "Recaption: " 6. Your answer must not contain words like "video" or "frame". Only enrich the given prompt. ## Examples - Example 1: Original Caption: a man and woman are walking down a hallway Recaption: a man and woman in business attire are seen walking down a hallway in a professional building, engaged in a serious discussion with the man holding a book and the woman holding a clipboard, reflecting a professional or academic setting. - Example 2: Your Current Task: You will be given a caption of a video, this caption is very short and simple, only containing the main entity and perhaps the simplest description of the background. Please take the provided caption and expand it to at least 40 words by providing additional details. Original Caption: {the input user prompt} Recaption: Table 5: Our full prompt for Prompt Recaptioning. # Your Role: Excellent Region Planner ## Objective: Analyze your input prompts and plan every object s reasonable region in the frame with bounding boxes. ## Process Steps 1. Analyze the given multi-object prompt, consider a reasonable layout. 2. Define square images with top-left at [0, 0] and bottom-right at [1, 1], and the output Box Format: [Top-left x, Top-left y, Width, Height] 3. Assign each sub-object to a specific region. You can start by splitting the original image square. 4. The corresponding regions do not need to be very specific as long as the region includes the sub-object and all regions never overlap 5. Output the result, and present every object and its region with a bounding box. ## Examples - Example 1 User prompt: A handsome young man and a lady with blonde hair are drinking coffee on a wooden table. Output: ["a handsome young man": "[0.5, 0, 0.5, 0.8]", "a lady with blonde hair": "[0, 0, 0.5, 0.8]", "a wooden table": "[0, 0.8, 1, 0.2]"] - Example 2: Your Current Task: Follow the steps closely and accurately output each sub-object s bounding box. Ensure adherence to the above output format. User prompt: {the input user prompt} Output: Table 6: Our full prompt for the LLM Spatial Decomposer A.2 Dataset Prompt Recaptioning In this section, we detailed our dataset prompt recaptioning process. We first select the top three caption models ranked highest in [57], namely Video-LLa MA [58], Video-Chat GPT [59] and Videollava [60], and have them generate captions of 40-50 words for each filtered video. We assume that different caption models may be suitable for different types of video input, so we collect outputs from various models to ensure a comprehensive effect. We then append these captions to the original prompt caption provided by Panda-70M. All collected prompts are fed into a local LLM (in our experiment, LLama-34), to consolidate the captions, extract common elements, and add details. The final unified caption, around 40-50 words in length, is used for training each filtered video. A.3 Ablations about Effect of Spatio-Temporal Compositionl Diffusion In this section, we provided detailed explanations and ablations compared with similar prompt decomposing and object composing diffusion methods, LVD [28], Training-Free Layout Control with Cross-Attention Guidance [61], and Video Director GPT [29] in table 7. The decomposition methods in [28] and [29] only isolate specific tokens from the original prompt. This approach struggles with complex attributes or multiple identical objects, making it difficult for the video generation model to understand numeracy and attribute binding, leading to significant performance degradation in these scenarios. In contrast, our decomposition method extracts subobjects and then uses global information for recaptioning, resulting in richer descriptions. Our generated frames are more natural, detailed, and semantically accurate. 4https://llama.meta.com/llama3/ Next, we have conducted ablation studies about the composition method from [61] and reported the results in the table below. The backward guidance approach in [61] tends to restrict objects within specified boxes, offering low flexibility and poor responsiveness to multiple or overlapping objects. In contrast, our model s local-global information fusion ensures that the final generated images are more harmonious and visually appealing, performing better in compositional generation, even in overlapping regions. Table 7: Ablation Studies for Spatio-Temporal Compositional Diffusion Method VBLIP-VQA VUnidet CLIP-SIM Ours w/ Decomposing in LVD [28] 0.5203 0.2139 0.9303 Ours w/ Decomposing in [29] 0.4982 0.2237 0.9178 Ours w/ Composing in Video Director GPT [61] 0.5112 0.1857 0.9073 Video Tetris (Ours) 0.5563 0.2350 0.9312 A.4 User Study To verify the effectiveness of our proposed Video Tetris, we conduct an extensive user study across various scenes and models. Users compared model pairs by selecting their preferred video from three options: method 1, method 2, and comparable results. As presented in fig. 7, our method (orange in left) obtains more user preferences than others (blue in right), which further proving its effectiveness. Figure 7: User Study about Comparision with Other Video Generation Methods A.5 Model Hyperparameters and Implementation Details In this section, we further detailed our model hyperparameters and our implementation details. The hyperparameters in section 3.2 and section 3.3 are shown in table 8. In training process, we randomly drop out 5% of text prompts for classifier-free guidance training. We trained our model with batch size = 2 and learning rate = 1e-5 on 4 A800 GPUs for 16k steps in total. A.6 More examples We provided more examples in the figures below. Table 8: Hyperparameters of Video Tetris Dynamic-Aware Data Filtering n0 4 optical flow score threshold s1 0.25 optical flow score threshold s2 0.75 Diffusion Training Parametrization ϵ Diffusion steps 1000 Noise scheduler Linear β0 0.0085 βT 0.0120 Sampler DDIM Steps 50 η 1.0 Reference Frame Attention 2D Conv input dim 4 2D Conv output dim 320 2D Conv kernel size 3 2D Conv padding 1 MLP hidden layers 1 MLP inner dim 320 MLP output dim 1024 l 2 Text prompt: A curious detective and a sneaky thief are solving a mystery. Text prompt: A cheerful farmer and a hardworking blacksmith are building a barn. Text prompt: A speedy train and a leisurely boat are traveling across the country. Text prompt: A dolphin, her father dolphine and a turtle explore an underwater city. Text prompt: A wizard, a hot girl, a pirate, and a robot walk into a bar. Text prompt: A mother fox, a baby fox and a father fox go on a camping trip. Figure 8: More qualitative results of Video Tetris. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: The main claims made in the abstract and introduction accurately reflect the paper s contributions and scope: a novel framework for compositional video genration. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: Yes, we disscuss the limitications of the work in section 5 Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [NA] Justification: This paper does not include theoretical results. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: Yes. We give detailed instructions about experiment setup in in section 4.1 and appendix A.5 Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: We provide open access to our code and data are available at https://github.com/Yang Ling0818/Video Tetris . Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: We have also give detailed instructions about experiment setup in section 4.1 and appendix A.5 Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [NA] Justification: Error bars are not reported because it would be too computationally expensive. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: We have give detailed information about experiment setup in section 4.1 and appendix A.5 Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: Our research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [Yes] Justification: We disscuss the limitications of the work in section 5 Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: This paper poses no such risks. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: The creators or original owners of code used in the paper are properly credited, and the license and terms of use are explicitly mentioned and properly respected. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [Yes] Justification: New assets introduced in the paper are well documented. We provide them as supplementary material. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: This paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: This paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.