# latent_diffusion_planning_for_imitation_learning__5ba58c64.pdf Latent Diffusion Planning for Imitation Learning Amber Xie 1 Oleh Rybkin 2 Dorsa Sadigh 1 Chelsea Finn 1 Recent progress in imitation learning has been enabled by policy architectures that scale to complex visuomotor tasks, multimodal distributions, and large datasets. However, these methods often rely on learning from large amount of expert demonstrations. To address these shortcomings, we propose Latent Diffusion Planning (LDP), a modular approach consisting of a planner which can leverage action-free demonstrations, and an inverse dynamics model which can leverage suboptimal data, that both operate over a learned latent space. First, we learn a compact latent space through a variational autoencoder, enabling effective forecasting of future states in image-based domains. Then, we train a planner and an inverse dynamics model with diffusion objectives. By separating planning from action prediction, LDP can benefit from the denser supervision signals of suboptimal and action-free data. On simulated visual robotic manipulation tasks, LDP outperforms state-of-the-art imitation learning approaches, as they cannot leverage such additional data. 1 1. Introduction Combining large-scale expert datasets and powerful imitation learning policies has been a promising direction for robot learning. Recent methods using transformer backbones or diffusion heads (Octo Model Team et al., 2024; Kim et al., 2024; Zhao et al., 2024; Chi et al., 2023) have capitalized on new robotics datasets pooled together from many institutions (Khazatsky et al., 2024; Open X-Embodiment Collaboration et al., 2023), showing potential for learning generalizable robot policies. However, this recipe is fundamentally limited by expert data, as robotics demonstration data can be challenging, time-consuming, and expensive to 1Stanford 2UC Berkeley. Correspondence to: Amber Xie . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). 1Project Website and Code: https://amberxie88. github.io/ldp/ collect. While it is often easier to collect in-domain data that is suboptimal or action-free, these methods are not designed to use such data, as they rely on directly modeling optimal actions. Prior works in offline RL, reward-conditioned policies, or imitation learning from suboptimal demonstrations attempt to leverage suboptimal trajectories, though they are still unable to utilize action-free data. Notably, these works often make restrictive assumptions like access to either reward labels (Kumar et al., 2020; Chen et al., 2021; Kumar et al., 2019a), the optimality of demonstrations (Beliaev et al., 2022), or similar metrics (Zhang et al., 2022), which can be impractical or noisy to label. Other works implicitly attempt to use unlabelled, suboptimal data via pretraining on such data and later fine-tuning the policy on the optimal data (Radosavovic et al., 2023; Wu et al., 2023b; Cui et al., 2024). While these approaches can potentially learn representations during pretraining, it does not necessarily improve planning capabilities of these methods. Our key idea is to take a modular approach, where we separate learning a video planner from learning an inverse dynamics model. Each one of these two components can leverage different types of data. For instance, a planner can benefit from action-free data, while an IDM can leverage unlabelled suboptimal data. While using a modular approach has been proposed in recent prior work (Du et al., 2023a; Black et al., 2023), prior approaches focus on highlevel decision making by forecasting subgoals, limiting its capabilities for closed loop re-planning in robotics tasks. These works also operate across images, which are highdimensional and expensive to generate. To create an efficient modular approach that can benefit from all forms of data (suboptimal, action-free, and optimal), we propose learning the planner and inverse-dynamics model over a learned, compact latent space, allowing for closed-loop robot policies. Our imitation learning objective consists of forecasting a dense trajectory of latent states, scaling up gracefully to vision-based domains without the computational complexities of video generation. We propose Latent Diffusion Planning (LDP), which learns a planner that can be trained on action-free data; and an inverse dynamics model (IDM) that can be trained on data that may be suboptimal. First, it trains a variational autoencoder Latent Diffusion Planning for Imitation Learning Figure 1. Latent Diffusion Planning. Left: LDP separates the control problem into forecasting future states with a diffusion-based planner, and extracting actions with a diffusion-based inverse dynamics model (IDM). This design enables training on heterogeneous sources of data, including suboptimal data and action-free data. Right: Unlike action imitation methods such as diffusion policy, LDP is based on forecasting a dense temporal sequence of latent states as well as actions. Using powerful diffusion models for both of these objectives enables LDP to have competitive performance to state-of-the-art imitation learning. Further, unlike prior work on forecasting subgoals, LDP predicts a dense temporal sequence of latent states, which enables scalable closed-loop planning. with an image reconstruction loss, producing compressed latent embeddings that are used by the planner and inverse dynamics model. Then, it learns an imitation learning policy through two components: (1) a planner, which consumes demonstration state sequences, which may be action-free, and (2) an inverse dynamics model, trained on in-domain, possibly suboptimal, environment interactions. As diffusion objectives have proven to be effective for imitation learning in robotics tasks (Chi et al., 2023), we use diffusion for both forecasting plans (planner) and extracting actions (IDM), which enables competitive performance. Our method is closed-loop and reactive, as planning over latent space is much faster than generating visually and physically consistent video frames. In summary, our main contributions are threefold: We propose a novel imitation learning algorithm, Latent Diffusion Planning, a simple, diffusion planning-based method comprised of a learned visual encoder, latent planner, and an inverse dynamics model. We show that Latent Diffusion Planning can be trained on suboptimal or action-free data, and improves from learning on such data in the regime where demonstration data is limited. We experimentally show that our method outperforms prior video planning-based work by leveraging temporally dense predictions in a latent space, which enables fast inference for closed-loop planning. 2. Related Work Imitation Learning in Robotics. One common approach to learning robot control policies is imitation learning, where policies are learned from expert-collected demonstration datasets. This is most commonly done via behavior cloning, which reduces policy learning to a supervised learning objective of mapping states to actions. Recently, Diffusion Policy (Chi et al., 2023) and Action Chunking with Transformers (Zhao et al., 2023) have shown successful results in complex manipulation tasks using action chunking and more expressive architectures. Diffusion models have also been successful in capturing multimodal human behavior (Pearce et al., 2023). Similarly, Behavior Transformer (Shafiullah et al., 2022) and VQ-Be T (Lee et al., 2024) improve the ability of policies to capture multimodal behaviors. In this work, we focus on forecasting a sequence of future states instead of actions, and use diffusion to capture multimodal trajectories. Learning from Unlabelled Suboptimal and Action-Free Data. Learning from suboptimal data has long been a goal of many robot learning methods, including reinforcement learning. A typical approach is offline reinforcement learning, which considers solving a Markov decision process from an offline dataset of states, actions, and reward (Levine et al., 2020; Kumar et al., 2020; Kostrikov et al., 2021; Hansen-Estruch et al., 2023; Yu et al., 2022). Particularly relevant are the approaches that use supervised learning conditioned on rewards (Schmidhuber, 2019; Kumar et al., 2019a; Chen et al., 2021). In this work, we want to leverage Latent Diffusion Planning for Imitation Learning suboptimal, reward-free data, such as play data or failed trajectories. In addition, we would like to avoid the additional complexity of annotating the data with rewards or training a value function which the offline RL methods rely on. Several works have also addressed learning from actionfree data, such as using inverse models (Torabi et al., 2018; Baker et al., 2022), latent action models (Edwards et al., 2019; Schmeckpeper et al., 2020; Bruce et al., 2024), or representation learning (Radosavovic et al., 2023; Wu et al., 2023b; Cui et al., 2024). In this work we focus on a simple recipe for robotic imitation learning that is naturally able to leverage action-free data through state forecasting. Diffusion and Image Prediction in Robot Learning. Diffusion models, due to their expressivity and training and sampling stability, have been applied to robot learning tasks. Diffusion has been used in offline reinforcement learning (Hansen-Estruch et al., 2023) and imitation learning (Chi et al., 2023). Diffuser (Janner et al., 2022) learns a denoising diffusion model on trajectories, including both states and actions, in a model-based reinforcement learning setting. Decision Diffuser (Ajay et al., 2023) extends Diffuser by showing compositionality over skills, rewards, and constraints, and instead diffuses over states and uses an inverse dynamics model to extract actions from the plan. Due to the complexity of modeling image trajectories, Diffuser and Decision Diffuser restrict their applications to low-dimensional states. To scale up to diffusing over higher-dimensional plans, Uni Pi (Du et al., 2023a; Ko et al., 2023) adapts video models for planning. Unlike works that rely on foundation models and video models for planning (Du et al., 2023b; Yang et al., 2024; Zhou et al., 2024), our method avoids computational and modeling complexities of generative video modeling by planning over latent embeddings instead. Previous works have used world models to plan over images in a compact latent space (Hansen et al., 2024; Hafner et al., 2019; 2020). In contrast with these works, we focus on single task imitation instead of reinforcement learning. Many prior works argue that state forecasting objectives are uniquely suitable for robotics to improve planning quality with trajectory optimization or reinforcement learning (Finn & Levine, 2017; Yang et al., 2023), by using the model directly to plan future states (Du et al., 2023b;a), as well as representation learning (Wu et al., 2023a; Radosavovic et al., 2023). We follow this line of work by proposing a planningbased method competitive to state-of-the-art robotic imitation learning that can leverage heterogeneous data sources. 3. Background Diffusion Models Diffusion models, such as Denoising Diffusion Probabilistic Models (DDPMs), are likelihoodbased generative models that learn an iterative denoising process from a Gaussian prior to a data distribution (Sohl Dickstein et al., 2015; Ho et al., 2020; Song et al., 2020). During training time, DDPMs are trained to reverse a single noising step. Then, at sampling time, to reverse the diffusion process, the model iteratively denoises a sample drawn from the known Gaussian prior. Diffusion models may also be conditioned on additional context. For example, text-to-image generative models are conditioned on text, Diffusion Policy is conditioned on visual observations, and Decision Diffuser can be conditioned on reward, skills, and constraints. Recent generative models have used Latent Diffusion Models, which trains a diffusion model in a learned, compressed latent space (Rombach et al., 2022; Peebles & Xie, 2023; Blattmann et al., 2023) to improve computational and memory efficiency. The latent space is typically learned via an autoencoder, with encoder E and decoder D trained to reconstruct x xˆ = D(E(x)). Instead of diffusing over x, the diffusion model is trained on diffusing over z = E(x). Imitation Learning In the imitation learning framework, we assume access to a dataset of expert demonstrations, D {(s0, x0, a0), . . . , (s T , x T , a T )}, generated by πE, an expert policy. si, xi, ai correspond to the state, image, and action at timestep i respectively. The imitation learning objective is to extract a policy πˆ(a|s, x) that most closely imitates πE. In robotics, this is typically approached through behavior cloning, which learns the mapping between states and actions directly via supervised learning. We consider single-task imitation, where the dataset corresponds to a single task. Diffusion Policy (Chi et al., 2023) is an instantiation of diffusion models for imitation learning that has shown success in simulated and real-world tasks. Diffusion Policy uses a DDPM objective to model the distribution of action sequences, conditioned on observations. The CNN instantiation uses a Conditional U-Net Architecture, based on the 1D Temporal CNN in (Janner et al., 2022), which encourages temporal consistency due to the inductive biases of convolutions. LDP s planner architecture is based on the CNN-based Diffusion Policy, though we forecast latent states instead of actions. Datasets of expert demonstrations often do not provide sufficient state distribution coverage to effectively solve a given task with imitation learning. However, there often exists additional data in the form of action-free or suboptimal data, which may consist of failed policy rollouts, play data, or miscellaneous environment interactions. Unfortunately, Latent Diffusion Planning for Imitation Learning behavior cloning assumes access to data annotated with optimal actions, so such additional data cannot be easily incorporated into training. 4. Latent Diffusion Planning Latent Diffusion Planning consists of three parts, as shown in Figure 1: (1) Training an image encoder via an image reconstruction loss, (2) learning an inverse dynamics model to extract actions at from pairs of latent states zt, zt+1, and (3) learning a planner to forecast future latents zt. Figure 2. After training the encoder, Latent Diffusion Planning trains two diffusion models. Top: We train a inverse dynamics model (IDM) with a diffusion objective to directly extract the actions that will be used for control from pairs of latent states. Bottom: We train a powerful latent diffusion model to forecast a chunk of future latent states. The planner and the IDM are used together to produce an action chunk, similar to (Chi et al., 2023). 4.1. Learning the Latent Space We circumvent planning over high-dimensional image observations by planning over a learned latent space. Similar to prior work in planning with world models (Watter et al., 2015; Ha & Schmidhuber, 2018; Hafner et al., 2020), we learn this latent space using an image reconstruction objective. Our planner thus becomes similar to video models that forecast image frames in a learned latent space (Yan et al., 2021; Hong et al., 2022; Blattmann et al., 2023). In this work, we train a variational autoencoder (Kingma Algorithm 1 Inference with Latent Diffusion Planning 1: Input: Encoder E, Planner ϵψ, IDM ϵξ, Planner Diffusion Timesteps Tp, IDM Diffusion Timesteps TIDM, Planning Horizon Hp, Action Horizon Ha 2: Observe initial state s0 and image x0; k = 0 3: while not done do 4: zk (E(xk), sk) // Diffuse over latent embedding plan 5: zˆk+1, ..., zˆk+Hp (0, I) 6: for t = Tp . . . 1 do 7: ϵˆ ϵψ(zˆk+1, ..., zˆk+Hp; zk, t) 8: Update zˆk+1, ..., zˆk+Hp using DDPM update with ϵˆ 9: end for // Diffuse over actions between latent embeddings 10: for i = 0 . . . Ha 1 do 11: aˆk+i (0, I) // Predict action for each timestep in action horizon 12: for t = TIDM . . . 1 do 13: ϵˆ ϵξ(aˆk+i; zˆk+i, zˆk+i+1, t) 14: Update aˆk+i using DDPM update with ϵˆ 15: end for 16: end for // Execute actions 17: for i = 0 . . . Ha 1 do 18: sk+i+1 env.step(sk+i, aˆk+i) 19: end for 20: k k + Ha 21: end while & Welling, 2014; Rezende et al., 2014) to obtain a latent encoder E and decoder D. Specifically, we optimize the βVAE (Higgins et al., 2017) objective, where x is our original image, z is our learned latent representation of the image, θ are the parameters for our decoder, ϕ are the parameters for our encoder, and β is the weight for the KL regularization term: LVAE(θ, ϕ; x, z, β) = Eqϕ(z | x)[log pθ(x | z)] βDKL(qϕ(z | x)||p(z)) (1) In practical scenarios, we may have a limited expert demonstration dataset, but much larger and diverse suboptimal or action-free datasets. In this phase of learning, we can make use of the visual information in such datasets for training a more robust latent encoder. Latent Diffusion Planning for Imitation Learning 4.2. Planner and Inverse Dynamics Model Our policy consists of two separate modules: (1) a planner over latent embeddings z, and (2) an inverse dynamics model similarly operating over the latent embeddings. The planner and IDM are both parameterized as DDPM models, motivated by the expressivity that diffusion models offer. The planner is conditioned on the current latent embedding, which consists of the concatenated latent image embedding and robot proprioception, and diffuses over a horizon of future embeddings. We use Diffusion Policy s Conditional U-Net architecture, with a CNN backbone. Concretely, we optimize the following objective: Lplanner(ψ, z) = Et,ϵ[||ϵψ(zˆk+1, . . . , zˆk+H; zk, t) ϵ||2] (2) where zk is the latent embedding at timestep k of the trajectory; zˆk+1, . . . , zˆk+H is the noised latent embedding sequence, with corresponding noise ϵ; H is the maximum horizon of the forecasted latent plan; t is the diffusion noise timestep; and ψ are the parameters of the planner diffusion model. Our inverse dynamics model is trained to reconstruct the action between a pair of states, conditioned on their associated latent embeddings. We use the MLPRes Net architecture from IDQL (Hansen-Estruch et al., 2023) to diffuse actions, as it is more lightweight. We optimize the loss: LIDM(ξ, z) = Et,ϵ[||ϵξ(aˆk; zk, zk+1, t) ϵ||2] (3) where zk is the latent embedding at timestep k of the trajectory; aˆk is the noised action, with corresponding noise ϵ; t is the diffusion noise timestep; and ξ are the parameters of the inverse dynamics diffusion model. Because our latent embedding is frozen from the learned VAE, the planner and IDM do not share parameters and can be trained separately (Figure 2). Then, at inference time, the two modules are combined to extract action sequences. First, the planner forecasts a future horizon of states via DDPM sampling (Alg. 1 Lines 6-9). Then, we use the inverse dynamics model to extract actions from latent embedding pairs produced by the planner, also via DDPM Diffusion (Alg. 1 Lines 12-15). Like Diffusion Policy, we employ receding-horizon control (Mayne & Michalska, 1988), and execute for a shorter horizon than the full forecasted horizon (Alg. 1 Lines 17-19). 5. Experiments We seek to answer the following questions: Does Latent Diffusion Planning leverage action-free data for improved planning? Is Latent Diffusion Planning comparable to state-ofthe-art imitation learning algorithms that leverage suboptimal data? Can Latent Diffusion Planning be an effective imitation learning method in a real-world robotics system, where there may be suboptimal or action-free data? 5.1. Experimental Setup Simulated Tasks We focus our experiments on 4 imagebased imitation learning tasks: (1) Robomimic Lift, (2) Robomimic Can, (3) Robomimic Square, and (4) ALOHA Sim Transfer Cube. Robomimic (Mandlekar et al., 2021) is a robotic manipulation and imitation benchmark, including the tasks Lift, Can, and Square. The Transfer Cube task is a simulated bimanual ALOHA task, in which one Viper X 6-Do F arm grabs a block and transfers it to the other arm (Zhao et al., 2023). To demonstrate the effectiveness of Latent Diffusion Planning, we assume a low demonstration data regime, such that additional suboptimal or action-free data can improve performance. For Can and Square, we use 100 out of the 200 demonstrations in the Robomimic datasets; for Lift, we use 3 demonstrations out of the 200 total; and for Transfer Cube, we use 25 demonstrations. To further emphasize the importance of suboptimal data, these demonstrations cover a limited state space of the environment. Our suboptimal data consists of 500 failed trajectories from an undertrained behavior cloning agent. Our action-free data consists of 100 demonstrations for Lift, Can, and Square from the Robomimic dataset, and 25 demonstrations for Cube. We evaluate the success rate out of 50 trials, using the best checkpoint from the last 5 saved checkpoints, with 2 seeds. Real World Task We create a real world implementation of the Robomimic Lift task, where the task is to pick up a red block from a randomly initialized position. We use a Frank Panda 7 degree of freedom robot arm, with a wrist-mounted Zed camera. We use the DROID setup (Khazatsky et al., 2024) and teleoperate via the Oculus Quest 2 headset. We use cartesian pose control. Latent Diffusion Planning for Imitation Learning Table 1. Leveraging Action-Free Data. LDP is able to leverage suboptimal and action-free data. We compare against DP baselines that leverage action-free data by using an IDM to relabel actions (DP-VPT) and video planning models that may use action-free data for the video planner (Uni Pi). We find LDP to perform better than both approaches, especially when combined with suboptimal data. Method Lift Can Square ALOHA Cube Average DP 0.60 0.00 0.63 0.01 0.48 0.00 0.32 0.00 0.51 DP-VPT 0.69 0.01 0.75 0.01 0.48 0.04 0.45 0.03 0.59 Uni Pi-OL + Action-Free 0.09 0.05 0.23 0.03 0.07 0.03 0.02 0.00 0.11 Uni Pi-CL + Action-Free 0.14 0.02 0.32 0.04 0.09 0.01 0.17 0.03 0.18 LDP 0.69 0.03 0.70 0.02 0.46 0.00 0.64 0.04 0.65 LDP + Action-Free 0.67 0.01 0.78 0.04 0.47 0.03 0.70 0.02 0.66 LDP + Action-Free + Subopt 1.00 0.00 0.98 0.00 0.83 0.01 0.97 0.01 0.95 We collect 82 demonstrations, collect 84 suboptimal trajectories, and 12 action-free demonstrations. Our suboptimal data consists of failed trajectories from policy evaluations. This is an effective way to reuse the data generated during iterations of training, that algorithms modeling actions, such as Diffusion Policy, cannot use. Action Free data may consist of kinesthetic demonstrations, human videos, or handheld demonstrations (Chi et al., 2024). In our case, we collect teleoperated demonstrations with actions removed. To evaluate our policies, we calculate the success rate across 45 evaluation trials. To thoroughly evaluate performance across the initial state space, we evaluate across a grid of 3x3 points, with 5 attempts per point. We evaluate 3 seeds per method. Baselines We consider two main categories of baselines: (1) Imitation learning with suboptimal or action-free data (DP, RC-DP, DP+Repr, DP PT + FT, DP-VPT), and (2) Video planning (Uni Pi-OL, Uni Pi-CL). Diffusion Policy (DP) is a state-of-the-art imitation learning algorithm. Reward-Conditioned Diffusion Policy (RC-DP) utilizes suboptimal actions by conditioning the policy on a binary value indicating whether the action chunk comes from optimal demonstrations or not. This method is inspired by reward-conditioned approaches (Kumar et al., 2019b; Chen et al., 2021). Diffusion Policy with Representation Learning (DP+Repr) uses a VAE pretrained on demonstration, suboptimal, and action-free data as the observation encoder. This is representative of methods that leverage suboptimal data through representation learning. Diffusion Policy Pretrain + Finetune (DP PT + FT) pretrains on suboptimal trajectories and finetunes on demos. This is representative of methods that leverage suboptimal data through learning trajectory-level features. Diffusion Policy with Video Pre Training (DP-VPT) trains an inverse dynamics model to relabel actionfree data, inspired by VPT (Baker et al., 2022) and BCO (Torabi et al., 2018). Open-Loop Uni Pi (Uni Pi-OL) is based off of Uni Pi (Du et al., 2023a), a video planner for robot manipulation. Uni Pi-OL generates a single video trajectory, extracts actions, and executes the actions in an open-loop fashion. We use a goal-conditioned behavior cloning agent to reach generated subgoals (Wen et al., 2024). Closed-Loop Uni Pi (Uni Pi-CL) is a modification that allows Uni Pi to perform closed-loop replanning over image chunks. Like LDP, Uni Pi-CL generates dense plans instead of waypoints, though in image space. We learn an inverse dynamics model to extract actions. 5.2. Imitation Learning with Action-Free Data In Table 1, we examine how action-free data can be used to improve imitation learning policies. Imitation learning policies that model actions, such as DP, are unable to natively use action-free data, while planning-based approaches can benefit from this additional data. One approach is to relabel action-free data using an inverse dynamics model. We find this to be effective for most tasks, showing that the reannotated actions are useful for policy improvement. However, we see that LDP is better able to leverage action-free data by directly using it for the planner, rather than generating possibly inaccurate actions to subsequently learn from. Like LDP, video planning methods can directly use actionfree data for improving the planner, which for Uni Pi is the Latent Diffusion Planning for Imitation Learning Table 2. Leveraging Suboptimal Data. Latent Diffusion Planning can utilize suboptimal data for improved performance across a suite of tasks. Video planning methods like Uni Pi consistently struggle, whereas reward-conditioned or pretraining DP approaches are able to improve DP policy performance. LDP is best able to use suboptimal data, especially when combined with action-free data. Method Lift Can Square ALOHA Cube Average DP 0.60 0.00 0.63 0.01 0.48 0.00 0.32 0.00 0.51 RC-DP 0.40 0.04 0.73 0.03 0.66 0.02 0.60 0.04 0.60 DP+Repr 0.66 0.04 0.61 0.01 0.44 0.02 0.25 0.03 0.49 DP PT + FT 0.52 0.02 0.67 0.01 0.57 0.03 0.78 0.00 0.64 Uni Pi-OL 0.12 0.06 0.28 0.02 0.07 0.01 0.00 0.00 0.12 Uni Pi-CL 0.12 0.02 0.30 0.02 0.10 0.04 0.15 0.07 0.17 LDP 0.69 0.03 0.70 0.02 0.46 0.00 0.64 0.04 0.65 LDP + Subopt 0.84 0.06 0.68 0.02 0.55 0.03 0.71 0.03 0.70 LDP + Action-Free + Subopt 1.00 0.00 0.98 0.00 0.83 0.01 0.97 0.01 0.95 video generation model. We see limited improvement when using action-free data, and both DP and LDP still strongly outperform Uni Pi-CL and Uni Pi-OL. Uni Pi methods particularly struggle with more complex tasks like Square and ALOHA Cube, implying that video planning is still ineffective for tasks that require more precise manipulation skills. Qualitatively, for Uni Pi-OL, we find that while the goalconditioned agent is able to follow goals effectively, the policy still struggles with the difficult parts of the task, such as grasping the object. Forecasting goals does not provide the dense supervision for exactly how to grasp an object, and furthermore, Uni Pi-OL does not support replanning when a grasp is missed. Uni Pi-CL is able to address this by dense image forecasting, and consistently outperforms Uni Pi-OL. However, this closed-loop method is not only slow, but faces issues with video generation, such as regenerating static frames during parts of the task with less movement, leading the agent to be stuck in certain positions. This is especially noticeable for the ALOHA Cube task, where the agent is often stuck right before picking up the cube. Compared to Uni Pi-OL and Uni Pi-CL, LDP is able to circumvent many of these issues due to its latent planning and dense forecasting. 5.3. Imitation Learning with Suboptimal Data In Table 2, we present imitation learning results with suboptimal data. First, LDP outperforms DP, which can only utilize data with optimal actions. We notice, especially, that LDP with suboptimal data typically improves further upon LDP, showing the potential of leveraging diverse data sources outside of the demonstration dataset. Next, RC-DP, a conditional variant of DP that utilizes suboptimal data, outperforms DP. By learning from suboptimal data, RC-DP can learn priors of robot motions while distinguishing optimal action sequences. We hypothesize that for the Can, Square, and ALOHA Cube tasks, the primitive motions of reaching toward or grasping the object, which are partially covered by the suboptimal dataset, provides a useful visuomotor prior for the policy. ALOHA Cube sees significant improvement, possibly because the larger action space of bimanual control benefits from additional reward-labelled data. Next, we explore using suboptimal data for feature-learning. DP + Repr uses suboptimal data for pretraining a visionencoder, and DP PT + FT for pretraining the visuomotor policy. DP + Repr only improves policy performance for the Lift task, implying that end-to-end training of the vision encoder learns stronger features for complex manipulation tasks. DP PT + FT is particularly successful for tasks that require more precise manipulation, such as Square and ALOHA Cube, implying that learned prior motions is a useful policy initialization. Both the success of RC-DP and DP PT + FT suggest that leveraging the suboptimal trajectories is a useful way to improve imitation learning results. LDP, which leverages suboptimal data for the latent encoder and IDM, has higher overall performance than these methods, averaging across the suite of simulated tasks. Next, we compare against Uni Pi, which plans over image subgoals (OL) or image chunks (CL). Due to the low demonstration data regime, learning effective and accurate video policies is difficult, and LDP strongly outperforms Uni Pi OL and Uni Pi-CL. In addition, we notice that Uni Pi-CL outperforms Uni Pi-OL for all tasks, implying that dense forecasting, even within the image domain, is more effective than goal-conditioned methods. Finally, we find that LDP with action-free and suboptimal data leads to the strongest performance. We find a significant improvement in all of the simulated tasks, including compared to the variants of LDP that only use either actionfree or suboptimal data. This suggests that the recipe for combining these two data sources can lead to a stronger Latent Diffusion Planning for Imitation Learning Table 3. Real World Results For the Franka Lift task, we find that LDP consistently outperforms DP. Suboptimal and action-free data, that DP is unable to use, further improves performance. Method Success Rate DP 69.6 4 LDP 73.3 7 LDP + Subopt 74.8 4 LDP + Action-Free 79.3 3 planner and a more robust inverse dynamics model, leading to the best performing model for these tasks. 5.4. Imitation Learning in the Real World Real world data is more expensive and time-consuming to collect; hence, examining the effect of easier-to-collect suboptimal and action-free data provides insights for scalable learning. In Table 3, we provide results on a Franka Lift Cube task. In this task, we examine the performance of DP, which can only leverage action-labeled data, with our method, which can use suboptimal and action-free data. Figure 3. Visualizations of Generated Plans. LDP produces dense, closed-loop plans. Here, we visualize decoded latents selected from an LDP trajectory for the Lift, Can, Square, ALOHA Cube, and Franka Lift tasks. We find that LDP is able to consistently outperform DP, especially with the addition of action-free data. For real-world systems, this is promising, as collecting high-quality demos can be difficult and time intensive, whereas utilizing suboptimal trajectories, often a byproduct of evaluating policies, or collecting action-free trajectories more efficiently, such as in (Chi et al., 2024), can be a scalable direction. 5.5. Ablation: LDP Hierarchical In an attempt to understand the effect of LDP s dense forecasting, we compare LDP with a hierarchical version of LDP (LDP Hierarchical). In this implementation, LDP Hierarchical plans over subgoals 4 steps apart, and extracts 4 actions between pairs of forecasted latent states. Thus, LDP Hierarchical is closed-loop, yet the planner operates at a slightly abstracted level compared to non-hierarchical LDP. In Table 4, we find that LDP outperforms LDP Hierarchical across the 3 Robomimic tasks. This suggests that dense forecasting is an important part and contribution of LDP. This also reflects Uni Pi results in Table 1 and Table 2 that show closed-loop planning outperforms open-loop planning. Table 4. Hierarchical Ablation. LDP s dense forecasting outperforms a hierarchical variant of LDP. Method Lift Can Square LDP 0.69 0.03 0.70 0.02 0.46 0.00 LDP Hier. 0.53 0.03 0.62 0.04 0.31 0.01 LDP + Subopt 0.84 0.06 0.68 0.02 0.55 0.03 LDP Hier. + Subopt 0.65 0.03 0.60 0.02 0.43 0.05 6. Discussion We presented Latent Diffusion Planning, a simple planningbased method for imitation learning. We show that our design using powerful diffusion models for latent state forecasting enables competitive performance with state-of-theart imitation learning. We further show this latent state forecasting objective enables us to easily leverage heterogeneous data sources. In the low-demonstration data imitation regime, LDP outperforms prior imitation learning work that does not leverage such additional data as effectively. Limitations. One limitation of the current approach is that the latent space for planning is simply learned with a variational autoencoder and might not learn the most useful features for control. Future work will explore different representation learning objectives. Further, our method requires diffusing over states, which incurs additional computational overhead as compared to diffusing actions. However, we expect continued improvements in hardware and inference speed will mitigate this drawback. Finally, we did not explore applying recent improvements in diffusion models (Peebles & Xie, 2023; Lipman et al., 2022), which may be important in large data regimes. Future work. We have validated in simulation and real the hypothesis that latent state forecasting can leverage heterogeneous data sources. Future work can also evaluate whether this can be used to further improve more complex Latent Diffusion Planning for Imitation Learning real-world tasks. One direction is to use a diverse dataset of human collected data, such as with handheld data collection tools (Young et al., 2021). Another approach would be to use autonomously collected robotic data (Bousmalis* et al., 2023). As these alternative data sources are easier to collect than demonstrations, they represent a different scaling paradigm that can outperform pure behavior cloning approaches. By presenting a method that can leverage such data, we believe this work makes a step toward more performant and general robot policies. Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here. Acknowledgments We would like to thank members of ILIAD and IRIS at Stanford for fruitful discussions and feedback. We are grateful to Archit Sharma, Annie Chen, Austin Patel, Hengyuan Hu, Jensen Gao, Joey Hejna, Mallika Parulekar, Mengda Xu, Philippe Hansen-Estruch, Shuang Li, Suvir Mirchandani, and Zeyi Liu for helpful conversations. Ajay, A., Du, Y., Gupta, A., Tenenbaum, J., Jaakkola, T., and Agrawal, P. Is conditional generative modeling all you need for decision-making?, 2023. URL https: //arxiv.org/abs/2211.15657. Baker, B., Akkaya, I., Zhokov, P., Huizinga, J., Tang, J., Ecoffet, A., Houghton, B., Sampedro, R., and Clune, J. Video pretraining (vpt): Learning to act by watching unlabeled online videos. Advances in Neural Information Processing Systems, 35:24639 24654, 2022. Beliaev, M., Shih, A., Ermon, S., Sadigh, D., and Pedarsani, R. Imitation learning by estimating expertise of demonstrators, 2022. URL https://arxiv.org/abs/ 2202.01288. Black, K., Nakamoto, M., Atreya, P., Walke, H., Finn, C., Kumar, A., and Levine, S. Zero-shot robotic manipulation with pretrained image-editing diffusion models, 2023. URL https://arxiv.org/abs/2310.10639. Blattmann, A., Rombach, R., Ling, H., Dockhorn, T., Kim, S. W., Fidler, S., and Kreis, K. Align your latents: Highresolution video synthesis with latent diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 22563 22575, 2023. Bousmalis*, K., Vezzani*, G., Rao*, D., Devin*, C., Lee*, A. X., Bauza*, M., Davchev*, T., Zhou*, Y., Gupta*, A., Raju, A., Laurens, A., Fantacci, C., Dalibard, V., Zambelli, M., Martins, M., Pevceviciute, R., Blokzijl, M., Denil, M., Batchelor, N., Lampe, T., Parisotto, E., Zołna, K., Reed, S., Colmenarejo, S. G., Scholz, J., Abdolmaleki, A., Groth, O., Regli, J.-B., Sushkov, O., Roth orl, T., Chen, J. E., Aytar, Y., Barker, D., Ortiz, J., Riedmiller, M., Springenberg, J. T., Hadsell , R., Nori , F., and Heess, N. Robocat: A self-improving foundation agent for robotic manipulation. Ar Xiv, 2023. Bruce, J., Dennis, M. D., Edwards, A., Parker-Holder, J., Shi, Y., Hughes, E., Lai, M., Mavalankar, A., Steigerwald, R., Apps, C., et al. Genie: Generative interactive environments. In Forty-first International Conference on Machine Learning, 2024. Chen, L., Lu, K., Rajeswaran, A., Lee, K., Grover, A., Laskin, M., Abbeel, P., Srinivas, A., and Mordatch, I. Decision transformer: Reinforcement learning via sequence modeling, 2021. URL https://arxiv.org/abs/ 2106.01345. Chi, C., Feng, S., Du, Y., Xu, Z., Cousineau, E., Burchfiel, B., and Song, S. Diffusion policy: Visuomotor policy learning via action diffusion. In Proceedings of Robotics: Science and Systems (RSS), 2023. Chi, C., Xu, Z., Pan, C., Cousineau, E., Burchfiel, B., Feng, S., Tedrake, R., and Song, S. Universal manipulation interface: In-the-wild robot teaching without in-the-wild robots. In Proceedings of Robotics: Science and Systems (RSS), 2024. Cui, Z. J., Pan, H., Iyer, A., Haldar, S., and Pinto, L. Dynamo: In-domain dynamics pretraining for visuo-motor control. ar Xiv preprint ar Xiv:2409.12192, 2024. Darcet, T., Oquab, M., Mairal, J., and Bojanowski, P. Vision transformers need registers, 2023. Du, Y., Yang, M., Dai, B., Dai, H., Nachum, O., Tenenbaum, J. B., Schuurmans, D., and Abbeel, P. Learning universal policies via text-guided video generation, 2023a. URL https://arxiv.org/abs/2302.00111. Du, Y., Yang, M., Florence, P., Xia, F., Wahid, A., Ichter, B., Sermanet, P., Yu, T., Abbeel, P., Tenenbaum, J. B., Kaelbling, L., Zeng, A., and Tompson, J. Video language planning, 2023b. URL https://arxiv.org/abs/ 2310.10625. Edwards, A., Sahni, H., Schroecker, Y., and Isbell, C. Imitating latent policies from observation. In International conference on machine learning, pp. 1755 1763. PMLR, 2019. Latent Diffusion Planning for Imitation Learning Finn, C. and Levine, S. Deep visual foresight for planning robot motion. In 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 2786 2793. IEEE, 2017. Ha, D. and Schmidhuber, J. World models. ar Xiv preprint ar Xiv:1803.10122, 2018. Hafner, D., Lillicrap, T., Fischer, I., Villegas, R., Ha, D., Lee, H., and Davidson, J. Learning latent dynamics for planning from pixels, 2019. URL https://arxiv. org/abs/1811.04551. Hafner, D., Lillicrap, T., Ba, J., and Norouzi, M. Dream to control: Learning behaviors by latent imagination, 2020. URL https://arxiv.org/abs/1912.01603. Hansen, N., Su, H., and Wang, X. Td-mpc2: Scalable, robust world models for continuous control, 2024. URL https://arxiv.org/abs/2310.16828. Hansen-Estruch, P., Kostrikov, I., Janner, M., Kuba, J. G., and Levine, S. Idql: Implicit q-learning as an actorcritic method with diffusion policies, 2023. URL https: //arxiv.org/abs/2304.10573. Higgins, I., Matthey, L., Pal, A., Burgess, C., Glorot, X., Botvinick, M., Mohamed, S., and Lerchner, A. beta-VAE: Learning basic visual concepts with a constrained variational framework. In International Conference on Learning Representations, 2017. URL https: //openreview.net/forum?id=Sy2fz U9gl. Ho, J., Jain, A., and Abbeel, P. Denoising diffusion probabilistic models, 2020. URL https://arxiv.org/ abs/2006.11239. Hong, W., Ding, M., Zheng, W., Liu, X., and Tang, J. Cogvideo: Large-scale pretraining for text-to-video generation via transformers. ar Xiv preprint ar Xiv:2205.15868, 2022. Janner, M., Du, Y., Tenenbaum, J. B., and Levine, S. Planning with diffusion for flexible behavior synthesis, 2022. URL https://arxiv.org/abs/2205.09991. Khazatsky, A., Pertsch, K., Nair, S., Balakrishna, A., Dasari, S., Karamcheti, S., Nasiriany, S., Srirama, M. K., Chen, L. Y., Ellis, K., et al. Droid: A large-scale in-the-wild robot manipulation dataset. ar Xiv preprint ar Xiv:2403.12945, 2024. Kim, M., Pertsch, K., Karamcheti, S., Xiao, T., Balakrishna, A., Nair, S., Rafailov, R., Foster, E., Lam, G., Sanketi, P., Vuong, Q., Kollar, T., Burchfiel, B., Tedrake, R., Sadigh, D., Levine, S., Liang, P., and Finn, C. Openvla: An open-source vision-language-action model. ar Xiv preprint ar Xiv:2406.09246, 2024. Kingma, D. P. and Welling, M. Auto-encoding variational bayes, 2014. URL https://arxiv.org/ abs/1312.6114. Ko, P.-C., Mao, J., Du, Y., Sun, S.-H., and Tenenbaum, J. B. Learning to Act from Actionless Videos through Dense Correspondences. ar Xiv:2310.08576, 2023. Kostrikov, I., Nair, A., and Levine, S. Offline reinforcement learning with implicit q-learning, 2021. URL https: //arxiv.org/abs/2110.06169. Kumar, A., Peng, X. B., and Levine, S. Reward-conditioned policies. ar Xiv preprint ar Xiv:1912.13465, 2019a. Kumar, A., Peng, X. B., and Levine, S. Reward-conditioned policies, 2019b. URL https://arxiv.org/abs/ 1912.13465. Kumar, A., Zhou, A., Tucker, G., and Levine, S. Conservative q-learning for offline reinforcement learning, 2020. URL https://arxiv.org/abs/2006.04779. Lee, S., Wang, Y., Etukuru, H., Kim, H. J., Shafiullah, N. M. M., and Pinto, L. Behavior generation with latent actions, 2024. URL https://arxiv.org/abs/ 2403.03181. Levine, S., Kumar, A., Tucker, G., and Fu, J. Offline reinforcement learning: Tutorial, review, and perspectives on open problems. ar Xiv preprint ar Xiv:2005.01643, 2020. Lipman, Y., Chen, R. T., Ben-Hamu, H., Nickel, M., and Le, M. Flow matching for generative modeling. ar Xiv preprint ar Xiv:2210.02747, 2022. Mandlekar, A., Xu, D., Wong, J., Nasiriany, S., Wang, C., Kulkarni, R., Fei-Fei, L., Savarese, S., Zhu, Y., and Mart ın-Mart ın, R. What matters in learning from offline human demonstrations for robot manipulation. In Conference on Robot Learning (Co RL), 2021. Mayne, D. Q. and Michalska, H. Receding horizon control of nonlinear systems. In Proceedings of the 27th IEEE Conference on Decision and Control, pp. 464 465. IEEE, 1988. Octo Model Team, Ghosh, D., Walke, H., Pertsch, K., Black, K., Mees, O., Dasari, S., Hejna, J., Xu, C., Luo, J., Kreiman, T., Tan, Y., Chen, L. Y., Sanketi, P., Vuong, Q., Xiao, T., Sadigh, D., Finn, C., and Levine, S. Octo: An open-source generalist robot policy. In Proceedings of Robotics: Science and Systems, Delft, Netherlands, 2024. Open X-Embodiment Collaboration, O Neill, A., Rehman, A., Gupta, A., Maddukuri, A., Gupta, A., Padalkar, A., Lee, A., Pooley, A., Gupta, A., Mandlekar, A., Jain, A., Tung, A., Bewley, A., Herzog, A., Irpan, A., Khazatsky, Latent Diffusion Planning for Imitation Learning A., Rai, A., Gupta, A., Wang, A., Kolobov, A., Singh, A., Garg, A., Kembhavi, A., Xie, A., Brohan, A., Raffin, A., Sharma, A., Yavary, A., Jain, A., Balakrishna, A., Wahid, A., Burgess-Limerick, B., Kim, B., Sch olkopf, B., Wulfe, B., Ichter, B., Lu, C., Xu, C., Le, C., Finn, C., Wang, C., Xu, C., Chi, C., Huang, C., Chan, C., Agia, C., Pan, C., Fu, C., Devin, C., Xu, D., Morton, D., Driess, D., Chen, D., Pathak, D., Shah, D., B uchler, D., Jayaraman, D., Kalashnikov, D., Sadigh, D., Johns, E., Foster, E., Liu, F., Ceola, F., Xia, F., Zhao, F., Frujeri, F. V., Stulp, F., Zhou, G., Sukhatme, G. S., Salhotra, G., Yan, G., Feng, G., Schiavi, G., Berseth, G., Kahn, G., Yang, G., Wang, G., Su, H., Fang, H.-S., Shi, H., Bao, H., Amor, H. B., Christensen, H. I., Furuta, H., Bharadhwaj, H., Walke, H., Fang, H., Ha, H., Mordatch, I., Radosavovic, I., Leal, I., Liang, J., Abou-Chakra, J., Kim, J., Drake, J., Peters, J., Schneider, J., Hsu, J., Vakil, J., Bohg, J., Bingham, J., Wu, J., Gao, J., Hu, J., Wu, J., Wu, J., Sun, J., Luo, J., Gu, J., Tan, J., Oh, J., Wu, J., Lu, J., Yang, J., Malik, J., Silv erio, J., Hejna, J., Booher, J., Tompson, J., Yang, J., Salvador, J., Lim, J. J., Han, J., Wang, K., Rao, K., Pertsch, K., Hausman, K., Go, K., Gopalakrishnan, K., Goldberg, K., Byrne, K., Oslund, K., Kawaharazuka, K., Black, K., Lin, K., Zhang, K., Ehsani, K., Lekkala, K., Ellis, K., Rana, K., Srinivasan, K., Fang, K., Singh, K. P., Zeng, K.-H., Hatch, K., Hsu, K., Itti, L., Chen, L. Y., Pinto, L., Fei-Fei, L., Tan, L., Fan, L. J., Ott, L., Lee, L., Weihs, L., Chen, M., Lepert, M., Memmel, M., Tomizuka, M., Itkina, M., Castro, M. G., Spero, M., Du, M., Ahn, M., Yip, M. C., Zhang, M., Ding, M., Heo, M., Srirama, M. K., Sharma, M., Kim, M. J., Kanazawa, N., Hansen, N., Heess, N., Joshi, N. J., Suenderhauf, N., Liu, N., Palo, N. D., Shafiullah, N. M. M., Mees, O., Kroemer, O., Bastani, O., Sanketi, P. R., Miller, P. T., Yin, P., Wohlhart, P., Xu, P., Fagan, P. D., Mitrano, P., Sermanet, P., Abbeel, P., Sundaresan, P., Chen, Q., Vuong, Q., Rafailov, R., Tian, R., Doshi, R., Mart in-Mart in, R., Baijal, R., Scalise, R., Hendrix, R., Lin, R., Qian, R., Zhang, R., Mendonca, R., Shah, R., Hoque, R., Julian, R., Bustamante, S., Kirmani, S., Levine, S., Lin, S., Moore, S., Bahl, S., Dass, S., Sonawani, S., Tulsiani, S., Song, S., Xu, S., Haldar, S., Karamcheti, S., Adebola, S., Guist, S., Nasiriany, S., Schaal, S., Welker, S., Tian, S., Ramamoorthy, S., Dasari, S., Belkhale, S., Park, S., Nair, S., Mirchandani, S., Osa, T., Gupta, T., Harada, T., Matsushima, T., Xiao, T., Kollar, T., Yu, T., Ding, T., Davchev, T., Zhao, T. Z., Armstrong, T., Darrell, T., Chung, T., Jain, V., Kumar, V., Vanhoucke, V., Zhan, W., Zhou, W., Burgard, W., Chen, X., Chen, X., Wang, X., Zhu, X., Geng, X., Liu, X., Liangwei, X., Li, X., Pang, Y., Lu, Y., Ma, Y. J., Kim, Y., Chebotar, Y., Zhou, Y., Zhu, Y., Wu, Y., Xu, Y., Wang, Y., Bisk, Y., Dou, Y., Cho, Y., Lee, Y., Cui, Y., Cao, Y., Wu, Y.-H., Tang, Y., Zhu, Y., Zhang, Y., Jiang, Y., Li, Y., Li, Y., Iwasawa, Y., Matsuo, Y., Ma, Z., Xu, Z., Cui, Z. J., Zhang, Z., Fu, Z., and Lin, Z. Open XEmbodiment: Robotic learning datasets and RT-X models. https://arxiv.org/abs/2310.08864, 2023. Oquab, M., Darcet, T., Moutakanni, T., Vo, H., Szafraniec, M., Khalidov, V., Fernandez, P., Haziza, D., Massa, F., El Nouby, A., Howes, R., Huang, P.-Y., Xu, H., Sharma, V., Li, S.-W., Galuba, W., Rabbat, M., Assran, M., Ballas, N., Synnaeve, G., Misra, I., Jegou, H., Mairal, J., Labatut, P., Joulin, A., and Bojanowski, P. Dinov2: Learning robust visual features without supervision, 2023. Pearce, T., Rashid, T., Kanervisto, A., Bignell, D., Sun, M., Georgescu, R., Macua, S. V., Tan, S. Z., Momennejad, I., Hofmann, K., and Devlin, S. Imitating human behaviour with diffusion models, 2023. URL https://arxiv. org/abs/2301.10677. Peebles, W. and Xie, S. Scalable diffusion models with transformers, 2023. URL https://arxiv.org/abs/ 2212.09748. Radosavovic, I., Shi, B., Fu, L., Goldberg, K., Darrell, T., and Malik, J. Robot learning with sensorimotor pretraining. In Conference on Robot Learning, pp. 683 693. PMLR, 2023. Rezende, D. J., Mohamed, S., and Wierstra, D. Stochastic backpropagation and approximate inference in deep generative models. In International conference on machine learning, pp. 1278 1286. PMLR, 2014. Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. High-resolution image synthesis with latent diffusion models, 2022. URL https://arxiv.org/ abs/2112.10752. Schmeckpeper, K., Xie, A., Rybkin, O., Tian, S., Daniilidis, K., Levine, S., and Finn, C. Learning predictive models from observation and interaction. In European Conference on Computer Vision, pp. 708 725. Springer, 2020. Schmidhuber, J. Reinforcement learning upside down: Don t predict rewards just map them to actions. ar Xiv preprint ar Xiv:1912.02875, 2019. Shafiullah, N. M. M., Cui, Z. J., Altanzaya, A., and Pinto, L. Behavior transformers: Cloning k modes with one stone, 2022. URL https://arxiv.org/abs/ 2206.11251. Sohl-Dickstein, J., Weiss, E. A., Maheswaranathan, N., and Ganguli, S. Deep unsupervised learning using nonequilibrium thermodynamics, 2015. URL https: //arxiv.org/abs/1503.03585. Latent Diffusion Planning for Imitation Learning Song, Y., Sohl-Dickstein, J., Kingma, D. P., Kumar, A., Ermon, S., and Poole, B. Score-based generative modeling through stochastic differential equations. ar Xiv preprint ar Xiv:2011.13456, 2020. Torabi, F., Warnell, G., and Stone, P. Behavioral cloning from observation. ar Xiv preprint ar Xiv:1805.01954, 2018. Watter, M., Springenberg, J., Boedecker, J., and Riedmiller, M. Embed to control: A locally linear latent dynamics model for control from raw images. Advances in neural information processing systems, 28, 2015. Wen, C., Lin, X., So, J., Chen, K., Dou, Q., Gao, Y., and Abbeel, P. Any-point trajectory modeling for policy learning, 2024. URL https://arxiv.org/abs/2401. 00025. Wu, P., Escontrela, A., Hafner, D., Abbeel, P., and Goldberg, K. Daydreamer: World models for physical robot learning. In Conference on robot learning, pp. 2226 2240. PMLR, 2023a. Wu, P., Majumdar, A., Stone, K., Lin, Y., Mordatch, I., Abbeel, P., and Rajeswaran, A. Masked trajectory models for prediction, representation, and control. In International Conference on Machine Learning, pp. 37607 37623. PMLR, 2023b. Yan, W., Zhang, Y., Abbeel, P., and Srinivas, A. Videogpt: Video generation using vq-vae and transformers. ar Xiv preprint ar Xiv:2104.10157, 2021. Yang, M., Du, Y., Ghasemipour, K., Tompson, J., Schuurmans, D., and Abbeel, P. Learning interactive real-world simulators. ar Xiv preprint ar Xiv:2310.06114, 2023. Yang, S., Walker, J., Parker-Holder, J., Du, Y., Bruce, J., Barreto, A., Abbeel, P., and Schuurmans, D. Video as the new language for real-world decision making, 2024. URL https://arxiv.org/abs/2402.17139. Young, S., Gandhi, D., Tulsiani, S., Gupta, A., Abbeel, P., and Pinto, L. Visual imitation made easy. In Conference on Robot Learning, pp. 1992 2005. PMLR, 2021. Yu, T., Kumar, A., Chebotar, Y., Hausman, K., Finn, C., and Levine, S. How to leverage unlabeled data in offline reinforcement learning. In Chaudhuri, K., Jegelka, S., Song, L., Szepesvari, C., Niu, G., and Sabato, S. (eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp. 25611 25635. PMLR, 17 23 Jul 2022. URL https://proceedings.mlr.press/ v162/yu22c.html. Zhang, S., Cao, Z., Sadigh, D., and Sui, Y. Confidenceaware imitation learning from demonstrations with varying optimality, 2022. URL https://arxiv.org/ abs/2110.14754. Zhao, T. Z., Kumar, V., Levine, S., and Finn, C. Learning fine-grained bimanual manipulation with low-cost hardware, 2023. URL https://arxiv.org/abs/ 2304.13705. Zhao, T. Z., Tompson, J., Driess, D., Florence, P., Ghasemipour, S. K. S., Finn, C., and Wahid, A. ALOHA unleashed: A simple recipe for robot dexterity. In 8th Annual Conference on Robot Learning, 2024. URL https: //openreview.net/forum?id=gvd XE7ik HI. Zhou, S., Du, Y., Chen, J., Li, Y., Yeung, D.-Y., and Gan, C. Robodreamer: Learning compositional world models for robot imagination, 2024. URL https://arxiv. org/abs/2404.12377. Latent Diffusion Planning for Imitation Learning A. Appendix A.1. Additional Ablations A.1.1. LDP WITH PRETRAINED EMBEDDINGS We evaluate the effectiveness of planning and extracting actions from pretrained embeddings with a DINOv2 ablation. In this approach, we replace our VAE embeddings with DINOv2 embeddings, with no other changes to the LDP architecture. The DINOv2 latent embedding is 384-dimensional. We found directly planning over the DINOv2 embeddings does not lead to good learned behaviors (0% success), which we hypothesize is due to the challenges of planning over a large latent space. Thus, as an alternative, we fix a random projection matrix, reducing the 384 dim feature space to 16 dim, matching LDP. We choose this, because this is a straightforward way of using frozen embeddings. It may be possible to plan over the large embeddings space with a much larger and complex model, or it may be possible to learn alternative ways to project DINOv2 embeddings to a lower-dimensional space, but this may lead to fundamental changes in our method, so we leave more complicated approaches to use pretrained embeddings for future work. We include results on Robomimic tasks in Table 5. Table 5. DINOv2 Ablation. Planning over frozen embeddings (DINOv2), even when projected to a lower dimensional space, has significantly lower success rate than planning over VAE embeddings. This suggests that the learned latent space is crucial to LDP s performance, and that a compact, well-structured latent space leads to optimal performance. Method Latent Dim Lift Can Square LDP w/ DINOv2 (frozen) 384 0.00 0.00 0.00 0.00 0.00 0.00 LDP w/ DINOv2 (frozen, randomly projected) 16 0.44 0.24 0.03 0.01 0.01 0.01 LDP + Action-Free + Subopt 16 1.00 0.00 0.98 0.00 0.83 0.01 A.1.2. UNIPI FROM PRETRAINED MODEL In our experimental results, we train our Uni Pi model from scratch on demonstration videos. To provide a comparison to training Uni Pi from pre-trained weights, we use the AVDC-THOR pretrained checkpoint (Ko et al., 2023). We finetune for an additional 50k steps at learning rate 1e-4. In Table 6 and Table 7, we find that the pretrained Uni Pi model does not substantially improve performance, likely because the pretrained model quality is limited. For Uni Pi-OL, there appears to be slightly improvement when pretraining from scratch, possibly because without closed-loop planning, the pretrained initialization may be helpful. Table 6. Uni Pi from Pretrained Checkpoint Method Lift Can Square Uni Pi-OL (from scratch) 0.09 0.03 0.27 0.01 0.07 0.01 Uni Pi-OL (from pretrain) 0.13 0.05 0.31 0.01 0.07 0.01 Uni Pi-CL (from scratch) 0.12 0.02 0.30 0.02 0.10 0.04 Uni Pi-CL (from pretrain) 0.13 0.05 0.23 0.01 0.08 0.02 Table 7. Uni Pi + Action-Free from Pretrained Checkpoint Method Lift Can Square Uni Pi-OL (from scratch) 0.11 0.03 0.25 0.03 0.05 0.03 Uni Pi-OL (from pretrain) 0.10 0.08 0.26 0.00 0.08 0.02 Uni Pi-CL (from scratch) 0.14 0.02 0.32 0.04 0.09 0.01 Uni Pi-CL (from pretrain) 0.11 0.03 0.25 0.01 0.10 0.00 A.2. Implementation Details Diffusion Policy We use a Jax reimplementation of the convolutional Diffusion Policy, which we verify can reproduce reported Robomimic benchmark results. Latent Diffusion Planning for Imitation Learning Uni Pi We use the open-source implementation of Uni Pi (Ko et al., 2023). For Uni Pi-OL and Uni Pi-CL, we predict 7 future frames. During training time, for Uni Pi-OL, the 7 future frames are evenly sampled from a training demonstration. For Uni Pi-CL, the 7 future frames are the next consecutive frames. The goal-conditioned behavior cloning agent is implemented as a goal-conditioned Diffusion Policy (Chi et al., 2023) agent, and it is trained on chunks of 16. The inverse dynamics model is based off of IDQL (Hansen-Estruch et al., 2023), and shares the same architecture as the IDM used in LDP. We train the video prediction models for 200k gradient steps with batch size 16. LDP The LDP VAE is adapted from Diffusion Transformer (Peebles & Xie, 2023). The planner is based directly off of the convolutional U-Net from Diffusion Policy (Chi et al., 2023), with modifications to plan across latent embeddings instead of action chunks. The IDM is based off of IDQL (Hansen-Estruch et al., 2023). Table 8. Diffusion Policy Architecture Hyperparameters Uni Pi-OL GCBC DP and LDP LDP - ALOHA Cube down dims [256, 512, 1024] [256, 512, 1024] [512, 1024, 2048] n diffusion steps 100 100 100 batch size 256 256 256 lr 1e-4 1e-4 1e-4 n grad steps 500k 500k 500k Table 9. IDM Architecture Hyperparameters Uni Pi-CL IDM LDP IDM n blocks 3 3 (Lift, Square, ALOHA Cube); 5 (Can) n diffusion steps 100 100 batch size 256 256 lr 1e-4 1e-4 n grad steps 500k 500k Table 10. VAE Architecture Hyperparameters VAE block out channels [128, 256, 256, 256, 256, 256] down block types [Down Encoder Block2D] x6 up block types [Up Decoder Block2D] x6 latent channels 4 Latent Dim (2, 2, 4) Lift KL Beta 1e-5 Can KL Beta 1e-6 Square KL Beta 1e-6 ALOHA Cube KL Beta 1e-7 n grad steps 300k A.3. Simulation Experiments A.3.1. ENVIRONMENT For all environments, we use an observation horizon of 1. For Robomimic tasks, we use the third-person view image. For ALOHA Cube and Franka Lift, we use the wrist camera. All images are 64x64. Latent Diffusion Planning for Imitation Learning A.3.2. DATASET SIZES We detail the design decisions for the number of expert, action-free, and suboptimal trajectories. For simulated robotics tasks, we kept consistent the action-free and suboptimal trajectories: 100 action-free trajectories for robomimic tasks and 500 suboptimal trajectories for all tasks. Number of Demonstrations Robomimic includes 200 proficient demonstrations per task, and ALOHA includes 50 demonstrations per task. To showcase the effectiveness of our method, we always take half the number of demonstrations 100 for Robomimic tasks, and 25 for ALOHA tasks. This is a consistent protocol we use to select the number of demonstrations. However, Lift is a particularly simple task, so in order to not saturate our baseline methods with excessive numbers of demos, we reduce the number of demonstrations to 3. Without this adjustment, Lift results would not be useful to discern the effect of different types of data, and furthermore, Lift results would all be saturated. Number of Action-Free Trajectories We chose to use the remaining half of the demonstrations as action-free trajectories. Hence, ALOHA uses 25 action-free trajectories, and Robomimic uses 100 action-free trajectories. Note again that Lift only used 3 expert demonstrations, but in order to keep our numbers as consistent as possible and avoid further confusion, we opted to use 100 action-free trajectories. A.3.3. METHOD DETAILS LDP VAE We train with additional action-free and suboptimal data. This VAE is shared across a single task for all variants of LDP, and including for DP + Repr. LDP + Subopt We train the inverse dynamics model with batches of 50% optimal, 50% suboptimal data. LDP + Action-Free We train only the planner on additional action-free demos. The IDM is trained only on expert demonstrations. LDP + Action-Free + Subopt We train only the planner on additional action-free demos. We train the inverse dynamics model with batches of 50% optimal, 50% suboptimal data. LDP Hierarchical The IDM is parameterized as a Conditional U-Net, like the planner, but the IDM is smaller and has down dims=[256, 512]. LDP DINOv2 In this variant, we plan over pretrained DINOv2 (Oquab et al., 2023; Darcet et al., 2023) embeddings instead of VAE embeddings. We use the Vi T-S variant and pad the 64x64 image to be 70x70, which matches the 14x14 patch size. The batch size is reduced to 128, but the remaining hyperparameters are identical to LDP. We further project the 384-dimensional embedding to 16-dimensional via a random matrix, in order for easier planning. Results are presented in Appendix A.1.1 DP + Repr We use the same VAE that LDP methods used. DP PT + FT We pretrain for 300k steps and finetune with a lower learning rate (1e-5) for 200k steps. DP-VPT We use an inverse dynamics model trained for 500k steps on curated data, as used in LDP, to relabel actions for the action-free data. Uni Pi We train two goal-conditioned agents and two inverse dynamics models and report the average success. For Uni Pi-OL evaluations, we predetermine the number of steps for the GCBC to reach each image subgoal based on the demonstration lengths. For Lift, evaluation episode lengths are 60 steps; Can is 140 steps; and Square is 160 steps. This maximum horizon is also enforced for Uni Pi-CL evaluations, for consistency.