# cubic_lstms_for_video_prediction__ace40a0b.pdf The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19) Cubic LSTMs for Video Prediction Hehe Fan, Linchao Zhu, Yi Yang Centre for Artificial Intelligence University of Technology Sydney, Australia {hehe.fan,linchao.zhu}@student.uts.edu.au, yi.yang@uts.edu.au Predicting future frames in videos has become a promising direction of research for both computer vision and robot learning communities. The core of this problem involves moving object capture and future motion prediction. While object capture specifies which objects are moving in videos, motion prediction describes their future dynamics. Motivated by this analysis, we propose a Cubic Long Short-Term Memory (Cubic LSTM) unit for video prediction. Cubic LSTM consists of three branches, i.e., a spatial branch for capturing moving objects, a temporal branch for processing motions, and an output branch for combining the first two branches to generate predicted frames. Stacking multiple Cubic LSTM units along the spatial branch and output branch, and then evolving along the temporal branch can form a cubic recurrent neural network (Cubic RNN). Experiment shows that Cubic RNN produces more accurate video predictions than prior methods on both synthetic and real-world datasets. Introduction Videos contain a large amount of visual information in scenes as well as profound dynamic changes in motions. Learning video representations, imagining motions and understanding scenes are fundamental missions in computer vision. It is indisputable that a model, which is able to predict future frames after watching several context frames, has the ability to achieve these missions (Srivastava, Mansimov, and Salakhutdinov 2015). Compared with video recognition (Zhu, Xu, and Yang 2017; Fan et al. 2017; 2018), video prediction has an innate advantage that it usually does not require external supervised information. Videos also provide a window for robots to understand the physical world. Predicting what will happen in future can help robots to plan their actions and make decisions. For example, action-conditional video prediction (Oh et al. 2015; Finn, Goodfellow, and Levine 2016; Ebert et al. 2017; Babaeizadeh et al. 2018) provides a physical understanding of the object in terms of the factors (e.g., forces) acting upon it and the long term effect (e.g., motions) of those factors. As video is a kind of spatio-temporal sequences, recurrent neural networks (RNNs), especially Long Short-Term Memory (LSTM) (Hochreiter and Schmidhuber 1997) and Gated Copyright c 2019, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Recurrent Unit (GRU) (Cho et al. 2014), have been widely applied to video prediction. One of the earliest RNN models for video prediction (Ranzato et al. 2014) directly borrows a structure from the language modeling literature (Bengio et al. 2003). It quantizes the space of frame patches as visual words and therefore can be seen as patch-level prediction. Later work (Srivastava, Mansimov, and Salakhutdinov 2015) builds encoder-decoder predictors using the fully connected LSTM (FC-LSTM). In addition to patchlevel prediction, it also learns to predict future frames at the feature level. Since the traditional LSTM (or GRU) units learn from one-dimensional vectors where the feature representation is highly compact and the spatial information is lost, both of these methods attempt to avoid directly predicting future video frames at the image level. To predict spatio-temporal sequences, convolutional LSTM (Conv LSTM) (Shi et al. 2015) modifies FC-LSTM by taking three-dimensional tensors as the input and replacing fully connected layers by convolutional operations. Conv LSTM has become an important component in several video prediction works (Finn, Goodfellow, and Levine 2016; Wang et al. 2017; Babaeizadeh et al. 2018; Lotter, Kreiman, and Cox 2017). Like most traditional LSTM units, FC-LSTM is designed to learn only one type of information, i.e., the dependency of sequences. Directly adapting FC-LSTM makes it difficult for Conv LSTM to simultaneously exploit the temporal information and the spatial information in videos. The convolutional operation in Conv LSTM has to process motions on one hand and capture moving objects on the other hand. Similarly, the state of Conv LSTM must be able to carry the motion information and object information at the same time. It could be insufficient to use only one convolution and one state for spatio-temporal prediction. In this paper, we propose a new unit for video prediction, i.e., the Cubic Long Short-Term Memory (Cubic LSTM) unit. This unit is equipped with two states, a temporal state and a spatial state, which are respectively generated by two independent convolutions. The motivation is that different kinds of information should be processed and carried by different operations and states. Cubic LSTM consists of three branches, which are built along the three axes in the Cartesian coordinate system. The temporal branch flows along the x-axis (temporal axis), on which the convolution aims to obtain and process motions. The temporal state is generated by this branch, which contains the motion information. The spatial branch flows along the z-axis (spatial axis), on which the convolution is responsible for capturing and analyzing moving objects. The spatial state is generated by this branch, which carries the spatial layout information about moving objects. The output branch generates intermediate or final prediction frames along the y-axis (output axis), according to the predicted motions provided by the temporal branch and the moving object information provided by the spatial branch. Stacking multiple Cubic LSTM units along the spatial branch and output branch can form a two-dimensional network. This two-dimensional network can further construct a three-dimensional network by evolving along the temporal axis. We refer to this three-dimensional network as the cubic recurrent neural network (Cubic RNN). Experiment shows that Cubic RNN produces highly accurate video predictions on the Moving-MNIST dataset (Srivastava, Mansimov, and Salakhutdinov 2015), Robotic Pushing dataset (Finn, Goodfellow, and Levine 2016) and KTH Action dataset (Sch uldt, Laptev, and Caputo 2004). Related work Video Prediction A number of prior works have addressed video prediction with different settings. They can essentially be classified as follows. Generation vs. Transformation. As mentioned above, video prediction can be classified as patch level (Ranzato et al. 2014), feature level (Ranzato et al. 2014; Srivastava, Mansimov, and Salakhutdinov 2015; Vondrick, Pirsiavash, and Torralba 2016a) and image level. For imagelevel prediction, the generation group of methods generates each pixel in frames (Shi et al. 2015; Reed et al. 2017). Some of these methods have difficulty in handling realworld video prediction due to the curse of dimensionality. The second group first predicts a transformation and then applies the transformation to the previous frame to generate a new frame (Jia et al. 2016; Jaderberg et al. 2015; Finn, Goodfellow, and Levine 2016; Vondrick and Torralba 2017). These types of methods can reduce the difficulty of predicting real-world future frames. In addition, some methods (Villegas et al. 2017; Denton and Birodkar 2017; Tulyakov et al. 2018) first decompose a video into a stationary content part and a temporally varying motion component by multiple loss functions, and then combine the predicted motion and stationary content to construct future frames. Short-term prediction vs. Long-term prediction. Video prediction can be classified as short-term (< 10 frames) prediction (Xue et al. 2016; Kalchbrenner et al. 2017) and long-term ( 10 frames) prediction (Oh et al. 2015; Shi et al. 2015; Vondrick, Pirsiavash, and Torralba 2016b; Finn, Goodfellow, and Levine 2016; Wang et al. 2017) according to the number of predicted frames. Most methods can usually undertake long-term predictions on virtual-word datasets (e.g., game-play video datasets (Oh et al. 2015)) or synthetic datasets (e.g., the Moving-MNIST dataset (Srivastava, Mansimov, and Salakhutdinov 2015)). Since realworld sequences are less deterministic, some of these methods are limited to making long-term predictions. For example, dynamic filter network (Jia et al. 2016) is able to predict ten frames on the Moving-MNIST dataset but three frames on the highway driving dataset. A special case of short-term prediction is the so-called next-frame prediction (Xue et al. 2016; Lotter, Kreiman, and Cox 2017), which only predicts one frame in the future. Single dependency vs. Multiple dependencies. Most methods need to observe multiple context frames before making video predictions. Other methods, by contrast, aim to predict the future based only on the understanding of a single image (Mottaghi et al. 2016; Mathieu, Couprie, and Le Cun 2016; Walker et al. 2016; Xue et al. 2016; Zhou and Berg 2016; Xiong et al. 2018). Unconditional prediction vs. Conditional prediction. The computer vision community usually focuses on unconditional prediction, in which the future only depends on the video itself. The action-conditional prediction has been widely explored in the robotic learning community, e.g., video game videos (Oh et al. 2015; Chiappa et al. 2017) and robotic manipulations (Finn, Goodfellow, and Levine 2016; Kalchbrenner et al. 2017; Ebert et al. 2017; Babaeizadeh et al. 2018; Reed et al. 2017). Deterministic model vs. Probabilistic model. One common assumption for video prediction is that the future is deterministic. Most video prediction methods therefore belong to the deterministic model. However, the real-world can be full of stochastic dynamics. The probabilistic model predicts multiple possible frames at one time. (Babaeizadeh et al. 2018; Fragkiadaki et al. 2017). The probabilistic method is also adopted in single dependency prediction because the precise motion corresponding to a single image is often stochastic and ambiguous (Xue et al. 2016). Long Short-Term Memory (LSTM) LSTM-based methods have been widely used in video prediction (Finn, Goodfellow, and Levine 2016; Wang et al. 2017; Ebert et al. 2017; Babaeizadeh et al. 2018; Lotter, Kreiman, and Cox 2017). The proposed Cubic LSTM can therefore be considered as a fundamental module for video prediction, which can be applied to many frameworks. Among multidimensional LSTMs, our Cubic LSTM is similar to Grid LSTM (Kalchbrenner, Danihelka, and Graves 2016). The difference is that Grid LSTM is built by full connections. Furthermore, all dimensions in Grid LSTM are equal. However, for Cubic LSTM, the spatial branch applies 5 5 convolutions while the temporal branch applies 1 1 convolutions. Our Cubic LSTM is also similar to Pyramid LSTM (Stollenga et al. 2015), which consists of six LSTMs along with three axises, to capture the biomedical volumetric image. The information flows dependently in the six LSTMs and the output of Pyramid LSTM is simply to add the hidden states of the six LSTMs. However, Cubic LSTM has three branches and information flows across the these branches. In this section, we first review Fully-Connected Long Short Term Memory (FC-LSTM) (Hochreiter and Schmidhuber 1997) and Convolutional LSTM (Conv LSTM) (Shi et al. 2015), and then describe the proposed Cubic LSTM (Cubic LSTM) unit in detail. LSTM is a special recurrent neural network (RNN) unit for modeling long-term dependencies. The key to LSTM is the cell state Ct which acts as an accumulator of the sequence or the temporal information. The information from every new input Xt will be integrated to Ct if the input it is activated. At the same time, the past cell state Ct 1 may be forgotten if the forget gate ft turns on. Whether Ct will be propagated to the hidden state Ht is controlled by the output gate ot. Usually, the cell state Ct and the hidden state Ht are jointly referred to as the internal LSTM state, denoted as (Ct, Ht). The updates of fully-connected LSTM (FC-LSTM) for the t-th time step can be formulated as follows: it = σ(Wi [Xt, Ht 1] + bi), ft = σ(Wf [Xt, Ht 1] + bf), ot = σ(Wo [Xt, Ht 1] + bo), ct = tanh(Wc [Xt, Ht 1] + bc), Ct = ft Ct + it ct, Ht = ot tanh(Ct), where , and [ ] denote the matrix product, the elementwise product and the concatenation operation, respectively. To generate it, ft, ot and ct in the FC-LSTM, a fullyconnected layer is first applied on the concatenation of the input Xt and the last hidden state Ht 1 with the form W [Xt, Ht 1] + b, where W = [Wi, Wf, Wo, Wc] and b = [bi, bf, bo, bc]. The intermediate result is then split into four parts and passed to the activation functions, i.e., σ and tanh. Lastly, the new state (Ct, Ht) is produced according to the gates and ct by element-wise product. In summary, FC-LSTM takes the current input Xt as input, the previous state (Ct 1, Ht 1) and generates the new state (Ct, Ht). It is parametrized by W, b. For simplicity, we reformulate FCLSTM as follow: FC LSTM : (Ct, Ht) = LSTM(Xt, (Ct 1, Ht 1); W, b; ). (2) The input Xt and the state (Ct, Ht) in FC-LSTM are all one-dimensional vectors, which cannot directly encode spatial information. Although we can use one-dimensional vectors to represent images by flattening the two-dimensional data (greyscale images) or the three-dimensional data (color images), such operation loses spatial correlations. To exploit spatial correlations for video prediction, Conv LSTM takes three-dimensional tensors as input and replaces the fully-connected layer (matrix product) in FC-LSTM with the convolutional layer. The updates for Conv LSTM can be written as follow: Conv LSTM : (Ct, Ht) = LSTM(Xt, (Ct 1, Ht 1); W, b; ), (3) where denotes the convolution operator and Xt, (Ct 1, Ht 1), (Ct, Ht) are all three-dimensional tensors with shape (height, width, channel). As FC-LSTMs are designed to learn only one type of information, directly adapting FC-LSTM makes it difficult for Conv LSTM to simultaneously process the temporal information and the spatial information in videos. To predict future spatio-temporal sequences, as previously noted, the convolution in Conv LSTM has to catch motions on one hand and capture moving objects on the other hand. Similarly, the state of Conv LSTM must be capable of storing motion information and visual information at the same time. Cubic LSTM To reduce the burden of prediction, the proposed Cubic LSTM unit processes the temporal information and the spatial information separately. Specifically, Cubic LSTM consists of three branches: a temporal branch, a spatial branch and a output branch. The temporal branch aim to obtain and process motions. The spatial branch is responsible for capturing and analyzing objects. The output branch generates predictions according to the predicted motion information and the moving object information. As shown in Figure 1(a), the unit is built along the three axes in a space Cartesian coordinate system. Along the x-axis (temporal axis), the convolution operation obtains the current motion information according to the input Xt,l, the previously motion information Ht 1,l and the previously object information H t,l 1. The current motion information is then used to update the previous temporal cell state Ct 1,l and produce the new motion information Ht,l. Along the z-axis (spatial axis), the convolution captures the current spatial layout of objects. This information is then used to rectify the previous spatial cell state C t,l 1 and generate the new object visual information H t,l. Along the y-axis (output axis), the output branch combines the current motion information and the current object information to generate an intermediate prediction for the input of the next Cubic LSTM unit or construct the final prediction frame. The topological diagram of Conv LSTM is shown in Figure 1(b). The updates of Cubic LSTM are formulated as Eq. (4), where l, and denote the spatial network layer, the spatial branch and the output branch respectively. Essentially, in a single Cubic LSTM, the temporal branch and the spatial branch are symmetrical. However, in experiment, we found that adopting 1 1 convolution for temporal branch achieves higher accuracy than adopting 5 5 convolution. This may indicate that motion should focus on temporal neighbors while object should focus on spatial neighbors. Their functions also depend on their positions and connections in RNNs. To deal with a sequence, the same unit Ct+1,2 Ht+1,2 Xt,l Ht,l Ht,l convolution o c f i o c Ht+1,l Ct+1,l Yt,l Ht,l+1 convolution convolution σ tanh tanh σ σ σ σ σ o c i f c o f i o c o c f i output (y-) axis temporal (x-) axis spatial (z-) axis Ct,l Ct+1,l Ct,l+1 Ht,l+1 Yt,l Figure 1: (a) 3D structure of the Cubic LSTM unit. (b) Topological diagram of the Cubic LSTM unit. (c) Two-spatial-layer RNN composed of Cubic LSTM units. The unit consists of three branches, a spatial (z-) branch for extracting and recognizing moving objects, a temporal (y-) branch for capturing and predicting motions, and an output (x-) branch for combining the first two branches to generate the predicted frames. temporal branch : (Ct,l, Ht,l) = LSTM(Xt,l, H t,l 1, (Ct 1,l, Ht 1,l); W, b; ), spatial branch : (C t,l, H t,l) = LSTM(Xt,l, Ht 1,l, (C t,l 1, H t,l 1); W , b ; ), output branch : Yt,l = W [Ht,l, H t,l] + b . is repeated along the temporal axis. Therefore, the parameters are shared along the temporal dimension. For the spatial axis, we stack multiple different units to form a multi-layer structure to better exploit spatial correlations and capture objects. The parameters along the spatial axis are different. At the end of spatial direction, rather than being discarded, the spatial state is used to initialize the starting spatial state at the next time step. Formally, the spatial states between two time steps are defined as follows: C t,1 = C t 1,L, H t,1 = H t 1,L, (5) where L > 1 is the number of layers along the spatial axis. We demonstrate a 2-spatial-layer RNN in Figure 1(c), which is the smallest network formed by Cubic LSTMs. Cubic RNN In this section, we introduce a new RNN architecture for video prediction, the cubic RNN (Cubic RNN). Cubic RNN is created by first stacking multiple Cubic LSTM units along the spatial axis and along the output axis, which forms a twodimensional network, and then evolving along the temporal branch, which forms a three-dimensional structure. In contrast to traditional RNN structures, Cubic RNN is capable of watching multiple adjacent frames at one time step along the spatial axis, which forms a sliding window. The size of the sliding window is equal to the number of time step t-1 time step t Figure 2: A Cubic RNN consisting of three spatial layers and two output layers, which can watch three frames at once. spatial layers. Suppose we have L spatial layers: Cubic RNN will view the previous Xt L+1, , Xt 1 frames to predict the t-th frame. The sliding window enables Cubic RNN to better capture the information about both motions and objects. An example of Cubic RNN is illustrated in Figure 2. Table 1: Results of Cubic RNN and state-of-the-art models on the Moving-MNIST dataset. The Cubic RNN (c y z) denotes that the Cubic RNN model has z spatial layer(s) and y output layer(s), and the channel size of its state is c. We report per-frame mean square error (MSE) and per-frame binary cross-entropy (BCE) of generated frames. Lower MSE or CE means better prediction accuracy. Models MSE BCE MNIST-2 MNIST-3 MNIST-2 FC-LSTM (Srivastava, Mansimov, and Salakhutdinov 2015) 118.3 162.4 483.2 Crid LSTM (Kalchbrenner, Danihelka, and Graves 2016) 111.6 157.8 419.5 Conv LSTM (128 4) (Shi et al. 2015) 103.3 142.1 367.0 Pyramid LSTM (Stollenga et al. 2015) 100.5 142.8 355.3 CDNA (Finn, Goodfellow, and Levine 2016) 97.4 138.2 346.6 DFN (Jia et al. 2016) 89.0 130.5 285.2 VPN baseline (Kalchbrenner et al. 2017) 70.0 125.2 110.1 Pred RNN with spatialtemporal memory (Wang et al. 2017) 74.0 118.2 118.5 Pred RNN + ST-LSTM (128 4) (Wang et al. 2017) 56.8 93.4 97.0 Cubic RNN (32 3 1) 111.5 158.4 386.3 Cubic RNN (32 1 3) 73.4 127.1 210.3 Cubic RNN (32 3 2) 59.7 110.2 121.9 Cubic RNN (32 3 3) 47.3 88.2 91.7 Experiments We evaluated the proposed Cubic LSTM unit on three video prediction datasets, Moving-MNIST dataset (Srivastava, Mansimov, and Salakhutdinov 2015), Robotic Pushing dataset (Finn, Goodfellow, and Levine 2016) and KTH Action dataset (Sch uldt, Laptev, and Caputo 2004), including synthetic and real-world video sequences. All models were trained using the ADAM optimizer (Kingma and Ba 2015) and implemented in Tensor Flow. We trained the models using eight GPUs in parallel and set the batch size to four for each GPU. Moving MNIST dataset The Moving MNIST dataset consists of 20 consecutive frames, 10 for the input and 10 for the prediction. Each frame contains two potentially overlapping handwritten digits moving and bouncing inside a 64 64 image. The digits are chosen randomly from the MNIST dataset and placed initially at random locations. Each digit is assigned a velocity whose direction is chosen uniformly on a unit circle and whose magnitude is also chosen uniformly at random over a fixed range. The size of the training set can be considerable large. We use the code provided by (Shi et al. 2015) to generate training samples on-the-fly. For the test, we followed (Wang et al. 2017) which evaluates methods on two settings, i.e., two moving digits (MNIST-2) and three moving digits (MNIST-3). We also followed (Wang et al. 2017) to evaluate predictions, in which both the mean square error (MSE) and binary cross-entropy (BCE) were used. Only simple prepossessing was done to convert pixel values into the range [0, 1]. Each state in the implementation of Cubic LSTM has 32 channels. The size of the spatial-convolutional kernel was set to 5 5. Both the temporal-convolutional kernel and output-convolutional kernel were set to 1 1. We used the MSE loss and BCE loss to train the models corresponding to the different evaluation metrics. We also adopted an encoder-decoder framework as (Srivastava, Mansimov, and MNIST-2 MNIST-3 spatial state temporal state motion area ground truth Figure 3: Prediction examples on the Moving MNIST dataset (top) and visualizations of the spatial hidden state (middle) and temporal hidden state (bottom). The spatial state provides the object information such as contours and appearances, while the temporal state provides the motion information of potential motion areas. The two states will be exploited to generate the future frame by the output branch. Salakhutdinov 2015; Shi et al. 2015), in which the initial states of the decoder network are copied from the last states of the encoder network. The inputs and outputs were fed and generated as Cubic RNN (Figure 2). All models were trained for 300K iterations with a learning rate of 10 3 for the first 150K iterations and a learning rate of 10 4 for the latter 150K iterations. Improvement by the spatial branch. Compared to Conv LSTM (Shi et al. 2015), Cubic LSTM has an additional spatial branch. To prove the improvement achieved by this branch, we compare two models, both of which have 3 Cubic LSTM units. The first model stacks units along the output axis forms, forming a structure of 1 spatial layer and 3 output layers. Since each state has 32 channels, we denote this structure as (32 3 1). The second model stacks the three units along the spatial axis, forming a structure of 3 spatial composing masks transformed images CDNA kernels kernel: stride: shape: gripper action t-2 gripper state t-2 5 5 2 32 32 32 5 5 1 32 32 32 5 5 1 32 32 32 3 3 2 16 16 64 5 5 1 16 16 64 5 5 1 16 16 64 3 3 2 8 8 128 1 1 1 8 8 128 5 5 1 8 8 128 3 3 2 16 16 64 5 5 1 16 16 64 3 3 2 32 32 32 5 5 1 32 32 32 3 3 2 64 64 11 channel softmax Cubic LSTM Cubic LSTM Cubic LSTM Cubic LSTM Cubic LSTM Cubic LSTM Cubic LSTM gripper action t-1 gripper state t-1 gripper action t-3 gripper state t-3 output (y-) axis spatial (z-) axis 1 2 3 4 5 6 7 8 9 10 Cubic LSTM + CDNA Conv LSTM + CDNA 1 2 3 4 5 6 7 8 9 10 F Novel objects Cubic LSTM + CDNA Conv LSTM + CDNA D &XELF/670 &'1$ E Seen objects Figure 4: (a): Architecture of our model for the Robotic Pushing dataset. The model is largely borrowed from the convolutional dynamic neural advection (CDNA) model proposed in (Finn, Goodfellow, and Levine 2016) and replaces Conv LSTMs in the CNDA model with Cubic LSTMs. Our model has three spatial layers, among which the convolutions and the deconvolutions are shared. (b)-(c): Frame-wise PSNR comparisons of Cubic LSTM + CDNA , Conv LSTM + CDNA (Finn, Goodfellow, and Levine 2016) and CNN + CDNA on the Robotic Pushing dataset. Higher PSNR means better prediction accuracy. layers and 1 output layer structure, denoted as (32 1 3). The results are listed in Table 1. On one hand, although both have 3 Cubic LSTM units, Cubic RNN (32 1 3) significantly outperforms Cubic RNN (32 3 1). As the spatial branch and the temporal branch in Cubic RNN (32 3 1) are identical, the model does not exploit spatial correlations well and only achieves similar accuracy to Conv LSTM (128 4). On the other hand, even though Cubic RNN (32 1 3) uses fewer parameters than Conv LSTM (128 4), it obtains better predictions. This experiment validates our belief that the temporal information and the spatial information should be processed separately. Comparison with other models. We report all the results from existing models on the Moving MNIST dataset (Table 1). We report two Cubic RNN settings. The first has three output layers and two spatial layers (32 3 2), and the second one has three output layers and three spatial layers (32 3 3). Cubic RNN (32 3 3) produces the best prediction. The prediction of Cubic RNN is improved by increasing the number of spatial layers. We illustrate two prediction examples produced by Cubic RNN (32 3 3) in the first row of Figure 3. The model is capable of generating accurate predictions for the two-digit case (MNIST-2). In the second and third rows of Figure 3, we visualize the spatial hidden states and the temporal states of the last Cubic LSTM unit in the (32 3 3) structure when it predicts the first frames of the two prediction examples. These states have 32 channels and we visualize each channel by the function σ( ) 255. As can be seen from the visualization, the spatial hidden state reflects the visual and the spatial information of the digits in the frame. The temporal hidden state suggests that it is likely to contain some motion areas . Compared with the relatively precise infor- mation of objects provided by the spatial hidden state, the motion areas provided by the temporal hidden state are somewhat rough. The output branch will apply these motion areas on the digits to generate the prediction. Robotic Pushing dataset Robotic Pushing (Finn, Goodfellow, and Levine 2016) is an action-conditional video prediction dataset which recodes 10 robotic arms pushing hundreds of objects in a basket. The dataset consists of 50,000 iteration sequences with 1.5 million video frames and two test sets. Objects in the first test set use two subsets of objects in the training set, which are so-called seen objects . The second test set involves two subsets of novel objects , which are not used during training. Each test set has 1,500 recoded sequences. In addition to RGB images, the dataset also provides the corresponding gripper poses by recoding its states and commanded actions, both of which are 5-dimensional vectors. We follow (Finn, Goodfellow, and Levine 2016) to center-crop and downsample images to 64 64, and use the Peak Signal to Noise Ratio (PSNR) (Mathieu, Couprie, and Le Cun 2016) to evaluate the prediction accuracy. Our model is illustrated in Figure 4(a). The model is largely borrowed from the convolutional dynamic neural advection (CDNA) model (Finn, Goodfellow, and Levine 2016) which is an encoder-decoder architecture that expands along the output direction and consists of several convolutional encoders, Conv LSTMs, deconvolutional decoders and CDNA kernels. We denote the CDNA model in (Finn, Goodfellow, and Levine 2016) as Conv LSTM + CDNA . Our model replaces Conv LSTMs in the CNDA model with Cubic LSTMs and expands the model along the spatial axis to form a three-spatial-layer architecture. The Seen objects Novel objects 6 7 8 9 10 6 7 8 9 10 CNN + CDNA Ground truth Cubic LSTM + CDNA Conv LSTM + CDNA (Finn et al.) Figure 5: Qualitative comparisons of Cubic LSTM + CDNA , Conv LSTM + CDNA (Finn, Goodfellow, and Levine 2016) and CNN + CDNA on the Robotic Pushing dataset. Our Cubic LSTM + CDNA method can generate clearer frames than others, especially for the videos with novel objects. convolutional encoders and the deconvolutional decoders are shared among these three spatial layers. We refer to our model as Cubic LSTM + CDNA . We also design a baseline model which replaces Conv LSTMs in the CNDA model with CNNs, denoted as CNN + CDNA . Since CNNs do not have internal state that flows along the temporal dimension, it cannot exploit the temporal information. In this experiment, all models were given three context frames before predicting 10 future frames and were trained for 100K iterations with the mean square error loss and the learning rate of 10 3. The results are shown in Figure 4(b) and Figure 4(c). The Conv LSTM + CDNA model only obtains similar accuracy to the CNN + CDNA model, which indicates that convolutions of Conv LSTMs in the CDNA model may mainly focus on the spatial information while neglecting the temporal information. The Cubic LSTM + CDNA model achieves the best prediction accuracy. A qualitative comparison of predicted video sequences is given in Figure 5. The Cubic LSTM + CDNA model generates sharper frames than other models. KTH Action dataset The KTH Action dataset (Sch uldt, Laptev, and Caputo 2004) is a real-world videos of people performing one of six actions (walking, jogging, running, boxing, handwaving, hand-clapping) against fairly uniform backgrounds. We compared our method with the Motion-Content Network (MCnet) (Villegas et al. 2017) and Disentangled Representation Net (Dr Net) (Denton and Birodkar 2017) on the KTH Action dataset. The qualitative and quantitative comparisons between Dr Net, MCnet and Cubic LSTM are shown in Figure 6. Al- t = 12 t = 15 t = 17 t = 21 t = 25 t = 27 t = 30 Ground truth Time Accuracy Figure 6: Qualitative and quantitative comparison of generated sequences between Dr Net, MCnet and Cubic LSTM. though Dr Net can generate relatively clear frames, the appearance and position of the person is slightly changed. Therefore, the PSNR accuracy is not quite high. For MCnet, the generated frames are a little distorted. Compared to Dr Net and MCnet, our Cubic LSTM model can predict more accurate frames and therefore achieves the highest accuracy. Conclusions In this work, we develop a Cubic LSTM unit for video prediction. The unit processes spatio-temporal information separately by a spatial branch and a temporal branch. This separation can reduce the video prediction burden for networks. The Cubic RNN is created by stacking multiple Cubic LSTMs and generates better predictions than prior models, which validates out belief that the spatial information and temporal information should be processed separately. References Babaeizadeh, M.; Finn, C.; Erhan, D.; Campbell, R. H.; and Levine, S. 2018. Stochastic variational video prediction. In ICLR. Bengio, Y.; Ducharme, R.; Vincent, P.; and Janvin, C. 2003. A neural probabilistic language model. JMLR 3:1137 1155. Chiappa, S.; Racani ere, S.; Wierstra, D.; and Mohamed, S. 2017. Recurrent environment simulators. In ICLR. Cho, K.; van Merrienboer, B.; G ulc ehre, C .; Bahdanau, D.; Bougares, F.; Schwenk, H.; and Bengio, Y. 2014. Learning phrase representations using RNN encoder-decoder for statistical machine translation. In EMNLP. Denton, E. L., and Birodkar, V. 2017. Unsupervised learning of disentangled representations from video. In NIPS. Ebert, F.; Finn, C.; Lee, A. X.; and Levine, S. 2017. Selfsupervised visual planning with temporal skip connections. In Co RL. Fan, H.; Chang, X.; Cheng, D.; Yang, Y.; Xu, D.; and Hauptmann, A. G. 2017. Complex event detection by identifying reliable shots from untrimmed videos. In ICCV. Fan, H.; Xu, Z.; Zhu, L.; Yan, C.; Ge, J.; and Yang, Y. 2018. Watching a small portion could be as good as watching all: Towards efficient video classification. In IJCAI. Finn, C.; Goodfellow, I. J.; and Levine, S. 2016. Unsupervised learning for physical interaction through video prediction. In NIPS. Fragkiadaki, K.; Huang, J.; Alemi, A.; Vijayanarasimhan, S.; Ricco, S.; and Sukthankar, R. 2017. Motion prediction under multimodality with conditional stochastic networks. Co RR abs/1705.02082. Hochreiter, S., and Schmidhuber, J. 1997. Long short-term memory. Neural Computation 9(8):1735 1780. Jaderberg, M.; Simonyan, K.; Zisserman, A.; and Kavukcuoglu, K. 2015. Spatial transformer networks. In NIPS. Jia, X.; Brabandere, B. D.; Tuytelaars, T.; and Gool, L. V. 2016. Dynamic filter networks. In NIPS. Kalchbrenner, N.; van den Oord, A.; Simonyan, K.; Danihelka, I.; Vinyals, O.; Graves, A.; and Kavukcuoglu, K. 2017. Video pixel networks. In ICML. Kalchbrenner, N.; Danihelka, I.; and Graves, A. 2016. Grid long short-term memory. In ICLR. Kingma, D. P., and Ba, J. 2015. Adam: A method for stochastic optimization. In ICLR. Lotter, W.; Kreiman, G.; and Cox, D. D. 2017. Deep predictive coding networks for video prediction and unsupervised learning. In ICLR. Mathieu, M.; Couprie, C.; and Le Cun, Y. 2016. Deep multiscale video prediction beyond mean square error. In ICLR. Mottaghi, R.; Bagherinezhad, H.; Rastegari, M.; and Farhadi, A. 2016. Newtonian image understanding: Unfolding the dynamics of objects in static images. In CVPR. Oh, J.; Guo, X.; Lee, H.; Lewis, R. L.; and Singh, S. P. 2015. Action-conditional video prediction using deep networks in atari games. In NIPS. Ranzato, M.; Szlam, A.; Bruna, J.; Mathieu, M.; Collobert, R.; and Chopra, S. 2014. Video (language) modeling: a baseline for generative models of natural videos. ar Xiv abs/1412.6604. Reed, S. E.; van den Oord, A.; Kalchbrenner, N.; Colmenarejo, S. G.; Wang, Z.; Chen, Y.; Belov, D.; and de Freitas, N. 2017. Parallel multiscale autoregressive density estimation. In ICML. Sch uldt, C.; Laptev, I.; and Caputo, B. 2004. Recognizing human actions: A local SVM approach. In ICPR. Shi, X.; Chen, Z.; Wang, H.; Yeung, D.; Wong, W.; and Woo, W. 2015. Convolutional LSTM network: A machine learning approach for precipitation nowcasting. In NIPS. Srivastava, N.; Mansimov, E.; and Salakhutdinov, R. 2015. Unsupervised learning of video representations using lstms. In ICML. Stollenga, M. F.; Byeon, W.; Liwicki, M.; and Schmidhuber, J. 2015. Parallel multi-dimensional lstm, with application to fast biomedical volumetric image segmentation. In NIPS. Tulyakov, S.; Liu, M.-Y.; Yang, X.; and Kautz, J. 2018. Mocogan: Decomposing motion and content for video generation. Villegas, R.; Yang, J.; Hong, S.; Lin, X.; and Lee, H. 2017. Decomposing motion and content for natural video sequence prediction. Vondrick, C., and Torralba, A. 2017. Generating the future with adversarial transformers. In CVPR. Vondrick, C.; Pirsiavash, H.; and Torralba, A. 2016a. Anticipating visual representations from unlabeled video. In CVPR. Vondrick, C.; Pirsiavash, H.; and Torralba, A. 2016b. Generating videos with scene dynamics. In NIPS. Walker, J.; Doersch, C.; Gupta, A.; and Hebert, M. 2016. An uncertain future: Forecasting from static images using variational autoencoders. In ECCV. Wang, Y.; Long, M.; Wang, J.; Gao, Z.; and Yu, P. S. 2017. Predrnn: Recurrent neural networks for predictive learning using spatiotemporal lstms. In NIPS. Xiong, W.; Luo, W.; Ma, L.; Liu, W.; and Luo, J. 2018. Learning to generate time-lapse videos using multi-stage dynamic generative adversarial networks. In CVPR. Xue, T.; Wu, J.; Bouman, K. L.; and Freeman, B. 2016. Visual dynamics: Probabilistic future frame synthesis via cross convolutional networks. In NIPS. Zhou, Y., and Berg, T. L. 2016. Learning temporal transformations from time-lapse videos. In ECCV. Zhu, L.; Xu, Z.; and Yang, Y. 2017. Bidirectional multirate reconstruction for temporal modeling in videos. In CVPR.