# threedimensional_trajectory_prediction_with_3dmotraj_dataset__e369e0d7.pdf Three-Dimensional Trajectory Prediction with 3DMo Traj Dataset Hao Zhou * 1 2 3 Xu Yang * 4 Mingyu Fan 5 Lu Qi 6 Xiangtai Li 7 Ming-Hsuan Yang 8 Fei Luo 1 3 With the growing interest in embodied and spatial intelligence, accurately predicting trajectories in 3D environments has become increasingly critical. However, no datasets have been explicitly designed to study 3D trajectory prediction. To this end, we contribute a 3D motion trajectory (3DMo Traj) dataset collected from unmanned underwater vehicles (UUVs) operating in oceanic environments. Mathematically, trajectory prediction becomes significantly more complex when transitioning from 2D to 3D. To tackle this challenge, we analyze the prediction complexity of 3D trajectories and propose a new method consisting of two key components: decoupled trajectory prediction and correlated trajectory refinement. The former decouples inter-axis correlations, thereby reducing prediction complexity and generating coarse predictions. The latter refines the coarse predictions by modeling their inter-axis correlations. Extensive experiments show that our method significantly improves 3D trajectory prediction accuracy and outperforms state-of-the-art methods. Both the 3DMo Traj dataset and the method are available at https://github.com/zhouhao94/3DMo Traj. 1. Introduction In recent years, the embodied (Wang et al., 2024; Kaur et al., 2023; Qi et al., 2024) and spatial intelligence (Huang et al., 2024) have witnessed rapid advancement, both of *Equal contribution 1School of Computing and Information Technology, Great Bay Institute for Advanced Study/Great Bay University, Dongguan, China. 2Tsinghua Shenzhen International Graduate School, Tsinghua University, Shenzhen, China. 3Dongguan Key Laboratory for Intelligence and Information Technology, Dongguan, China. 4State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China. 5Institute of Artificial Intelligence, Donghua University, Shanghai, China. 6Insta360 Research, Shenzhen, China. 7Nanyang Technological University, Singapore. 8University of California Merced, Merced, USA. Correspondence to: Fei Luo . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). 2D trajectory prediction Parameters: 3D trajectory prediction Parameters: Decoupled 3D trajectory prediction Parameters: Decoupling the prediction along different axes. Figure 1: Illustrations of different trajectory predictions and their prediction complexities. The 2D, 3D, and decoupled 3D predictions for each point correspond to optimizing one 2D, one 3D, and three 1D Gaussian distributions, which involve optimizing 5, 9, and 6 free parameters, respectively. which emphasize the ability to understand, reason about, and manipulate information within a three-dimensional (3D) space (Zhu et al., 2025; Ishida, 2024). Such advancement highlights the critical role of 3D trajectory prediction, which enables systems to effectively navigate and interact with dynamic environments by anticipating future movements in space. For example, in unmanned systems, accurate 3D motion path prediction for unmanned underwater vehicles (UUVs) (Cao et al., 2022) and unmanned aerial vehicles (UAVs) (Zhang et al., 2023; Huang et al., 2023) helps them avoid obstacles and plan efficient navigation routes in ocean and air environments. Despite the significant importance of 3D trajectory prediction for embodied and spatial intelligence, most current trajectory prediction tasks, including pedestrian (Mohamed et al., 2020; Mangalam et al., 2020; Zhou et al., 2021) and vehicle (Gao et al., 2020; Zhou et al., 2023; Shi et al., 2022) trajectory prediction, focus primarily on 2D trajectories due to the scarcity of publicly available 3D trajectory datasets. To this end, we contribute a 3D motion trajectory (3DMo- Three-Dimensional Trajectory Prediction Traj) dataset, which consists of 3D trajectories collected from UUVs across eight distinct oceanic scenarios. There are two reasons for collecting trajectories from UUVs: (1) the trajectories of multiple UUVs exhibit complex interactions arising from formation movements, and (2) ocean currents introduce fluctuations in UUV trajectories, significantly increasing the challenge of accurate prediction. The 3DMo Traj dataset is annotated with frame-wise intentions to facilitate future research. Existing 2D trajectory prediction methods have shown that incorporation of intentions (Rasouli et al., 2019; Liu et al., 2020; Mangalam et al., 2021; Choi et al., 2021; Zhao et al., 2021) can significantly improve prediction accuracy. Specifically, the trajectories in 3DMo Traj are annotated with both static and motion intentions. The static intention is defined as the octant the agent is expected to reach by the end of the prediction horizon, whereas the motion intention represents the mean velocity changes between the observed and predicted trajectories. Static intentions provide insight into the agent s final spatial goal, while motion intentions capture the agent s dynamic state transitions. Aside from the scarcity of available datasets, one major challenge impeding progress in 3D trajectory prediction is its substantially higher prediction complexity. As illustrated in Figure 1, predicting a 2D trajectory is generally equivalent to optimizing 2D Gaussian distributions with 5 parameters to approximate the possible locations of each 2D point (Mohamed et al., 2020). In contrast, 3D trajectory prediction requires optimizing a 3D Gaussian distribution with 9 parameters for each 3D point. As a result, the complexity of predicting one 3D point is approximately twice that of one 2D point. Furthermore, the prediction complexity of a 3D scenario involving N points is linearly related to 9 N, significantly hindering the prediction of complex scenarios containing multiple 3D agents. Therefore, reducing the optimized parameters for each 3D point is important. To address the increased prediction complexity of 3D trajectory, we propose a novel 3D trajectory prediction method comprising two key components: decoupled trajectory prediction and correlated trajectory refinement. Optimizing 3D Gaussian distribution can be decomposed into optimizing three individual 1D Gaussian distributions and modeling their correlations; please refer to Section A in the Appendix for detailed mathematical proofs. We adopt a divide-andconquer strategy. Specifically, 3D trajectory prediction is first decoupled along three independent axes to mitigate the impact of inter-axis correlations, followed by modeling these correlations. In this way, the prediction of each 3D point with 9 parameters is first simplified as the optimization of three 1D Gaussian distributions with 6 free parameters, as illustrated in Figure 1, thereby reducing the overall prediction complexity. Therefore, the decoupled trajectory pre- diction part employs three independent decoders to predict future trajectories along the x-, y-, and z-axes, producing a coarse prediction. The correlated trajectory refinement part then models the inter-axis correlations of the coarse predictions and generates offsets to refine them. Extensive experiments conducted on the 3DMo Traj dataset demonstrate the superior performance of our proposed method for 3D trajectory prediction. In conclusion, the contributions of this work are as follows: We propose a 3D trajectory dataset named 3DMo Traj, which is collected from UUVs in ocean environments. The dataset includes frame-wise annotations for both motion and static intentions. We analyze the increased prediction complexity of 3D trajectories compared to 2D trajectories and propose a 3D trajectory prediction method that consists of two components: decoupled trajectory prediction and correlated trajectory refinement. The former decouples the inter-axis correlations to generate coarse predictions. The latter models the inter-axis correlations to refine these coarse predictions. We conduct extensive experiments on the 3DMo Traj dataset. The experimental results verify the superior performance of our method in 3D trajectory prediction, establishing a solid baseline for future research. 2. Related Work Trajectory Prediction Datasets. Some trajectory datasets are collected from top-view cameras, such as ETH/UCY (Lerner et al., 2007; Pellegrini et al., 2009), PWRD, and NYGC (Zhou et al., 2011), which focus on pedestrian trajectory prediction. Other datasets, such as SDD (Robicquet et al., 2016) and in D (Bock et al., 2020), are collected using flying drones and include trajectories of vehicles, bicycles, and pedestrians. With advances in sensor technology, high-quality datasets such as Apollo (Ma et al., 2019), Argoverse (Chang et al., 2019; Wilson et al.), and Waymo (Ettinger et al., 2021) use sensors, such as Li DAR, radar, high-definition maps, and localization modules, to record trajectories. Recognizing the importance of agents intentions in prediction, datasets like PIE (Rasouli et al., 2019) and LOKI (Girase et al., 2021) include frame-wise intention annotations. PIE focuses on short-term pedestrian intentions, while LOKI provides long-term intentions for hybrid road users. Although these datasets offer valuable resources for trajectory prediction, they mainly focus on 2D scenarios. To date, no publicly available datasets are specifically designed for 3D trajectory prediction. Trajectory Prediction Methods. These methods in 2D scenarios can be broadly categorized into RNN-based methods, one-shot prediction methods, and intention-conditioned methods. RNN-based methods iteratively predict future trajectory points. Representative works include S-GAN (Gupta Three-Dimensional Trajectory Prediction 0 -200 -200 -400 AUV-1 AUV-2 AUV-3 0 0 -500 -200 AUV-1 AUV-2 AUV-3 (a) Orthographic Projection -300 -200 -100 0 100 200 300 400 North (m) AUV-1 AUV-2 AUV-3 -100 0 100 200 300 400 500 600 North (m) AUV-1 AUV-2 AUV-3 (b) Front View -300 -200 -100 0 100 200 300 400 North (m) AUV-1 AUV-2 AUV-3 -100 0 100 200 300 400 500 600 North (m) AUV-1 AUV-2 AUV-3 (c) Top View -200 -100 0 100 200 300 400 500 600 East (m) AUV-1 AUV-2 AUV-3 -200 0 200 400 600 800 1000 1200 1400 East (m) AUV-1 AUV-2 AUV-3 (d) Side View Figure 2: Orthographic projections, along with front, top, and side views of 3D trajectories for two typical underwater scenarios. Please refer to Figure 9 in the Appendix for the rest of the scenarios. et al., 2018), PIE (Rasouli et al., 2019), STGAT (Huang et al., 2019), Trajectron++(Salzmann et al., 2020), and SRLSTM(Zhang et al., 2020). One-shot prediction methods predict the entire future trajectory in a single step. Prominent approaches include SSTGCNN (Mohamed et al., 2020), MSRL (Wu et al., 2023), MRGTraj (Peng et al., 2023), SImplicit (Mohamed et al., 2022), and Flow Chain (Maeda & Ukita, 2023). Intention-conditioned methods predict future intentions and then generate trajectories conditioned on these predictions. Notable examples include LBEBM (Pang et al., 2021), TNT (Zhao et al., 2021), NSP-SFM (Yue et al., 2022), and Memo Net (Xu et al., 2022). Besides, interaction modeling techniques such as pooling (Mangalam et al., 2020), attention (Salzmann et al., 2020), and graph neural networks (Mohamed et al., 2020) are studied to capture the dynamics of multi-agent systems. Recent research also addresses real-world challenges, including domain adaptation (Feng et al., 2024) and observation length shift (Xu & Fu, 2024), which commonly arise in practical deployment. Furthermore, large language models and unsupervised learning applied in time-series tasks (Sun et al., 2024; Yang & Hong, 2022) also inspire research in trajectory prediction. Trajectory prediction in 3D scenarios primarily focuses on predicting UAV trajectories. Current research on UAV trajectory prediction largely relies on traditional methods, such as Gaussian process regression (Xie & Chen, 2022), Kalman filters (Wang et al., 2018), and principal component analysis (Zhang et al., 2023), as well as simple learningbased methods, including LSTMs (Zhong et al., 2022) and CNNs (Liu et al., 2021). However, these methods struggle to predict 3D trajectories in complex scenarios due to the high prediction complexity of 3D trajectories. 3. 3D Motion Trajectory Dataset The 3DMo Traj dataset comprises eight scenarios featuring three-dimensional (3D) trajectories collected from unmanned underwater vehicles (UUVs) operating in ocean environments. The 3D trajectories are recorded using positioning devices installed on the UUVs. Each scenario includes three UUVs following predefined formations. We Figure 3: Illustration of annotations for motion and static intentions. Motion intention is determined by the velocities of the observed and predicted trajectories, while static intention is based on the location of the trajectory s endpoint. present the orthographic projections and the front, top, and side views of the 3D trajectories for two representative scenarios in Figure 2. Notably, in addition to the complex 3D motion patterns of each UUV, the trajectories also exhibit fluctuations caused by ocean currents, which introduce significant challenges to the prediction task. This section provides an overview of the 3DMo Traj dataset and analyzes its statistical characteristics. 3.1. Dataset Construction To enhance the diversity of 3D motion behaviors, we collect trajectory data using three agents over an extended period. Dataset Preprocess. The sampling rate for the trajectory points is 10 fps, resulting in a very short time interval between consecutive points and, thus, minimal spatial separation. To address this issue, we first downsample the trajectory points to a sample rate of 2 fps. Since each scenario includes only three agents, we further divide the trajectory data into short fragments, each containing 200 data frames. We assign unique IDs to agents in each fragment to enhance the representation of agents in the dataset. The fragments for each scenario are then randomly divided into training, validation, and test sets with a 1:1:1 ratio. Dataset Annotation. Intentions play a crucial role in trajectory prediction. We annotate intentions in the dataset to support further research in 3D trajectory prediction. Three-Dimensional Trajectory Prediction Scenario #1 #2 #3 #4 #5 #6 #7 #8 Frame number 765 1744 1010 1558 1861 895 3379 2597 Distance(m) 0.707 0.282 0.482 0.212 0.505 0.185 0.569 0.287 0.491 0.202 0.784 0.333 0.707 0.251 0.514 0.273 Velocity(m/s) 1.413 0.564 0.964 0.424 1.009 0.369 1.139 0.573 0.982 0.405 1.567 0.666 1.415 0.501 1.028 0.546 Acceleration(m/s2) 0.003 1.373 0.002 1.046 0.005 0.937 0.002 1.361 0.003 1.045 -0.001 1.683 0.001 1.433 0.001 1.212 Curvature(1/m) 0.184 2.428 0.419 3.494 0.215 2.784 0.486 3.591 0.185 2.692 0.074 0.518 0.030 0.300 0.407 2.259 Motion intention (velocity change) :679, :1028, :1494, :1955, :997, :1002, :1735, :1350, :1559, :1981, :948, :1182, :3168, :4810, :2562, :3045, Static intention (Endpoint octant) #1:390,#2:104, #3:722,#4:5, #5:288,#6:138, #7:445,#8:55 #1:161,#2:16, #3:144,#4:95, #5:1051,#6:850, #7:2141,#8:441 #1:74,#2:509, #3:669,#4:491, #5:11,#6:398, #7:312,#8:351 #1:175,#2:1333, #3:509,#4:975, #5:57,#6:713, #7:196,#8:378 #1:301,#2:902, #3:826,#4:854, #5:300,#6:422, #7:911,#8:697 #1:324,#2:524, #3:291,#4:657, #5:145,#6:218, #7:47,#8:294 #1:788,#2:594, #3:682,#4:951, #5:1772,#6:878, #7:2579,#8:1264 #1:1363,#2:1243, #3:2182,#4:786, #5:412,#6:474, #7:556,#8:294 Table 1: Statistics information of the 3DMo Traj dataset. For motion intention, notations , , and respectively represent speed up, speed down, and constant speed. For static intention, notation #N denotes the endpoints in N-th octant. Unlike datasets such as PIE (Rasouli et al., 2019) and LOKI (Girase et al., 2021), which annotate intentions based solely on future trajectories, we propose utilizing both observed and future trajectories. Specifically, we compare the average velocity of the observed trajectory (vo) and the future trajectory (vf) to label samples with motion intention speed up, slow down , or constant speed , as shown in Figure 3. A trajectory sample is labeled as speed up if vf vo > δ, speed down if vf vo < δ, and constant speed if |vf vo| δ, with δ set to 0.1 by default. Additionally, we annotate future location intention, referred to as static intention, based on the endpoint of each trajectory sample, as illustrated in Figure 3. Specifically, we define the current point as the origin, with the velocity direction at that point serving as the x-axis. Using the Frenet-Serret frame, we establish a local coordinate system, within which the endpoints are classified into one of eight octants. This octant-based labeling provides trajectory samples with static intention. As a result, our dataset includes annotations for both motion and static intentions. Dataset Usage. Following the setting in datasets ETH (Pellegrini et al., 2009) and UCY (Lerner et al., 2007), we adopt a sliding window technique to generate samples from the dataset. Each sample contains 20 frames of trajectory points, with the first 8 frames (4 seconds) as the observed trajectory and the remaining 12 frames (6 seconds) as the ground truth future trajectory. Furthermore, we recommend evaluating prediction methods using a leave-one-out cross-validation strategy. Specifically, the model is trained and validated on the training and validation sets of seven scenarios and tested on the test set of the remaining scenario. This strategy helps mitigate validation bias. 3.2. Dataset Statistics To demonstrate the practicality of the proposed dataset for evaluating 3D trajectory prediction, we analyze its characteristics across several dimensions. A detailed summary of the dataset s statistical properties is presented in Table 1. Basic information. Firstly, we present the number of frames for each processed scenario in the second row. Our dataset contains over 13000 frames, offering sufficient data for model training and validation. Motion information. Secondly, we present the average distance, velocity, and acceleration of the trajectory points in the third to fifth rows. The average distance across all scenarios ranges from 0.4 to 0.8 meters, with corresponding average velocities between 0.9 and 1.6 meters per second, which are relatively high speeds for UUVs in ocean environments. Additionally, the significant standard deviations in velocity and acceleration indicate frequent velocity changes among agents in our dataset. The distributions of distance, velocity, and acceleration, presented in the Appendix, further illustrate this variability. These characteristics make the dataset well-suited for evaluating the ability of prediction methods to handle complex velocity dynamics. Curvature information. Thirdly, we calculate the average 3D curvature of the trajectory points, reported in the sixth row. The lowest average curvature is 0.074 in scenario 1, while the highest is 0.486 in scenario 4. Notably, the standard deviation of curvature in each scenario is approximately ten times its mean value. The curvature distributions across all scenarios, visualized in the Appendix, further support this variability. These statistics indicate that the trajectories in our dataset exhibit a wide range of curvatures, providing a robust benchmark for evaluating the prediction ability of methods on complex curvilinear paths. Intention information. Finally, the statistics on intention information are presented in the seventh and eighth rows. Motion intentions are evenly distributed across the labels speed up, slow down, and constant speed. Similarly, static intentions are uniformly distributed across eight different octants. These balanced distributions ensure that our dataset offers a clear distinction in intention labeling. 4. 3D Trajectory Prediction Method In this section, we address 3D trajectory prediction in two steps: Decoupling inter-axis correlations by independently predicting the trajectory along each axis, and Modeling inter-axis correlations to generate offsets that refine the initial predictions. The former reduces the high prediction Three-Dimensional Trajectory Prediction MLP Encoder (x) Key Point Predictor (x) Trajectory Predictor (x) MLP Encoder (y) Key Point Predictor (y) Trajectory Predictor (y) MLP Encoder (z) Key Point Predictor (z) Trajectory Predictor (z) Input Trajectories Decoupled Trajectory Prediction Baseline Encoder 3D Coordinate System ) ) + 1 Correlated Trajectory Refinement (SCA-LSTM) Predicted Trajectories 3D Coordinate System !"#$( !"#$- :(8 6(; 9 051 !"#$( !"#$% !"#$& !"#$' 2,4 6(B 6%B 6&B 6'B 6B( 6B% 6B& 6B' State Correlation (SC) State Aggregation (SA) Element-wise Addition > Sigmoid Concatenation Element-wise Multiplication Att Attention Figure 4: The framework of our proposed 3D trajectory prediction network. In addition to the baseline encoder from LBEBM, our method consists of two key components: decoupled trajectory prediction and correlated trajectory refinement. The former independently predicts the future trajectories along each axis, thereby decoupling inter-axis correlations. The latter generates offsets to refine these predictions by modeling their inter-axis correlations. complexity of 3D trajectories, providing a coarse prediction. The latter refines these predictions by capturing their interaxis correlations. Problem Formulation. Let pt i = (xt i, yt i, zt i) R3 represent the position of agent i at time t. The observed trajectory of agent i is defined as po i = {pt i | t = 1, ..., to}, and the ground truth future trajectory is pf i = {pt i | t = to + 1, ..., to + tf}, where to and tf are the time horizons of the observed and future trajectories, respectively. The objective of 3D trajectory prediction is to use the observed trajectory po i as input to predict a future trajectory ˆpf i = {ˆpt i | t = to+1, ..., to+tp} that closely approximates the ground truth trajectory pf i . 4.1. Decoupled Trajectory Prediction We adopt the LBEBM method (Pang et al., 2021) as our baseline. To decouple the inter-axis correlations of 3D trajectories, we replace LBEBM s decoder with three independent decoders to predict trajectories separately along the x-, y-, and z-axes. LBEBM is a key point-conditioned prediction method that first predicts future key points and then generates the future trajectories conditioned on those points. We follow the same overall procedure as LBEBM but decouple the prediction process into three parallel streams for each axis. The detailed process of our decoupled prediction is illustrated in Figure 4. Let F e i represent the encoded feature of po i , extracted by LBEBM while accounting for interactions and other elements. We first employ three key point predictors to predict key points along different axes: ˆνk i = Φν(F e i ), ν {x, y, z}, (1) where Φν( ) and ˆνk i denote the MLP key point predictor and the predicted key points for the ν-axis, respectively. We then extract features from the predicted key points: F k i = κ([ˆxk i ; ˆyk i ; ˆzk i ]), (2) where κ( ) is an MLP encoder, [ ; ; ] denotes concatenation, and F k i is the extracted feature. Finally, we concatenate F k i with F e i and feed this concatenated feature into three trajectory predictors to generate the future trajectory for each axis: ˆνf i =Ψν([F k i ;F e i ]), ν {x, y, z}, (3) where Ψν( ) and ˆνf i represent the MLP predictor and the predicted future trajectory for the ν-axis, respectively. 4.2. Correlated Trajectory Refinement. We design a novel LSTM-based method, State-Correlation and Aggregation LSTM (SCA-LSTM), to model the interaxis correlations of the coarse predictions and refine them. Three-Dimensional Trajectory Prediction An overview of the SCA-LSTM is shown in Figure 4. SCALSTM employs a centralized architecture comprising a center LSTM, node LSTMs, a state correlation (SC) module, and a state aggregation (SA) module. The center LSTM extracts summarized trajectory features across different axes, while the node LSTM extracts trajectory features for each axis. The SC module captures the correlation between the center and node LSTMs, using it to update the node features. Meanwhile, the SA module aggregates the state features from the node LSTMs to update those of the center LSTM. Inputs Representation. Instead of directly using raw trajectories, we extract their embeddings as inputs for SCALSTM. For the node LSTMs, we first concatenate the predicted future trajectory with the observed trajectory for each axis. Then, we calculate and concatenate the trajectory, velocity, and acceleration of each axis. This concatenated information is passed through MLP encoder to extract embeddings: F ν i = Ων([νu i ; vνu i ; aνu i ]), ν {x, y, z}, (4) where νu i , vνu i , and aνiu represent the trajectory, velocity, and acceleration on the ν-axis, respectively. Ων is the MLP encoder for the ν-axis. F ν i represents the extracted feature, which is the input to the node LSTM for the ν-axis. For the center LSTM, we apply maxand average-pooling operations to combine the input representations from all axes: F m i = Max Pooling(F x i , F y i , F z i ), F a i = Avg Pooling(F x i , F y i , F z i ), F c i = [F m i ; F a i ], (5) where F c i is the extracted input feature for the center LSTM. The Center and Node LSTMs. The center and node LSTMs use vanilla LSTMs to extract features from the input representations: cct i,hct i =LSTMc(F c i [t], hct 1, cct 1; Θc), cνt i,hνt i =LSTMν(F ν i [t], hνt 1, cνt 1; Θν), ν {x,y,z}, (6) where LSTMc, Θc, cct i, and hct i represent the center LSTM, its parameters, cell state, and hidden state, while LSTMν, Θν, cνt i, and hνt i denote the node LSTM for the ν-axis and its components. For more details on the vanilla LSTM, please refer to Section B in the Appendix. The main difference between the LSTMs used in our method and vanilla LSTMs is the inclusion of the SC and SA modules, which refine their cell and hidden states. The SC Module. The SC module takes two inputs: the hidden states and cell states from the center and node LSTMs. The output is the refined cell states for the node LSTMs. Mathematically, the SC module is formulated as follows: ˆcνt,l+1 i = Mc( ˆhc t,l i , ˆ hν t,l i ) + ˆcνt,l i ν {x, y, z}, (7) where ˆcνt,l+1 i denotes the refined cell states for the ν-axis after l + 1 iterations, and Mc represents the message correlation function. We design the message correlation function with a gated selection mechanism to adaptively model correlations and refine the node LSTM s cell states. Specifically, we first extract the input features for the gate from hidden states: F hc,l i = Whc ˆhc t,l i , F hν,l i = Whν ˆ hν t,l i , ν {x, y, z}, (8) where Whc and Whν are linear transformations. Then, we construct a correlation gate using the features F c,l i and F ν,l i such that: gc,ν i =σ(Wcν[F hc,l i ; F hν,l i ] + bcν), ν {x, y, z}, (9) where gc,ν i is the gate vector for the ν-axis, Wcν and bcν are linear transformation and bias, and σ refers to the Sigmoid function. Using the gate vector, Mc is formulated as: Mc( ˆhc t,l i , ˆ hν t,l i )=Wcngc,ν i ˆhc t,l i , ν {x, y, z}, (10) where Wcn is a linear transformation. The SA Module. The SA module adopts a similar strategy to the SC module but operates in reverse, aggregating the hidden states from node LSTMs to update the center LSTM. Mathematically, the SA module is formulated as: ˆcct,l+1 i = X ν {x,y,z} Mν( ˆhc t,l i , ˆ hν t,l i ) + ˆcct,l i , (11) where ˆcct,l+1 i is the refined cell state of the center LSTM, and Mν is the message aggregation function for the ν-axis. To adaptively aggregate the hidden states of the node LSTMs, we design a message aggregation function that incorporates axis-wise attention and an aggregation gate. The attention mechanism assigns appropriate weights to different axes, while the aggregation gate extracts key features for effective integration. To build the attention and aggregation gate, we first extract a global feature F g,l i from the hidden states of all LSTMs: F g,l i = ([Wgc ˆhc t,l i ; Wgx ˆ hx t,l i ; Wgy ˆ hy t,l i ; Wgz ˆ hz t,l i ]), (12) where Wgc, Wgx, Wgy, and Wgz are linear transformations. We then project the global embedding into different LSTMs projection spaces: F gc,l i = Wgc F g,l i , F gν,l i = WgνF g,l i , ν {x, y, z}, (13) where F gc,l i is for the center LSTM, and F gν,l i is for the node LSTM of the ν-axis, with Wgc and Wgν being linear transformations. Using these projected features, the attention weight can be computed as follows: αc,ν = exp(Wq F gc,l i (Wk F gν,l i )T ) P β {x,y,z} exp(Wq F gc,l i (Wk F gβ,l i )T ) , ν {x,y,z}, (14) Three-Dimensional Trajectory Prediction Scenario #1 #2 #3 #4 #5 #6 #7 #8 Mean SSTGCNN (Mohamed et al., 2020) 2.86/5.19 1.05/1.47 1.08/1.65 2.23/3.83 1.15/1.34 2.25/4.00 4.49/8.28 2.18/3.30 2.16/3.63 MSRL (Wu et al., 2023) 3.72/5.42 0.62/0.73 0.56/0.61 1.69/2.05 1.85/2.50 1.54/2.84 1.14/1.73 0.90/1.15 1.50/2.13 Flow Chain (Maeda & Ukita, 2023) 1.44/3.20 0.61/0.99 0.62/1.02 0.90/1.76 0.54/0.84 1.31/2.52 1.18/2.53 0.93/1.84 0.94/1.84 PECNet (Mangalam et al., 2020) 0.79/1.05 0.70/1.29 0.74/1.26 1.29/2.46 0.41/0.58 1.16/1.65 0.92/1.53 1.22/2.21 0.90/1.50 LBEBM (Pang et al., 2021) 0.72/0.98 0.52/0.80 0.64/1.11 0.94/1.87 0.35/0.57 1.02/1.67 1.27/2.35 1.25/2.42 0.84/1.47 NPSN* (Bae et al., 2022) 0.75/0.83 0.69/1.21 0.71/0.96 0.83/1.28 0.34/0.40 0.97/1.20 0.71/0.93 1.03/1.50 0.75/1.04 Traj CLIP (Yao et al., 2024) 0.56/0.94 0.40/0.69 0.37/0.70 1.32/2.81 0.33/0.58 0.79/1.27 1.04/2.02 0.85/1.69 0.71/1.34 Causal HTP (Chen et al., 2021) 0.69/1.29 0.45/0.78 0.45/0.78 0.79/1.40 0.49/0.87 1.11/1.95 0.97/1.96 0.74/1.40 0.71/1.30 MS-TIP (Nath et al., 2024) 0.62/1.16 0.61/1.16 0.58/1.15 0.78/1.42 0.57/1.11 0.84/1.50 0.77/1.44 0.79/1.52 0.70/1.31 MRGTraj (Peng et al., 2023) 0.63/1.29 0.34/0.54 0.33/0.55 0.94/1.89 0.42/0.83 0.99/1.73 1.06/2.44 0.84/1.63 0.69/1.36 S-Implicit (Mohamed et al., 2022) 0.54/0.87 0.43/0.72 0.47/0.84 1.13/2.31 0.40/0.68 0.85/1.38 0.62/1.06 0.98/1.92 0.68/1.22 Our 0.36/0.51 0.37/0.60 0.48/0.86 0.81/1.69 0.28/0.44 0.69/1.10 1.12/1.95 0.55/0.99 0.58/1.02 Table 2: Comparison with state-of-the-art methods on the 3DMo Traj dataset under best-of-20 evaluation setting. : SSTGCNN uses predicted Gaussian distributions to sample future trajectories. *: NPSN here adopts PECNet as the baseline. MRGTraj S-Implicit Our (a) Sample #1: linear trajectories (easy) MRGTraj S-Implicit Our (b) Sample #2: curved trajectories (medium) MRGTraj S-Implicit Our (c) Sample #3: fluctuating trajectories (hard) Figure 5: Visual comparisons between our method with MRGTraj and S-Implicit on 3D trajectory prediction using samples featuring linear, curved, and fluctuating trajectories denoting easy, medium, and hard prediction difficulties. The first to the fourth rows are the orthographic projections and front, top, and side views of visualized results. Dotted lines with blue, green, and red colors represent the observed trajectories, the ground-truth future trajectories, and the predicted future trajectories. where ( )T denotes transposition, and αc,ν is the attention weight for the ν-axis. Additionally, the aggregation gate is defined as: gν,c i =σ(Wνc[F hc,l i ; F hν,l i ] + bνc), ν {x, y, z}, (15) where gν,c i represents the gate vector for the ν-axis, while Wνc and bνc are linear transformations and bias terms. After that, the message aggregation function Mν is defined as: Mν( ˆhc t,l i , ˆ hν t,l i )=αc,νWncgν,c i ˆ hν t,l i , ν {x, y, z}, (16) where Wnc is a linear transformation. Finally, after a round of refinements through the SA and SC modules, the refined cell states are used to update the hidden states via the output gate of the vanilla LSTM, as shown in Equation 30 in the Appendix. Using these updated hidden states, the LSTMs predict the offsets νf i , ν {x, y, z} to refine the coarse predictions on each axis. 4.3. Loss Function Aside from the specialized loss of the baseline method LBEBM, the loss for our method is defined as follows: i {1,...,N} ν {x,y,z} (Lk,ν i + Lf,ν i + Lo,ν i ), Lk,ν i = L2(νk i , ˆνk i ), Lf,ν i = L2(νf i , ˆνf i ), Lo,ν i = L2( νf i , νf i ˆνf i ). where νk i and νf i denote the ground truth key points and future trajectory on the ν-axis, respectively, and L2 represents the mean squared error. 5. Experiments 5.1. Implementation Details Our method uses observed 8 frames to predict future 12 frames. The key points predicted are the 3rd, 6th, 9th, and Three-Dimensional Trajectory Prediction Metric Iteration Number # 1 2 3 4 ADE 0.62 0.67 0.58 0.63 FDE 1.05 1.21 1.02 1.07 Table 3: Analysis of different iterations of the SC and SA on the 3DMo Traj dataset. Metric Layer Number # 1 2 3 4 ADE 0.58 0.64 0.64 0.65 FDE 1.02 1.07 1.06 1.10 Table 4: Analysis of different layers of the SCA-LSTM on the 3DMo Traj dataset. Components Performance BD DD VL SL ADE FDE 1 0.84 1.47 2 0.68 1.19 3 0.66 1.05 4 0.58 1.02 Table 5: Ablation study on our method using the 3DMo Traj dataset. BD refers to our baseline s decoder, DD represents our decoupled trajectory prediction part, VL stands for the vanilla LSTM, and SL refers to our correlated trajectory refinement part, SCA-LSTM. 12th frames of the future trajectory. The primary components of our approach are implemented using MLP or LSTM. Training is conducted with a batch size of 70 for 100 epochs, employing the Adam optimizer with an initial learning rate of 0.0005. The learning rate decayed by 0.5 after the 5th epoch. Results are evaluated using (Average Displacement Error) ADE and (Final Displacement Error) FDE metrics. 5.2. Comparisons to State-of-the-art Methods We present the performance of our proposed method and benchmark it against state-of-the-art methods on the 3DMo Traj dataset, as detailed in Table 2. We modify all compared methods to accept 3D observed trajectories as inputs and predict 3D future trajectories. Additionally, we adjust their data augmentation strategies, such as rotation, flipping, and reversing, to their 3D counterparts for consistency. The experimental results indicate that our method outperforms all evaluated approaches in 3D trajectory prediction. Specifically, our method achieves a 14.6% improvement in the ADE metric, reducing it from 0.68 to 0.58, and a 16.3% enhancement in the FDE metric, lowering it from 1.22 to 1.02, compared to the second-best method, S-Implicit. While NPSN shows an FDE metric close to ours, our ADE metric surpasses it by 22.7%, decreasing from 0.75 to 0.58. Furthermore, we visualize some samples compared to leading methods MRGTraj and S-Implicit, as shown in Figure 5. The three visualized samples illustrate the ability of these methods to predict linear, curved, and fluctuating trajectories with increasing difficulty levels. The visual results indicate: 1) all compared methods perform well on the xand y-axes, reflecting their original 2D design; 2) our method surpasses MRGTraj and S-Implicit in 3D trajectory prediction, particularly along the z-axis; and 3) sample #3 validates our method s marked improvement in predicting complex, fluc- Baseline Our Baseline Our Figure 6: Visual comparisons before and after applying our proposed components on the baseline method. Rows one to four denote different views of the visualized results. tuating trajectories, which pose a significant challenge for methods designed for 2D prediction. 5.3. Hyperparameter Analysis Training our proposed method involves two key hyperparameters. One is the iterations of the SC and SA modules, and the other is the layers of the SAC-LSTM. This subsection analyzes them one after another. Iterations of the SC and SA modules. Each iteration of the SC and SA modules refines the cell and hidden states of the LSTM. We conduct experiments with counts of {1, 2, 3, 4} to choose the optimal number of iterations. The SAC-LSTM layer is fixed at 1 in this experiment to eliminate confounding variables. As shown in Table 3, our method performs best when the iteration number is set to 3. Layers of the SCA-LSTM. When utilizing multiple SCALSTM layers, each layer refines the previous layer s output and serves as input for the subsequent one. Different numbers of SAC-LSTM layers can produce varying outcomes. To evaluate the effect of layer depth, we experiment with SAC-LSTM layers of {1, 2, 3, 4}, as shown in Table 4. The iteration count for the SC and SA modules is fixed at 3, as determined from the previous experiment. The experimental results confirm that using a single SAC-LSTM layer achieves the best performance. 5.4. Ablation Study and Discussion We conduct an ablation study to evaluate our method s core components, i.e., decoupled trajectory prediction and correlated trajectory refinement. The results, presented in Table 5, reveal several vital insights. First, variants 1 and 2 reveal that the decoupled trajectory prediction strategy outperforms the LBEBM decoder by 19.0%/19.0% in ADE/FDE, highlighting its effectiveness. Second, variants 2 and 3 indicate that using vanilla LSTM for refinement provides a performance Three-Dimensional Trajectory Prediction Orthographic projection Top view Front view Side view Figure 7: A representative failure case generated by our method on the 3DMo Traj dataset. Dataset ETH/UCY SDD Method BL Our BL Our ADE 0.22 0.21 9.20 8.98 FDE 0.40 0.38 16.47 15.93 Table 6: Discussion of the generalization ability of our method for 2D trajectory prediction using the ETH/UCY and SDD datasets. BL refers to the baseline LBEBM. boost of 2.9%/11.8%, confirming that trajectory refinement can further improve prediction accuracy. Finally, variants 3 and 4 show that our proposed SCA-LSTM outperforms the vanilla LSTM by 12.1%/2.9%, validating that SCA-LSTM, which models inter-axis correlations, significantly improves trajectory refinement. We also visualize two typical samples before and after applying our method, as shown in Figure 6. While the baseline performs well in predicting trajectories along the xand yaxes, it struggles with the additional z-axis of 3D trajectories. Our method enhances the baseline s performance, particularly in predicting the z-axis component, further demonstrating the effectiveness of our core components. Furthermore, to validate the generalization ability of the proposed method for 2D trajectory prediction, we conduct experiments on the pedestrian trajectory datasets ETH/UCY and SDD, as shown in Table 6. Our method outperforms the baseline by 4.5% and 5.0% on the ETH/UCY dataset and by 2.4% and 3.3% on the SDD dataset. These experimental results demonstrate that our decoupled prediction strategy can also improve the prediction accuracy of 2D trajectory. 5.5. Model Efficiency Analysis We analyze the model efficiency of the proposed method by comparing its parameters, computational cost, and runtime performance with several state-of-the-art methods, as illustrated in Table 7. All models in this experiment are tested on an NVIDIA 2080 Ti GPU using an input size of 70 8 3, where 70 represents the number of agents predicted simultaneously-exceeding the agent count in most real-world applications. The results demonstrate that our method achieves the best performance with a relatively good model efficiency, making it suitable for deployment on embedded robotic systems. Additionally, with an inference speed exceeding 12 FPS, our method meets the real-time Method Paramters (M) FLOPs (G) Speed (s) ADE/FDE MSRL 0.59 0.12 0.09 1.50 / 2.13 LBEBM 1.24 0.09 0.05 0.84 / 1.47 NPSN 0.22 0.14 1.29 0.75 / 1.04 Causal HTP 0.04 0.16 2.54 0.71 / 1.30 MRGTraj 4.36 20.04 0.06 0.69 / 1.36 Our 3.41 0.24 0.08 0.58 / 1.02 Table 7: Comparison with state-of-the-art methods in terms of parameters, computational costs, and inference speed on the 3DMo Traj dataset. All methods are tested using an NVIDIA 2080 Ti GPU with an input size of 70 8 3. decision-making requirements of robotics applications. 5.6. Failure Case Analysis We present a representative failure case in Figure 7, which illustrates that our method struggles when predicting trajectories with multiple sharp bends in short time intervals. Incorporating more advanced interaction modeling techniques or integrating 3D point cloud maps could provide additional structural constraints, thereby improving this case. Nevertheless, as our primary focus is reducing the prediction complexity of 3D trajectories, we adopt a simple interaction modeling strategy and do not use 3D point cloud maps, leading to suboptimal performance in this case. Exploring stronger interaction modeling methods and incorporating 3D environmental representations will be important directions for future work. 6. Conclusion In this paper, we present a large-scale 3D trajectory dataset, 3DMo Traj, which includes frame-wise annotations for both static and dynamic intentions. This dataset facilitates research on 3D trajectory prediction. Additionally, we propose a novel 3D trajectory prediction method that addresses the increased prediction complexity encountered when moving from 2D to 3D trajectories. Experiments on the 3DMo Traj dataset demonstrate that our method outperforms all state-of-the-art models tested, validating its effectiveness in reducing prediction complexity and enhancing prediction accuracy in 3D. In future work, we aim to collect a 3D trajectory dataset of unmanned aerial vehicles (UAVs) operating in air environments to further validate our method and advance research in 3D trajectory prediction. Three-Dimensional Trajectory Prediction Acknowledgements This work is supported in part by the National Natural Science Foundation of China under grant No. 62202308, in part by the Guangdong Research Team for Communication and Sensing Integrated with Intelligent Computing under project No. 2024KCXTD047, in part by the Youth Innovation Promotion Association CAS under grant No. 2022133, in part by the Science and Technology Commission of Shanghai under grant No. 24ZR1400400, and in part by the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korean Government (MSIT) (No. RS-2024-00457882, National AI Research Lab Project). Impact Statement The work presented in this paper aims to advance the field of 3D trajectory prediction. Our work may have potential societal impacts in various industries, including autonomous driving, drone operations, and robotics. While these impacts are still in their early stages, we believe that there is no need to specifically highlight any particular ethical issues or societal consequences at this point. Bae, I., Park, J.-H., and Jeon, H.-G. Non-probability sampling network for stochastic human trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6477 6487, 2022. Bock, J., Krajewski, R., Moers, T., Runde, S., Vater, L., and Eckstein, L. The ind dataset: A drone dataset of naturalistic road user trajectories at german intersections. In 2020 IEEE Intelligent Vehicles Symposium (IV), pp. 1929 1934, 2020. doi: 10.1109/IV47402.2020.9304839. Cao, X., Ren, L., and Sun, C. Dynamic target tracking control of autonomous underwater vehicle based on trajectory prediction. IEEE Transactions on Cybernetics, 53 (3):1968 1981, 2022. Chang, M.-F., Lambert, J., Sangkloy, P., Singh, J., Bak, S., Hartnett, A., Wang, D., Carr, P., Lucey, S., Ramanan, D., et al. Argoverse: 3d tracking and forecasting with rich maps. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8748 8757, 2019. Chen, G., Li, J., Lu, J., and Zhou, J. Human trajectory prediction via counterfactual analysis. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9824 9833, 2021. Choi, C., Malla, S., Patil, A., and Choi, J. H. Drogon: A trajectory prediction model based on intention-conditioned behavior reasoning. In Conference on Robot Learning, pp. 49 63. PMLR, 2021. Ettinger, S., Cheng, S., Caine, B., Liu, C., Zhao, H., Pradhan, S., Chai, Y., Sapp, B., Qi, C. R., Zhou, Y., et al. Large scale interactive motion forecasting for autonomous driving: The waymo open motion dataset. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9710 9719, 2021. Feng, L., Bahari, M., Amor, K. M. B., Zablocki, É., Cord, M., and Alahi, A. Unitraj: A unified framework for scalable vehicle trajectory prediction. In European Conference on Computer Vision, pp. 106 123. Springer, 2024. Gao, J., Sun, C., Zhao, H., Shen, Y., Anguelov, D., Li, C., and Schmid, C. Vectornet: Encoding hd maps and agent dynamics from vectorized representation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 11525 11533, 2020. Girase, H., Gang, H., Malla, S., Li, J., Kanehara, A., Mangalam, K., and Choi, C. Loki: Long term and key intentions for trajectory prediction. In 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 9783 9792, 2021. doi: 10.1109/ICCV48922.2021.00966. Gupta, A., Johnson, J., Fei-Fei, L., Savarese, S., and Alahi, A. Social gan: Socially acceptable trajectories with generative adversarial networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2255 2264, 2018. Huang, D., Zhang, Z., Fang, X., He, M., Lai, H., and Mi, B. Stif: A spatial temporal integrated framework for end-to-end micro-uav trajectory tracking and prediction with 4-d mimo radar. IEEE Internet of Things Journal, 10(21):18821 18836, 2023. Huang, W., Wang, C., Li, Y., Zhang, R., and Fei-Fei, L. Rekep: Spatio-temporal reasoning of relational keypoint constraints for robotic manipulation. In 8th Annual Conference on Robot Learning, 2024. Huang, Y., Bi, H., Li, Z., Mao, T., and Wang, Z. Stgat: Modeling spatial-temporal interactions for human trajectory prediction. In Proceedings of the IEEE/CVF international conference on computer vision, pp. 6272 6281, 2019. Ishida, S. Spatial reasoning and planning for deep embodied agents. Ph D thesis, University of Oxford, 2024. Kaur, D. P., Singh, N. P., and Banerjee, B. A review of platforms for simulating embodied agents in 3d virtual environments. Artificial Intelligence Review, 56(4):3711 3753, 2023. Three-Dimensional Trajectory Prediction Lerner, A., Chrysanthou, Y., and Lischinski, D. Crowds by example. In Computer graphics forum, volume 26, pp. 655 664. Wiley Online Library, 2007. Liu, B., Adeli, E., Cao, Z., Lee, K.-H., Shenoi, A., Gaidon, A., and Niebles, J. C. Spatiotemporal relationship reasoning for pedestrian intent prediction. IEEE Robotics and Automation Letters, 5(2):3485 3492, 2020. Liu, Y., Wang, H., Fan, J., Wu, J., and Wu, T. Controloriented uav highly feasible trajectory planning: A deep learning method. Aerospace Science and Technology, 110:106435, 2021. Ma, Y., Zhu, X., Zhang, S., Yang, R., Wang, W., and Manocha, D. Trafficpredict: Trajectory prediction for heterogeneous traffic-agents. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 6120 6127, 2019. Maeda, T. and Ukita, N. Fast inference and update of probabilistic density estimation on trajectory prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 9795 9805, 2023. Mangalam, K., Girase, H., Agarwal, S., Lee, K.-H., Adeli, E., Malik, J., and Gaidon, A. It is not the journey but the destination: Endpoint conditioned trajectory prediction. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part II 16, pp. 759 776. Springer, 2020. Mangalam, K., An, Y., Girase, H., and Malik, J. From goals, waypoints & paths to long term human trajectory forecasting. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 15233 15242, 2021. Mohamed, A., Qian, K., Elhoseiny, M., and Claudel, C. Social-stgcnn: A social spatio-temporal graph convolutional neural network for human trajectory prediction. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 14424 14432, 2020. Mohamed, A., Zhu, D., Vu, W., Elhoseiny, M., and Claudel, C. Social-implicit: Rethinking trajectory prediction evaluation and the effectiveness of implicit maximum likelihood estimation. In European Conference on Computer Vision, pp. 463 479. Springer, 2022. Nath, A., Kabra, P., Gupta, I., Singh, P., et al. Ms-tip: Imputation aware pedestrian trajectory prediction. In Forty-first International Conference on Machine Learning, 2024. Pang, B., Zhao, T., Xie, X., and Wu, Y. N. Trajectory prediction with latent belief energy-based model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11814 11824, 2021. Pellegrini, S., Ess, A., Schindler, K., and Van Gool, L. You ll never walk alone: Modeling social behavior for multi-target tracking. In 2009 IEEE 12th international conference on computer vision, pp. 261 268. IEEE, 2009. Peng, Y., Zhang, G., Shi, J., Li, X., and Zheng, L. Mrgtraj: A novel non-autoregressive approach for human trajectory prediction. IEEE Transactions on Circuits and Systems for Video Technology, 2023. Qi, Z., Dong, R., Zhang, S., Geng, H., Han, C., Ge, Z., Yi, L., and Ma, K. Shapellm: Universal 3d object understanding for embodied interaction. In European Conference on Computer Vision, pp. 214 238. Springer, 2024. Rasouli, A., Kotseruba, I., Kunic, T., and Tsotsos, J. Pie: A large-scale dataset and models for pedestrian intention estimation and trajectory prediction. In 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 6261 6270, 2019. doi: 10.1109/ICCV.2019.00636. Robicquet, A., Sadeghian, A., Alahi, A., and Savarese, S. Learning social etiquette: Human trajectory understanding in crowded scenes. In Leibe, B., Matas, J., Sebe, N., and Welling, M. (eds.), Computer Vision ECCV 2016, pp. 549 565, Cham, 2016. Springer International Publishing. Salzmann, T., Ivanovic, B., Chakravarty, P., and Pavone, M. Trajectron++: Dynamically-feasible trajectory forecasting with heterogeneous data. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XVIII 16, pp. 683 700. Springer, 2020. Shi, S., Jiang, L., Dai, D., and Schiele, B. Motion transformer with global intention localization and local movement refinement. Advances in Neural Information Processing Systems, 35:6531 6543, 2022. Sun, C., Li, H., Li, Y., and Hong, S. TEST: Text prototype aligned embedding to activate LLM s ability for time series. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview. net/forum?id=Tuh4n ZVb0g. Wang, B., Liu, D., Wang, W., and Peng, X. A hybrid approach for uav flight data estimation and prediction based on flight mode recognition. Microelectronics Reliability, 84:253 262, 2018. Wang, T., Mao, X., Zhu, C., Xu, R., Lyu, R., Li, P., Chen, X., Zhang, W., Chen, K., Xue, T., et al. Embodiedscan: A holistic multi-modal 3d perception suite towards embodied ai. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 19757 19767, 2024. Three-Dimensional Trajectory Prediction Wilson, B., Qi, W., Agarwal, T., Lambert, J., Singh, J., Khandelwal, S., Pan, B., Kumar, R., Hartnett, A., Pontes, J. K., et al. Argoverse 2: Next generation datasets for self-driving perception and forecasting. In Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2). Wu, Y., Wang, L., Zhou, S., Duan, J., Hua, G., and Tang, W. Multi-stream representation learning for pedestrian trajectory prediction. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pp. 2875 2882, 2023. Xie, G. and Chen, X. Efficient and robust online trajectory prediction for non-cooperative unmanned aerial vehicles. Journal of Aerospace Information Systems, 19(2):143 153, 2022. Xu, C., Mao, W., Zhang, W., and Chen, S. Remember intentions: Retrospective-memory-based trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6488 6497, 2022. Xu, Y. and Fu, Y. Adapting to length shift: Flexilength network for trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15226 15237, 2024. Yang, L. and Hong, S. Unsupervised time-series representation learning with iterative bilinear temporal-spectral fusion. In International conference on machine learning, pp. 25038 25054. PMLR, 2022. Yao, P., Zhu, Y., Bi, H., Mao, T., and Wang, Z. Trajclip: Pedestrian trajectory prediction method using contrastive learning and idempotent networks. In The Thirty-eighth Annual Conference on Neural Information Processing Systems, 2024. Yue, J., Manocha, D., and Wang, H. Human trajectory prediction via neural social physics. In European conference on computer vision, pp. 376 394. Springer, 2022. Zhang, J., Shi, Z., Zhang, A., Yang, Q., Shi, G., and Wu, Y. Uav trajectory prediction based on flight state recognition. IEEE Transactions on Aerospace and Electronic Systems, 60(3):2629 2641, 2023. Zhang, P., Xue, J., Zhang, P., Zheng, N., and Ouyang, W. Social-aware pedestrian trajectory prediction via states refinement lstm. IEEE transactions on pattern analysis and machine intelligence, 44(5):2742 2759, 2020. Zhao, H., Gao, J., Lan, T., Sun, C., Sapp, B., Varadarajan, B., Shen, Y., Shen, Y., Chai, Y., Schmid, C., et al. Tnt: Target-driven trajectory prediction. In Conference on Robot Learning, pp. 895 904. PMLR, 2021. Zhong, G., Zhang, H., Zhou, J., Zhou, J., and Liu, H. Shortterm 4d trajectory prediction for uav based on spatiotemporal trajectory clustering. IEEE Access, 10:93362 93380, 2022. Zhou, B., Wang, X., and Tang, X. Random field topic model for semantic region analysis in crowded scenes from tracklets. In CVPR 2011, pp. 3441 3448, 2011. doi: 10.1109/CVPR.2011.5995459. Zhou, H., Ren, D., Xia, H., Fan, M., Yang, X., and Huang, H. Ast-gnn: An attention-based spatio-temporal graph neural network for interaction-aware pedestrian trajectory prediction. Neurocomputing, 445:298 308, 2021. Zhou, Z., Wang, J., Li, Y.-H., and Huang, Y.-K. Querycentric trajectory prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 17863 17873, 2023. Zhu, H., Yang, H., Wang, Y., Yang, J., Wang, L., and He, T. SPA: 3d spatial-awareness enables effective embodied representation. In The Thirteenth International Conference on Learning Representations, 2025. Three-Dimensional Trajectory Prediction This supplementary file provides more details to demonstrate the benefits of our proposed dataset and method. Prediction complexity analysis. Details of vanilla LSTM. The motion trajectory visualization. The distance distributions for different scenarios. The velocity distributions for different scenarios The acceleration distributions for different scenarios. The curvature distributions for different scenarios. A. Prediction Complexity Analysis The prediction complexity increases nearly twofold when transitioning from 2D (x, y) to 3D (x, y, z) trajectories. Predicting 2D trajectory is equivalent to optimizing 2D Gaussian distributions to approximate the possible locations of each 2D point: f(x, y) = 1 2π |Σxy|1/2 exp 1 2 x µx y µy Σ 1 xy where Σxy denotes the covariance matrix for the 2D Gaussian distribution and can be represented as follows: Σxy = σ2 x ρxyσxσy ρxyσxσy σ2 y Therefore, the prediction of each 2D point involves optimizing 5 free parameters µx, µy, σ2 x, σ2 y, ρxy . Similarly, predicting a 3D point is equivalent to optimizing a 3D Gaussian distribution: f(x, y, z) = 1 (2π)3/2 |Σxyz|1/2 exp 2 x µx y µy z µz Σ 1 xyz x µx y µy z µz where Σxyz is the covariance matrix for the 3D Gaussian distribution and can be represented as follows: σ2 x ρxyσxσy ρxzσxσz ρxyσxσy σ2 y ρyzσyσz ρxzσxσz ρyzσyσz σ2 z This implies that there are 9 parameters µx, µy, µz, σ2 x, σ2 y, σ2 z, ρxy, ρyz, ρxz to be optimized for predicting each 3D point. Consequently, the complexity of predicting one 3D point is approximately twice that of one 2D point. For a 3D Gaussian distribution, by substituting the covariance matrix Σxyz into Equation 20, we obtain its expanded form: f(x, y, z)= 1 (2π)3/2 |Σxyz|1/2 exp σ2x + (y µy)2 σ2y + (z µz)2 2ρxy (x µx)(y µy) σxσy 2ρxz (x µx)(z µz) σxσz 2ρyz (y µy)(z µz) The exponential term of Equation 22 can be decomposed into two multiplicative components, such that Equation 22 can be expressed as follows: f(x, y, z)= 1 σ2xσ2yσ2z exp 1 σ2x + (y µy)2 σ2y + (z µz)2 | {z } Part I |Σxyz|1/2 exp 1 2ρxy (x µx)(y µy) σxσy + 2ρxz (x µx)(z µz) σxσz + 2ρyz (y µy)(z µz) | {z } Part II Three-Dimensional Trajectory Prediction where Part I and Part II denote the independent and correlation parts, respectively. Part I corresponds to the distribution of each variable within its own dimension, while Part II captures the interdependencies between variables in the 3D Gaussian distribution. This decomposition underscores that the overall Gaussian distribution is the product of independent Gaussian distributions (Part I) and a correction factor that accounts for inter-axis correlations (Part II). Additionally, the independent part can further be expressed as the product of three 1D Gaussian distributions: 2πσ2ν exp (xν µν)2 Therefore, we propose a divide-and-conquer strategy for predicting 3D trajectories. This approach first decouples the prediction process along three independent axes, mitigating the impact of inter-axis correlations. Subsequently, the inter-axis correlations are modeled to refine the predictions. In this way, the prediction of each 3D point with 9 parameters is first simplified to optimize three 1D Gaussian distributions with 6 free parameters, thereby reducing the overall prediction complexity. B. Details of Vanilla LSTM LSTM is a Recurrent Neural Network (RNN) designed to process sequential data such as time series, natural language text, etc. Traditional RNNs struggle with long sequences due to the vanishing or exploding gradient problem. LSTM addresses this by incorporating a gating mechanism that helps capture long-term dependencies effectively. Figure 8: Illustration of LSTM cell. Vanilla LSTM refers to the basic form of LSTM, composed of a forget gate, input gate, output gate, and cell state, as illustrated in Figure 8. At each time step t, the LSTM receives an input vector xt, a previous hidden state ht 1, and a previous cell state ct 1. Then the four modules compute as follows: Forget Gate: The forget gate decides what information to remove from the cell state. It takes the previous hidden state ht 1 and current input xt, and computes a forget ratio: ft = σ(Wf [ht 1, xt] + bf), (25) ft: Forget gate activation vector at time t, σ: Sigmoid activation function, Wf: Weight matrix for the forget gate, bf: Bias vector for the forget gate, ht 1: Hidden state from the previous time step, xt: Input vector at the current time step. Three-Dimensional Trajectory Prediction Input Gate: The input gate determines what new information to add to the cell state. The gate activation is computed as: it = σ(Wi [ht 1, xt] + bi), (26) and the candidate memory content is: ct = tanh(Wc [ht 1, xt] + bc), (27) it: Input gate activation vector, ct: Candidate cell state vector, Wi, Wc: Weight matrices for the input gate and cell state, bi, bc: Bias vectors for the input gate and cell state, tanh: Hyperbolic tangent activation function. Cell State Update: The cell state is updated using information from the forget and input gates: ct = ft ct 1 + it ct, (28) ct: Cell state vector at time t, ct 1: Cell state vector from the previous time step, : Element-wise multiplication. Output Gate: The output gate decides the current hidden state (and the output of the LSTM). The gate activation is: ot = σ(Wo [ht 1, xt] + bo), (29) and the hidden state is computed as: ht = ot tanh(ct), (30) ot: Output gate activation vector, ht: Hidden state vector at time t, Wo: Weight matrix for the output gate, bo: Bias vector for the output gate. In summary, the LSTM performs the following operations at each time step: 1. Inputs the previous hidden state ht 1 and current input xt. 2. Uses the forget gate to decide which past information to discard. 3. Updates the cell state with the input gate and candidate memory. 4. Computes the current hidden state using the output gate. Three-Dimensional Trajectory Prediction C. The Motion Trajectory Visualization 0 -200 -200 -300 -400 AUV-1 AUV-2 AUV-3 East (m) North (m) 0 0 -100 -200 AUV-1 AUV-2 AUV-3 100 0 0 -100 -100 AUV-1 AUV-2 AUV-3 0 0 -200 -200 -400 AUV-1 AUV-2 AUV-3 0 0 -200 -100 AUV-1 AUV-2 AUV-3 200 0 0 -100 -200 AUV-1 AUV-2 AUV-3 (a) Orthographic Projection -400 -300 -200 -100 0 100 200 300 400 500 North (m) AUV-1 AUV-2 AUV-3 -100 0 100 200 300 400 500 600 North (m) AUV-1 AUV-2 AUV-3 -100 -50 0 50 100 150 200 250 300 350 North (m) AUV-1 AUV-2 AUV-3 -300 -200 -100 0 100 200 300 400 500 600 North (m) AUV-1 AUV-2 AUV-3 -100 -50 0 50 100 150 200 250 300 North (m) AUV-1 AUV-2 AUV-3 -100 0 100 200 300 400 500 600 700 800 North (m) AUV-1 AUV-2 AUV-3 (b) Front View -400 -300 -200 -100 0 100 200 300 400 500 North (m) AUV-1 AUV-2 AUV-3 -100 0 100 200 300 400 500 600 North (m) AUV-1 AUV-2 AUV-3 -100 -50 0 50 100 150 200 250 300 350 North (m) AUV-1 AUV-2 AUV-3 -300 -200 -100 0 100 200 300 400 500 600 North (m) AUV-1 AUV-2 AUV-3 -100 -50 0 50 100 150 200 250 300 North (m) AUV-1 AUV-2 AUV-3 -100 0 100 200 300 400 500 600 700 800 North (m) AUV-1 AUV-2 AUV-3 (c) Top View -250 -200 -150 -100 -50 0 50 East (m) AUV-1 AUV-2 AUV-3 -100 -50 0 50 100 150 200 250 300 350 400 East (m) AUV-1 AUV-2 AUV-3 -100 -50 0 50 100 150 200 250 300 350 East (m) AUV-1 AUV-2 AUV-3 -200 -100 0 100 200 300 400 East (m) AUV-1 AUV-2 AUV-3 -100 0 100 200 300 400 500 600 East (m) AUV-1 AUV-2 AUV-3 -100 -50 0 50 100 150 200 250 300 350 400 East (m) AUV-1 AUV-2 AUV-3 (d) Side View Figure 9: Orthographic projections, along with front, top, and side views of 3D trajectories for the remaining five underwater scenarios in the 3DMo Traj dataset. Three-Dimensional Trajectory Prediction D. The Distance Distributions for Different Scenarios (a) Scenario #1 (b) Scenario #2 (c) Scenario #3 (d) Scenario #4 (e) Scenario #5 (f) Scenario #6 (g) Scenario #7 (h) Scenario #8 Figure 10: The distance distributions for scenarios #1 through #8. E. The Velocity Distributions for Different Scenarios (a) Scenario #1 (b) Scenario #2 (c) Scenario #3 (d) Scenario #4 (e) Scenario #5 (f) Scenario #6 (g) Scenario #7 (h) Scenario #8 Figure 11: The velocity distributions for scenarios #1 through #8. Three-Dimensional Trajectory Prediction F. The Acceleration Distributions for Different Scenarios (a) Scenario #1 (b) Scenario #2 (c) Scenario #3 (d) Scenario #4 (e) Scenario #5 (f) Scenario #6 (g) Scenario #7 (h) Scenario #8 Figure 12: The acceleration distributions for scenarios #1 through #8. G. The Curvature Distributions for Different Scenarios (a) Scenario #1 (b) Scenario #2 (c) Scenario #3 (d) Scenario #4 (e) Scenario #5 (f) Scenario #6 (g) Scenario #7 (h) Scenario #8 Figure 13: The curvature distributions for scenarios #1 through #8.