# zeroflow_scalable_scene_flow_via_distillation__406d521a.pdf Published as a conference paper at ICLR 2024 Zero Flow: SCALABLE SCENE FLOW VIA DISTILLATION Kyle Vedder1 Neehar Peri2 Nathaniel Chodosh2 Ishan Khatri3 Eric Eaton1 Dinesh Jayaraman1 Yang Liu4 Deva Ramanan2 James Hays5 1University of Pennsylvania 2Carnegie Mellon University 3Motional 4Lawrence Livermore National Laboratory 5Georgia Tech Scene flow estimation is the task of describing the 3D motion field between temporally successive point clouds. State-of-the-art methods use strong priors and test-time optimization techniques, but require on the order of tens of seconds to process full-size point clouds, making them unusable as computer vision primitives for real-time applications such as open world object detection. Feedforward methods are considerably faster, running on the order of tens to hundreds of milliseconds for full-size point clouds, but require expensive human supervision. To address both limitations, we propose Scene Flow via Distillation, a simple, scalable distillation framework that uses a label-free optimization method to produce pseudo-labels to supervise a feedforward model. Our instantiation of this framework, Zero Flow, achieves state-of-the-art performance on the Argoverse 2 Self-Supervised Scene Flow Challenge while using zero human labels by simply training on large-scale, diverse unlabeled data. At test-time, Zero Flow is over 1000 faster than label-free state-of-the-art optimization-based methods on full-size point clouds (34 FPS vs 0.028 FPS) and over 1000 cheaper to train on unlabeled data compared to the cost of human annotation ($394 vs $750,000). To facilitate further research, we release our code, trained model weights, and high quality pseudo-labels for the Argoverse 2 and Waymo Open datasets at https://vedder.io/zeroflow. 1 INTRODUCTION Scene flow estimation is an important primitive for open-world object detection and tracking (Najibi et al., 2022; Zhai et al., 2020; Baur et al., 2021; Huang et al., 2022; Erçelik et al., 2022). As an example, Najibi et al. (2022) generates supervisory boxes for an open-world Li DAR detector via offline object extraction using high quality scene flow estimates from Neural Scene Flow Prior (NSFP) (Li et al., 2021b). Although NSFP does not require human supervision, it takes tens of seconds to run on a single full-size point cloud pair. If NSFP were both high quality and real-time, its estimations could be directly used as a runtime primitive in the downstream detector instead of relegated to an offline pipeline. This runtime feature formulation is similar to Zhai et al. (2020) s use of scene flow from Flow Net3D (Liu et al., 2019) as an input primitive for their multi-object tracking pipeline; although Flow Net3D is fast enough for online processing of subsampled point clouds, its supervised feedforward formulation requires significant in-domain human annotations. Broadly, these exemplar methods are representative of the strengths and weakness of their class of approach. Supervised feedforward methods use human annotations which are expensive to annotate1. To amortize these costs, human annotations are typically done on consecutive observations, severely limiting the structural diversity of the annotated scenes (e.g. a 15 second sequence from an Autonomous Vehicle typically only covers a single city block); due to costs and labeling difficulty, large-scale labels are also rarely even available outside of Autonomous Vehicle domains. Test-time optimization techniques circumvent the need for human labels by relying on hand-built priors, but they are too slow for online scene flow estimation2. Corresponding email: kvedder@seas.upenn.edu 1At $0.10 / cuboid / frame, the Argoverse 2 (Wilson et al., 2021) train split cost $750,000 to label; Zero Flow s pseudo-labels cost $394 at current cloud compute prices. See Supplemental E for details. 2NSFP (Li et al., 2021b) takes more than 26 seconds and Chodosh (Chodosh et al., 2023) takes more than 35 seconds per point cloud pair on the Argoverse 2 (Wilson et al., 2021) train split. See Supplemental E for details. Published as a conference paper at ICLR 2024 102 103 104 Runtime (ms) Threeway EPE (m) Real Time at 10Hz Zero Flow 1x (Full) Zero Flow 3x (Full) Zero Flow 5x (Full) Fast Flow3D (Full) Zero Flow XL 3x (Full) Fast Flow3D XL (Full) NSFP (Full) Chodosh (Full) Gojcic (20.0k points) Sim2Real (8.2k points) Ego Flow (8.2k points) PPWC (8.2k points) Flow Step3D (8.2k points) Figure 1: We plot the error and run-time of recent scene flow methods on the Argoverse 2 Sensor dataset (Wilson et al., 2021), along with the size of the point cloud prescribed in the method s evaluation protocol. Our method Zero Flow 3X (Zero Flow trained on 3 pseudo-labeled data) outperforms its teacher (NSFP, Li et al. (2021b)) while running over 1000 faster, and Zero Flow XL 3X (Zero Flow with a larger backbone trained on 3 pseudo-labeled data) achieves state-of-the-art. Methods that use any human labels are plotted with , and zero-label methods are plotted with . We propose Scene Flow via Distillation (SFv D), a simple, scalable distillation framework that uses a label-free optimization method to produce pseudo-labels to supervise a feedforward model. SFv D generates a new class of scene flow estimation methods that combine the strengths of optimizationbased and feedforward methods with the power of data scale and diversity to achieve fast run-time and superior accuracy without human supervision. We instantiate this pipeline into Zero-Label Scalable Scene Flow (Zero Flow), a family of methods that, motivated by real-world applications, can process full-size point clouds while providing high quality scene flow estimates. We demonstrate the strength of Zero Flow on Argoverse 2 (Wilson et al., 2021) and Waymo Open (Sun et al., 2020), notably achieving state-of-the-art on the Argoverse 2 Self-Supervised Scene Flow Challenge (Figure 1). Our primary contributions include: We introduce a simple yet effective distillation framework, Scene Flow via Distillation (SFv D), which uses a label-free optimization method to produce pseudo-labels to supervise a feedforward model, allowing us to surpass the performance of slow optimization-based approaches at the speed of feedforward methods. Using SFv D, we present Zero-Label Scalable Scene Flow (Zero Flow), a family of methods that produce fast, state-of-the-art scene flow on full-size clouds, with methods running over 1000 faster than state-of-the-art optimization methods (29.33 ms for Zero Flow 1X vs 35,281.4 ms for Chodosh) on real point clouds, while being over 1000 cheaper to train compared to the cost of human annotations ($394 vs $750,000). We release high quality flow pseudo-labels (representing 7.1 GPU months of compute) for the popular Argoverse 2 (Wilson et al., 2021) and Waymo Open (Sun et al., 2020) autonomous vehicle datasets, alongside our code and trained model weights, to facilitate further research. 2 BACKGROUND AND RELATED WORK Given point clouds Pt at time t and Pt+1 at time t + 1, scene flow estimators predict ˆFt,t+1, a 3D vector for each point in Pt that describes how it moved from t to t + 1 (Dewan et al., 2016). Performance is traditionally measured using the Endpoint Error (EPE) between the predicted flow Published as a conference paper at ICLR 2024 Figure 2: Scene Flow vectors describe where the point on an object at time t will end up on the object at t + 1. In this example, ground truth flow vector A, associated with a point in the upper left concave corner of the object at t has no nearby observations at t + 1 due to occlusion of the concave feature. The ground truth flow vector B, associated with a point on the face of the object at t, does not directly match with any observed point on the object at t + 1 due to observational sparsity. Thus, point matching between t and t + 1 alone is insufficient to generate ground truth flow. ˆFt,t+1 and ground truth flow F t,t+1 (Equation 1): EPE (Pt) = 1 Pt ˆFt,t+1(p) F t,t+1(p) 2 . (1) Unlike next token prediction in language (Radford et al., 2018) or next frame prediction in vision (Weng et al., 2021), future observations do not provide ground truth scene flow (Figure 2). To simply evaluate scene flow estimates, ground truth motion descriptions must be provided by an oracle, typically human annotation of real data (Sun et al., 2020; Wilson et al., 2021) or the generator of synthetic datasets (Mayer et al., 2016; Zheng et al., 2023). Recent scene flow estimation methods either train feedforward methods via supervision from human annotations (Liu et al., 2019; Behl et al., 2019; Tishchenko et al., 2020; Kittenplon et al., 2021; Wu et al., 2020; Puy et al., 2020; Li et al., 2021a; Jund et al., 2021; Gu et al., 2019; Battrawy et al., 2022; Wang et al., 2022), perform human-designed test-time surrogate objective optimization over hand-designed representations (Pontes et al., 2020; Eisenberger et al., 2020; Li et al., 2021b; Chodosh et al., 2023), or learn from self-supervision from human-designed surrogate objectives (Mittal et al., 2020; Baur et al., 2021; Gojcic et al., 2021; Dong et al., 2022; Li et al., 2022). Supervised feedforward methods are efficient at test-time; however, they require costly human annotations at train-time. Both test-time optimization and self-supervised feedforward methods seek to address this problem by optimizing or learning against label-free surrogate objectives, e.g. Chamfer distance (Pontes et al., 2020), cycle-consistency (Mittal et al., 2020), and various hand-designed rigidity priors (Dewan et al., 2016; Pontes et al., 2020; Li et al., 2022; Chodosh et al., 2023; Baur et al., 2021; Gojcic et al., 2021). Self-supervised methods achieve faster inference by forgoing expensive test-time optimization, but do not match the quality of optimization-based methods (Chodosh et al., 2023) and tend to require human-designed priors via more sophisticated network architectures compared to supervised methods (Baur et al., 2021; Gojcic et al., 2021; Kittenplon et al., 2021). In practice, this makes them slower and more difficult to train. In contrast to existing work, we take advantage of the quality of optimization-based methods as well as the efficiency and architectural simplicity of supervised networks. Our approach, Zero Flow, uses label-free optimization methods (Li et al., 2021b) to produce pseudo-labels to supervise a feedforward model (Jund et al., 2021), similar to methods used for distillation in other domains Yang et al. (2023). Pseudo-label Distill Label-Free Optimization Method Slow Teacher Pseudo-labeled Supervised Feed Forward Model Fast Student Figure 3: The Scene Flow via Distillation (SFv D) framework, which describes a new class of scene flow methods that produce high quality, human label-free flow at the speed of feedforward networks. Published as a conference paper at ICLR 2024 We propose Scene Flow via Distillation (SFv D), a simple, scalable distillation framework that creates a new class of scene flow estimators by using a label-free optimization method to produce pseudo-labels to supervise a feedforward model (Figure 3). While conceptually simple, efficiently instantiating SFv D requires careful construction; most online optimization methods and feedforward architectures are unable to efficiently scale to full-size point clouds (Section 3.1). Based on our scalability analysis, we propose Zero-Label Scalable Scene Flow (Zero Flow), a family of scene flow models based on SFv D that produces fast, state-of-the-art scene flow estimates for full-size point clouds without any human labels (Algorithm 1). Zero Flow uses Neural Scene Flow prior (NSFP) (Li et al., 2021b) to generate high quality, label-free pseudo-labels on full-size point clouds (Section 3.2) and Fast Flow3D (Jund et al., 2021) for efficient inference (Section 3.3). 3.1 SCALING SCENE FLOW VIA DISTILLATION TO LARGE POINT CLOUDS Popular AV datasets including Argoverse 2 (Wilson et al. (2021), collected with dual Velodyne VLP-32 sensors) and Waymo Open (Sun et al. (2020), collected with a proprietary lidar sensor and subsampled) have full-size point clouds with an average of 52,000 and 79,000 points per frame, respectively, after ground plane removal (Supplemental A, Figure 6). For practical applications, sensors such as the Velodyne VLP-128 in dual return mode produce up to 480,000 points per sweep (Vel, 2019) and proprietary sensors at full resolution can produce well over 1 million points per sweep. Thus, scene flow methods must be able to process many points in real-world applications. Unfortunately, most existing methods focus strictly on scene flow quality for toy-sized point clouds, constructed by randomly subsampling full point clouds down to 8,192 points (Jin et al., 2022; Tishchenko et al., 2020; Wu et al., 2020; Kittenplon et al., 2021; Liu et al., 2019; Li et al., 2021b). As we are motivated by real-world applications, we instead target scene flow estimation for the full-sized point cloud, making architectural efficiency of paramount importance. As an example of stark differences between feedforward architectures, Fast Flow3D (Jund et al., 2021), which uses a Point Pillar-style encoder (Lang et al., 2019), can process 1 million points in under 100 ms on an NVIDIA Tesla P1000 GPU (making it real-time for a 10Hz Li DAR), while methods like Flow Net3D (Liu et al., 2019) take almost 4 seconds to process the same point cloud. We design our approach to efficiently process full-size point clouds. For SFv D s pseudo-labeling step, speed is less of a concern; pseudo-labeling each point cloud pair is offline and highly parallelizable. High-quality methods like Neural Scene Flow Prior (NSFP, Li et al. (2021b)) require only a modest amount of GPU memory (under 3GB) when estimating scene flow on point clouds with 70K points, enabling fast and low-cost pseudo-labeling using a cluster of commodity GPUs; as an example, pseudo-labeling the Argoverse 2 train split with NSFP is over 1000 cheaper than human annotation (Supplemental E). The efficiency of SFv D s student feedforward model is critical, as it determines both the method s test-time speed and its training speed (faster training enables scaling to larger datasets), motivating models that can efficiently process full-size point clouds. 3.2 NEURAL SCENE FLOW PRIOR IS A SLOW TEACHER Neural Scene Flow Prior (NSFP, Li et al. (2021b)) is an optimization-based approach to scene flow estimation. Notably, it does not use ground truth labels to generate high quality flows, instead relying upon strong priors in its learnable function class (determined by the coordinate network s architecture) and optimization objective (Equation 2). Point residuals are fit per point cloud pair Pt, Pt+1 at test-time by randomly initializing two MLPs; one to describe the forward flow ˆF + from Pt to Pt+1, and one to describe the reverse flow ˆF from Pt + ˆFt,t+1 to Pt in order to impose cycle consistency. The forward flow ˆF + and backward flow ˆF are optimized jointly to minimize Truncated Chamfer(Pt + ˆF +, Pt+1) + Truncated Chamfer(Pt + ˆF + + ˆF , Pt) , (2) where Truncated Chamfer is the standard Chamfer distance with per-point distances above 2 meters set to zero to reduce the influence of outliers. NSFP is able to produce high-quality scene flow estimations due to its choice of coordinate network architecture and use of cycle consistency constraint. The coordinate network s learnable function class is expressive enough to fit the low frequency signal of residuals for moving objects while restrictive enough to avoid fitting the high frequency noise from Truncated Chamfer, and the cycle Published as a conference paper at ICLR 2024 consistency constraint acts as a local smoothness regularizer for the forward flow, as any shattering effects in the forward flow are penalized by the backwards flow. NSFP provides high quality estimates on full-size point clouds (Figure 1), so we select NSFP for Zero Flow s pseudo-label step of SFv D. 3.3 FASTFLOW3D IS A FAST STUDENT Fast Flow3D (Jund et al., 2021) is an efficient feedforward method that learns using human supervisory labels F t,t+1 and per-point foreground / background class labels. Fast Flow3D s loss minimizes a variation of the End-Point Error (Equation 1) that reduces the importance of annotated background points, thus minimizing p Pt σ(p) ˆFt,t+1(p) F t,t+1(p) 2 (3) where σ(p) = 1 if p Foreground 0.1 if p Background . (4) Fast Flow3D s architecture is a Point Pillars-style encoder (Lang et al., 2019), traditionally used for efficient Li DAR object detection (Vedder & Eaton, 2022), that converts the point cloud into a birdseye-view pseudoimage using infinitely tall voxels (pillars). This pseudoimage is then processed with a 4 layer U-Net style backbone. The encoder of the U-Net processes the Pt and Pt+1 pseudoimage separately, and the decoder jointly processes both pseudoimages. A small MLP is used to decode flow for each point in Pt using the point s coordinate and its associated pseudoimage feature. As discussed in Section 3.1, Fast Flow3D s architectural design choices make fast even on full-size point clouds. While most feedforward methods are evaluated using a standard toy evaluation protocol with subsampled point clouds, Fast Flow3D is able to scale up to full resolution point clouds while maintaining real-time performance and emitting competitive quality scene flow estimates using human supervision, making it a good candidate for the distillation step of SFv D. In order to train Fast Flow3D using pseudo-labels, we replace the foreground / background scaling function (Equation 4) with a simple uniform weighting (σ( ) = 1), which collapses to Average EPE; see Supplemental B for experiments with other weighting schemes. Additionally, we depart from Fast Flow3D s problem setup in two minor ways: we delete ground points using dataset provided maps, a standard pre-processing step (Chodosh et al., 2023), and use the standard scene flow problem setup of predicting flow between two frames (Section 2) instead of predicting future flow vectors in meters per second. Algorithm 1 describes our approach, with details specified in Section 4.1. In order to take advantage of the unlabeled data scaling of SFv D, we expand Fast Flow3D to a family of models by designing a higher capacity backbone, producing Fast Flow3D XL. This larger backbone halves the size of each pillar to quadruple the pseudoimage area, doubles the size of the pillar embedding, and adds an additional layer to maintain the network s receptive field in metric space; as a result, the total parameter count increases from 6.8 million to 110 million. Algorithm 1 Zero Flow 1: D collection of unlabeled point cloud pairs Training Data 2: for Pt, Pt+1 D do Parallel For 3: F t,t+1 Teacher NSFP(Pt, Pt+1) SFv D Pseudo-label Step 4: for epoch epochs do 5: for Pt, Pt+1, F t,t+1 D do SFv D s Distill Step 6: l Equation 3(Student Fast Flow3Dθ(Pt, Pt+1), F t,t+1) 7: θ θ updated w.r.t. l 4 EXPERIMENTS Zero Flow provides a family of fast, high quality scene flow estimators. In order to validate this family and understand the impact of components in the underlying Scene Flow via Distillation framework, we perform extensive experiments on the Argoverse 2 (Wilson et al., 2021) and Waymo Open (Sun et al., 2020) datasets. We compare to author implementations of NSFP (Li et al., 2021b) and Chodosh et al. (2023), implement Fast Flow3D (Jund et al., 2021) ourselves (no author implementation is available), and use Chodosh et al. (2023) s implementations for all other baselines. Published as a conference paper at ICLR 2024 As discussed in Chodosh et al. (2023), downstream applications typically rely on good quality scene flow estimates for foreground points. Most scene flow methods are evaluated using average Endpoint Error (EPE, Equation 1); however, roughly 80% of real-world point clouds are background, causing average EPE to be dominated by background point performance. To address this, we use the improved evaluation metric proposed by Chodosh et al. (2023), Threeway EPE: Threeway EPE(Pt) = Avg EPE(p Pt : p Background) (Static BG) EPE(p Pt : p Foreground F t,t+1(p) 0.5m/s) (Static FG) EPE(p Pt : p Foreground F t,t+1(p) > 0.5m/s) (Dynamic FG) . (5) 4.1 HOW DOES ZEROFLOW PERFORM COMPARED TO PRIOR ART ON REAL POINT CLOUDS? The overarching promise of Zero Flow is the ability to build fast, high quality scene flow estimators that improve with the the availability of large-scale unlabeled data. Does Zero Flow deliver on this promise? How does it compare to state-of-the-art methods? To characterize the Zero Flow family s performance, we use Argoverse 2 to perform scaling experiments along two axes: dataset size and student size. For our standard size configuration, we use the Argoverse 2 Sensor train split and the standard Fast Flow3D architecture, enabling head-to-head comparisons against the fully supervised Fast Flow3D as well as other baseline methods. For our scaled up dataset (denoted 3X in all experiments), we use the Argoverse 2 Sensor train split and concatenate a roughly twice as large set of unannotated frame pairs from the Argoverse 2 Li DAR dataset, uniformly sampled from its 20,000 sequences to maximize data diversity. For our scaled up student architecture (denoted XL in all experiments), we use the XL backbone described in Section 3.3. For details on the exact dataset construction and method hyperparameters, see Supplemental A Table 1: Quantitative results on the Argoverse 2 Sensor validation split using the evaluation protocol from Chodosh et al. (2023). The methods used in this paper, shown in the first two blocks of the table, are trained and evaluated on point clouds within a 102.4m 102.4m area centered around the ego vehicle (the settings for the Argoverse 2 Self-Supervised Scene Flow Challenge) . However, following the protocol of Chodosh et al. (2023), all methods report error on points in the 70m 70m area centered around the ego vehicle. Runtimes are collected on an NVIDIA V100 with a batch size of 1 (Peri et al., 2023). Fast Flow3D, Zero Flow 1X, and Zero Flow 3X have identical feedforward architectures and thus share the same real-time runtime; Fast Flow3D XL, Zero Flow XL 1X, and Zero Flow XL 3X have identical feedforward architectures and thus share the same runtime. Methods with an * have performance averaged over 3 training runs (see Supplemental C for details). Underlined methods require human supervision. Runtime (ms) Point Cloud Threeway Dynamic Static Static Subsampled Size EPE FG EPE FG EPE BG EPE Fast Flow3D* (Jund et al., 2021) 29.33 2.38 Full Point Cloud 0.071 0.186 0.021 0.006 Zero Flow 1X* (Ours) Full Point Cloud 0.088 0.231 0.022 0.011 Zero Flow 3X (Ours) Full Point Cloud 0.064 0.164 0.017 0.011 Zero Flow 5X (Ours) Full Point Cloud 0.056 0.140 0.017 0.011 Fast Flow3D XL 260.61 1.21 Full Point Cloud 0.055 0.139 0.018 0.007 Zero Flow XL 1X (Ours) Full Point Cloud 0.070 0.178 0.019 0.013 Zero Flow XL 3X (Ours) Full Point Cloud 0.054 0.131 0.018 0.012 NSFP w/ Motion Comp (Li et al., 2021b) 26, 285.0 18, 139.3 Full Point Cloud 0.067 0.131 0.036 0.034 Chodosh et al. (Chodosh et al., 2023) 35, 281.4 20, 247.7 Full Point Cloud 0.055 0.129 0.028 0.008 Odometry Full Point Cloud 0.198 0.583 0.010 0.000 ICP (Chen & Medioni, 1992) 523.11 169.34 Full Point Cloud 0.204 0.557 0.025 0.028 Gojcic (Gojcic et al., 2021) 6, 087.87 1, 690.56 20000 0.083 0.155 0.064 0.032 Sim2Real (Jin et al., 2022) 99.35 13.88 8192 0.157 0.229 0.106 0.137 Ego Flow (Tishchenko et al., 2020) 2, 116.34 292.32 8192 0.205 0.447 0.079 0.090 PPWC (Wu et al., 2020) 79.43 2.20 8192 0.130 0.168 0.092 0.129 Flow Step3D (Kittenplon et al., 2021) 687.54 3.13 8192 0.161 0.173 0.132 0.176 As shown in Table 1, Zero Flow is able to leverage scale to deliver superior performance. While Zero Flow 1X loses a head-to-head competition against the human-supervised Fast Flow3D on both Argoverse 2 (Table 1) and Waymo Open (Table 2), scaling the distillation process to additional unlabeled data provided by Argoverse 2 enables Zero Flow 3X to significantly surpass the performance of both methods just by training on more pseudo-labled data. Zero Flow 3X even surpasses the performance of its own teacher, NSFP, while running in real-time! Zero Flow s pipeline also benefits from scaling up the student architecture. We modify Zero Flow s architecture with the much larger XL backbone, and show that our Zero Flow XL 3X is able to Published as a conference paper at ICLR 2024 combine the power of dataset and model scale to outperform all other methods, including significantly outperform its own teacher. Our simple approach achieves state-of-the-art on both the Argoverse 2 validation split and Argoverse 2 Self-Supervised Scene Flow Challenge. Table 2: Quantitative results on Waymo Open using the evaluation protocol from Chodosh et al. (2023). Runtimes are scaled to approximate the performance on a V100 (Li et al., 2020). Both Fast Flow3D and Zero Flow 1X have identical feedforward architectures and thus share the same runtime. Underlined methods require human supervision. Runtime (ms) Point Cloud Threeway Dynamic Static Static Subsampled Size EPE FG EPE FG EPE BG EPE Zero Flow 1X (Ours) 21.66 0.48 Full Point Cloud 0.092 0.216 0.015 0.045 Fast Flow3D (Jund et al., 2021) Full Point Cloud 0.078 0.195 0.015 0.024 Chodosh (Chodosh et al., 2023) 93, 752.3 76, 786.1 Full Point Cloud 0.041 0.073 0.013 0.039 NSFP Li et al. (2021b) 90, 999.1 74, 034.9 Full Point Cloud 0.100 0.171 0.022 0.108 ICP (Chen & Medioni, 1992) 302.70 157.61 Full Point Cloud 0.192 0.498 0.022 0.055 Gojcic Gojcic et al. (2021) 501.69 54.63 20000 0.059 0.107 0.045 0.025 Ego Flow (Tishchenko et al., 2020) 893.68 86.55 8192 0.183 0.390 0.069 0.089 Sim2Real (Jin et al., 2022) 72.84 14.79 8192 0.166 0.198 0.099 0.201 PPWC (Wu et al., 2020) 101.43 5.48 8192 0.132 0.180 0.075 0.142 Flow Step3D (Kittenplon et al., 2021) 872.02 6.24 8192 0.169 0.152 0.123 0.232 4.2 HOW DOES ZEROFLOW SCALE? Section 4.1 demonstrates that Zero Flow can leverage scale to capture state-of-the-art performance. However, it s difficult to perform extensive model tuning for large training runs, so predictable estimates of performance as a function of dataset size are critical (Open AI, 2023). Does Zero Flow s performance follow predictable scaling laws? We train Zero Flow and Fast Flow3D on sequence subsets / supersets of the Argoverse 2 Sensor train split. Figure 4 shows Zero Flow and Fast Flow3D s validation Threeway EPE both decrease roughly logarithmically, and this trend appears to hold for XL backbone models as well. 1% 10% 20% 50% 100% 200% 300% 500% Dataset Size as Percentage of AV2 Sensor Train Split 0.14 0.16 0.18 0.20 0.22 Threeway EPE (m) Data without Human Labels Ours Ours XL Fast Flow3D Fast Flow3D XL Figure 4: Empirical scaling laws for Zero Flow. We report Argoverse 2 validation split Threeway EPE as a percentage of the Argoverse 2 train split used, on a log10-log10 scale, trained to convergence. Threeway EPE performance of Zero Flow scales logarithmically with the amount of training data. Empirically, Zero Flow adheres to predictable scaling laws that demonstrate more data (and more parameters) are all you need to get better performance. This makes Zero Flow a practical pipeline for building scene flow foundation models (Bommasani et al., 2021) using the raw point cloud data that exists today in the deployment logs of Autonomous Vehicles and other deployed systems. 4.3 HOW DOES DATASET DIVERSITY INFLUENCE ZEROFLOW S PERFORMANCE? In typical human annotation setups, a point cloud sequence is given to the human annotator. The human generates box annotations in the first frame, and then updates the pose of those boxes as the objects move through the sequence, introducing and removing annotations as needed. This process is much more efficient than annotating disjoint frame pairs, as it amortizes the time spent annotating Published as a conference paper at ICLR 2024 most objects in the sequence. This is why most human annotated training datasets (e.g. Argoverse 2 Sensor, Waymo Open) are composed of contiguous sequences. However, contiguous frames have significant structural similarity; in the 150 frames (15 seconds) of an Argoverse 2 Sensor sequence, the vehicle typically observes no more than a city block s worth of unique structure. Zero Flow, which requires zero human labels, does not have this constraint on its pseudo-labels; NSFP run on non-sequential frames is no more expensive than NSFP run on non-sequential frames, enabling Zero Flow to train on a more diverse dataset. How does dataset diversity impact performance? To understand the impact of data diversity, we train a version of Zero Flow 1X and Zero Flow 2X only on the diverse subset of our Argoverse 2 Li DAR data selected by uniformly sampling 12 frame pairs from each of the 20,000 unique sequences (Table 3). Table 3: Comparison between Zero Flow trained on Argoverse 2 Sensor dataset versus the more diverse, unlabeled Argoverse 2 Li DAR subset described in Section 4.1. Diverse training datasets result in non-trivial performance improvements. Threeway Dynamic Static Static EPE FG EPE FG EPE BG EPE Fast Flow3D* (Jund et al., 2021) 0.071 0.186 0.021 0.006 Zero Flow 1X (AV2 Sensor Data)* 0.088 0.231 0.022 0.011 Zero Flow 1X (AV2 Li DAR Subset Data) 0.082 0.218 0.018 0.009 Zero Flow 2X (AV2 Li DAR Subset Data) 0.072 0.184 0.022 0.011 Dataset diversity has a non-trivial impact on performance; Zero Flow, by virtue of being able to learn across non-contiguous frame pairs, is able to see more unique scene structure and thus learn to better to extract motion in the presence of the unique geometries of the real world. 4.4 HOW DO THE NOISE CHARACTERISTICS OF ZEROFLOW COMPARE TO OTHER METHODS? Zero Flow distills NSFP into a feedforward model from the Fast Flow3D family. Section 4.1 highlights the average performance of Zero Flow across Threeway EPE catagories, but what does the error distribution look like? (a) Fast Flow3D, Log (b) Zero Flow 1X, Log (c) Zero Flow 3X, Log (d) Zero Flow XL 3X, Log (e) NSFP, Log (f) Chodosh, Log (g) Fast Flow3D, Abs. (h) Zero Flow 1X, Abs. (i) Zero Flow 3X, Abs. (j) Zero Flow XL 3X, Abs. (k) NSFP, Abs. (l) Chodosh, Abs. Figure 5: Normalized frame birds-eye-view heatmaps of endpoint residuals for Chamfer Distance, as well as the outputs for NSFP and Chodosh on moving points (points with ground truth speed above 0.5m/s). Perfect predictions would produce a single central dot. Top row shows the frequency on a log10 color scale, bottom row shows the frequency on an absolute color scale. Qualitatively, methods with better quantitative results have tighter residual distributions. See Supplemental F for details. To answer this question, we plot birds-eye-view flow vector residuals of NSFP, Chodosh, Fast Flow3D, and several members of the Zero Flow family on moving objects from the Argoverse 2 validation dataset, where the ground truth is rotated vertically and centered at the origin to present all vectors in the same frame (Figure 5; see Supplemental F for details on construction). Qualitatively, these plots show that error is mostly distributed along the camera ray and distributional tightness (log10 plots) roughly corresponds to overall method performance. Published as a conference paper at ICLR 2024 Overall, these plots provide useful insights to practitioners and researchers, particularly for consumption in downstream tasks; as an example, open world object extraction (Najibi et al., 2022) requires the ability to threshold for motion and cluster motion vectors together to extract the entire object. Decreased average EPE is useful for this task, but understanding the magnitude and distribution of flow vectors is needed to craft good extraction heuristics. 4.5 HOW DOES TEACHER QUALITY IMPACT ZEROFLOW S PERFORMANCE? As shown in Section 4.1 (Chodosh et al., 2023) has superior Threeway EPE over NSFP on both Argoverse 2 and Waymo Open. Can a better performing teacher lead a better version of Zero Flow? To understand the impact of a better teacher, we train Zero Flow on Argoverse 2 using superior quality flow vectors from Chodosh et al. (2023), which proposes a refinement step to NSFP lablels to provide improvements to flow vector quality (Table 4). Zero Flow trained on Chodosh refined pseudo-labels provides no meaningful quality improvement over NSFP pseudo-labels (as discussed in Supplemental C, the Threeway EPE difference is within training variance for Zero Flow). These results also hold for our ablated speed scaled version of Zero Flow in Supplemental B. Since increasing the quality of the teacher over NSFP provides no noticeable benefit, can we get away with using a significantly faster but lower quality teacher to replace NSFP, e.g. the commonly used self-supervised proxy of Truncated Chamfer? To understand if NSFP is necessary, we train Zero Flow on Argoverse 2 using pseudo-labels from the nearest neighbor, truncated to 2 meters as with Truncated Chamfer. The residual distribution of Truncated Chamfer is shown in Supplemental F, Figure 10a. Zero Flow trained on Truncated Chamfer pseudo-labels performs significantly worse than NSFP, motivating the use of NSFP as a teacher. Table 4: Comparison between Zero Flow trained on Argoverse 2 using NSFP pseudo-labels, Zero Flow using Chodosh et al. (2023) pseudo-labels, and Zero Flow using Truncated Chamfer. Methods with an * have performance averaged over 3 training runs (see Supplemental C for details). The minor quality improvement of Chodosh pseudo-labels does not lead to a meaningful difference in performance, while the significant degradation of Truncated Chamfer leads to significantly worse performance. Threeway Dynamic Static Static EPE FG EPE FG EPE BG EPE Zero Flow 1X (NSFP pseudo-labels)* 0.088 0.231 0.022 0.011 Zero Flow 1X (Chodosh et al. (2023) pseudo-labels) 0.085 0.234 0.018 0.004 Zero Flow 1X (Truncated Chamfer pseudo-labels) 0.105 0.226 0.049 0.040 5 CONCLUSION Our scene flow approach, Zero-Label Scalable Scene Flow (Zero Flow), produces fast, state-of-the-art scene flow without human labels via our conceptually simple distillation pipeline. But, more importantly, we present the first practical pipeline for building scene flow foundation models (Bommasani et al., 2021) using the raw point cloud data that exists today in the deployment logs at Autonomous Vehicle companies and other deployed robotics systems. Foundational models in other domains like language (Brown et al., 2020; Open AI, 2023) and vision (Kirillov et al., 2023; Rajeswaran et al., 2022) have enabled significant system capabilities with little or no additional domain-specific fine-tuning (Wang et al., 2023; Ma et al., 2022; 2023). We posit that a scene flow foundational model will enable new systems that can leverage high quality, general scene flow estimates to robustly reason about object dynamics even in foreign or noisy environments. Limitations and Future Work. Zero Flow inherits the biases of its pseudo-labels. Unsurprisingly, if the pseudo-labels consistently fail to estimate scene flow for certian objects, our method will also be unable to predict scene flow for those objects; however, further innovation in model architecture, loss functions, and pseudo-labels may yield better performance. In order to enable further work on Scene Flow via Distillation-based methods, we release3 our code, trained model weights, and NSFP flow pseudo-labels, representing 3.6 GPU months for Argoverse 2 and 3.5 GPU months for Waymo Open. 3https://vedder.io/zeroflow Published as a conference paper at ICLR 2024 Acknowledgements. The research presented in this paper was partially supported by the DARPA SAIL-ON program under contract HR001120C0040, the DARPA Sh ELL program under agreement HR00112190133, the Army Research Office under MURI grant W911NF20-1-0080, and the CMU Center for Autonomous Vehicle Research. This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC5207NA27344. Ramy Battrawy, René Schuster, Mohammad-Ali Nikouei Mahani, and Didier Stricker. RMS-Flow Net: Efficient and Robust Multi-Scale Scene Flow Estimation for Large-Scale Point Clouds. In Int. Conf. Rob. Aut., pp. 883 889. IEEE, 2022. Stefan Andreas Baur, David Josef Emmerichs, Frank Moosmann, Peter Pinggera, Björn Ommer, and Andreas Geiger. SLIM: Self-supervised Li DAR scene flow and motion segmentation. In Int. Conf. Comput. Vis., pp. 13126 13136, 2021. Aseem Behl, Despoina Paschalidou, Simon Donné, and Andreas Geiger. Pointflownet: Learning representations for rigid motion estimation from point clouds. In Int. Conf. Comput. Vis., pp. 7962 7971, 2019. Michael Black. Novelty in science: A guide to reviewers. https://medium.com/@black_ 51980/novelty-in-science-8f1fd1a0a143, 2022. Rishi Bommasani, Drew A. Hudson, Ehsan Adeli, Russ Altman, Simran Arora, Sydney von Arx, Michael S. Bernstein, Jeannette Bohg, Antoine Bosselut, Emma Brunskill, Erik Brynjolfsson, S. Buch, Dallas Card, Rodrigo Castellon, Niladri S. Chatterji, Annie S. Chen, Kathleen A. Creel, Jared Davis, Dora Demszky, Chris Donahue, Moussa Doumbouya, Esin Durmus, Stefano Ermon, John Etchemendy, Kawin Ethayarajh, Li Fei-Fei, Chelsea Finn, Trevor Gale, Lauren E. Gillespie, Karan Goel, Noah D. Goodman, Shelby Grossman, Neel Guha, Tatsunori Hashimoto, Peter Henderson, John Hewitt, Daniel E. Ho, Jenny Hong, Kyle Hsu, Jing Huang, Thomas F. Icard, Saahil Jain, Dan Jurafsky, Pratyusha Kalluri, Siddharth Karamcheti, Geoff Keeling, Fereshte Khani, O. Khattab, Pang Wei Koh, Mark S. Krass, Ranjay Krishna, Rohith Kuditipudi, Ananya Kumar, Faisal Ladhak, Mina Lee, Tony Lee, Jure Leskovec, Isabelle Levent, Xiang Lisa Li, Xuechen Li, Tengyu Ma, Ali Malik, Christopher D. Manning, Suvir P. Mirchandani, Eric Mitchell, Zanele Munyikwa, Suraj Nair, Avanika Narayan, Deepak Narayanan, Benjamin Newman, Allen Nie, Juan Carlos Niebles, Hamed Nilforoshan, J. F. Nyarko, Giray Ogut, Laurel Orr, Isabel Papadimitriou, Joon Sung Park, Chris Piech, Eva Portelance, Christopher Potts, Aditi Raghunathan, Robert Reich, Hongyu Ren, Frieda Rong, Yusuf H. Roohani, Camilo Ruiz, Jack Ryan, Christopher R e, Dorsa Sadigh, Shiori Sagawa, Keshav Santhanam, Andy Shih, Krishna Parasuram Srinivasan, Alex Tamkin, Rohan Taori, Armin W. Thomas, Florian Tramèr, Rose E. Wang, William Wang, Bohan Wu, Jiajun Wu, Yuhuai Wu, Sang Michael Xie, Michihiro Yasunaga, Jiaxuan You, Matei A. Zaharia, Michael Zhang, Tianyi Zhang, Xikun Zhang, Yuhui Zhang, Lucia Zheng, Kaitlyn Zhou, and Percy Liang. On the opportunities and risks of foundation models. Ar Xiv, 2021. Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam Mc Candlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language Models are Few-Shot Learners. In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 1877 1901, 2020. Yang Chen and Gérard Medioni. Object modelling by registration of multiple range images. Img. Vis. Comput., 10(3):145 155, 1992. Nathaniel Chodosh, Deva Ramanan, and Simon Lucey. Re-Evaluating Li DAR Scene Flow for Autonomous Driving. ar Xiv preprint, 2023. Ayush Dewan, Tim Caselitz, Gian Diego Tipaldi, and Wolfram Burgard. Rigid scene flow for 3d lidar scans. In Int. Conf. Intel. Rob. Sys., pp. 1765 1770. IEEE, 2016. Published as a conference paper at ICLR 2024 Guanting Dong, Yueyi Zhang, Hanlin Li, Xiaoyan Sun, and Zhiwei Xiong. Exploiting Rigidity Constraints for Li DAR Scene Flow Estimation. In IEEE Conf. Comput. Vis. Pattern Recog., pp. 12776 12785, 2022. Marvin Eisenberger, Zorah Lahner, and Daniel Cremers. Smooth shells: Multi-scale shape registration with functional maps. In IEEE Conf. Comput. Vis. Pattern Recog., pp. 12265 12274, 2020. Emeç Erçelik, Ekim Yurtsever, Mingyu Liu, Zhijie Yang, Hanzhen Zhang, Pınar Topçam, Maximilian Listl, Yılmaz Kaan Çaylı, and Alois Knoll. 3D Object Detection with a Self-supervised Lidar Scene Flow Backbone. In Shai Avidan, Gabriel Brostow, Moustapha Cissé, Giovanni Maria Farinella, and Tal Hassner (eds.), Computer Vision ECCV 2022, pp. 247 265, Cham, 2022. Springer Nature Switzerland. Zan Gojcic, Or Litany, Andreas Wieser, Leonidas J Guibas, and Tolga Birdal. Weakly supervised learning of rigid 3d scene flow. In IEEE Conf. Comput. Vis. Pattern Recog., pp. 5692 5703, 2021. Xiuye Gu, Yijie Wang, Chongruo Wu, Yong Jae Lee, and Panqu Wang. Hplflownet: Hierarchical permutohedral lattice flownet for scene flow estimation on large-scale point clouds. In IEEE Conf. Comput. Vis. Pattern Recog., pp. 3254 3263, 2019. Shengyu Huang, Zan Gojcic, Jiahui Huang, and Konrad Schindler Andreas Wieser. Dynamic 3D Scene Analysis by Point Cloud Accumulation. In European Conference on Computer Vision, ECCV, 2022. Zhao Jin, Yinjie Lei, Naveed Akhtar, Haifeng Li, and Munawar Hayat. Deformation and Correspondence Aware Unsupervised Synthetic-to-Real Scene Flow Estimation for Point Clouds. In IEEE Conf. Comput. Vis. Pattern Recog., pp. 7233 7243, 2022. Philipp Jund, Chris Sweeney, Nichola Abdo, Zhifeng Chen, and Jonathon Shlens. Scalable Scene Flow From Point Clouds in the Real World. IEEE Robotics and Automation Letters, 12 2021. Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. Alexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alexander C. Berg, Wan-Yen Lo, Piotr Dollár, and Ross Girshick. Segment Anything. ar Xiv:2304.02643, 2023. Yair Kittenplon, Yonina C Eldar, and Dan Raviv. Flowstep3d: Model unrolling for self-supervised scene flow estimation. In IEEE Conf. Comput. Vis. Pattern Recog., pp. 4114 4123, 2021. Alex Lang, Sourabh Vora, Holger Caesar, Lubing Zhou, Jiong Yang, and Oscar Beijbom. Point Pillars: Fast Encoders for Object Detection From Point Clouds. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12689 12697, 2019. Mengtian Li, Yu-Xiong Wang, and Deva Ramanan. Towards streaming perception. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part II 16, pp. 473 488. Springer, 2020. Ruibo Li, Guosheng Lin, Tong He, Fayao Liu, and Chunhua Shen. HCRF-Flow: Scene flow from point clouds with continuous high-order CRFs and position-aware flow embedding. In IEEE Conf. Comput. Vis. Pattern Recog., pp. 364 373, 2021a. Ruibo Li, Chi Zhang, Guosheng Lin, Zhe Wang, and Chunhua Shen. Rigid Flow: Self-Supervised Scene Flow Learning on Point Clouds by Local Rigidity Prior. In IEEE Conf. Comput. Vis. Pattern Recog., pp. 16959 16968, 2022. Xueqian Li, Jhony Kaesemodel Pontes, and Simon Lucey. Neural Scene Flow Prior. Advances in Neural Information Processing Systems, 34, 2021b. Xingyu Liu, Charles R Qi, and Leonidas J Guibas. Flow Net3D: Learning Scene Flow in 3D Point Clouds. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Published as a conference paper at ICLR 2024 Nikola Lopac, Irena Jurdana, Adrian Brneli c, and Tomislav Krljan. Application of Laser Systems for Detection and Ranging in the Modern Road Transportation and Maritime Sector. Sensors, 22(16), 2022. ISSN 1424-8220. Yecheng Jason Ma, Shagun Sodhani, Dinesh Jayaraman, Osbert Bastani, Vikash Kumar, and Amy Zhang. VIP: Towards Universal Visual Reward and Representation via Value-Implicit Pre-Training. ar Xiv preprint ar Xiv:2210.00030, 2022. Yecheng Jason Ma, William Liang, Vaidehi Som, Vikash Kumar, Amy Zhang, Osbert Bastani, and Dinesh Jayaraman. LIV: Language-Image Representations and Rewards for Robotic Control. ar Xiv preprint ar Xiv:2306.00958, 2023. N. Mayer, E. Ilg, P. Häusser, P. Fischer, D. Cremers, A. Dosovitskiy, and T. Brox. A Large Dataset to Train Convolutional Networks for Disparity, Optical Flow, and Scene Flow Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2016. Himangi Mittal, Brian Okorn, and David Held. Just Go With the Flow: Self-Supervised Scene Flow Estimation. In IEEE Conf. Comput. Vis. Pattern Recog., June 2020. Mahyar Najibi, Jingwei Ji, Yin Zhou, Charles R. Qi, Xinchen Yan, Scott Ettinger, and Dragomir Anguelov. Motion Inspired Unsupervised Perception and Prediction in Autonomous Driving. European Conference on Computer Vision (ECCV), 2022. Open AI. Gpt-4 technical report, 2023. Neehar Peri, Mengtian Li, Benjamin Wilson, Yu-Xiong Wang, James Hays, and Deva Ramanan. An empirical analysis of range for 3d object detection. ar Xiv preprint ar Xiv:2308.04054, 2023. Jhony Kaesemodel Pontes, James Hays, and Simon Lucey. Scene flow from point clouds with or without learning. In Int. Conf. 3D Vis., pp. 261 270. IEEE, 2020. Gilles Puy, Alexandre Boulch, and Renaud Marlet. Flot: Scene flow on point clouds guided by optimal transport. In Eur. Conf. Comput. Vis., pp. 527 544. Springer, 2020. Alec Radford, Karthik Narasimhan, Tim Salimans, and Ilya Sutskever. Improving language understanding by generative pre-training. 2018. Aravind Rajeswaran, Vikash Kumar, Chelsea Finn, and Abhinav Gupta. R3M: A Universal Visual Representation for Robot Manipulation. Conference on Robot Learning (Co RL) 2022, 03 2022. Pei Sun, Henrik Kretzschmar, Xerxes Dotiwalla, Aurelien Chouard, Vijaysai Patnaik, Paul Tsui, James Guo, Yin Zhou, Yuning Chai, Benjamin Caine, Vijay Vasudevan, Wei Han, Jiquan Ngiam, Hang Zhao, Aleksei Timofeev, Scott Ettinger, Maxim Krivokon, Amy Gao, Aditya Joshi, Yu Zhang, Jonathon Shlens, Zhifeng Chen, and Dragomir Anguelov. Scalability in Perception for Autonomous Driving: Waymo Open Dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020. Ivan Tishchenko, Sandro Lombardi, Martin R Oswald, and Marc Pollefeys. Self-supervised learning of non-rigid residual flow and ego-motion. In Int. Conf. 3D Vis., pp. 150 159. IEEE, 2020. Kyle Vedder and Eric Eaton. Sparse Point Pillars: Maintaining and Exploiting Input Sparsity to Improve Runtime on Embedded Systems. International Conference on Intelligent Robots and Systems (IROS), 2022. Velodyne Lidar Alpha Prime. Velodyne Lidar, 11 2019. Guanzhi Wang, Yuqi Xie, Yunfan Jiang, Ajay Mandlekar, Chaowei Xiao, Yuke Zhu, Linxi Fan, and Anima Anandkumar. Voyager: An Open-Ended Embodied Agent with Large Language Models. ar Xiv preprint ar Xiv: Arxiv-2305.16291, 2023. Jun Wang, Xiaolong Li, Alan Sullivan, Lynn Abbott, and Siheng Chen. Point Motion Net: Point-Wise Motion Learning for Large-Scale Li DAR Point Clouds Sequences. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 4418 4427, 2022. Published as a conference paper at ICLR 2024 Xinshuo Weng, Jianren Wang, Sergey Levine, Kris Kitani, and Nicholas Rhinehart. Inverting the pose forecasting pipeline with spf2: Sequential pointcloud forecasting for sequential pose forecasting. In Conference on robot learning, pp. 11 20. PMLR, 2021. Benjamin Wilson, William Qi, Tanmay Agarwal, John Lambert, Jagjeet Singh, Siddhesh Khandelwal, Bowen Pan, Ratnesh Kumar, Andrew Hartnett, Jhony Kaesemodel Pontes, Deva Ramanan, Peter Carr, and James Hays. Argoverse 2: Next Generation Datasets for Self-driving Perception and Forecasting. In Proceedings of the Neural Information Processing Systems Track on Datasets and Benchmarks (Neur IPS Datasets and Benchmarks 2021), 2021. Wenxuan Wu, Zhi Yuan Wang, Zhuwen Li, Wei Liu, and Li Fuxin. Pointpwc-net: Cost volume on point clouds for (self-) supervised scene flow estimation. In Eur. Conf. Comput. Vis., pp. 88 107. Springer, 2020. Bo Yang, Ishan Khatri, Michael Happold, and Chulong Chen. Adcnet: Learning from raw radar data via distillation, 2023. Guangyao Zhai, Xin Kong, Jinhao Cui, Yong Liu, and Zhen Yang. Flow MOT: 3D Multi-Object Tracking by Scene Flow Association. Ar Xiv, abs/2012.07541, 2020. Yang Zheng, Adam W. Harley, Bokui Shen, Gordon Wetzstein, and Leonidas J. Guibas. Point Odyssey: A Large-Scale Synthetic Dataset for Long-Term Point Tracking. In ICCV, 2023.