# deepinteraction_3d_object_detection_via_modality_interaction__5406ef28.pdf Deep Interaction: 3D Object Detection via Modality Interaction Zeyu Yang1 Jiaqi Chen1 Zhenwei Miao2 Wei Li3 Xiatian Zhu4 Li Zhang1 1Fudan University 2Alibaba DAMO Academy 3S-Lab, NTU 4University of Surrey https://github.com/fudan-zvg/Deep Interaction Existing top-performance 3D object detectors typically rely on the multi-modal fusion strategy. This design is however fundamentally restricted due to overlooking the modality-specific useful information and finally hampering the model performance. To address this limitation, in this work we introduce a novel modality interaction strategy where individual per-modality representations are learned and maintained throughout for enabling their unique characteristics to be exploited during object detection. To realize this proposed strategy, we design a Deep Interaction architecture characterized by a multi-modal representational interaction encoder and a multi-modal predictive interaction decoder. Experiments on the large-scale nu Scenes dataset show that our proposed method surpasses all prior arts often by a large margin. Crucially, our method is ranked at the first position at the highly competitive nu Scenes object detection leaderboard. 1 Introduction 3D object detection is critical for autonomous driving by localizing and recognizing decisionsensitive objects in a 3D world. For reliable object detection, Li DAR and camera sensors have been simultaneously deployed to provide point clouds and RGB images for more stronger perception. The two modalities exhibit naturally strong complementary effects due to their different perceiving characteristics. Point clouds offer necessary localization and geometry information at low resolution, whilst images give rich appearance information at high resolution. Therefore, information fusion across modalities becomes particularly crucial for strong 3D object detection performance. Existing multi-modal 3D objection detection methods typically adopt a modality fusion strategy (Figure 1(a)) by combining individual per-modality representations into a single hybrid representation. For example, Point Painting [38] and its variants [39, 46, 43] aggregate category scores or semantic features from the image space into the 3D point cloud space. Auto Align [10] and VFF [23] similarly integrate image representations into the 3D grid space. Latest alternatives [24, 30, 26] merge the image and point cloud features into a joint bird s-eye view (BEV) representation. This fusion approach is, however, structurally restricted due to its intrinsic limitation of potentially dropping off a large fraction of modality-specific representational strengths due to largely imperfect information fusion into a unified representation. To overcome the aforementioned limitations, in this work a novel modality interaction strategy (Figure 1(b)) for multi-modal 3D object detection is introduced. Our key idea is that, instead of deriving a fused single representation, we learn and maintain two modality-specific representations throughout to enable inter-modality interaction so that both information exchange and modalityspecific strengths can be achieved spontaneously. Our strategy is implemented by formulating a Li Zhang (lizhangfd@fudan.edu.cn) is the corresponding author with School of Data Science, Fudan University. 36th Conference on Neural Information Processing Systems (Neur IPS 2022). (a) Multi-modality fusion (b) Multi-modality interaction H box & class H ... Image feature F Li DAR BEV H box & class H ... Image feature F Fusion block H Detection head Encoder Figure 1: Schematic strategy comparison. (a) Existing multi-modality fusion based 3D detection: Fusing individual per-modality representations into a single hybrid representation and from which the detection results are further decoded. (b) Our multi-modality interaction based 3D detection: Maintaining two modality-specific representations throughout the whole pipeline with both representational interaction in the encoder and predictive interaction in the decoder. Deep Interaction architecture. It starts by mapping 3D point clouds and 2D multi-view images into Li DAR BEV feature and image perspective feature with two separate feature backbones in parallel. Subsequently, an encoder interacts the two features for progressive information exchange and representation learning in a bilateral manner. To fully exploit per-modality representations, a decoder/head is further designed to conduct multi-modal predictive interaction in a cascaded manner. Our contributions are summarized as follows: (i) We propose a novel modality interaction strategy for multi-modal 3D object detection, with the aim to resolve a fundamental limitation of previous modality fusion strategy in dropping the unique perception strengths per modality. (ii) To implement our proposed strategy, we formulate a Deep Interaction architecture with a multi-modal representational interaction encoder and a multi-modal predictive interaction decoder. (iii) Extensive experiments on the nu Scenes dataset show that our Deep Interaction yields new state of the art for multi-modality 3D object detection and achieves the first position at the highly competitive nu Scenes leaderboard. 2 Related work 3D object detection with single modality Automated driving vehicles are generally equipped with both Li DAR and multiple surround-view cameras. But many previous methods perform 3D object detection by exploiting data captured from only a single form of sensor. For camera-based 3D object detection, since depth information is not directly accessible from RGB images, some previous works [17, 40, 37] lift 2D features into a 3D space by conducting depth estimation, followed by performing object detection in the 3D space. Another line of works [41, 28, 25, 31, 19] resort to the detection Transformer [5] architecture. They leverage 3D object queries and 3D-2D correspondence to incorporate 3D computation into the detection pipelines. Despite the rapid progress of camerabased approaches, the state-of-the-art of 3D object detection is still dominated by Li DAR-based methods. Most of Li DAR-based detectors quantify point clouds into regular grid structures such as voxels [47, 44], pillars [22] or range images [2, 14, 6] before processing them. Due to the sampling characteristics of Li DAR, these grids are naturally sparse and hence fit the Transformer design. So a number of approaches [32, 13] have applied the Transformer for point cloud feature extraction. Differently, several methods use the Transformer decoder or its variants as their detection head [1, 42]. 3DETR [33] adopts a complete Transformer encoder-decoder architecture with less priori in design. Due to intrinsic limitations with either sensor, these methods are largely limited in performance. Multi-modality fusion for 3D object detection Leveraging the perception data from both camera and Li DAR sensors usually leads to improved performance. This has emerged as a promising direction. Existing 3D detection methods typically perform multi-modal fusion at one of the three stages: raw input, intermediate feature, and object proposal. For example, Point Painting [38] is the pioneering input fusion method [39, 18, 46]. The main idea is to decorate the 3D point clouds with the category scores or semantic features from the 2D instance segmentation network. Whilst 4D-Net [35] placed the fusion module in the point cloud feature extractor for allowing the point cloud features to dynamically attend to the image features. Im Vote Net [36] injects visual information into a set of 3D seed points abstracted from raw point clouds. The proposal based fusion methods [21, 7] ... Sampling ... Sampling (a) Representational interaction from image to Li DAR (b) Representational interaction from Li DAR to image coord. coord. Figure 2: Illustration of the multi-modal representational interactions. Given two modality-specific representations, the image-to-Li DAR feature interaction (a) spread the visual signal in the image representation to the Li DAR BEV representation, and the Li DAR-to-image feature interaction (b) takes cross-modal relative contexts from Li DAR representation to enhance the image representations. keep the feature extraction of two modalities independently and aggregate multi-modal features via proposals or queries at the detection head. The first two categories of methods take a unilateral fusion strategy with bias to 3D Li DAR modality due to the superiority of point clouds in distance and spatial perception. Instead, the last category fully ignores the intrinsic association between the two modalities in representation. As a result, all above previous methods fail to fully exploit both modalities, in particular their strong complementary nature. Besides, a couple of concurrent works have explored fusion of the two modalities in a shared representation space [30, 26]. They conduct view transformation in the same way [34] as in the camera-only approach. This design is however less effective in exploiting the spatial cues of point clouds during view transformation, potentially compromising the quality of camera BEV representation. This gives rise to an extra need of calibrating such misalignment in network capacity. In this work, we address the aforementioned limitations in all previous solutions with a novel multi-modal interaction strategy. The key insight behind our approach is that we maintain two modality-specific feature representations and conduct representational and predictive interaction for maximally exploring their complementary benefits whilst preserving their respective strengths. We present a novel modality interaction framework, dubbed Deep Interaction, for multi-modal (3D point clouds and 2D multi-camera images) 3D object detection. In contrast to all prior arts, we learn two representations specific for 3D Li DAR and 2D image modalities respectively, whilst conducting multi-modal interaction through both model encoding and decoding. An overview of Deep Interaction is shown in Figure 1(b). It consists of two main components: An encoder with multi-modal representational interaction (Section 3.1), and a decoder with multi-modal predictive interaction (Section 3.2). 3.1 Encoder: Multi-modal representational interaction Unlike conventional modality fusion strategy that often aggregates multi-modal inputs into a hybrid feature map, individual per-modality representations are learned and maintained via multi-modal representational interaction within our encoder. Specifically, our encoder is formulated as a multiinput-multi-output (MIMO) structure: Taking as input two modality-specific scene representations which are independently extracted by Li DAR and image backbones, and producing two refined representations as output. Overall, it is composed by stacking multiple encoder layers each with (I) multi-modal representational interaction (MMRI), (II) intra-modal representational learning (IML), and (III) representational integration. (I) Multi-modal representational interaction (MMRI) Each encoder layer takes the representations of two modalities, i.e., the image perspective representation hc and the Li DAR BEV representation hp, as input. Our multi-modal representational interaction aims to exchange the neighboring context in a bilateral cross-modal manner, as shown in Figure 2. It consists of two steps: (i) Cross-modal correspondence mapping and sampling To define cross-modality adjacency, we first need to build the pixel-to-pixel(s) correspondence between the representations hp and hc. To that end, we construct dense mappings between the image coordinate frame c and the BEV coordinate frame p (Mp c and Mc p). From image to Li DAR BEV coordinate frame Mc p : R2 2R2 (Figure 2(a)): We first project each point (x, y, z) in 3D point cloud to multi-camera images to form a sparse depth map dsparse, followed by depth completion [20] leading to a dense depth map ddense. We further utilize ddense to back-project each pixel in the image space into the 3D point space. This results in the corresponding 3D coordinate (x, y, z), given an image pixel (i, j) with depth d[i,j] dense. Next, (x, y) is used to locate the corresponding BEV coordinate (ip, jp). We denote the above mapping as T(i, j) = (ip, jp). We obtain this correspondence via (2k + 1) (2k + 1) sized neighbor sampling as Mc p(i, j) = {T(i + i, j + j)| i, j [ k, +k]}. From Li DAR BEV to image coordinate frame Mp c : R2 2R2 (Figure 2(b)): Given a coordinate (ip, jp) in BEV, we first obtain the Li DAR points {(x, y, z)} within the pillar corresponding to (ip, jp). Then we project these 3D points into camera image coordinate frame {(i, j)} according to the camera intrinsics and extrinsics. This correspondence is obtained as: Mp c(ip, jp) = {(i, j)}. (ii) Attention-based feature interaction For an image feature point as query q = h[ic,jc] c , its crossmodality neighbors Nq, denoted as Nq = h[Mc p(ic,jc)] p , are used as the key k and value v for cross-attention learning: fϕc p (hc, hp)[i,j] = X k,v Nq softmax qk where h[i,j] denotes indexing the element at location (i, j) on the 2D representation h. This is image-to-Li DAR representational interaction. The other way around, given a Li DAR BEV feature point as query q = h[ip,jp] p , we similarly obtain its cross-modality neighbors as Nq = h[Mp c(ip,jp)] c . The same process (Eq. (1)) is applied for realizing Li DAR-to-image representational interaction fϕp c (hc, hp). (II) Intra-modal representational learning (IML) Concurrently, we conduct intra-modal representational learning complementary to multi-modal interaction. The same local attention as defined in Eq. (1) is consistently applied. For either modality, we use a k k grid neighborhood as the key and value. Formally, we denote fϕc c(hc) for image representation and fϕp p(hp) for Li DAR representation. (III) Representational integration Each layer ends up by integrating the two outputs per modality: h p = FFN(Concat(FFN(Concat(hp p p , hc p p )), hp)), h c = FFN(Concat(FFN(Concat(hc c c , hp c c )), hc)), (2) where FFN specifies a feed-forward network, and Concat denotes element-wise concatenation. 3.2 Decoder: Multi-modal predictive interaction Beyond considering the multi-modal interaction at the representation-level, we further introduce a decoder with multi-modal predictive interaction (MMPI) to maximize the complementary effects in prediction. As depicted in Figure 3(a), our core idea is to enhance the 3D object detection of one modality conditioned on the other modality. In particular, the decoder is built by stacking multiple multi-modal predictive interaction layers, within which predictive interactions are formulated in an alternative and progressive manner. Similar to the decoder of DETR [5], we cast the 3D object detection as a set prediction problem. Here, we define a set of N object queries {Qn}N n=1 and the resulting N object predictions {(bn, cn)}N n=1, where bn and cn denote the predicted bounding box and category for the n-th prediction. Ro I Feats. ... Mul-Norm-Re LU (b) Multi-modal predictive interaction layer (MMPI) self-attention Linear layer MMPI-Li DAR ... ... ... (a) Predictive interaction decoder Figure 3: Illustration of our multi-modal predictive interaction. Our predictive interaction decoder (a) generates predictions via (b) progressively interacting with two modality-specific representations. Multi-modal predictive interaction layer (MMPI) For the l-th decoding layer, the set prediction is computed by taking the object queries n Q(l 1) n o N n=1 and the bounding box predictions n b(l 1) n o N n=1 from previous layer as inputs and enabling interaction with the intensified image h p or Li DAR h c representations (h c if l is odd, h p if l is even). We formulate the multi-modal predictive interaction layer (Figure 3(b)) for specific modality as follows: (I) Multi-modal predictive interaction on image representation (MMPI-image) Taking as input 3D object proposals n b(l 1) n o N n=1 and object queries n Q(l 1) n o N n=1 generated by the previous layer, this layer leverages the image representation h c for further prediction refinement. To integrate the previous predictions n b(l 1) n o N n=1, we first extract N Region of Interest (Ro I) [15] features {Rn}N n=1 from the image representation h c, where Rn RS S C is the extracted Ro I feature for the n-th query, (S S) is Ro I size, and C is the number of channels of Ro I features. Specifically, for each 3D bounding box, we project it onto image representation h c to get the 2D convex polygon and take the minimum axis-aligned circumscribed rectangle. We then design a multi-modal predictive interaction operator that maps n Q(l 1) n o N n=1 into the parameters of a series of 1 1 convolutions and then applies them consecutively to {Rn}N n=1; The resulted interactive representation is further used to obtain the updated object query Ql n N n=1. (II) Multi-modal predictive interaction on Li DAR representation (MMPI-Li DAR) This layer shares the same design as the above except that it takes as input Li DAR representation instead. With regards to the Ro I for Li DAR representation, we project the 3D bounding boxes from previous layer to the Li DAR BEV representation h p and take the minimum axis-aligned rectangle. It is worth mentioning that due to the scale of objects in autonomous driving scenarios is usually tiny in the BEV coordinate frame, we enlarge the scale of the 3D bounding box by 2 times. The shape of Ro I features cropped from the Li DAR BEV representation h p is also set to be S S C. Here C is the C is the number of channels of Ro I features as well as the height of BEV representation. The multi-modal predictive interaction layer on Li DAR representation is stacked on the above image counterpart. For object detection, a feed-forward network is appended on the Ql n N n=1 for each multi-modal predictive interaction layer to infer the locations, dimensions, orientations and velocities. During training, the matching cost and loss function as [1] are applied. 4 Experiments 4.1 Experimental setup Dataset We evaluate our approach on the nu Scenes dataset [3]. It provides point clouds from 32-beam Li DAR and images with a resolution of 1600 900 from 6 surrounding cameras. The total of 1000 scenes, where each sequence is roughly 20 seconds long and annotated every 0.5 second, is Table 1: Comparison with state-of-the-art methods on the nu Scenes test set. Metrics: m AP(%), NDS(%). L and C represent Li DAR and camera, respectively. denotes test-time augmentation is used. denotes that test-time augmentation and model ensemble both are applied for testing. Method Modality Backbones validation test Image Li DAR m AP NDS m AP NDS BEVDet4D [17] C Swin-Base - 42.1 54.5 45.1 56.9 BEVFormer [25] C V99 - - - 48.1 56.9 Ego3RT [31] C V99 - 47.8 53.4 42.5 47.9 Polar Former [19] C V99 - 50.0 56.2 49.3 57.2 Center Point [45] L - Voxel Net 59.6 66.8 60.3 67.3 Focals Conv [8] L - Voxel Net-Focals Conv 61.2 68.1 63.8 70.0 Transfusion-L [1] L - Voxel Net 65.1 70.1 65.5 70.2 Large Kernel3D [9] L - Voxel Net-Large Kernel3D 63.3 69.1 65.3 70.5 FUTR3D [7] L+C R101 Voxel Net 64.5 68.3 - - Point Augmenting [39] L+C DLA34 Voxel Net - - 66.8 71.0 MVP [46] L+C DLA34 Voxel Net 67.1 70.8 66.4 70.5 Auto Align V2 [10] L+C CSPNet Voxel Net 67.1 71.2 68.4 72.4 Trans Fusion [1] L+C R50 Voxel Net 67.5 71.3 68.9 71.6 BEVFusion [26] L+C Swin-Tiny Voxel Net 67.9 71.0 69.2 71.8 BEVFusion [30] L+C Swin-Tiny Voxel Net 68.5 71.4 70.2 72.9 Deep Interaction-base L+C R50 Voxel Net 69.9 72.6 70.8 73.4 Focals Conv-F [8] L+C R50 Voxel Net-Focals Conv 67.1 71.5 70.1 73.6 Large Kernel3D-F [9] L+C R50 Voxel Net-Large Kernel - - 71.1 74.2 Deep Interaction-large L+C Swin-Tiny Voxel Net 72.6 74.4 74.1 75.5 BEVFusion-e [30] L+C Swin-Tiny Voxel Net 73.7 74.9 75.0 76.1 Deep Interaction-e L+C Swin-Tiny Voxel Net 73.9 75.0 75.6 76.3 officially split into train/val/test set with 700/150/150 scenes. For the 3D object detection task, 1.4M objects in scenes are annotated with 3D bounding boxes and classified into 10 categories: car, truck, bus, trailer, construction vehicle, pedestrian, motorcycle, bicycle, barrier, and traffic cone. Metric Mean average precision (m AP) [12] and nu Scenes detection score (NDS) [3] are used as the evaluation metric of 3D detection performance. The final m AP is computed by averaging over the distance thresholds of 0.5m, 1m, 2m, 4m across 10 classes. NDS is a weighted average of m AP and other attribute metrics, including translation, scale, orientation, velocity, and other box attributes. 4.2 Implementation details Our implementation is based on the public code base mmdetection3d [11]. For the image branch backbone, we use a simple Res Net-50 [16] and initialize it from the instance segmentation model Cascade Mask R-CNN [4] pretrained on COCO [27] and then nu Image [3], which is same as Transfusion [1]. To save the computation cost, we rescale the input image to 1/2 of its original size before feeding into the network, and freeze the weights of image branch during training. The voxel size is set to (0.075m, 0.075m, 0.2m), and the detection range is set to [ 54m, 54m] for X and Y axis and [ 5m, 3m] for Z axis. The representational interaction encoder is composed by stacking two representational interaction layers. For the multi-modal predictive interaction decoder, we use 5 cascaded decoder layers. We set the query number to 200 for training and testing and use the same query initialization method as Transfusion [1]. The above configuration is termed Deep Interaction-base. We also adopt another two widely used settings for online submission, i.e., test-time augmentation (TTA) and model ensemble. In the following, we refer to the two settings as Deep Interaction-large and Deep Interaction-e respectively. In particular, Deep Interaction-large uses Swin-Tiny [29] as image backbone, and doubles the number of channel for each convolution block in Li DAR backbone. The voxel size of Deep Interaction-large LIDAR BEV Car Truck Construction vehicle Motorcycle Bicycle Pedestrian Traffic cone FRONT_RIGHT Bus Trailor Barrier Figure 4: Qualitative results on nu Scenes val set. In Li DAR BEV (right), green boxes are the ground-truth and blue boxes are the predictions. Best viewed when zooming in. is set to [0.5m, 0.5m, 0.2m]. Following the common practice, we use double flipping and rotation with yaw angles [0 , 6.25 , 12.5 ] for test-time augmentation. Deep Interaction-e ensembles multiple Deep Interaction-large models with input Li DAR BEV grid size between [0.5m, 0.5m] and [1.5m, 1.5m]. For data augmentation, following Trans Fusion [1] we adopt random rotation with a range of r [ π/4, π/4], random scaling with a factor of r [0.9, 1.1], random translation with standard deviation 0.5 in three axis, and random horizontal flipping. We also use the class-balanced resampling in CBGS [48] to balance the class distribution for nu Scenes. Following [1], we adopt a two stage training recipe. We take Trans Fusion-L [1] as our Li DAR-only baseline. We use Adam optimizer with one-cycle learning rate policy, with max learning rate 1 10 3, weight decay 0.01 and momentum 0.85 to 0.95, following CBGS [48]. Our Li DAR-only baseline is trained for 20 epochs and Li DAR-image fusion for 6 epochs with batch size of 16 using 8 NVIDIA V100 GPUs. 4.3 Comparison to the state of the art Performance We compare with state-of-the-art alternatives on the nu Scenes test set. As shown in Table 1, Deep Interaction achieves new state-of-the-art performance under all settings. The base variant without TTA and model ensemble, Deep Interaction-base, with a simple Res Net-50 image backbone, surpasses all the prior arts as well as the concurrent work BEVFusion [30] even with a Swin-Tiny image backbone. Deep Interaction-large beats the closest rival Large Kernel3DF [9] with the same TTA and test time augmentation (single model) by a considerable margin. Our ensemble version Deep Interaction-e achieves the first rank among all the solutions on the nu Scenes leaderboard. These results verify the performance advantages of our multi-modal interaction approach. Per-category results are shown in Appendix A.1. Qualitative results are provided in Figure 4 and Appendix A.4. Run time We compare inference speed tested on NVIDIA V100, A6000 GPUs and A100 separately. As shown in Table 2, our method achieves the best performance while running faster than alternative painting-based [39] and query-based [7] fusion approaches. This validates superior trade-off between detection performance and inference speed achieved by our method. As found in [39], feature extraction for multi-view high resolution camera images contributes the most of overall latency in a multi-modal 3D detector. Indeed, from Table 3(c) we observe that increasing the number of decoder layers only brings negligible extra latency, which concurs with the same conclusion. 4.4 Ablations on the decoder Multi-modal predictive interaction layer vs. DETR [5] decoder layer In Table 3(a) we evaluate the design of decoder layer by comparing our multi-modal predictive interaction (MMPI) with DETR [5] decoder layer. Note the DETR decoder layer means the conventional Transformer deocder layer is used to aggregate multi-modal information same as in Transfusion [1]. We further test a mixing design: using vanilla DETR decoder layer for aggregating features in Li DAR representation and our MMPI for aggregating features in image representation (second row). It is evident that MMPI Table 2: Run time comparison. If not specified, the performance and efficiency are evaluated on the nu Scenes val set. Method m AP(%) NDS(%) FPS(A100) FPS(A6000) FPS(V100) Point Augmenting [39] 66.8 (Test) 71.0 (Test) 1.4 2.8 2.3 FUTR3D [7] 64.2 68.0 4.5 2.3 1.8 Transfusion [1] 67.5 71.3 6.2 5.5 3.8 Deep Interaction 69.9 72.6 4.9 3.1 2.6 Table 3: Ablation studies on the decoder. The m AP and NDS are evaluated on the nu Scenes val set. Li DAR Image m AP NDS DETR [5] DETR [5] 68.6 71.6 DETR [5] MMPI 69.3 72.1 MMPI MMPI 69.9 72.6 (a) The type of decoder layer. Modality m AP NDS Fully Li DAR 69.2 72.2 Li DAR-image alternating 69.9 72.6 (b) Single vs. multiple representations. # of decoder layers m AP NDS FPS 1 (Li DAR-only) 65.1 70.1 8.7 2 69.5 72.3 2.8 3 69.7 72.5 2.7 4 69.8 72.5 2.7 5 69.9 72.6 2.6 6 69.7 72.1 2.5 (c) The number of decoder layers. Train Inference m AP NDS 200 69.9 72.6 200 300 70.1 72.7 400 70.0 72.6 200 69.7 72.5 300 300 69.9 72.6 400 70.0 72.6 (d) The number of queries. is significantly superior over DETR by improving 1.3% m AP and 1.0% NDS, with combinational flexibility in design. How many representations/modalities? In Table 3(b), we evaluate the effect of using different numbers of representations/modalities in decoding. We compare our MMPI using both representations in an alternating manner with a variant using Li DAR representation in all decoder layers. It is observed that using both representations is beneficial, verifying our design consideration. Number of decoder layers As shown in Table 3(c), increasing the number of decoder layers up to 5 layers can consistently improve the performance whilst introducing negligible latency. Li DAR-only denotes the Transfusion-L [1] baseline. Number of queries Since our query embeddings are initialized in a non-parametric and inputdependent manner as in [1], the number of queries are adjustable during inference. In Table 3(d), we evaluate different combinations of query numbers for training and test. Overall, the performance is stable over different choices with 200/300 for training/test as the best setting. 4.5 Ablations on the encoder Multi-modal representational interaction vs. fusion To precisely demonstrate the superiority of our multi-modal representational interaction, we compare a naive version of our Deep Interaction with conventional representational fusion strategy as presented in Transfusion [1]. We limit our Deep Interaction using the same number of encoder and decoder layers as [1] for a fair comparison. Table 4(c) shows that our representational interaction is clearly more effective. Encoder design We ablate the design of our encoder with focus on multi-modal representational interaction (MMRI) and intra-modal representational learning (IML). We have a couple of observations from Table 4(a): (1) Our MMRI can significantly improve the performance over IML; (2) MMRI and IML can work well together for further performance gain. As seen from Table 4(b), stacking our encoder layers for iterative MMRI is beneficial. Objects in the image Before interaction After interaction Figure 5: Illustrations of the heatmaps predicted from BEV representations before and after representational interactions. All samples are from the nu Scenes val split. (a) A case of the occluded tiny object. (b) A case of small object at long distance. (c) An example of two adjacent barriers which are connected together in Li DAR point clouds and thus it is difficult to have their instance-level understanding without the help of visual clues. Qualitative results of representational interaction To gain more insight about the effect of our multi-modal representational interaction (MMRI), we visualize the heatmaps of challenging cases. We observe from Figure 5 that without the assistance of MMRI, some objects cannot be detected when using Li DAR only (the middle column). The locations of these objects are highlighted by red circles in the heatmap and white bounding boxes in the RGB image below. Table 4: Ablation studies on the representational interaction encoder. The m AP and NDS are evaluated on the nu Scenes val set. IML MMRI m AP NDS 68.1 71.9 69.5 72.5 69.9 72.6 (a) Encoder design. IML: Intra-modal learning; MMRI: Multi-modal representational interaction. # of encoder layers m AP NDS w/o 66.4 70.7 1 67.7 71.2 2 69.9 72.6 (b) The number of encoder layers. Method m AP NDS Representational fusion 67.5 71.3 Representational interaction (Ours) 68.7 71.9 (c) Our representational interaction vs. conventional representational fusion (e.g., Transfusion [1]). Table 5: Evaluation on different Li DAR backbones. The m AP and NDS are evaluated on the nu Scenes val set. Methods Modility m AP NDS Point Pillars [22] L 46.2 59.1 +Transfusion-L [1] L 54.5 62.7 +Transfusion [1] L+C 58.3 64.5 +Deep Interaction L+C 60.0 65.6 (a) Comparison between pillar-based methods. Methods Modility m AP NDS Voxel Net [47] L 52.6 63.0 +Transfusion-L [1] L 65.1 70.1 +Transfusion [1] L+C 67.5 71.3 +Deep Interaction L+C 69.9 72.6 (b) Comparison between voxel-based methods. Table 6: Comparison with the Li DAR-only baseline Transfusion-L [1] on nu Scenes val split. Method m AP NDS Car Truck C.V. Bus T.L. B.R. M.T. Bike Ped. T.C. Transfusion-L [1] 65.1 70.1 86.5 59.6 25.4 74.4 42.2 74.1 72.1 56.0 86.6 74.1 Deep Interaction 69.9 72.6 88.5 64.4 30.1 79.2 44.6 76.4 79.0 67.8 88.9 80.0 Concretely, the sample (a) suggests that camera information is helpful to recover partially obscured tiny objects with sparse observation in the point cloud. The sample (b) shows a representative case where some distant objects can be recognized successfully due to the help of visual information. From the sample (c), we observe that the centers of some barriers yield a more distinct activation in the heatmap after representational interaction. This is probably due to that it is too difficult to locate the boundaries of several consecutive barriers from Li DAR point clouds only. 4.6 Ablation on Li DAR backbones We examine the generality of our framework with two different Li DAR backbones: Point Pillars [22] and Voxel Net [47]. For Point Pillars, we set the voxel size to (0.2m, 0.2m) while keeping the remaining settings same as Deep Interaction-base. For fair comparison, we use the same number of queries as Trans Fusion [1]. As shown in Table 5, due to the proposed multi-modal interaction strategy, Deep Interaction exhibits consistent improvements over Li DAR-only baseline using either backbone (by 5.5% m AP for voxel-based backbone, and 4.4% m AP for pillar-based backbone). This manifests the generality of our Deep Interaction across varying point cloud encoder. 4.7 Performance breakdown To demonstrate more fine-grained performance analysis, we compare our Deep Interaction and our Li DAR-only baseline Transfusion [1] at the category level in terms of m AP on nu Scenes val set. We can see from Table 6 that our fusion approach achieves remarkable improvements on all the categories, especially on tiny or rare object categories (+11.8% m AP for bicycle, +6.9% m AP for motorcycle, and +5.9% m AP for traffic cone). 5 Conclusion In this work, we have presented a novel 3D object detection method Deep Interaction for exploring the intrinsic multi-modal complementary nature. This key idea is to maintain two modality-specific representations and establish interactions between them for both representation learning and predictive decoding. This strategy is designed particularly to resolve the fundamental limitation of existing unilateral fusion approaches that image representation are insufficiently exploited due to their auxiliary-source role treatment. Extensive experiments demonstrate our proposed Deep Interaction yields new state of the art on the nu Scenes benchmark dataset and achieves the first position at the highly competitive nu Scenes 3D object detection leaderboard. Acknowledgement This work was supported in part by National Natural Science Foundation of China (Grant No. 6210020439), Lingang Laboratory (Grant No. LG-QS-202202-07), Natural Science Foundation of Shanghai (Grant No. 22ZR1407500), Science and Technology Innovation 2030 - Brain Science and Brain-Inspired Intelligence Project (Grant No. 2021ZD0200204) and Shanghai Municipal Science and Technology Major Project (Grant No. 2018SHZDZX01). [1] Xuyang Bai, Zeyu Hu, Xinge Zhu, Qingqiu Huang, Yilun Chen, Hongbo Fu, and Chiew-Lan Tai. Transfusion: Robust lidar-camera fusion for 3d object detection with transformers. In CVPR, 2022. 2, 5, 6, 7, 8, 9, 10 [2] Alex Bewley, Pei Sun, Thomas Mensink, Dragomir Anguelov, and Cristian Sminchisescu. Range conditioned dilated convolutions for scale invariant 3d object detection. ar Xiv preprint, 2020. 2 [3] Holger Caesar, Varun Bankiti, Alex H Lang, Sourabh Vora, Venice Erin Liong, Qiang Xu, Anush Krishnan, Yu Pan, Giancarlo Baldan, and Oscar Beijbom. nuscenes: A multimodal dataset for autonomous driving. In CVPR, 2020. 5, 6 [4] Zhaowei Cai and Nuno Vasconcelos. Cascade r-cnn: high quality object detection and instance segmentation. IEEE TPAMI, 2019. 6 [5] Nicolas Carion, Francisco Massa, Gabriel Synnaeve, Nicolas Usunier, Alexander Kirillov, and Sergey Zagoruyko. End-to-end object detection with transformers. In ECCV, 2020. 2, 4, 7, 8 [6] Yuning Chai, Pei Sun, Jiquan Ngiam, Weiyue Wang, Benjamin Caine, Vijay Vasudevan, Xiao Zhang, and Drago Anguelov. To the point: Efficient 3d object detection in the range image with graph convolution kernels. In CVPR, 2021. 2 [7] Xuanyao Chen, Tianyuan Zhang, Yue Wang, Yilun Wang, and Hang Zhao. Futr3d: A unified sensor fusion framework for 3d detection. ar Xiv preprint, 2022. 2, 6, 7, 8 [8] Yukang Chen, Yanwei Li, Xiangyu Zhang, Jian Sun, and Jiaya Jia. Focal sparse convolutional networks for 3d object detection. ar Xiv preprint, 2022. 6 [9] Yukang Chen, Jianhui Liu, Xiaojuan Qi, Xiangyu Zhang, Jian Sun, and Jiaya Jia. Scaling up kernels in 3d cnns. ar Xiv preprint, 2022. 6, 7 [10] Zehui Chen, Zhenyu Li, Shiquan Zhang, Liangji Fang, Qinghong Jiang, Feng Zhao, Bolei Zhou, and Hang Zhao. Autoalign: Pixel-instance feature aggregation for multi-modal 3d object detection. ar Xiv preprint, 2022. 1, 6 [11] MMDetection3D Contributors. MMDetection3D: Open MMLab next-generation platform for general 3D object detection. https://github.com/open-mmlab/mmdetection3d, 2020. 6 [12] Mark Everingham, Luc Van Gool, Christopher KI Williams, John Winn, and Andrew Zisserman. The pascal visual object classes (voc) challenge. IJCV, 2010. 6 [13] Lue Fan, Ziqi Pang, Tianyuan Zhang, Yu-Xiong Wang, Hang Zhao, Feng Wang, Naiyan Wang, and Zhaoxiang Zhang. Embracing single stride 3d object detector with sparse transformer. ar Xiv preprint, 2021. 2 [14] Lue Fan, Xuan Xiong, Feng Wang, Naiyan Wang, and Zhao Xiang Zhang. Rangedet: In defense of range view for lidar-based 3d object detection. In ICCV, 2021. 2 [15] Kaiming He, Georgia Gkioxari, Piotr Dollár, and Ross Girshick. Mask r-cnn. In ICCV, 2017. 5 [16] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, 2016. 6 [17] Junjie Huang, Guan Huang, Zheng Zhu, and Dalong Du. Bevdet: High-performance multicamera 3d object detection in bird-eye-view. ar Xiv preprint, 2021. 2, 6 [18] Tengteng Huang, Zhe Liu, Xiwu Chen, and Xiang Bai. Epnet: Enhancing point features with image semantics for 3d object detection. In ECCV, 2020. 2 [19] Yanqin Jiang, Li Zhang, Zhenwei Miao, Xiatian Zhu, Jin Gao, Weiming Hu, and Yu-Gang Jiang. Polarformer: Multi-camera 3d object detection with polar transformers. ar Xiv preprint, 2022. 2, 6 [20] Jason Ku, Ali Harakeh, and Steven L Waslander. In defense of classical image processing: Fast depth completion on the cpu. In CRV, 2018. 4 [21] Jason Ku, Melissa Mozifian, Jungwook Lee, Ali Harakeh, and Steven L Waslander. Joint 3d proposal generation and object detection from view aggregation. In IROS, 2018. 2 [22] Alex H Lang, Sourabh Vora, Holger Caesar, Lubing Zhou, Jiong Yang, and Oscar Beijbom. Pointpillars: Fast encoders for object detection from point clouds. In CVPR, 2019. 2, 10 [23] Yanwei Li, Xiaojuan Qi, Yukang Chen, Liwei Wang, Zeming Li, Jian Sun, and Jiaya Jia. Voxel field fusion for 3d object detection. In CVPR, 2022. 1 [24] Yingwei Li, Adams Wei Yu, Tianjian Meng, Ben Caine, Jiquan Ngiam, Daiyi Peng, Junyang Shen, Bo Wu, Yifeng Lu, Denny Zhou, et al. Deepfusion: Lidar-camera deep fusion for multi-modal 3d object detection. ar Xiv preprint, 2022. 1 [25] Zhiqi Li, Wenhai Wang, Hongyang Li, Enze Xie, Chonghao Sima, Tong Lu, Qiao Yu, and Jifeng Dai. Bevformer: Learning bird s-eye-view representation from multi-camera images via spatiotemporal transformers. ar Xiv preprint, 2022. 2, 6 [26] Tingting Liang, Hongwei Xie, Kaicheng Yu, Zhongyu Xia, Zhiwei Lin, Yongtao Wang, Tao Tang, Bing Wang, and Zhi Tang. BEVFusion: A Simple and Robust Li DAR-Camera Fusion Framework. ar Xiv preprint, 2022. 1, 3, 6 [27] T. Y. Lin, M. Maire, S. Belongie, J. Hays, and C. L. Zitnick. Microsoft coco: Common objects in context. In ECCV, 2014. 6 [28] Yingfei Liu, Tiancai Wang, Xiangyu Zhang, and Jian Sun. Petr: Position embedding transformation for multi-view 3d object detection. ar Xiv preprint, 2022. 2 [29] Ze Liu, Yutong Lin, Yue Cao, Han Hu, Yixuan Wei, Zheng Zhang, Stephen Lin, and Baining Guo. Swin transformer: Hierarchical vision transformer using shifted windows. In ICCV, 2021. 6 [30] Zhijian Liu, Haotian Tang, Alexander Amini, Xingyu Yang, Huizi Mao, Daniela Rus, and Song Han. Bevfusion: Multi-task multi-sensor fusion with unified bird s-eye view representation. ar Xiv preprint, 2022. 1, 3, 6, 7 [31] Jiachen Lu, Zheyuan Zhou, Xiatian Zhu, Hang Xu, and Li Zhang. Learning ego 3d representation as ray tracing. In ECCV, 2022. 2, 6 [32] Jiageng Mao, Yujing Xue, Minzhe Niu, Haoyue Bai, Jiashi Feng, Xiaodan Liang, Hang Xu, and Chunjing Xu. Voxel transformer for 3d object detection. In CVPR, 2021. 2 [33] Ishan Misra, Rohit Girdhar, and Armand Joulin. An End-to-End Transformer Model for 3D Object Detection. In ICCV, 2021. 2 [34] Jonah Philion and Sanja Fidler. Lift, splat, shoot: Encoding images from arbitrary camera rigs by implicitly unprojecting to 3d. In ECCV, 2020. 3 [35] AJ Piergiovanni, Vincent Casser, Michael S Ryoo, and Anelia Angelova. 4d-net for learned multi-modal alignment. In CVPR, 2021. 2 [36] Charles R Qi, Xinlei Chen, Or Litany, and Leonidas J Guibas. Imvotenet: Boosting 3d object detection in point clouds with image votes. In CVPR, 2020. 2 [37] Cody Reading, Ali Harakeh, Julia Chae, and Steven L. Waslander. Categorical depth distributionnetwork for monocular 3d object detection. In CVPR, 2021. 2 [38] Sourabh Vora, Alex H Lang, Bassam Helou, and Oscar Beijbom. Pointpainting: Sequential fusion for 3d object detection. In CVPR, 2020. 1, 2 [39] Chunwei Wang, Chao Ma, Ming Zhu, and Xiaokang Yang. Pointaugmenting: Cross-modal augmentation for 3d object detection. In CVPR, 2021. 1, 2, 6, 7, 8 [40] Yan Wang, Wei-Lun Chao, Divyansh Garg, Bharath Hariharan, Mark Campbell, and Kilian Weinberger. Pseudo-lidar from visual depth estimation: Bridging the gap in 3d object detection for autonomous driving. In CVPR, 2019. 2 [41] Yue Wang, Vitor Campagnolo Guizilini, Tianyuan Zhang, Yilun Wang, Hang Zhao, and Justin Solomon. Detr3d: 3d object detection from multi-view images via 3d-to-2d queries. In Co RL, 2022. 2 [42] Yue Wang and Justin M. Solomon. Object dgcnn: 3d object detection using dynamic graphs. In Neur IPS, 2021. 2 [43] Shaoqing Xu, Dingfu Zhou, Jin Fang, Junbo Yin, Bin Zhou, and Liangjun Zhang. Fusion Painting: Multimodal fusion with adaptive attention for 3d object detection. ITSC, 2021. 1 [44] Yan Yan, Yuxing Mao, and Bo Li. Second: Sparsely embedded convolutional detection. Sensors, 2018. 2 [45] Tianwei Yin, Xingyi Zhou, and Philipp Krahenbuhl. Center-based 3d object detection and tracking. In CVPR, 2021. 6 [46] Tianwei Yin, Xingyi Zhou, and Philipp Krähenbühl. Multimodal virtual point 3d detection. Neur IPS, 2021. 1, 2, 6 [47] Yin Zhou and Oncel Tuzel. Voxelnet: End-to-end learning for point cloud based 3d object detection. In CVPR, 2018. 2, 10 [48] Benjin Zhu, Zhengkai Jiang, Xiangxin Zhou, Zeming Li, and Gang Yu. Class-balanced grouping and sampling for point cloud 3d object detection. ar Xiv preprint, 2019. 7 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] See Section 4. (b) Did you describe the limitations of your work? [Yes] See supplementary material. (c) Did you discuss any potential negative societal impacts of your work? [Yes] See supplementary material. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] See Section 4. (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] See Section 4. 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [No] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]