# how2comm_communicationefficient_and_collaborationpragmatic_multiagent_perception__040121c9.pdf How2comm: Communication-Efficient and Collaboration-Pragmatic Multi-Agent Perception Dingkang Yang1,2 Kun Yang1 Yuzheng Wang1 Jing Liu1 Zhi Xu1 Rongbin Yin5 Peng Zhai1,2 Lihua Zhang1,2,3,4 1Academy for Engineering and Technology, Fudan University 2Cognition and Intelligent Technology Laboratory (CIT Lab) 3Engineering Research Center of AI and Robotics, Ministry of Education 4AI and Unmanned Systems Engineering Research Center of Jilin Province 5FAW (Nanjing) Technology Development Company Ltd {dkyang20, kungyang20, pzhai, lihuazhang}@fudan.edu.cn Multi-agent collaborative perception has recently received widespread attention as an emerging application in driving scenarios. Despite the advancements in previous efforts, challenges remain due to various dilemmas in the perception procedure, including communication redundancy, transmission delay, and collaboration heterogeneity. To tackle these issues, we propose How2comm, a collaborative perception framework that seeks a trade-off between perception performance and communication bandwidth. Our novelties lie in three aspects. First, we devise a mutual information-aware communication mechanism to maximally sustain the informative features shared by collaborators. The spatial-channel filtering is adopted to perform effective feature sparsification for efficient communication. Second, we present a flow-guided delay compensation strategy to predict future characteristics from collaborators and eliminate feature misalignment due to temporal asynchrony. Ultimately, a pragmatic collaboration transformer is introduced to integrate holistic spatial semantics and temporal context clues among agents. Our framework is thoroughly evaluated on several Li DAR-based collaborative detection datasets in real-world and simulated scenarios. Comprehensive experiments demonstrate the superiority of How2comm and the effectiveness of all its vital components. The code will be released at https://github.com/ydk122024/How2comm. 1 Introduction Precise perception of complex and changeable driving environments is essential to ensure the safety and reliability of intelligent agents [25, 46], e.g., autonomous vehicles (AVs). With the emergence of learning-based technologies, remarkable single-agent perception systems are extensively explored for several in-vehicle tasks, such as instance segmentation [19, 58] and object detection [26, 49]. Nevertheless, single-agent perception suffers from various shortcomings due to the isolated view, such as unavoidable occlusions [55], restricted detection ranges [56], and sparse sensor observations [57]. Recently, multi-agent collaborative perception [39, 54] has provided promising solutions as an emerging application for vehicle-to-vehicle/everything (V2V/X) communications. The impressive studies [7, 14, 16, 17, 31, 32, 33, 38, 40, 51] have progressively presented to aggregate valuable information and complementary perspectives among on-road agents, resulting in a more Equal contributions . The two first authors thank Runsheng Xu for providing constructive suggestions. *Corresponding authors. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). (c) Infrastructure side image (d) Fused point cloud (a) Aligned point cloud fusion (b) Non-aligned point cloud fusion Figure 1: (a) and (b) show the point cloud fusion results in the absence and presence of transmission delay, respectively. (c) and (d) show the vanilla image and fused point cloud of the collaborative perception scene containing an ego vehicle (red circle) and an infrastructure, respectively. precise perception. Despite recent advancements, challenges remain due to unpreventable dilemmas, including communication redundancy, transmission delay, and collaboration heterogeneity. Communication Redundancy. The dominant patterns for reducing communication overhead are summarized as feature compression [14, 33, 40] and spatial filtering [7, 8]. The former assumes that agents share all spatial areas indiscriminately, which dramatically wastes bandwidth. The latter overly relies on confidence maps to highlight gullible locations and fails to consider spatially holistic information. Moreover, these methods invariably cause losses of transmitted valuable information. Transmission Delay. Figures 1(a)&(b) present the point cloud fusion results from an ego vehicle and an infrastructure in the time-synchronous and time-asynchronous cases, respectively. The inevitable transmission delay causes position misalignment of fast-moving objects within the green circles, potentially harming subsequent collaboration performance. Although several delay-aware strategies [13, 40, 53] are proposed to tackle this issue, they either suffer from performance bottlenecks [13, 40] or introduce massive computation costs [53], leading to sub-optimal solutions. Collaboration Heterogeneity. Figures 1(c)&(d) show the typical collaboration scenario involving two agents and the fused point cloud. Intuitively, Li DAR configuration discrepancies (e.g., different Li DAR densities, distributions, reflectivities, and noise interference) across agents potentially cause collaboration heterogeneity within the feature space. In this case, the orange box contains the common perception region of both agents, which facilitates bridging the feature-level gap caused by sensor configuration discrepancies [37, 40]. The magenta box contains the exclusive perception region of the infrastructure, which provides complementary information for the ego vehicle and compensates for the occluded view. Fusing valuable spatial semantics from these two perception regions facilitates comprehensive and pragmatic perception. However, most previous methods [14, 16, 32, 33, 40] integrate collaborator-shared features via per-agent/location message fusion to enhance ego representations, whose collaboration processes could be vulnerable since the advantages of distinct perception regions from heterogeneous agents are not considered holistically. Moreover, the current single-frame perception paradigm faces the challenges of 3D point cloud sparsity and localization errors, increasing the difficulty of building a robust multi-agent perception system. Motivated by the above observations, we propose How2comm, an end-to-end collaborative perception framework to address the existing issues jointly. Through three novel components, How2comm advances towards a reasonable trade-off between perception performance and communication bandwidth. Specifically, (i) we first design a mutual information-aware communication mechanism to maximally preserve the beneficial semantics from vanilla characteristics in the transmitted messages of collaborators. In this case, spatial-channel message filtering is introduced to determine how to use less bandwidth for efficient communication. (ii) Second, we present a flow-guided delay compensation strategy to predict the future features of collaborators by mining contextual dependencies in sequential frames. Our ingenious strategy determines how to dynamically compensate for the delay s impact and explicitly accomplish temporal alignment. (iii) Furthermore, we construct a spatio-temporal collaboration transformer (STCFormer) module to integrate perceptually comprehensive information from collaborators and temporally valuable clues among agents. Our unified transformer structure determines how to achieve pragmatic collaboration, contributing to a more robust collaborative perception against localization errors and feature discrepancies. How2comm is systematically evaluated on various collaborative 3D object detection datasets, including DAIR-V2X [52], V2XSet [40], and OPV2V [41]. Quantitative experiments demonstrate that our framework significantly outperforms previous state-of-the-art (SOTA) methods under the bandwidth-limited noisy setting. Systematic analyses confirm the robustness of How2comm against distinct collaboration noises. Shared Encoder Message Filtering Flow Generator Message Packing Message Delivery Detection Decoders Spatio-Temporal Collaboration Transformer (STCFormer) Ego Vehicle Infrastructure Temporal Context Projected Li DAR Projected Li DAR Figure 2: How2comm overview. The projected features of all agents are obtained via a shared encoder. Upon receiving requests {Rt0 i,s, Rt0 i,c} from the ego vehicle, the collaborator (i.e., infrastructure) shares the sparse feature F t0 j , feature flow Ot0 j , and scale matrix St0 j via the message filtering and flow generator. After that, the ego vehicle predicts the future feature Zt j and adopts the importance maps {It i, It j} to get the exclusive and common maps, which decouple Zt j into Zt j,e and Zt j,c. Finally, our STCFormer fuses the temporal context and decoupled spatial features to output Ft i used for detection. 2 Related Work Multi-Agent Communication. Benefiting from rapid advances in learning-based technologies [12, 21, 22, 23, 24, 42, 43, 44, 45, 47, 48, 50], communication has played an essential role in constructing robust and stable multi-agent systems. Although early works provided heuristic insights into information sharing among different agents through predefined protocols and centralized paradigms [4, 28, 29], these efforts are typically difficult to generalize into challenging scenarios. Recently, several learning-driven communication strategies have been proposed to accommodate diverse scenario applications. For instance, Vain [6] utilized an attentional neural structure to specify what information needs to be shared in agent interactions. ATOC [9] introduced the recurrent unit to decide whom the agent communicates with by receiving local observations and action intentions from other agents. Tar MAC [2] designed a reinforcement learning-oriented architecture to learn communication from task-specific rewards. In comparison, we focus on Li DAR-based collaborative 3D object detection tasks. For more challenging driving scenarios, we design the mutual information supervision and attention-guiding mechanism to achieve efficient communication across agents. Collaborative Perception. Collaborative perception is only in its infancy as a promising application of multi-agent systems. Several impressive approaches have been designed to facilitate the overall perception performance of AVs. The mainstream works [7, 14, 16, 33, 37, 38, 40] followed the intermediate collaboration pattern to balance average precision and bandwidth overhead. Specifically, When2com [16] introduced a handshake mechanism to determine when to communicate with collaborators. V2VNet [33] employed a fully connected graph network to aggregate feature representations shared by agents. After that, Disco Net [14] proposed a knowledge distillation framework to supervise the intermediate collaboration through an early collaboration of full views. V2X-Vi T [40] designed distinct attention interactions to facilitate adaptive information fusion among heterogeneous agents. Where2comm [7] aimed to transmit perceptually critical information via sparse spatial confidence maps. However, these methods invariably ignore valuable historical clues and lead to sub-optimal solutions. In this paper, we propose a novel collaboration transformer to jointly capture spatial semantics and temporal dynamics among agents, resulting in a more pragmatic collaboration. 3 Methodology 3.1 Problem Formulation In this paper, we seek to develop a communication-efficient and collaboration-pragmatic multi-agent system to enhance the perception ability of the ego agent. Figure 2 illustrates the proposed system framework, which accommodates different agents (e.g., AVs and infrastructures). Consider N agents in a driving scene, let Xt i be the local point cloud observation of the i-th agent (i.e., ego agent) and Yi be the corresponding ground-truth supervision. The objective of How2comm is to maximize the Li DAR-based 3D detection performance ℜ( ) under a total communication budget B: ℜ(B) = arg max θ, Fj i ð(Ψθ(Xt i, { F t0 j }N j=1), Yi), s.t. X j | F t0 j | B, (1) where ð( , ) denotes the perception evaluation metric and Ψθ is the perception system parameterized by θ. F t0 j is the message transmitted from the j-th agent to the i-th agent at time delay τ-aware moment t0, where t0 = t τ. The remainder of Section 3 details the major components. 3.2 Metadata Conversion and Feature Extraction In the initial stage of collaboration, we build a communication graph [14, 40] where one agent is selected as the ego agent and the other connected agents act as collaborators. Upon receiving the broadcast metadata (e.g., poses and extrinsic) from the ego agent, the collaborators project their local observations to the ego agent s coordinate system. Moreover, ego-motion compensation [35] synchronizes each agent s historical frames. The shared Point Pillar [11] encoder fenc( ) converts the point cloud of the i-th agent at timestamp t into the bird s-eye-view (BEV) features as F t i = fenc(Xt i) RH W C, where H, W, C denote height, width, and channel, respectively. 3.3 Mutual Information-aware Communication Previous attempts to reduce the required transmission bandwidth relied heavily on autoencoders [14, 33, 40] or confidence maps [7, 8], which are one-sided as they only consider information compression over spatial positions or channels. To this end, we design a mutual information-aware communication (MIC) mechanism to select the most informative messages from space and channels to save precious bandwidth. MIC consists of two core parts as follows. Spatial-channel Message Filtering. Stemming from solid evidence in signal picking, we first introduce the CBAM [34]-like spatial-channel attention queries to assist each agent in sharing their salient features. The spatial query At0 i,s = σ(ω3 3[φa(F t0 i ); φm(F t0 i )]) RH W 1 reflects what spatial locations on delayed feature F t0 i are informative, where [ ; ] is the concatenation, σ is the sigmoid activation, φa/m( ) denote average and max pooling functions, and ω3 3( ) is the 2D 3 3 convolution operation. The channel query At0 i,c = σ(ω1 1(φa(F t0 i )) + ω1 1(φm(F t0 i ))) R1 1 C reflects which channels in F t0 i are semantically meaningful. φa/m( ) in the spatial and channel queries are applied to the channel and spatial dimensions, respectively. Then, the ego agent indicates the supplementary messages required to improve local perception performance by broadcasting request queries Rt0 i,s/Rt0 i,c = 1 At0 i,s/At0 i,c. The j-th collaborator then aggregates the requests with its attention queries to obtain a spatial-channel binary message filtering matrix as follows: Mt0 j = fsel(ω1 1[Rt0 i,s; At0 j,s] ω1 1[Rt0 i,c; At0 j,c]) {0, 1}H W C , (2) where fsel( ) is a threshold-based selection function and is the element-wise multiplication. Ultimately, the selected feature map is obtained as F t0 j = F t0 j Mt0 j , which provides spatialchannel sparse, yet perceptually critical information. Mutual Information Maximization Supervision. Most existing works ignore the potential loss of valuable information due to feature compression. To overcome this dilemma, we maximally sustain the local critical semantics in the corresponding vanilla feature F t0 j on the selected regions of the transmitted features F t0 j by mutual information estimation. Since we only focus on maximizing the mutual information rather than getting its precise value, a stable estimator [5] is utilized to build the objective supervision based on the Jensen-Shannon divergence. Formally, the mutual information between two random variables X and Z is estimated as follows: ˆI(JSD) ϱ (X, Z) = Ep(x,z)[ log(1 + e Tϱ(x,z))] Ep(x)p(z)[log(1 + e Tϱ(x,z))], (3) where Tϱ : X Z R is a statistics network parameterized by ϱ. In our case, the mutual information supervision of all collaborators within the communication link is defined as follows: Lmul = 1 N 1 j {1,...,N},j =i ˆI(JSD) ϱ (F t0 j , F t0 j ). (4) Time Embedding Temporal Cross-Attention Feed Forward Feed Forward Adaptive Late Fusion Spatial Attention Spatial Attention 2D Conv 2D Conv 2D Conv Global Pooling Target Query Cross Attention Position Embedding 1D Conv Global Pooling (a) (b) (c) Figure 3: (a) The overall architecture of the proposed STCFormer. (b) and (c) show the structure of the temporal cross-attention (TCA) and exclusive spatial attention (ESA) modules, respectively, which contain the computational flow and the major dimensional transformations of features. 3.4 Flow-guided Delay Compensation We present a flow-guided delay compensation (FDC) strategy to eliminate the two-sided fusion error in feature-level collaboration due to temporal asynchrony. Existing solutions relied on received historical features [13] and produced large errors under severe delay [53], leading to performance bottlenecks. To tackle the issues, we adopt the philosophy of feature flow to predict the collaborators future features for temporal alignment with the ego representation. The details are as follows. Flow Generation and Warping. Due to the uncertain delay between the ego agent and collaborators, FDC predicts the feature flow Ot0 j for a fixed time interval and scale matrix St0 j based on the j-th agent s historical frames. Specifically, as Figure 2 shows, features {F t0 k j , ..., F t0 j } are concatenated in channel dimension and entered to a generator fflow( ) to output Ot0 j RH W 2 and St0 j RH W 1. Then the j-th agent sends {Ot0 j , St0 j } with prediction ability and sparse feature F t0 j to the ego agent. The ego agent estimates predicted collaborator features as Zt j = fwarp( F t0 j , (t t0) Ot0 j ) St0 j , where fwarp( ) is the bilinear warping function applied to all positions and channels [60, 61], and is the scalar multiplication. The temporally aligned features are passed to the STCFormer. Self-supervised Training Pattern. Self-supervised learning is employed to train the flow generator fflow( ) since the existing datasets [40, 41, 52] lack the motion annotations. Concretely, we first form the training group {F t0 k j , ..., F t0 j , F t j }, where {F t0 k j , ..., F t0 j } is a continuous feature sequence, and F t j is considered as the ground truth feature. Subsequently, we predict the feature Zt j as Zt j = fwarp(F t0 j , (t t0) Ot0 j ) St0 j . Since the optimization objective of fflow( ) is to increase the similarity between F t j and Zt j, we formulate the self-supervised loss function Lflow based on the cosine similarity [53] as follows: Lflow = 1 N 1 j {1,...,N},j =i (1 F t j Zt j F t j 2 F Zt j 2 F ), (5) where 2 F is the squared Frobenius norm. 3.5 Spatio-Temporal Collaboration Transformer To efficiently mitigate collaboration heterogeneity, we propose a spatio-temporal collaboration transformer (STCFormer) to jointly integrate the decoupled spatial semantics and temporal dynamics among agents. From Figure 3(a), the core contributions of STCFormer lie in the following three customized modules, where the other basic components follow the choice of the vanilla transformer [30]. Temporal Cross-Attention. To bridge the detection gap regarding fast-moving objects due to point cloud sparsity, we capture historical context clues across agents to reinforce the current representation via a temporal cross-attention (TCA) module. The core is to perform Query-Key Value-like attention operations by projecting the ego feature F t i and merged temporal features E = [ F t0 j , F t0 i , ..., F t i ] into different subspaces via three 2D convolutional layers ω3 3( ). In Figure 3(b), the branches of Key&Value Ek/v ω1(φa(ω3 3(E))) ω3 3(E) share the same structure but separate weights, where a 1D temporal convolution ω1( ) with global average pooling φa( ) provides temporal correlations. φa( ) is applied to the spatial dimension to shrink the feature map. The computation of TSA can be shown as: Ht i = ω3 3(softmax(ω3 3(F t i ) ET k ) Ev) RH W C. (6) Decoupled Spatial Attention. To comprehensively integrate the distinct spatial semantics from the collaborators, we facilitate pragmatic message fusion with a feature decoupling perspective inspired by the observation in Figure 1(d). Formally, an importance generator fgen( ) is employed to generate the importance maps of the ego feature F t i and the estimated collaborator feature Zt j as It i/It j = σ(φm(fgen(F t i /Zt j))) [0, 1]H W . The importance maps reflect the perceptually critical level of each pixel in the features. Then, the j-th agent spatially decouples the feature Zt j via candidate maps Et j = (1 It i) It j and Ct j = It i It j. Intuitively, Et j and Ct j depict the collaborators exclusive and common perception regions relative to the ego agent, respectively. The exclusive and common collaborator features are obtained as Zt j,e = fsel(Et j) Zt j and Zt j,c = fsel(Ct j) Zt j. Then, we present two spatial attention modules based on deformable cross-attention [59] to aggregate the decoupled exclusive and common features, which share the same structure but different weights. Here the exclusive spatial attention (ESA) is taken as an example (see Figure 3(c)), and its input comprises Ht i and Zt j,e. An importance-aware query initialization is first designed to guide ESA to focus on the potential foreground objects. Specifically, we obtain the element-wise summation of the importance maps as It = PN j=1 It j and extract Nq target queries from the salient locations in It. The attention scores are learned from the initial queries via a linear layer and the softmax function. Subsequently, a linear layer learns an offset map for each input feature, providing the 2D spatial offset { qv |1 v Nv} for each query q. We sample the keypoints based on the learned offset maps and extract these keypoints features to form the attending feature. The cross-attention layer aggregates multiple collaborators features to output the enhanced feature for each query q as: v=1 softmax(Wf Ht i (q))Zt j,e(q + qv)], (7) where u indexs the attention head, and Wu/f denotes the learnable parameters. Then, the filling operation fills ESA(q) into Ht i based on the initial positions of the queries and outputs F t i,e. Similarly, the enhanced common feature F t i,c is obtained via the common spatial attention (CSA). Adaptive Late Fusion. The adaptive late fusion (ALF) module is presented to effectively fuse the exclusive and common representations {F t i,e, F t i,c} for incorporating their perceptual advantages. Formally, we obtain two weight maps as Gt i,e/Gt i,c = ω1 1(F t i,e/F t i,c), and apply the softmax function to produce the normalized weight maps as Gt i,e/Gt i,c = softmax(Gt i,e/Gt i,c). The learned Gt i,e and Gt i,c reflect the complementary perception contributions of {F t i,e, F t i,c} at each spatial location. Therefore, we adaptively activate the perceptually critical information of each representation by a weighted summation. The refined feature map is obtained as Ft i = Gt i,e F t i,e + Gt i,c F t i,c. 3.6 Detection Decoders and Objective Optimization Two detection decoders {f r dec( ), f c dec( )} are employed to convert the output fused representation Ft i into the prediction results. The regression result represents the position, size, and yaw angle of the predefined box at each location, which is Y(t) i,r = f r dec(Ft i ) RH W 7. The classification result is Y(t) i,c = f c dec(Ft i ) RH W 2, revealing the confidence value of each bounding box to be an object. For objective optimization, we leverage the smooth absolute error loss for regression (denoted as Lreg) and the focal loss [15] for classification (denoted as Lcla). In total, we formulate the overall objective function as follows: Lall = Lreg + Lcla + Lmul + Lflow. 12 14 16 18 20 Communication Volume (log2) 32 34 36 38 40 42 44 46 AP0.7 on DAIR-V2X 12 14 16 18 20 Communication Volume (log2) 30 35 40 45 50 55 60 65 AP0.7 on V2XSet 12 14 16 18 20 Communication Volume (log2) 55.0 57.5 60.0 62.5 65.0 67.5 70.0 72.5 AP0.7 on OPV2V How2comm (Ours) Where2comm Co BEVT Disco Net V2X-Vi T V2VNet When2com Late Fusion Figure 4: Collaborative perception performance comparison of How2comm and Where2comm [7] on the DAIR-V2X, V2XSet, and OPV2V datasets with varying communication volumes. 0.0 0.1 0.2 0.3 0.4 0.5 Localization Error Std (m) AP0.7 on DAIR-V2X 0.0 0.1 0.2 0.3 0.4 0.5 Localization Error Std (m) AP0.7 on V2XSet 0.0 0.1 0.2 0.3 0.4 0.5 Localization Error Std (m) AP0.7 on OPV2V How2comm (Ours) Where2comm Disco Net V2X-Vi T V2VNet When2com No Fusion Figure 5: Robustness to the localization error on the DAIR-V2X, V2XSet, and OPV2V datasets. 4 Experiments 4.1 Datasets and Implementation Details Multi-Agent 3D Detection Datasets. To evaluate the performance of How2comm on the collaborative perception task, we conduct extensive experiments on three multi-agent datasets, including DAIRV2X [52], V2XSet [40], and OPV2V [41]. DAIR-V2X [52] is a real-world vehicle-to-infrastructure perception dataset containing 100 realistic scenarios and 18,000 data samples. Each sample collects the labeled Li DAR point clouds of a vehicle and an infrastructure. The training/validation/testing sets are split in a ratio of 5:2:3. V2XSet [40] is a simulated dataset supporting V2X perception, co-simulated by Carla [3] and Open CDA [36]. It includes 73 representative scenes with 2 to 5 connected agents and 11,447 3D annotated Li DAR point cloud frames. The training/validation/testing sets are 6,694, 1,920, and 2,833 frames, respectively. OPV2V [41] is a large-scale simulated dataset for multi-agent V2V perception, comprising 10,914 Li DAR point cloud frames with 3D annotation. The training/validation/testing splits include 6,764, 1,981, and 2,169 frames, respectively. Evaluation Metrics. We adopt the Average Precision (AP) at Intersection-over-Union (Io U) thresholds of 0.5 and 0.7 to evaluate the 3D object detection performance. Also, the calculation format of communication volume in [7] is used to count the message size by byte in the log scale with base 2. 4.2 Quantitative Evaluation Table 1: Performance comparison on the DAIRV2X [52], V2XSet [40], and OPV2V [41] datasets. The results are reported in AP@0.5/0.7. Models DAIR-V2X V2XSet OPV2V AP@0.5/0.7 AP@0.5/0.7 AP@0.5/0.7 No Fusion 50.03/43.57 60.60/40.20 68.71/48.66 Late Fusion 48.93/34.06 54.92/30.75 79.62/59.48 When2com [16] 46.64/32.49 65.06/41.87 70.64/53.73 F-Cooper [1] 49.77/35.21 71.48/46.92 75.27/63.05 Att Fuse [41] 50.86/38.30 70.85/48.66 79.14/64.52 V2VNet [33] 52.18/38.62 79.09/49.25 77.45/62.10 Disco Net [14] 51.44/40.01 79.83/54.06 81.08/64.85 V2X-Vi T [40] 51.68/39.97 83.64/61.41 80.61/66.42 Co BEVT [38] 56.08/41.45 81.07/54.33 81.59/67.50 Where2comm [7] 59.34/43.53 82.02/53.38 82.75/67.29 How2comm (ours) 62.36/47.18 84.05/67.01 85.42/72.24 Experimental Settings. We build all the models using the Pytorch toolbox [18] and train them on Tesla V100 GPUs with the Adam optimizer [10]. The learning rate is set to 2e-3 and decays exponentially by 0.1 every 15 epochs. The training settings on the DAIR-V2X [52], V2XSet [40], and OPV2V [41] datasets include: the training epochs are {30, 40, 40}, and batch sizes are {2, 1, 1}. The height and width resolution of the feature encoder fenc( ) is 0.4 m. The selection function fsel( ) has a threshold of 0.01. The flow generator fflow( ) leverages the multi-scale backbone to extract multi-grained representations and an extra encoder to produce Ot0 j and St0 j . We implement the importance generator fgen( ) and statistical network Tϱ( ) with the classification decoder in [11] and following [27], respectively. The keypoint number Nv is 9, and the attention head is 8. Two 1 1 convolutional layers 0.0 0.2 0.4 0.6 0.8 1.0 Heading Error Std ( ) AP0.7 on DAIR-V2X 0.0 0.2 0.4 0.6 0.8 1.0 Heading Error Std ( ) AP0.7 on V2XSet 0.0 0.2 0.4 0.6 0.8 1.0 Heading Error Std ( ) AP0.7 on OPV2V How2comm (Ours) Where2comm Disco Net V2X-Vi T V2VNet When2com No Fusion Figure 6: Robustness to the heading error on the DAIR-V2X, V2XSet, and OPV2V datasets. 0.0 0.1 0.2 0.3 0.4 Transmission Dealy (s) AP0.7 on DAIR-V2X 0.0 0.1 0.2 0.3 0.4 Transmission Dealy (s) AP0.7 on V2XSet 0.0 0.1 0.2 0.3 0.4 Transmission Dealy (s) AP0.7 on OPV2V How2comm (Ours) Where2comm Disco Net V2X-Vi T V2VNet When2com No Fusion Figure 7: Robustness to the transmission delay on the DAIR-V2X, V2XSet, and OPV2V datasets. are utilized to implement the detection decoders. Under the default noise settings, the transmission delay τ is set to 100 ms, and the localization and heading errors of the collaborators are sampled from a Gaussian distribution with standard deviations of 0.2 m and 0.2 , respectively. All experiments are constrained to 1 MB bandwidth consumption to reflect the narrow communication channels in real V2V/X scenarios [20]. Detection Performance Comparison. Table 1 compares the detection performance of the proposed How2comm with various models on three datasets under default noise settings. We consider two typical baselines. No Fusion is a single-agent perception pattern that only uses the local observations. Late Fusion integrates predicted boxes across agents and produces results with non-maximum suppression. Moreover, the existing SOTAs are comprehensively considered, including When2com [16], F-Cooper [1], Att Fuse [41], V2VNet [33], Disco Net [14], V2X-Vi T [40], Co BEVT [38], and Where2comm [7]. Intuitively, How2comm outperforms previous methods in the real-world (DAIRV2X [52]) and simulated datasets, demonstrating the superiority of our model and its robustness to various realistic noises. In particular, the SOTA performance of AP@0.7 on the DAIR-V2X and OPV2V is improved by 8.4% and 7.0%, respectively. Compared to previous per-agent/location message fusion efforts [14, 16, 33, 40], How2comm simultaneously considers the decoupled spatial semantics and temporal dynamics among agents, resulting in a more precise perception. Comparison of Communication Volume. Figure 4 presents the performance comparison results with distinct bandwidth consumptions. Concretely, the orange and blue curves denote the detection precision of our How2comm and Where2comm under varying communication volumes, respectively. (i) How2comm keeps superior to Where2comm across all the communication choices, i.e., How2comm achieves a better performance-bandwidth trade-off with spatial-channel filtering than Where2comm. (ii) Moreover, our framework seeks comparable performances as the SOTAs by consuming less bandwidth. The noteworthy improvements demonstrate that the proposed communication mechanism filters invalid semantics and maintains performance by mutual information maximization. Robustness to Localization and Heading Errors. We verify the detection performance of How2comm under varying pose errors of collaborators in Figures 5 and 6 following the noise settings in [40]. Specifically, the localization and heading errors are sampled from Gaussian distributions with a standard deviation of σxyz [0, 0.5] m and [0 , 1.0 ], respectively. As shown in the figures, the performance of all intermediate collaboration models consistently deteriorates due to feature map misalignment as the pose errors increase. Noticeably, How2comm is superior to the previous SOTA models and No Fusion across three datasets under all error levels, while some models (e.g., V2VNet and When2com) are even weaker than No Fusion when the error exceeds 0.2 m and 0.2 on the DAIR-V2X dataset. This comparison demonstrates the robustness of How2comm against collaboration pose noises. One reasonable explanation is that our framework captures perceptually critical and holistic information across heterogeneous agents via the tailored STCFormer. Robustness to Transmission Delay. As noted in Figure 1(b), temporal asynchrony due to transmission delay results in two-sided fusion errors and harms the collaboration procedure. We analyze the sensitivity of existing models to varying delays (i.e., from 0 to 400 ms) in Figure 7. Noticeably, all the intermediate fusion methods inevitably degrade with increasing transmission delay due to misleading feature matching. Nevertheless, How2comm maintains higher precision than SOTAs at all delay levels across all three datasets and improves AP@0.7 by 7.5% than No Fusion on DAIR-V2X under a severe delay (400 ms). This robustness to the transmission delay proves that How2comm accomplishes temporal alignment of two-sided features by predicting the collaborators future features. 4.3 Ablation Studies We perform thorough ablation studies on all datasets to understand the necessity of the different designs and strategies in How2comm. Table 2 shows the following vital observations. Table 2: Ablation study results of the proposed components and strategies on all the datasets. w/o stands for the without. Components/Strategies DAIR-V2X V2XSet OPV2V AP@0.5/0.7 AP@0.5/0.7 AP@0.5/0.7 Full Framework 62.36/47.18 84.05/67.01 85.42/72.24 Rationality of Communication Mechanism w/o Spatial Query 61.25/46.33 83.14/66.15 84.52/71.40 w/o Channel Query 61.76/46.52 83.70/66.58 85.23/71.76 w/o Mutual Information 60.61/46.04 82.53/65.75 84.06/70.69 Effect of Delay Compensation Strategy w/o Scale Matrix 62.17/46.86 83.61/66.63 85.08/71.54 w/o Self-supervised Training 60.55/45.77 82.84/65.78 84.20/70.36 Importance of STCFormer w/o Temporal Cross-Attention 60.13/45.92 83.09/65.88 84.27/70.85 w/o Exclusive Spatial Attention 59.46/45.06 82.65/65.70 83.81/71.19 w/o Common Spatial Attention 61.24/46.63 83.48/66.41 84.39/71.44 w/o Adaptive Late Fusion 61.93/46.77 83.72/66.69 84.86/71.64 Impact of Keypoint Number Nv Nv = 6 61.68/46.05 82.82/66.15 84.67/71.38 Nv = 9 (Default) 62.36/47.18 84.05/67.01 85.42/72.24 Nv = 12 62.13/46.57 84.12/66.59 85.28/72.01 Necessity of Decoupled Design w/o Decoupled Design 60.08/45.36 82.41/65.38 84.02/71.15 Rationality of Communication Mechanism. (i) The spatial and channel attention queries are first removed separately to perform incomplete message filtering. The decreased performance implies that both query patterns contribute to sharing sparse yet salient features among agents. (ii) There is a significant degradation in the detection results on all datasets when the communication mechanism lacks mutual information supervision. A plausible deduction is that our supervision mitigates the loss of valuable information due to spatialchannel feature filtering. Effect of Delay Compensation Strategy. (i) Here, we remove the scale matrix to verify its effect. The poor results show that appropriate scaling of predicted features promotes effective temporal alignment. (ii) Furthermore, the performance drop caused by the self-supervised training removal shows the importance of imposing motion representation supervision for collaborator-shared features. Importance of STCFormer. STCFormer is evaluated in three dimensions. (i) We first find that temporal cross-attention provides beneficial gains due to performance deterioration when discarded. It is because the meaningful temporal clues in historical frames bridge the single-frame detection gap. (ii) Then, exclusive and common spatial attention modules are removed separately to explore the impact on performance. The consistently decreased results on each dataset suggest that integrating distinct spatial semantics is indispensable for pragmatic collaboration. (iii) Finally, the adaptive late fusion is replaced by pixel-wise addition. The gain decline suggests that our fusion paradigm provides new insights for aggregating the perceptual advantages of distinct spatial semantics. Impact of Keypoint Number Nv. Empirically, we set the variable number of keypoints for the decoupled spatial attention modules to perform experiments and find that 9 keypoints achieve the most competitive detection performance. Conversely, too few keypoints may cause valuable semantics loss and too many keypoints may cause performance bottlenecks due to accumulated errors. The above finding inspires us to determine the appropriate keypoint number that samples more rich visual clues and captures more pragmatic collaboration information from collaborators. Necessity of Decoupled Design. Ultimately, we set Et j and Ct j to 1 while maintaining two spatial attention branches in the STCFormer to justify the decoupled design. Without explicitly defining exclusive and common features, spatial attention would indiscriminately sample the entire region due to the lack of guidance from prior perception information, which may introduce excessive collaborator noises and impede bridging to collaboration heterogeneity. 4.4 Qualitative Evaluation Visualization of Detection Results. To illustrate the perception performance of different models, Figure 8 shows the detection visualizations of two challenging scenarios from the DAIR-V2X (a) V2X-Vi T (b) Co BEVT (c) Where2comm (d) How2comm (ours) Figure 8: Detection visualization comparison in real-world scenarios from the DAIR-V2X dataset. Green and red boxes denote the ground truths and detection results, respectively. (a) GT bounding boxes (b) Exclusive spatial attention (c) Common spatial attention Figure 9: Visualization of learned exclusive and common spatial attention maps. The red and blue point clouds are derived from the ego vehicle and infrastructure, respectively. In (b)&(c), cyan cross marker denotes the target query in the deformable cross-attention. The 9 sampled keypoints are represented by the forestgreen dots whose colors reflect the corresponding attention weights. dataset [52] under default noise settings. How2comm axiomatically achieves more robust and accurate detection compared to previous SOTA models, including V2X-Vi T [40], Co BEVT [38], and Where2comm [7]. Concretely, our method produces more predicted bounding boxes well aligned with the ground truths. The merits may lie in two aspects. (i) Our delay compensation strategy mitigates feature misalignment due to temporal asynchrony, improving detection performance. (ii) The proposed STCFormer provides effective temporal context clues for detecting fast-moving objects and fuses meaningful information from nearby agents to compensate for the occluded perspective. Visualization of Spatial Attention Maps. We show visualizations of exclusive and common spatial attention (ESA/CSA) in Figure 9 to justify the effectiveness of our feature decoupling philosophy. (i) Intuitively, from Figure 9(b), ESA effectively samples keypoints at the exclusive perception region from the infrastructure, providing complementary information for the ego agent to promote perception ability. (ii) In Figure 9(c), CSA mitigates the detection gap due to collaboration heterogeneity by sampling keypoints that reasonably focus on the common perception region between two agents. 5 Conclusion and Limitation This paper presents How2comm, a novel collaborative perception framework to tackle existing issues jointly. How2comm maximizes the beneficial semantics in filtered features and accomplishes temporal alignment via the feature flow estimation. Moreover, our STCFormer holistically aggregates the spatial semantics and temporal dynamics among agents. Extensive experiments on several multi-agent datasets show the effectiveness of How2comm and the necessity of all its components. Limitation and Future Work. The current work only exploits the short-term historical frames. In future work, we plan to expand the utilization of temporal information to long-term point cloud sequences. Also, we will explore optimizing the feature flow prediction with uncertainty estimation. Acknowledgment. This work is supported in part by the National Key R&D Program of China under Grant 2021ZD0113503 and in part by the Shanghai Municipal Science and Technology Major Project under Grant 2021SHZDZX0103. [1] Qi Chen, Xu Ma, Sihai Tang, Jingda Guo, Qing Yang, and Song Fu. F-cooper: Feature based cooperative perception for autonomous vehicle edge computing system using 3D point clouds. In Proceedings of the 4th ACM/IEEE Symposium on Edge Computing, pages 88 100, 2019. 7, 8 [2] Abhishek Das, Théophile Gervet, Joshua Romoff, Dhruv Batra, Devi Parikh, Mike Rabbat, and Joelle Pineau. Tarmac: Targeted multi-agent communication. In International Conference on Machine Learning (ICML), pages 1538 1546. PMLR, 2019. 3 [3] Alexey Dosovitskiy, German Ros, Felipe Codevilla, Antonio Lopez, and Vladlen Koltun. Carla: An open urban driving simulator. In Conference on Robot Learning, pages 1 16. PMLR, 2017. 7 [4] Jakob Foerster, Ioannis Alexandros Assael, Nando De Freitas, and Shimon Whiteson. Learning to communicate with deep multi-agent reinforcement learning. Advances in Neural Information Processing Systems (Neur IPS), 29, 2016. 3 [5] R Devon Hjelm, Alex Fedorov, Samuel Lavoie-Marchildon, Karan Grewal, Phil Bachman, Adam Trischler, and Yoshua Bengio. Learning deep representations by mutual information estimation and maximization. ar Xiv preprint ar Xiv:1808.06670, 2018. 4 [6] Yedid Hoshen. Vain: Attentional multi-agent predictive modeling. Advances in Neural Information Processing Systems (Neur IPS), 30, 2017. 3 [7] Yue Hu, Shaoheng Fang, Zixing Lei, Yiqi Zhong, and Siheng Chen. Where2comm: Communication-efficient collaborative perception via spatial confidence maps. Advances in Neural Information Processing Systems (Neur IPS), 35:4874 4886, 2022. 1, 2, 3, 4, 7, 8, 10 [8] Yue Hu, Yifan Lu, Runsheng Xu, Weidi Xie, Siheng Chen, and Yanfeng Wang. Collaboration helps camera overtake lidar in 3d detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 9243 9252, 2023. 2, 4 [9] Jiechuan Jiang and Zongqing Lu. Learning attentional communication for multi-agent cooperation. Advances in Neural Information Processing Systems (Neur IPS), 31, 2018. 3 [10] Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), 2015. 7 [11] Alex H Lang, Sourabh Vora, Holger Caesar, Lubing Zhou, Jiong Yang, and Oscar Beijbom. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 12697 12705, 2019. 4, 7 [12] Yuxuan Lei, Dingkang Yang, Mingcheng Li, Shunli Wang, Jiawei Chen, and Lihua Zhang. Textoriented modality reinforcement network for multimodal sentiment analysis from unaligned multimodal sequences. ar Xiv preprint ar Xiv:2307.13205, 2023. 3 [13] Zixing Lei, Shunli Ren, Yue Hu, Wenjun Zhang, and Siheng Chen. Latency-aware collaborative perception. In Proceedings of the European Conference on Computer Vision (ECCV), pages 316 332. Springer, 2022. 2, 5 [14] Yiming Li, Shunli Ren, Pengxiang Wu, Siheng Chen, Chen Feng, and Wenjun Zhang. Learning distilled collaboration graph for multi-agent perception. Advances in Neural Information Processing Systems (Neur IPS), 34:29541 29552, 2021. 1, 2, 3, 4, 7, 8 [15] Tsung-Yi Lin, Priya Goyal, Ross Girshick, Kaiming He, and Piotr Dollár. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 2980 2988, 2017. 6 [16] Yen-Cheng Liu, Junjiao Tian, Nathaniel Glaser, and Zsolt Kira. When2com: Multi-agent perception via communication graph grouping. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 4106 4115, 2020. 1, 2, 3, 7, 8 [17] Guiyang Luo, Hui Zhang, Quan Yuan, and Jinglin Li. Complementarity-enhanced and redundancy-minimized collaboration network for multi-agent perception. In Proceedings of the ACM International Conference on Multimedia (ACM MM), pages 3578 3586, 2022. 1 [18] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems (Neur IPS), 32, 2019. 7 [19] Charles R Qi, Hao Su, Kaichun Mo, and Leonidas J Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 652 660, 2017. 1 [20] Hang Qiu, Pohan Huang, Namo Asavisanu, Xiaochen Liu, Konstantinos Psounis, and Ramesh Govindan. Autocast: Scalable infrastructure-less cooperative perception for distributed collaborative driving. 2022. 8 [21] Linhao Qu, Shaolei Liu, Manning Wang, and Zhijian Song. Transmef: A transformer-based multi-exposure image fusion framework using self-supervised multi-task learning. In Proceedings of the AAAI Conference on Artificial Intelligence, pages 2126 2134, 2022. 3 [22] Linhao Qu, Siyu Liu, Xiaoyu Liu, Manning Wang, and Zhijian Song. Towards label-efficient automatic diagnosis and analysis: a comprehensive survey of advanced deep learning-based weakly-supervised, semi-supervised and self-supervised techniques in histopathological image analysis. Physics in Medicine & Biology, 2022. 3 [23] Linhao Qu, Xiaoyuan Luo, Shaolei Liu, Manning Wang, and Zhijian Song. Dgmil: Distribution guided multiple instance learning for whole slide image classification. In Medical Image Computing and Computer-Assisted Intervention (MICAAI), pages 24 34. Springer, 2022. 3 [24] Linhao Qu, Manning Wang, Zhijian Song, et al. Bi-directional weakly supervised knowledge distillation for whole slide image classification. Advances in Neural Information Processing Systems (Neur IPS), 35:15368 15381, 2022. 3 [25] Tika Ram and Khem Chand. Effect of drivers risk perception and perception of driving tasks on road safety attitude. Transportation Research Part F: Traffic Psychology and Behaviour, 42:162 176, 2016. 1 [26] Joseph Redmon, Santosh Divvala, Ross Girshick, and Ali Farhadi. You only look once: Unified, real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 779 788, 2016. 1 [27] Eduardo Hugo Sanchez, Mathieu Serrurier, and Mathias Ortner. Learning disentangled representations via mutual information estimation. In Proceedings of the European Conference on Computer Vision (ECCV), pages 205 221. Springer, 2020. 7 [28] Amanpreet Singh, Tushar Jain, and Sainbayar Sukhbaatar. Learning when to communicate at scale in multiagent cooperative and competitive tasks. ar Xiv preprint ar Xiv:1812.09755, 2018. 3 [29] Sainbayar Sukhbaatar, Rob Fergus, et al. Learning multiagent communication with backpropagation. Advances in Neural Information Processing Systems (Neur IPS), 29, 2016. 3 [30] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in Neural Information Processing Systems (Neur IPS), 30, 2017. 5 [31] Binglu Wang, Lei Zhang, Zhaozhong Wang, Yongqiang Zhao, and Tianfei Zhou. Core: Cooperative reconstruction for multi-agent perception. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 8710 8720, October 2023. 1 [32] Tianhang Wang, Guang Chen, Kai Chen, Zhengfa Liu, Bo Zhang, Alois Knoll, and Changjun Jiang. Umc: A unified bandwidth-efficient and multi-resolution based collaborative perception framework. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 8187 8196, October 2023. 1, 2 [33] Tsun-Hsuan Wang, Sivabalan Manivasagam, Ming Liang, Bin Yang, Wenyuan Zeng, and Raquel Urtasun. V2vnet: Vehicle-to-vehicle communication for joint perception and prediction. In Proceedings of the European Conference on Computer Vision (ECCV), pages 605 621. Springer, 2020. 1, 2, 3, 4, 7, 8 [34] Sanghyun Woo, Jongchan Park, Joon-Young Lee, and In So Kweon. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), pages 3 19, 2018. 4 [35] Pengxiang Wu, Siheng Chen, and Dimitris N Metaxas. Motionnet: Joint perception and motion prediction for autonomous driving based on bird s eye view maps. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 11385 11395, 2020. 4 [36] Runsheng Xu, Yi Guo, Xu Han, Xin Xia, Hao Xiang, and Jiaqi Ma. Opencda: an open cooperative driving automation framework integrated with co-simulation. In 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), pages 1155 1162. IEEE, 2021. 7 [37] Runsheng Xu, Jinlong Li, Xiaoyu Dong, Hongkai Yu, and Jiaqi Ma. Bridging the domain gap for multi-agent perception. In IEEE International Conference on Robotics and Automation (ICRA), pages 6035 6042. IEEE, 2023. 2, 3 [38] Runsheng Xu, Zhengzhong Tu, Hao Xiang, Wei Shao, Bolei Zhou, and Jiaqi Ma. Cobevt: Cooperative bird s eye view semantic segmentation with sparse transformers. In Conference on Robot Learning (Co RL), 2022. 1, 3, 7, 8, 10 [39] Runsheng Xu, Xin Xia, Jinlong Li, Hanzhao Li, Shuo Zhang, Zhengzhong Tu, Zonglin Meng, Hao Xiang, Xiaoyu Dong, Rui Song, et al. V2v4real: A real-world large-scale dataset for vehicle-to-vehicle cooperative perception. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 13712 13722, 2023. 1 [40] Runsheng Xu, Hao Xiang, Zhengzhong Tu, Xin Xia, Ming-Hsuan Yang, and Jiaqi Ma. V2x-vit: Vehicle-to-everything cooperative perception with vision transformer. In Proceedings of the European Conference on Computer Vision (ECCV), pages 107 124. Springer, 2022. 1, 2, 3, 4, 5, 7, 8, 10 [41] Runsheng Xu, Hao Xiang, Xin Xia, Xu Han, Jinlong Li, and Jiaqi Ma. Opv2v: An open benchmark dataset and fusion pipeline for perception with vehicle-to-vehicle communication. In Proceedings of the International Conference on Robotics and Automation (ICRA), pages 2583 2589. IEEE, 2022. 2, 5, 7, 8 [42] Dingkang Yang, Zhaoyu Chen, Yuzheng Wang, Shunli Wang, Mingcheng Li, Siao Liu, Xiao Zhao, Shuai Huang, Zhiyan Dong, Peng Zhai, and Lihua Zhang. Context de-confounded emotion recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 19005 19015, June 2023. 3 [43] Dingkang Yang, Shuai Huang, Haopeng Kuang, Yangtao Du, and Lihua Zhang. Disentangled representation learning for multimodal emotion recognition. In Proceedings of the 30th ACM International Conference on Multimedia (ACM MM), page 1642 1651, 2022. 3 [44] Dingkang Yang, Shuai Huang, Yang Liu, and Lihua Zhang. Contextual and cross-modal interaction for multi-modal speech emotion recognition. IEEE Signal Processing Letters, 29:2093 2097, 2022. 3 [45] Dingkang Yang, Shuai Huang, Shunli Wang, Yang Liu, Peng Zhai, Liuzhen Su, Mingcheng Li, and Lihua Zhang. Emotion recognition for multiple context awareness. In Proceedings of the European Conference on Computer Vision (ECCV), volume 13697, pages 144 162. Springer, 2022. 3 [46] Dingkang Yang, Shuai Huang, Zhi Xu, Zhenpeng Li, Shunli Wang, Mingcheng Li, Yuzheng Wang, Yang Liu, Kun Yang, Zhaoyu Chen, Yan Wang, Jing Liu, Peixuan Zhang, Peng Zhai, and Lihua Zhang. Aide: A vision-driven multi-view, multi-modal, multi-tasking dataset for assistive driving perception. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 20459 20470, October 2023. 1 [47] Dingkang Yang, Haopeng Kuang, Shuai Huang, and Lihua Zhang. Learning modality-specific and -agnostic representations for asynchronous multimodal language sequences. In Proceedings of the 30th ACM International Conference on Multimedia (ACM MM), page 1708 1717, 2022. 3 [48] Dingkang Yang, Yang Liu, Can Huang, Mingcheng Li, Xiao Zhao, Yuzheng Wang, Kun Yang, Yan Wang, Peng Zhai, and Lihua Zhang. Target and source modality co-reinforcement for emotion understanding from asynchronous multimodal sequences. Knowledge-Based Systems, page 110370, 2023. 3 [49] Kun Yang, Jing Liu, Dingkang Yang, Hanqi Wang, Peng Sun, Yanni Zhang, Yan Liu, and Liang Song. A novel efficient multi-view traffic-related object detection framework. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1 5, 2023. 1 [50] Kun Yang, Peng Sun, Jieyu Lin, Azzedine Boukerche, and Liang Song. A novel distributed task scheduling framework for supporting vehicular edge intelligence. In IEEE International Conference on Distributed Computing Systems (ICDCS), pages 972 982, 2022. 3 [51] Kun Yang, Dingkang Yang, Jingyu Zhang, Mingcheng Li, Yang Liu, Jing Liu, Hanqi Wang, Peng Sun, and Liang Song. Spatio-temporal domain awareness for multi-agent collaborative perception. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pages 23383 23392, October 2023. 1 [52] Haibao Yu, Yizhen Luo, Mao Shu, Yiyi Huo, Zebang Yang, Yifeng Shi, Zhenglong Guo, Hanyu Li, Xing Hu, Jirui Yuan, et al. Dair-v2x: A large-scale dataset for vehicle-infrastructure cooperative 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 21361 21370, 2022. 2, 5, 7, 8, 10 [53] Haibao Yu, Yingjuan Tang, Enze Xie, Jilei Mao, Jirui Yuan, Ping Luo, and Zaiqing Nie. Vehicle-infrastructure cooperative 3d object detection via feature flow prediction. ar Xiv preprint ar Xiv:2303.10552, 2023. 2, 5 [54] Haibao Yu, Wenxian Yang, Hongzhi Ruan, Zhenwei Yang, Yingjuan Tang, Xu Gao, Xin Hao, Yifeng Shi, Yifeng Pan, Ning Sun, et al. V2x-seq: A large-scale sequential dataset for vehicle-infrastructure cooperative perception and forecasting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5486 5495, 2023. 1 [55] Xiaoding Yuan, Adam Kortylewski, Yihong Sun, and Alan Yuille. Robust instance segmentation through reasoning about multi-object occlusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 11141 11150, 2021. 1 [56] Zhenxun Yuan, Xiao Song, Lei Bai, Zhe Wang, and Wanli Ouyang. Temporal-channel transformer for 3d lidar-based video object detection for autonomous driving. IEEE Transactions on Circuits and Systems for Video Technology, 32(4):2068 2078, 2021. 1 [57] Zixu Zhang and Jaime F Fisac. Safe occlusion-aware autonomous driving via game-theoretic active perception. ar Xiv preprint ar Xiv:2105.08169, 2021. 1 [58] Hengshuang Zhao, Jianping Shi, Xiaojuan Qi, Xiaogang Wang, and Jiaya Jia. Pyramid scene parsing network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 2881 2890, 2017. 1 [59] Xizhou Zhu, Weijie Su, Lewei Lu, Bin Li, Xiaogang Wang, and Jifeng Dai. Deformable detr: Deformable transformers for end-to-end object detection. In International Conference on Learning Representations (ICLR), 2021. 6 [60] Xizhou Zhu, Yujie Wang, Jifeng Dai, Lu Yuan, and Yichen Wei. Flow-guided feature aggregation for video object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pages 408 417, 2017. 5 [61] Xizhou Zhu, Yuwen Xiong, Jifeng Dai, Lu Yuan, and Yichen Wei. Deep feature flow for video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 2349 2358, 2017. 5