# multimodal_object_reidentification_via_sparse_mixtureofexperts__b88861bf.pdf Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Yingying Feng * 1 Jie Li * 2 Chi Xie 3 Lei Tan 4 Jiayi Ji 2 4 We present MFRNet, a novel network for multimodal object re-identification that integrates multi-modal data features to effectively retrieve specific objects across different modalities. Current methods suffer from two principal limitations: (1) insufficient interaction between pixellevel semantic features across modalities, and (2) difficulty in balancing modality-shared and modality-specific features within a unified architecture. To address these challenges, our network introduces two core components. First, the Feature Fusion Module (FFM) enables finegrained pixel-level feature generation and flexible cross-modal interaction. Second, the Feature Representation Module (FRM) efficiently extracts and combines modality-specific and modality-shared features, achieving strong discriminative ability with minimal parameter overhead. Extensive experiments on three challenging public datasets (RGBNT201, RGBNT100, and MSVR310) demonstrate the superiority of our approach in terms of both accuracy and efficiency, with 8.4% m AP and 6.9% accuracy improved in RGBNT201 with negligible additional parameters. The code is available at https: //github.com/stone96123/MFRNet. 1. Introduction Object Re-Identification (Re ID) focuses on retrieving specific object across non-overlapping camera views using known information. Traditionally, single-modal Re ID relies heavily on RGB images, which face significant challenges under adverse conditions such as poor lighting, shadows, and low image resolutions (Li et al., 2020; Zhang et al., *Equal contribution 1School of Computer Science and Engineering, Northeastern University, Shenyang, China. 2School of Informatics, Xiamen University, Xiamen, China. 3Tongji University, Shanghai, China. 4National University of Singapore, Singapore. Correspondence to: Lei Tan . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). Transformer Transformer Transformer Transformer Transformer (a) Current one-stream network (b) Current dual-stream network (c) Our proposed MFRNet Interaction Interaction Interaction Interaction RE RE RE RE Figure 1. Comparison between our proposed MFRNet and current mainstream structure. (a) and (b) show the current mainstream one-stream and two-stream networks, respectively. Besides the limitation of neglecting the pixel-level alignment characteristics of multi-modal images, current methods also face the challenge of balancing the modality-specific and modality-shared representation in a unified network. (c) shows our proposed MFRNet and RE refers to the representation expert. MFRNet inherits the idea of Mo E and extends it with multi-modal fusion (FFM) and representation (FRM), allowing it to achieve fine-grained interaction and efficient representation for multi-modal data. 2021; Liu et al., 2024; Tan et al., 2024b). These limitations frequently result in the extraction of misleading features and a subsequent loss of discriminative information (Zheng et al., 2021; Wang et al., 2022). To address these challenges, multi-modal Re ID (Shi et al., 2023; Lu et al., 2023; Shi et al., 2024; Yang et al., 2024) has emerged as a promising solution, leveraging complementary information from various modalities to enhance feature representation and achieve more acceptable identification ability in complex scenarios. With the widespread application of Vision Transformer (Vi T)(Alexey, 2020; Ji et al., 2025), numerous Vi T-based Re ID methods have emerged from general-purpose (He et al., 2021; Tan et al., 2022; 2024a) to multi-modal scenarios (Zhang et al., 2024a; Wang et al., 2024c). However, as illustrated in Figure 1, current mainstream methods face two fundamental limitations: insufficient interaction and feature imbalance. The insufficient interaction manifests in existing approaches that mainly emphasize interaction Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts at high-level semantic feature spaces (Wang et al., 2024a; Zhang et al., 2024a). These methods overlook the inherent pixel-to-pixel alignment characteristics in multi-modal images, resulting in inadequate fine-grained interaction between multi-modal information. The feature imbalance stems from the difficulty of balancing modality-shared and modality-specific features (Wang et al., 2022; Zhang et al., 2024a). A single parameter-sharing Vi T backbone, though parameter-efficient, struggles to preserve modality-specific discriminative features. Alternatively, employing separate Vi T backbones for different modalities can better preserve modal characteristics but introduce significant parameter redundancy and computational overhead. To address the insufficient interaction and feature imbalance challenges, we propose a Modality Fusion and Representation Network (MFRNet). Our design incorporates mixtureof-experts (Jacobs et al., 1991) paradigm into both generators and feature extractors, achieving pixel-level crossmodal interaction and balanced feature learning. MFRNet consists of two key components: Feature Fusion Module (FFM) with mixture-of-generators for fine-grained interaction, and Feature Representation Module (FRM) with mixture of representation experts for balanced feature extraction. The FFM tackles insufficient interaction through its mixture-of-generators design. Inspired by RLE (Tan et al., 2024c) that cross-spectral transformation exhibits locally linear characteristics with surface-dependent variations, FFM employs multiple simple generators rather than complex unified structures. Through the mixture-of-generators mechanism, diverse tokens spontaneously select the most suitable generator for different modalities, locations, and attribute tokens. Additionally, FFM achieves fine-grained information exchange through weighted feature fusion, where fusion weights are learned via a generative network. The FRM addresses feature imbalance through a mixture of representation experts structure. It dynamically routes tokens from different modalities to various experts, enabling efficient parameter utilization while preserving both modality-shared and modality-specific features. Through dynamic expert activation, different experts in FRM adaptively focus on various shared visual attributes while maintaining necessary modality-specific representations. Notably, this design enables our framework to handle missing modalities without specific architectural modifications or training procedures. The contributions of this paper are summarized as follows: We propose a Modality Fusion and Representation Network (MFRNet) for multi-modal object reidentification, which inherits the idea of a sparse mixture of experts and extends it with multi-modal fusion and representation. We introduce a Feature Fusion Module (FFM) and a Feature Representation Module (FRM). The former aims to achieve fine-grained interaction between multimodal inputs, while the latter aims to achieve efficient and balanced feature extraction between modalityshared and modality-specific representations. Extensive experiments on three public multi-modal object Re ID datasets including RGBNT201, RGBNT100, and MSVR310, verifying the superior performance of MFRNet. 2. Related Work 2.1. Multi-modal Person Re ID Human-centric computer vision tasks (Shen & Tang, 2024; Shen et al., 2025) have long been a central focus within the research community. Unlike single-modal object reidentification (Re ID), multi-modal person Re ID includes more information and is suitable for a wider range of applications. Existing methods (Park et al., 2021; Zheng et al., 2023; 2022; He et al., 2023; Tan et al., 2023; Zhang & Wang, 2023) for multi-modal object Re ID mainly focus on how to effectively integrate information from multiple modalities and alleviate the cross-modal heterogeneous issue. For example, Zheng et al. (Zheng et al., 2021) utilize the complementary advantages of multiple modalities to propose PFNet for learning effective multi-modal features. To boost modality-specific representations, Wang et al. (Wang et al., 2022) proposed the IEEE method, which includes important information exchange, feature enrichment, and intra-class discrepancy maximization mechanisms. To effectively utilize the relationship of modalities to reduce the gap between different modalities, Guo et al. (Guo et al., 2022) proposed the GAFNet model to fuse multiple data sources. Additionally, the vision transformers (Vi T) (Alexey, 2020; Radford et al., 2021; Pan et al., 2022; 2023; Crawford et al., 2023; Wang et al., 2024a) model has achieved good results in many fields, promoting the development of Vi T-based multimodal person Re ID algorithms. Wang et al. (Wang et al., 2024b) explored the influence of global and local features of Vi T and proposed the GLTrans model. Wang et al. (Wang et al., 2024c) explored the domain traits of unlabeled test data and proposed a heterogeneous test-time training framework based on Vi T to improve generalization performance. Yu et al. (Yu et al., 2024) proposed the RSCNet model based on the idea of token sparsification in Vi T to alleviate the multi-modal heterogeneity problem. Zhang et al. (Zhang et al., 2024a) proposed the EDITOR model, which selects diverse tokens from Vi T to mitigate the effect of irrelevant backgrounds and reduce the gap between modalities. Despite the good performance of these methods, they tend to employ high-level semantic multi-modal feature fusion, which overlooks the fine-grained spatial alignment characteristics of multi-modal data. Additionally, they struggle Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts to efficiently extract modality-shared and modality-specific multi-modal representation within a unified network. Therefore, we propose MFRNet, which adopts a mixture of generator experts to leverage the pixel-to-pixel alignment of multi-modal data and introduces a mixture of representation experts to adaptively extract modality-shared features while retaining modality-specific features. 2.2. Mixtures of Experts The Mixture of Experts (Mo E) (Jacobs et al., 1991) originates from Ensemble Learning (Ganaie et al., 2022) and can effectively enhance model performance. Notably, after Lepikhin et al. (Lepikhin et al., 2020) combined Mo E with the transformer framework and proposed the Gshard model, it garnered significant attention across various fields, leading to the development of more models incorporating Mo E technology (Chen & Wang, 2025; Dai et al., 2024; Gui et al., 2024; Lin et al., 2024; Liu et al., 2025; Mustafa et al., 2022; Li et al., 2025). Subsequently, to facilitate the implementation and flexible design of Mo E, Hwang et al. (Hwang et al., 2023) open-sourced the Tutel codebase. Additionally, to reduce training and inference costs, Chowdhury et al. (Chowdhury et al., 2023) proposed p Mo E, which can filter label-irrelevant patches and route similar classdiscriminative patches to the same expert. Recently, many works have attempted to introduce Mo E into Re ID. Among them, Li et al. (Li et al., 2023) proposed the MPC model, which leverages Mo E to learn multiple pseudo-label spaces, thereby flexibly handling different variations. Xu et al. (Xu et al., 2022) proposed the META model based on Mo E to better exploit the domain-invariant characteristics of multimodal data. Kuang et al. (Kuang et al., 2024) proposed the Mi Ke Co Co model, which incorporates multiple experts with unique perspectives into CLIP and fully leverages high-level semantic knowledge for comprehensive feature representation. Additionally, Li et al. (Li et al., 2022) explored an orthogonal direction for domain generalization research and proposed the GMo E model by combining Vi T and Mo E technologies. Based on the results of the aforementioned studies, we leveraged the advantages of Mo E technology and utilized the Tutel framework to design and implement Mo E techniques to enhance model performance. 3. Methodology As shown in Figure 2, the proposed MFRNet employs a CLIP pre-trained vanilla Vi T-base as its basic backbone network. On this basis, our network comprises two principal components, i.e., the feature fusion module (FFM) and the feature representation module (FRM). The FFM introduces the pixel-level feature generation for flexible crossmodal interaction, which is also suitable for modal-missing scenarios, while the FRM inherits the idea of the sparse Mixture-of-Experts (Mo E) to adapt to diverse modal input at an extremely low cost and performs joint optimization of modality-specific and modality-shared features. 3.1. Feature Fusion Module In general, a given image can be divided into patches of the same size and mapped to a fixed feature dimension through patchify operations. Then, we have the image feature as: IM = {[Icls M ], I1 M, I2 M, , In M} where M R, N, T, (1) where R, N, T denote RGB, NIR, and TIR modalities respectively. Given that multi-modal images maintain pixel-level alignment, our feature fusion module (FFM) is designed to enable effective interaction and mutual enhancement between different modality features. As illustrated by the RLE (Tan et al., 2024c) and AMML (Zhang et al., 2024b), the cross-spectral transformation can be considered as a local linear transformation, while the linear factor will be varied across different surfaces. It shows that the transformation between different spectra is quite simple but variant. Therefore, inspired by the idea of the Mixture-of-experts strategy and the above observation, FFM attempts to combine multiple simple generators to adapt to diverse input tokens. In addition, it is interesting to find that this kind of fusion strategy of FFM also works well for modalitymissing scenarios in a modality-complementing way. Specifically, as shown in Figure 2, taking the RGB feature IR for example, it can be re-generated with NIR and TIR features as: Ig R = λ IR + w N R (IN) g N R (IN) + w T R(IT ) g T R(IT ), (2) where λ is hyper-parameters to determine the proportion of information retained by the modality itself, w N R ( ) and w T R( ) generate the weighting coefficients for the features generated from NIR and TIR, respectively. And g N R ( ) and g T R( ) represent the reconstructed RGB features from the NIR and TIR modality, respectively. The re-generated features of NIR and TIR are similar. For the weight coefficient function w( )1, we define it as follows: w(I) = fc Avg Pool(I) , Rn d R1 d R1 1, (3) where I Rn d represents image feature from specific modality, Avg Pool represents global average pooling, and fc represents a linear layer. To normalize the weight coefficients, we apply the Softmax operation on the weights of each reconstructed modality feature like 1We omit the subscript for modality for simplification. Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Linear Projection Weighted Sum Multi-modal Feature Fusion Vision Transformer Blocks Multi-modal Feature Representation Multi-Head Attention Cosine Router Figure 2. Overall architecture of MFRNet. MFRNet is built from the basic Transformer blocks inherited from vanilla Vi T, with two new modules (feature fusion module and feature representation Module) added to adapt to multi-modal object re-identification tasks. In the blue box, the Feature Fusion Module (FFM) employs multiple simple generators to adaptively provide fine-grained interaction information, while in the green box, the Feature Representation Module (FRM) employs diverse representation experts to extract and combine modality-specific and modality-shared features. Here, GE refers to the generation experts in FFM, while RE represents the representation experts in FRM. Notably, the figure illustrates the NIR+RGB TIR interaction, with the same interaction applying to NIR and RGB. Softmax w N R (IN), w T R(IT ) . The weight generation function is modality-specific, and there are total 6 functions for 3 modalities. For the feature generation function g( ), as mentioned before, the transformation between different modality images is simple but variant. Compared to complex structures, simple generators work well (Zhang et al., 2024b). Therefore, we apply the mixture-of-experts (Mo E) architecture (Li et al., 2022) to combine multiple simple generators. We define the simplest generation expert gf( ) as: gf(I) = conv2 (drop (conv1(I))) . (4) Then we enhance it using sparse Mo E technology. Specifically, we categorize feature generation into two types based on the different modalities: {IN R, IT R, IR N} and {IT N, IR T , IN T }. Since each of the three modalities can generate features for the other two modalities, we define two moe layers. Finally, the feature generation function is as follows: Iexperts = {gf (fci (I)) | for i [1, K]} Srouter = Softmax( W F 1 IW F e τ W F 1 I W F e ) Gate(X) = Topk(Srouter) g(I) = Gate(I) Iexperts, where fci refers to the linear layer for uth expert, Srouter represents cosine router which indicates routing scores for selecting experts. indicate the L2 normalization, and τ indicates the learnable temperature scaler. W F 1 is the weight of linear projection which aims to project the input token to a low-rank subspace, while W F e is the expert embedding that transforms the feature to final scoring distribution. Note that k = 1 represents only the highest-scoring expert is selected for feature generation each time. As stated before, although do not have any specific training, we find that FFM can also work well for modality-missing scenarios in a modality-complementing way. Specifically, for the modality missing scenarios during testing, we develop conditional reconstructed rules for modality M as: S A w S M(IS) g(IS) if M present P S A w S M(Is) g(IS) if M missing, (6) where A denotes available modalities. For example, if RGB is missing, its features are generated based on NIR and TIR as follows: IR = w N R (IN) g(IN) + w T R(IT ) g(IT ). Here, g(IN) denotes the features for RGB generated from NIR, g(IT ) represents the features for RGB generated from TIR, w N R (IN) and w T R(IT ) are the respective weights. 3.2. Feature Representation Module Existing approaches face fundamental challenges in balancing modality-shared and modality-specific information. While employing separate networks for each modality preserves modality-specific characteristics, this introduces significant structural and computational redundancy. Conversely, using a fully shared single network for all modalities exclusively captures shared representations across modalities. Hence, we propose the Feature Representation Module (FRM), which integrates a Mixture-of-Experts architecture to preserve valuable modality-specific information with parameter-efficient implementation. With the reconstructed multi-modal image features Ig = (Ig R, Ig N, Ig T ), we use a vision transformer (Vi T) to extract fundamental features. Within this module, we aim to capture modality-specific feature representations within a unified architectural framework. Specifically, we modify the residual attention layer based on Rep Adapter as follows: To extract unique features for different modality data, we employ Rep Adapter (Luo et al., 2023), which is designed to rapidly adapts to vision tasks by inserting adapter layers between key layers. Specifically, Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts the residual attention layer can be represented as: Y = X + MH-Attn (Layer Norm(X)), Z = Y + RAda (Layer Norm(Y )), (7) where X and Z are the input and the output of the current layer, MH-Attn( ) represents the multi-head attention layer, RAda( ) represents the Rep Adapter layer. The formula of the Rep Adapter layer is as: RAda(Ylnorm) = Ylnorm + conv B(drop (conv A (Ylnorm))), (8) where Ylnorm refers to the Y after the Layer Norm layer, conv A and conv B represent 1D convolutions with kernel size of 1 1 and groups set to 1. Similarly, to better adapt to multi-modal data, we use RAda as the expert and enhance it by the Mo E (Li et al., 2022) layer. SAda router = Softmax( W R 1 Ylnorm W R e τ W R 1 Ylnorm W R e ), Gate Ada(Ylnorm) = Topk(SAda router), where τ = 0.01 and k = 1. It is important to note that since the number of modalities is 3, the number of experts should be greater than 3. W R 1 is the weight of the linear projection, which aims to project the input token to a lowrank subspace, while W R e is the expert embedding that transforms the feature to the final scoring distribution. Based on the above derivation, using lmoe Ada to represent the Mo E-enhanced Rep Adapter, the modality-specific features are as follows: Z = Y + lmoe Ada(Ylnorm). (10) Here, the auxiliary loss for Mo E is denoted as LAda aux. 3.3. Objective Function As shown in Figure 2, our model consists of three main loss functions: losses for the Vi T backbone, auxiliary losses for the multi-modal feature completion module, and auxiliary losses for the modality-specific feature representation module. For the losses for the Vi T backbone, we follow previous research by using label smoothing cross-entropy loss (Szegedy et al., 2016) and triplet loss (Hermans et al., 2017) for optimizing the representation. Besides the basic losses, following the (Riquelme et al., 2021; Li et al., 2022), we define two auxiliary losses: important loss Limp and load loss Lload to balance the frequency being activated for the expert. The important loss Limp to encourage a balanced usage of different experts across tokens and load loss Lload to encourage balanced assignment across experts. Specifically, the importance of expert e in a batch of images Ib is defined as the normalized routing weight corresponding to expert i summed over images as: Impe(Ib) = X I Ib Se router. (11) The important loss Limp is given by the squared coefficient of variation of the importance distribution over experts as: Limp(Ib) = ( STD(Imp(Ib)) MEAN(Imp(Ib)))2. (12) Meanwhile, the load loss Lload can be given as: Lload(Ib) = STD(load(I)) MEAN(load(I)), with loade(Ib) = P I Ib pe(I), and pe(I) = 1 Φ(ηk (W1IWe)e Herein, the ηk refers to the Kth largest entry after softmax while Φ indicates the cumulative distribution function of a Gaussian distribution. Therefore, the whole auxiliary losses Laux is formulated as: Laux = Limp + Lload Finally, the total loss function of MFRNet is given as follows: L = LV i T ce + LV i T tri + (Lmoe1 aux + Lmoe2 aux ) + LAda aux. (15) 4. Experiments 4.1. Implementation Datasets. We evaluate our model performance on three public multi-modal object Re ID datasets. Specifically, RGBNT201 (Zheng et al., 2021) is a multi-modal person Re ID dataset, which includes 4,787 aligned RGB, NIR, and TIR images from 201 identities. RGBNT100 (Li et al., 2020) is a large-scale multi-modal vehicle Re ID dataset comprising 17,250 image triples. It covers a wide range of challenging visual conditions, making it suitable for assessing the robustness of vehicle Re ID methods. MSVR310 (Zheng et al., 2022) is a small-scale multi-modal vehicle Re ID dataset that includes 2,087 high-quality image triples captured across diverse environments and time spans, providing complex visual challenges for vehicle Re ID evaluation. Evaluation protocols. To assess the performance of our method, we utilize the mean Average Precision (m AP) and Cumulative Matching Characteristics (CMC) at Rank-K (K = 1, 5, 10). These metrics are standard in this field and provide a comprehensive evaluation of our model s effectiveness. Additionally, we present the number of parameters and FLOPs to analyze the complexity of our model. Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Table 1. Performance comparison on RGBNT201. The best and second results are in bold and underlined, respectively. Methods m AP R-1 R-5 R-10 MUDeep (Qian et al., 2017) 23.8 19.7 33.1 44.3 HACNN (Li et al., 2018) 21.3 19.0 34.1 42.8 MLFN (Chang et al., 2018) 26.1 24.2 35.9 44.1 PCB (Sun et al., 2018) 32.8 28.1 37.4 46.9 OSNet (Zhou et al., 2019) 25.4 22.3 35.1 44.7 CAL (Rao et al., 2021) 27.6 24.3 36.5 45.7 HAMNet (Li et al., 2020) 27.7 26.3 41.5 51.7 PFNet (Zheng et al., 2021) 38.5 38.9 52.0 58.4 IEEE (Wang et al., 2022) 47.5 44.4 57.1 63.6 DENet (Zheng et al., 2023) 42.4 42.2 55.3 64.5 Uni Cat (Crawford et al., 2023) 57.0 55.7 - - HTT (Wang et al., 2024c) 71.1 73.4 83.1 87.3 EDITOR (Zhang et al., 2024a) 66.5 68.3 81.1 88.2 RSCNet (Yu et al., 2024) 68.2 72.5 - - TOP-Re ID (Wang et al., 2024a) 72.3 76.6 84.7 89.4 Ours 80.7 83.6 91.9 94.1 Implementation details. Our model is implemented using the Py Torch toolbox, and experiments are conducted on an NVIDIA V100 GPU. We utilize pre-trained CLIP (Radford et al., 2021) as the visual encoder. Images are resized to 256 128 for RGBNT201 and 128 256 for RGBNT100 and MSVR310. For data augmentation, we apply random horizontal flipping, cropping, and erasing following (Zhong et al., 2020). The mini-batch size is set to 128 for RGBNT100, and 64 for RGBNT201 and MSVR310, with corresponding sampling strategies for each dataset. We employ the Adam optimizer with an initial learning rate of 3.5e 4 and the learning rate of the visual encoder is 5e 6. The total number of training epochs is set to 45 for RGBNT201 and RGBNT100, and 50 for MSVR310. 4.2. Comparison with State-of-the-Art Methods multi-modal Person Re ID. We compare the proposed MFRNet with several single-spectral and multi-modal methods on the RGBNT201 dataset, as shown in Table 1. It is evident that multi-modal methods generally outperform single-spectral methods by leveraging complementary multimodal information. Among current multi-modal methods, TOP-Re ID achieves the best performance, with 72.3% m AP, 76.6% R-1, 84.7% R-5, and 89.4% R-10. Our proposed method improves these metrics by 8.4%, 6.9%, 7.2%, and 4.7%, which demonstrates the superior performance of MFRNet for multi-modal person Re ID. Notably, we use the CLIP-based Vi T to reproduce TOP-Re ID, and the results show that our method surpasses TOP-Re ID by approximately 9.7% m AP and 9.6% R-1. multi-modal Vehicle Re ID. Table 2 shows the comparison Table 2. Performance comparison on RGBNT100 and MSVR310. The best and second results are in bold and underlined, respectively. Methods RGBNT100 MSVR310 m AP R-1 m AP R-1 PCB (Sun et al., 2018) 57.2 83.5 23.2 42.9 MGN (Wang et al., 2018) 58.1 83.1 26.2 44.3 DMML (Chen et al., 2019) 58.5 82.0 19.1 31.1 Bo T (Luo et al., 2019) 78.0 95.1 23.5 38.4 OSNet (Zhou et al., 2019) 75.0 95.6 28.7 44.8 Circle Loss (Sun et al., 2020) 59.4 81.7 22.7 34.2 HRCN (Zhao et al., 2021) 67.1 91.8 23.4 44.2 Trans Re ID (He et al., 2021) 75.6 92.9 18.4 29.6 AGW (Ye et al., 2022) 73.1 92.7 28.9 46.9 HAMNet (Li et al., 2020) 74.5 93.3 27.1 42.3 PFNet (Zheng et al., 2021) 68.1 94.1 23.5 37.4 GAFNet (Guo et al., 2022) 74.4 93.4 - - Gra FT (Yin et al., 2023) 76.6 94.3 - - GPFNet (He et al., 2023) 75.0 94.5 - - PHT (Pan et al., 2023) 79.9 92.7 - - Uni Cat (Crawford et al., 2023) 79.4 96.2 - - CCNet (Zheng et al., 2022) 77.2 96.3 36.4 55.2 HTT (Wang et al., 2024c) 75.7 92.6 - - TOP-Re ID (Wang et al., 2024a) 81.2 96.4 35.9 44.6 EDITOR (Zhang et al., 2024a) 82.1 96.4 39.0 49.3 RSCNet (Yu et al., 2024) 82.3 96.6 39.5 49.6 Ours 88.2 97.4 50.6 64.8 results of our proposed MFRNet with other methods on the RGBNT100 and MSVR310 datasets. For the RGBNT100 dataset, Trans Re ID achieves the best performance among single-spectral methods, while RSCNet significantly outperforms it among multi-modal methods. However, for the MSVR310 dataset, AGW achieves the best performance among single-spectral methods, while RSCNet has the highest m AP and CCNet has the highest R-1 among multi-modal methods. Notably, our proposed method achieves significant improvements on both datasets, increasing m AP by 5.9% and R-1 by 0.8% on RGBNT100, and increasing m AP by 11.1% and R-1 by 9.6% on MSVR310. This demonstrates the superior performance of MFRNet for multi-modal vehicle re-identification. Evaluation on Missing-spectral Scenarios. In real-world application scenarios, persistent modality availability cannot be guaranteed. Therefore, exploring the performance of models under the condition of missing modalities helps to explore the application boundaries of the model. Therefore, we also evaluate the performance of MFRNet under different modality-missing scenarios on the RGBNT201 dataset. As shown in Table 3, PCB achieves the highest average m AP and R-1 among single-spectral methods. TOP-Re ID Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Table 3. Performance of missing-modality settings on RGBNT201. M (X) means missing the X image modality. The best and second results are in bold and underlined, respectively. Methods M (RGB) M (NIR) M (TIR) M (RGB+NIR) M (RGB+TIR) M (NIR+TIR) Average m AP R-1 m AP R-1 m AP R-1 m AP R-1 m AP R-1 m AP R-1 m AP R-1 MUDeep (Qian et al., 2017) 19.2 16.4 20.0 17.2 18.4 14.2 13.7 11.8 11.5 6.5 12.7 8.5 15.9 12.9 HACNN (Li et al., 2018) 12.5 11.1 20.5 19.4 16.7 13.3 9.2 6.2 6.3 2.2 14.8 12.0 13.3 10.7 MLFN (Chang et al., 2018) 20.2 18.9 21.1 19.7 17.6 11.1 13.2 12.1 8.3 3.5 13.1 9.1 15.6 12.4 PCB (Sun et al., 2018) 23.6 24.2 24.4 25.1 19.9 14.7 20.6 23.6 11.0 6.8 18.6 14.4 19.7 18.1 OSNet (Zhou et al., 2019) 19.8 17.3 21.0 19.0 18.7 14.6 12.3 10.9 9.4 5.4 13.0 10.2 15.7 12.9 PFNet (Zheng et al., 2021) - - 31.9 29.8 25.5 25.8 - - - - 26.4 23.4 - - DENet (Zheng et al., 2023) - - 35.4 36.8 33.0 35.4 - - - - 32.4 29.2 - - TOP-Re ID (Wang et al., 2024a) 54.4 57.5 64.3 67.6 51.9 54.5 35.3 35.4 26.2 26.0 34.1 31.7 44.4 45.4 Ours 64.7 65.2 72.3 76.1 51.6 49.5 41.4 43.4 27.3 27.9 37.2 35.6 49.1 49.6 Table 4. Comparison of computational cost with recent methods. We show the best result in bold. Methods Params(M) Flops(G) HTT (Wang et al., 2024c) 85.6 33.1 EDITOR (Zhang et al., 2024a) 117.5 38.6 TOP-Re ID (Wang et al., 2024a) 278.2 34.5 Ours 57.1 22.1 Table 5. Comparison with different modules. We show the best score in bold. Index Modules Metrics Params FLOPs FFM FRM m AP R-1 R-5 R-10 M G 1 69.2 76.3 84.8 89.5 57.2 22.1 2 75.1 78.1 87.4 92.0 60.8 23.5 3 77.8 80.9 88.6 92.2 53.5 20.7 4 80.7 83.5 91.9 94.1 57.1 22.1 outperforms it with an average performance of 44.4% m AP and 45.4% R-1 among multi-modal methods. Our proposed method improves these metrics by 4.7% m AP and 4.2% R-1 over TOP-Re ID, validating the adaptability of our model to modality missing scenarios. It is worth noting that compared to the TOP-Re ID which has a specific training phase for the modality missing condition, our MFRNet framework notably trains without explicit task-specific architectural modifications or dedicated training strategies, yet shows its intrinsic capability to compensate for the missing modality. 4.3. Ablation Study To explore the effectiveness of different components in our model, we conduct comprehensive ablation studies on the RGBNT201 dataset, as shown in Table 5. Feature Fusion Module. We use a shared vision transformer combined with label smoothing cross-entropy and triplet loss as our base framework. Building on this, we integrate the Feature Fusion Module (FFM), which boosts the performance on the m AP, R-1, R-5, and R-10 metrics by 5.9%, 1.8%, 2.6%, and 2.5% respectively. This shows that FFM effectively enhances overall model performance by generating and interacting multi-modal features using the fine-grained alignment of multi-modal data. However, Table 6. Performance analysis under different numbers of experts for FRM. We show the best score in bold. Number m AP R-1 R-5 R-10 Average 3 77.6 78.6 87.9 92.3 84.1 6 77.8 80.9 88.6 92.2 84.9 9 75.1 76.7 87.8 92.3 83.0 Table 7. Performance analysis of FRM at different locations within Vi T. We show the best score in bold. Layer m AP R-1 R-5 R-10 Average 10,11,12 71.6 75.5 87.1 90.8 81.2 10,12 75.7 77.6 87.6 91.7 83.1 12 77.8 80.9 88.6 92.2 84.9 integrating the FFM module increases the model s parameters and FLOPs by 3.6M and 1.4G, respectively, therefore further efforts are needed to reduce them. Feature Representation Module. Building on the aforementioned base framework, we introduce the Feature Representation Module (FRM). Experimental results show that integrating the FRM module into the baseline model improves m AP, R-1, R-5, and R-10 by 8.6%, 4.6%, 3.8%, and 2.7%, respectively. Notably, the model s parameters and FLOPs decrease by 3.7M and 1.4G as the FRM module uses Rep Adapter to build Mo E representations, reducing computational cost with two convolutional layers. This demonstrates that the FRM module enhances model accuracy by effectively representing shared and specific features of multi-modal data while using fewer parameters. MFRNet. MFRNet, which integrates the FFM and FRM modules, achieves 80.7% m AP, 83.5% R-1, 91.9% R-5, and 94.1% R-10. In addition, MFRNet reduces the parameters by 0.1M compared to the baseline framework, while being on par with the baseline model on FLOPS. Consequently, the design of effective multi-modal feature fusion and representation mechanisms enables more precise multi-modal object re-identification. Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Table 8. Performance analysis under different experts number for FFM. We show the best score in bold. Number m AP R-1 R-5 R-10 Average 1 78.7 81.0 90.8 93.5 86.0 2 74.9 78.6 86.8 90.8 82.8 3 80.7 83.5 91.9 94.1 87.5 6 76.9 80.4 88.2 90.6 84.0 9 79.2 82.3 90.7 93.5 86.4 10 76.4 80.1 88.0 92.0 84.1 11 74.1 78.9 87.2 92.2 83.1 12 74.1 77.8 86.8 91.1 82.4 4.4. More Analysis Analysis of model computational complexity. We compared computational costs across three major recent works. As shown in Table 4, our method achieves the lowest Params and FLOPs, measuring 57.1M and 22.1G, respectively. Experts Number in FRM. We explore the model s performance with different numbers of experts in the FRM, where experts are responsible for extracting modality-specific features. As shown in Table 6, the performance is optimal when the number of experts is 6. Although more experts may make sense in maintaining modality-specific representations, this may also limit the network s learning of modality-shared knowledge. The Location of FRM. We further explore the model s performance when the FRM module is placed in different layers of the Vi T. As shown in Table 7, using the FRM only in the final layer is sufficient to capture modality-specific features for multi-modal data, achieving optimal performance. Expert visualization of FRM. To visually observe the allocation of multiple experts in the FRM when extracting multi-modal features, we conduct visualizations on the RGBNT201 test set. The numbers in Figure 3 represent the index of the six experts. We also utilize similar colors to mark patches that are focused on by the same expert. It can be observed that Expert 0 concentrates on extracting pedestrian features in the NIR modality, while Experts 1, 2, and 3 focus on pedestrian regions in the RGB and TIR modalities. Additionally, Experts 4 and 5 tend to extract background features across all modalities. As our initial hypothesis predicted, semantically similar content achieves knowledge sharing through selecting analogous experts, while modality-specific representations that resist fusion maintain their distinct characteristics via dedicated expert allocation. Experts Number in FFM. Similarly, we analyze the model s performance with different numbers of experts in the FFM, where experts are responsible for generating cross- Figure 3. Visualization of expert in FRM. As expected earlier, similar content can share their knowledge by selecting similar experts to reduce redundancy, while for modality-specific representations that are difficult to fuse, their knowledge can also be retained by new experts. Table 9. Performance analysis under different λ for FFM. We show the best score in bold. Scale of λ m AP R-1 R-5 R-10 Average 0.0 4.7 2.4 9.8 18.4 8.8 0.3 79.9 82.1 89.2 92.8 86.0 0.5 80.7 83.5 91.9 94.1 87.5 0.7 74.8 77.8 86.1 90.4 82.3 1.0 77.5 81.5 88.8 92.5 85.1 modal features. As shown in Table 8, the model achieves the highest average performance when the number is 3. Parameter λ of FFM. For the parameter λ in the FFM, we analyze the model s performance at different values. As shown in Table 9, the model performs the worst when λ = 0, while achieving the highest average accuracy at λ = 0.5. The Location of FFM. We evaluate the approach by inserting FFM into the 3rd, 6th, and 9th layers of the network. Additionally, we evaluate the post-fusion method by positioning FFM after the Vi T feature encoding. As shown in Table 10, the pre-fusion method proves to be the most effective. We deem that high-level semantic representation may already lose most of the fine-grained perception and alignment characteristics, demonstrating the limitation of current methods that use the high-level semantic fusion strategy. Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Table 10. Performance analysis under different locations for FFM. We show the best score in bold. Location m AP R-1 R-5 R-10 Average 0(Before Vi T) 80.7 83.5 91.9 94.1 87.5 3 50.6 50.5 62.9 70.7 58.7 6 74.0 78.7 86.8 91.1 82.6 9 75.3 78.3 86.8 90.1 82.6 12(After Vi T) 76.7 79.7 86.8 92.0 83.8 5. Conclusion In this paper, we propose MFRNet, a novel framework for multi-modal object Re ID that enhances feature fusion and representation through the mixture of experts (Mo E). Specifically, we introduce a multi-modal feature fusion module (FFM) to enable fine-grained cross-modal interaction. Following this, the designed feature representation module (FRM) extracts modality-specific and modality-shared features through a parameter-efficient expert architecture. Extensive experiments on the RGBNT201, RGBNT100, and MSVR310 datasets demonstrate the effectiveness and efficiency of our model. Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none of which we feel must be specifically highlighted here. Alexey, D. An image is worth 16x16 words: Transformers for image recognition at scale. ar Xiv preprint ar Xiv: 2010.11929, 2020. Chang, X., Hospedales, T. M., and Xiang, T. Multi-level factorisation net for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018. Chen, G., Zhang, T., Lu, J., and Zhou, J. Deep meta metric learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019. Chen, Y. and Wang, L. emoe-tracker: Environmental moebased transformer for robust event-guided object tracking. IEEE Robotics and Automation Letters, 10(2):1393 1400, 2025. Chowdhury, M. N. R., Zhang, S., Wang, M., Liu, S., and Chen, P.-Y. Patch-level routing in mixture-of-experts is provably sample-efficient for convolutional neural networks. In Proceedings of the 40th International Conference on Machine Learning, volume 202, pp. 6074 6114. PMLR, 2023. Crawford, J., Yin, H., Mc Dermott, L., and Cummings, D. Unicat: Crafting a stronger fusion baseline for multimodal re-identification. ar Xiv preprint ar Xiv:2310.18812, 2023. Dai, D., Deng, C., Zhao, C., Xu, R., Gao, H., Chen, D., Li, J., Zeng, W., Yu, X., Wu, Y., et al. Deepseekmoe: Towards ultimate expert specialization in mixture-of-experts language models. ar Xiv preprint ar Xiv:2401.06066, 2024. Ganaie, M. A., Hu, M., Malik, A. K., Tanveer, M., and Suganthan, P. N. Ensemble deep learning: A review. Engineering Applications of Artificial Intelligence, 115: 105151, 2022. Gui, Y., Chen, M., Su, Y., Luo, G., and Yang, Y. Eegmamba: Bidirectional state space model with mixture of experts for eeg multi-task classification. ar Xiv preprint ar Xiv:2407.20254, 2024. Guo, J., Zhang, X., Liu, Z., and Wang, Y. Generative and attentive fusion for multi-spectral vehicle re-identification. In 2022 7th International Conference on Intelligent Computing and Signal Processing (ICSP), pp. 1565 1572, 2022. He, Q., Lu, Z., Wang, Z., and Hu, H. Graph-based progressive fusion network for multi-modality vehicle reidentification. IEEE Transactions on Intelligent Transportation Systems, 24(11):12431 12447, 2023. He, S., Luo, H., Wang, P., Wang, F., Li, H., and Jiang, W. Transreid: Transformer-based object re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 15013 15022, 2021. Hermans, A., Beyer, L., and Leibe, B. In defense of the triplet loss for person re-identification. ar Xiv preprint ar Xiv:1703.07737, 2017. Hwang, C., Cui, W., Xiong, Y., Yang, Z., Liu, Z., Hu, H., Wang, Z., Salas, R., Jose, J., Ram, P., Chau, H., Cheng, P., Yang, F., Yang, M., and Xiong, Y. Tutel: Adaptive mixture-of-experts at scale. In Proceedings of Machine Learning and Systems, volume 5, pp. 269 287. Curan, 2023. Jacobs, R. A., Jordan, M. I., Nowlan, S. J., and Hinton, G. E. Adaptive mixtures of local experts. Neural Computation, 3(1):79 87, 1991. Ji, J., Wang, H., Wu, C., Ma, Y., Sun, X., and Ji, R. Jm3d & jm3d-llm: Elevating 3d representation with joint multimodal cues. IEEE Transactions on Pattern Analysis and Machine Intelligence, 47(4):2475 2492, 2025. Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Kuang, Z., Zhang, H., Ye, M., Yang, B., Liu, Y., Huang, Y., Ding, X., and Li, H. Unity in diversity: Multiexpert knowledge confrontation and collaboration for generalizable vehicle re-identification. ar Xiv preprint ar Xiv:2407.07351, 2024. Lepikhin, D., Lee, H., Xu, Y., Chen, D., Firat, O., Huang, Y., Krikun, M., Shazeer, N., and Chen, Z. Gshard: Scaling giant models with conditional computation and automatic sharding. ar Xiv preprint ar Xiv:2006.16668, 2020. Li, B., Shen, Y., Yang, J., Wang, Y., Ren, J., Che, T., Zhang, J., and Liu, Z. Sparse mixture-of-experts are domain generalizable learners. ar Xiv preprint ar Xiv:2206.04046, 2022. Li, H., Li, C., Zhu, X., Zheng, A., and Luo, B. Multi-spectral vehicle re-identification: A challenge. Proceedings of the AAAI Conference on Artificial Intelligence, 34(07): 11345 11353, 2020. Li, W., Zhu, X., and Gong, S. Harmonious attention network for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2018. Li, X., Li, Q., Liang, F., and Wang, W. Multi-granularity pseudo-label collaboration for unsupervised person reidentification. Computer Vision and Image Understanding, 227:103616, 2023. Li, Y., Jiang, S., Hu, B., Wang, L., Zhong, W., Luo, W., Ma, L., and Zhang, M. Uni-moe: Scaling unified multimodal llms with mixture of experts. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2025. Lin, B., Tang, Z., Ye, Y., Cui, J., Zhu, B., Jin, P., Huang, J., Zhang, J., Pang, Y., Ning, M., et al. Moe-llava: Mixture of experts for large vision-language models. ar Xiv preprint ar Xiv:2401.15947, 2024. Liu, T., Liu, H., Shang, F., Yu, L., Han, T., and Wan, L. Completed feature disentanglement learning for multimodal mris analysis. IEEE Journal of Biomedical and Health Informatics, 2025. Liu, X., Zhang, P., Yu, C., Qian, X., Yang, X., and Lu, H. A video is worth three views: Trigeminal transformers for video-based person re-identification. IEEE Transactions on Intelligent Transportation Systems, 25(9):12818 12828, 2024. Lu, H., Zou, X., and Zhang, P. Learning progressive modality-shared transformers for effective visibleinfrared person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence, 37(2):1835 1843, 2023. Luo, G., Huang, M., Zhou, Y., Sun, X., Jiang, G., Wang, Z., and Ji, R. Towards efficient visual adaption via structural re-parameterization. ar Xiv preprint ar Xiv:2302.08106, 2023. Luo, H., Gu, Y., Liao, X., Lai, S., and Jiang, W. Bag of tricks and a strong baseline for deep person re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2019. Mustafa, B., Riquelme, C., Puigcerver, J., Jenatton, R., and Houlsby, N. Multimodal contrastive learning with limoe: the language-image mixture of experts. In Advances in Neural Information Processing Systems, volume 35, pp. 9564 9576. Curran Associates, Inc., 2022. Pan, W., Wu, H., Zhu, J., Zeng, H., and Zhu, X. H-vit: Hybrid vision transformer for multi-modal vehicle reidentification. In Artificial Intelligence, pp. 255 267, Cham, 2022. Springer Nature Switzerland. Pan, W., Huang, L., Liang, J., Hong, L., and Zhu, J. Progressively hybrid transformer for multi-modal vehicle re-identification. Sensors, 23(9), 2023. Park, H., Lee, S., Lee, J., and Ham, B. Learning by aligning: Visible-infrared person re-identification using crossmodal correspondences. In 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 12026 12035, 2021. Qian, X., Fu, Y., Jiang, Y.-G., Xiang, T., and Xue, X. Multi-scale deep learning architectures for person reidentification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017. Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748 8763. PMLR, 2021. Rao, Y., Chen, G., Lu, J., and Zhou, J. Counterfactual attention learning for fine-grained visual categorization and re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 1025 1034, 2021. Riquelme, C., Puigcerver, J., Mustafa, B., Neumann, M., Jenatton, R., Susano Pinto, A., Keysers, D., and Houlsby, N. Scaling vision with sparse mixture of experts. Advances in Neural Information Processing Systems, 34: 8583 8595, 2021. Shen, F. and Tang, J. Imagpose: A unified conditional framework for pose-guided person generation. Advances in neural information processing systems, 37:6246 6266, 2024. Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Shen, F., Jiang, X., He, X., Ye, H., Wang, C., Du, X., Li, Z., and Tang, J. Imagdressing-v1: Customizable virtual dressing. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 39, pp. 6795 6804, 2025. Shi, J., Zhang, Y., Yin, X., Xie, Y., Zhang, Z., Fan, J., Shi, Z., and Qu, Y. Dual pseudo-labels interactive selftraining for semi-supervised visible-infrared person reidentification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 11218 11228, 2023. Shi, J., Yin, X., Zhang, Y., Zhang, Z., Xie, Y., and Qu, Y. Learning commonality, divergence and variety for unsupervised visible-infrared person re-identification. In The Thirty-eighth Annual Conference on Neural Information Processing Systems, 2024. Sun, Y., Zheng, L., Yang, Y., Tian, Q., and Wang, S. Beyond part models: Person retrieval with refined part pooling (and a strong convolutional baseline). In Proceedings of the European Conference on Computer Vision (ECCV), 2018. Sun, Y., Cheng, C., Zhang, Y., Zhang, C., Zheng, L., Wang, Z., and Wei, Y. Circle loss: A unified perspective of pair similarity optimization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020. Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016. Tan, L., Dai, P., Ji, R., and Wu, Y. Dynamic prototype mask for occluded person re-identification. In Proceedings of the 30th ACM international conference on multimedia, pp. 531 540, 2022. Tan, L., Dai, P., Ye, Q., Xu, M., Wu, Y., and Ji, R. Spectral aware softmax for visible-infrared person reidentification. ar Xiv preprint ar Xiv:2302.01512, 2023. Tan, L., Dai, P., Chen, J., Cao, L., Wu, Y., and Ji, R. Partformer: Awakening latent diverse representation from vision transformer for object re-identification. ar Xiv preprint ar Xiv:2408.16684, 2024a. Tan, L., Xia, J., Liu, W., Dai, P., Wu, Y., and Cao, L. Occluded person re-identification via saliency-guided patch transfer. In Proceedings of the AAAI conference on artificial intelligence, volume 38, pp. 5070 5078, 2024b. Tan, L., Zhang, Y., Han, K., Dai, P., Zhang, Y., WU, Y., and Ji, R. Rle: A unified perspective of data augmentation for cross-spectral re-identification. In Advances in Neural Information Processing Systems, volume 37, pp. 126977 126996, 2024c. Wang, G., Yuan, Y., Chen, X., Li, J., and Zhou, X. Learning discriminative features with multiple granularities for person re-identification. In Proceedings of the 26th ACM International Conference on Multimedia, pp. 274 282, 2018. Wang, Y., Liu, X., Zhang, P., Lu, H., Tu, Z., and Lu, H. Topreid: Multi-spectral object re-identification with token permutation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(6):5758 5766, 2024a. Wang, Y., Zhang, P., Wang, D., and Lu, H. Other tokens matter: Exploring global and local features of vision transformers for object re-identification. Computer Vision and Image Understanding, 244:104030, 2024b. Wang, Z., Li, C., Zheng, A., He, R., and Tang, J. Interact, embed, and enlarge: Boosting modality-specific representations for multi-modal person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence, 36 (3):2633 2641, 2022. Wang, Z., Huang, H., Zheng, A., and He, R. Heterogeneous test-time training for multi-modal person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence, 38(6):5850 5858, 2024c. Xu, B., Liang, J., He, L., and Sun, Z. Mimic embedding via adaptive aggregation: Learning generalizable person re-identification. In European Conference on Computer Vision, pp. 372 388. Springer, 2022. Yang, B., Chen, J., and Ye, M. Shallow-deep collaborative learning for unsupervised visible-infrared person reidentification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16870 16879, 2024. Ye, M., Shen, J., Lin, G., Xiang, T., Shao, L., and Hoi, S. C. H. Deep learning for person re-identification: A survey and outlook. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(6):2872 2893, 2022. Yin, H., Li, J., Schiller, E., Mc Dermott, L., and Cummings, D. Graft: Gradual fusion transformer for multimodal re-identification. ar Xiv preprint ar Xiv:2310.16856, 2023. Yu, Z., Huang, Z., Hou, M., Pei, J., Yan, Y., Liu, Y., and Sun, D. Representation selective coupling via token sparsification for multi-spectral object re-identification. IEEE Transactions on Circuits and Systems for Video Technology, 2024. Zhang, G., Zhang, P., Qi, J., and Lu, H. Hat: Hierarchical aggregation transformers for person re-identification. In Proceedings of the 29th ACM International Conference on Multimedia, pp. 516 525, 2021. Multi-Modal Object Re-Identification via Sparse Mixture-of-Experts Zhang, P., Wang, Y., Liu, Y., Tu, Z., and Lu, H. Magic tokens: Select diverse tokens for multi-modal object reidentification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 17117 17126, 2024a. Zhang, Y. and Wang, H. Diverse embedding expansion network and low-light cross-modality benchmark for visibleinfrared person re-identification. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2153 2162, 2023. Zhang, Y., Yan, Y., Lu, Y., and Wang, H. Adaptive middle modality alignment learning for visible-infrared person reidentification. International Journal of Computer Vision, pp. 1 21, 2024b. Zhao, J., Zhao, Y., Li, J., Yan, K., and Tian, Y. Heterogeneous relational complement for vehicle re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), pp. 205 214, 2021. Zheng, A., Wang, Z., Chen, Z., Li, C., and Tang, J. Robust multi-modality person re-identification. Proceedings of the AAAI Conference on Artificial Intelligence, 35(4): 3529 3537, 2021. Zheng, A., Zhu, X., Ma, Z., Li, C., Tang, J., and Ma, J. Multi-spectral vehicle re-identification with crossdirectional consistency network and a high-quality benchmark. ar Xiv preprint ar Xiv:2208.00632, 2022. Zheng, A., He, Z., Wang, Z., Li, C., and Tang, J. Dynamic enhancement network for partial multi-modality person re-identification. ar Xiv preprint ar Xiv:2305.15762, 2023. Zhong, Z., Zheng, L., Kang, G., Li, S., and Yang, Y. Random erasing data augmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 34(07):13001 13008, 2020. Zhou, K., Yang, Y., Cavallaro, A., and Xiang, T. Omniscale feature learning for person re-identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), 2019.