# connecting_multimodal_contrastive_representations__14c7a02e.pdf Connecting Multi-modal Contrastive Representations Zehan Wang1 Yang Zhao2 Xize Cheng1 Haifeng Huang1 Jiageng Liu1 Li Tang1 Linjun Li1 Yongqi Wang1 Aoxiong Yin1 Ziang Zhang1 Zhou Zhao1, 3 1Zhejiang University 2Byte Dance 3Shanghai AI Laboratory {wangzehan01}@zju.edu.cn Multi-modal Contrastive Representation (MCR) learning aims to encode different modalities into a semantically aligned shared space. This paradigm shows remarkable generalization ability on numerous downstream tasks across various modalities. However, the reliance on massive high-quality data pairs limits its further development on more modalities. This paper proposes a novel training-efficient method for learning MCR without paired data called Connecting Multi-modal Contrastive Representations (C-MCR). Specifically, given two existing MCRs pre-trained on (A, B) and (B, C) modality pairs, we project them to a new space and use the data from the overlapping modality B to aligning the two MCRs in the new space. Meanwhile, since the modality pairs (A, B) and (B, C) are already aligned within each MCR, the connection learned by overlapping modality can also be transferred to non-overlapping modality pair (A, C). To unleash the potential of C-MCR, we further introduce a semantic-enhanced interand intra-MCR connection method. We first enhance the semantic consistency and completion of embeddings across different modalities for more robust alignment. Then we utilize the inter-MCR alignment to establish the connection, and employ the intra-MCR alignment to better maintain the connection for inputs from non-overlapping modalities. To demonstrate the effectiveness of C-MCR, we take the field of audio-visual and 3Dlanguage learning as examples. Specifically, we connect CLIP and CLAP via texts to derive audio-visual representations, and integrate CLIP and ULIP via images for 3D-language representations. Remarkably, without using any paired data, C-MCR for audio-visual achieves state-of-the-art performance on audio-image retrieval, audio-visual source localization, and counterfactual audio-image recognition tasks. Furthermore, C-MCR for 3D-language also attains advanced zero-shot 3D point cloud classification accuracy on Model Net40. Our project page is available at https://c-mcr.github.io/C-MCR/ 1 Introduction Multi-modal Contrastive Representation (MCR) learning aims to map inputs from different modalities to a shared representation space. With the impressive generalization performance of vision-language contrastive pre-training models [1, 2, 3, 4] demonstrated on various downstream tasks [5, 6, 7, 8, 9, 10], learning MCR spaces between multiple modalities has become a promising area of research, attracting increasing attention [11, 12, 13, 14, 15]. However, the generalization ability of MCR primarily benefits from the accessibility of massive data pairs from the web. For modalities where obtaining semantically matching data pairs is significantly more costly, the representations directly learned from limited data pairs are unreliable. On the other Corresponding author. 37th Conference on Neural Information Processing Systems (Neur IPS 2023). hand, these modality pairs with little direct paired data often have a large number of paired data with the same intermediate modality. For example, although audio-visual data are often vague, paired data of audio-language and language-image are sufficient and semantically explicit. Similarly, while 3D point-language pairs are rare, 3D point-image and image-language data are extensive. Consider that there are already many MCRs between modalities with sufficient paired data. In this paper, we propose Connecting Multi-modal Contrastive Representations (C-MCR), a novel trainingefficient and paired-data-free MCR learning method that extends the learned alignment knowledge in existing MCRs to more modalities. With regard to the overlapping modality, its representations in two MCRs are just different data views sharing the same inherent semantics. So we can take them as positive pairs to connect different MCRs. As modalities within each MCR are semantically aligned, the connections built from overlapping modalities can also be applied to non-overlapping modalities. The advantages of our C-MCR are two-fold: (1) Flexible. C-MCR enables MCR learning on modalities with limited paired data. More importantly, C-MCR treats each learned MCR space as a node and the overlapping modalities between different MCRs as links. Connecting the various isolated MCRs greatly extends the obtained multi-modal alignment knowledge, and discovers generalized contrastive representations of broader modalities. (2) Training-Efficient. Since C-MCR simply reprojects the learned representations into a new space, only two simple projectors are learnable during training. The training parameters and costs for connecting existing MCRs are very small. However, two factors impede the acquisition of a robust and transferable connection: Firstly, embeddings in MCR spaces are incapable of comprehensively reflecting all the semantic information of the input, and this loss of meaning would be inherited and amplified, thereby compromising the robustness of the connection. Secondly, as discussed in [16], MCR spaces exhibit a modality gap phenomenon, i.e., the embeddings of different modalities are located in two completely separate regions in each MCR space. This poses a challenge for maintaining the connection based on overlapping modality while facing inputs from non-overlapping modalities. Considering the above challenges, we propose a semantic-enhanced interand intra-MCR connection method. During training, the copious amounts of easily accessible unpaired unimodal data are first encoded into embeddings in two MCR spaces. We inject Gaussian noise into all the embeddings to mitigate the semantic bias, enhance the semantic completeness, and improve robustness. For directly quantifying the modality gap and the relationship between non-overlapping modalities, we exploit the inherent multi-modal alignment in MCR spaces to cluster semantic consistent embeddings and bridge different modalities. With the above strategies, we align the semantic-enhanced embeddings across different MCR spaces in a contrastive manner to establish the connection. To preserve the connection for inputs from non-overlap modalities, we realign the semantic similar embeddings across modalities within each MCR space to alleviate the modality gap. Our main contributions are summarized as follows: (1) We propose Connecting Multi-modal Contrastive Representations (C-MCR), a novel paired-datafree and training-efficient method for MCR learning. By connecting existing MCR spaces with simple projectors, we can mine the multi-modal alignment knowledge in existing MCR space, and extend MCRs on more modalities that lack large-scale high-quality data pairs. (2) We further propose a semantic-enhanced interand intra-MCR connection method to unleash our C-MCR. This approach establishes a transferable connection between two MCR spaces via overlapping modality and maintains it for non-overlapping modalities. (3) To demonstrate the effectiveness of C-MCR, we connect the CLIP and CLAP through texts to acquire audio-visual representations, and interage CLIP and 3D-image MCR space (ULIP) via images for 3D-language representations. Remarkably, without requiring any pair data or fine-tuning, C-MCR for audio-visual achieves state-of-the-art performance on six datasets across three downstream audiovisual tasks. Furthermore, C-MCR for 3D-language also attains advanced zero-shot 3D point cloud classification accuracy on Model Net40. 2 Related Work Multi-modal Contrastive Representation learning. Multi-modal contrastive representation focuses on learning separate unimodal encoders for different modalities, which can map inputs from different modalities into a shared representation space. These models are pre-trained on large-scale paired data Figure 1: The pipeline of connecting CLIP and CLAP using our C-MCR. During training, we take text as input and encode it with frozen CLAP and CLIP text encoders, respectively. Audio(image) memory is generated by encoding lots of unimodal audio(image) data by pre-trained audio(image) encoder. Semantic enhancement enriches the semantic consistency and completion of embeddings. Then two projectors map embeddings to a new shared representation space where interand intra MCR alignment establishes and maintains a stable connection between CLAP and CLIP. During inference, the audio and image are inputted to the corresponding encoder and projector. using a contrastive loss. Recent vision-language contrastive pre-training models, such as CLIP [1] and ALIGN [2], demonstrate impressive zero-shot retrieval and classification performance and remarkable generalization capability across diverse downstream tasks [5, 6, 7, 8, 9, 10]. Inspired by the success of vision-language models, contrastive representation learning across more modalities has garnered increasing attention. CLAP [12, 11] construct a contrastive language-audio pre-training model by collecting large-scale audio-text pairs from diverse data sources. ULIP [17, 18] collect and generate 3D-image-text triplet data via 3D rendering and image captioning, and learn an extra 3D encoder for existing vision-language space. Audio CLIP [19] and WAV2CLIP [20] leverage the pre-trained CLIP image encoder and acquire audio-visual representations by training on audio-image pairs from Audio Set [21] and VGGSound [22], respectively. For certain modality pairs, such as audio-visual and 3D-language, the pre-training model s generalization capability is restricted by ambiguous or limited paired data. Our proposed method introduces a novel method for better contrastive representation learning on these modalities. Audio-Visual Learning. Audio-visual learning [23] aims to exploit the relationship between audio and visual modalities, which is an essential part of intelligent multi-modal perception research. Previous methods primarily focus on learning specific audio-visual downstream tasks (such as retrieval [24, 25, 26], localization [27, 28, 29, 30, 31, 32, 33, 34, 35, 36], or generation [37, 38, 39, 40, 41, 42, 43, 44]) within limited domains, based on the manually cleaned small-scale datasets. Recently, several large-scale audio-image datasets [21, 22] collected from the web are proposed. However, these datasets contain many noisy image-audio pairs due to the ambiguous nature of both images and audio and the presence of non-visible sounds and silent objects in videos. Consequently, the generalization ability of audio-visual contrastive representation [19, 20] learned from these datasets is limited. Our C-MCR reduces the need for larger-scale high-quality data pairs. By extending the knowledge in CLIP and CLAP models, we acquire powerful audio-visual contrastive representations that exhibit powerful generalization capabilities across various downstream tasks. 3D-language Learning. 3D vision is an important way for robots to perceive the rich semantic and spatial information of the real world. The 3D-language learning including recognition [45, 46, 47, 48], localization [49, 50, 51, 52, 53], question-answer [54, 55] and general conversation [56, 57], have attracted increasing attentions. 3D-language contrastive representations are vital for the further development of 3D-language learning. However, due to the scarcity of 3D-language paired data, the development of 3D-language representation is limited. Recent ULIP [17, 18] focus on generating 3D-image-text triplet data, but they are still limited by the relatively low quality of training datasets. C-MCR gets rid of the dependence on 3D-language pairing data, and instead connects the reliable 3D-visual representation of ULIP and the visual-language representation of CLIP via images to obtain a more robust 3D-language contrastive representation. In this section, we take connecting CLIP and CLAP for audio-visual, as an example to introduce C-MCR. As depicted in Figure 1 (a), we utilize two projectors to connect CLIP and CLAP through texts. Before delving into our method, we first introduce the mathematical formulations and revisit the multi-modal contrastive learning in Section 3.1. Then we discuss our semantic enhancement approach for robust representation alignment in Section 3.2. This is followed by the inter-MCR alignment to establish the connection between CLIP and CLAP in Section 3.3, and the intra-MCR alignment to ensure the connection can be maintained for image-audio inputs in Section 3.4. 3.1 Background Problem formulation. For text inputs, the embeddings obtained by CLIP and CLAP encoder can be denoted as t I Rc and t A Rd respectively. Our C-MCR method aims to leverage the inherent consistency between t I and t A and the multi-modal alignment in MCR to learn two projectors f1( ) and f2( ) that map the representations from CLIP and CLAP to a new shared representation space. The connection between CLIP and CLAP, learned from texts, can effectively accommodate audio and image inputs. Multi-modal Contrastive Learning. Given N paired instances from two different modalities, we map the i-th pair to L2-normalized embeddings xi and zi via two encoders. Multi-modal contrastive learning aims to maximize the cosine similarity between xi and zi and minimize the cosine similarity between xi and zj where i = j. The contrastive loss can be formulated as: log exp(sim(xi, zi)/τ) PN j=1 exp(sim(xi, zj)/τ) + log exp(sim(zi, xi)/τ) PN j=1 exp(sim(zi, xj)/τ) where τ is the temperature parameter and the sim( , ) is the operator for cosine distance. The contrastive loss is based on multi-modal data pairs, and the generalization of the learned representation relies on the scale and quality of the data pairs. To extend contrastive representation learning on more modalities, we propose using overlapping modalities to connect two learned MCR spaces and extend the learned multi-modal alignment knowledge to non-overlapping modalities. 3.2 Semantic Enhancement To achieve more robust and comprehensive alignment, we enhance the semantics from two perspectives: inter-modality semantic consistency and intra-modality semantic completion. Inter-modality Semantic Consistency. CLIP and CLAP have already learned shared image-text and audio-text representations. To better quantify the modality gap in the MCR space and directly explore the correlation between audio and image, we first utilize the inherent modality alignment properties of CLIP and CLAP to generate semantically consistent embeddings across modalities. Specifically, we encode massive unpaired images and audio using the CLIP image encoder and CLAP audio encoder, respectively. All the obtained image embeddings are served as image memory V = {v1, v2, ..., v N} and the audio embeddings are audio memory A = {a1, a2, ..., a M}, where N, M indicate the number of images and audios. Considering i-th text embeddings t I i and t A i , we can generate image embedding v I i and audio embedding a A i that are semantically similar to i-th text. exp(sim(t I i , vk)/τ1) PN j=1 exp(sim(t I i , vj)/τ1) vk; a A i = exp(sim(t A i , ak)/τ1) PM j=1 exp(sim(t A i , aj)/τ1) ak (2) The τ1 is the temperature hyperparameter. By dynamically absorbing information from memories based on semantic similarity to the text embeddings t I i and t A i , we can generate more diverse and accurate semantically-consistent embeddings v I i and a A i . Intra-modality Semantic Completion. The semantics in the original input data are often complex, and some information is inevitably lost when encoding it into the MCR space. When connecting and aligning existing representation spaces, this loss and bias of meaning will be inherited and amplified, affecting the robustness of alignment. To enhance the semantic completeness of each embedding, we propose to serve Gaussian noise as an information augmentation method. Specifically, we add zero-mean Gaussian noises into the embeddings and re-normalize them to the unit hypersphere: t I = Normalize(t I + θ1); v I = Normalize(v I + θ2) t A = Normalize(t A + θ3); a A = Normalize(a A + θ4) (3) where noise items θ1, θ2 Rc and θ3, θ4 Rd are sampled from zero-mean gaussian distribution with variance σ2, and they are not learnable. Since the MCRs are L2 normalized, all embeddings are distributed on a unit sphere. As illustrated in Figure 1 (c), each embedding can be viewed as a point on the unit sphere s surface. The addition of Gaussian noise can transform the point into a small sphere, and re-normalizing projects the small sphere onto a circle on the surface of the unit sphere. Hence, aligning two embeddings with noise forces the model to acquire the ability to align all the embeddings within the two circles. In the MCR space, the closer two embeddings are to each other, the more similar their semantics are. Embeddings within the same circle share similar general semantics, and the semantics represented by the circle are more comprehensive and robust than the original embedding. 3.3 Inter-MCR Alignment To establish the connection between two MCRs, we project the semantic-enhanced embeddings from CLIP and CLAP space to a new shared space via two learnable projectors f1( ) and f2( ), respectively. ˆt I = f1( t I); ˆv I = f1( v I); ˆt A = f2( t A); ˆa A = f2( a A) (4) In the newly projected space, our objective is to ensure that embeddings with similar semantics from different MCR spaces are in close proximity to each other. The (t I i ,t A i ) from the same text is naturally semantic consistent, and it can be considered as a ground-truth pair label. Besides, there is pseudo consistency in (v I i ,t I i ) and (a A i ,t A i ) due to the multi-modal alignment properties in CLIP and CLAP. Thus the (ˆv I,ˆa A) derived from (t I i ,t A i ) can be viewed as a pseudo pair label. For a robust and stable connection of the two MCR, we propose to align both (ˆt I,ˆt A) and (ˆv I,ˆa A). The text-text contrastive loss Lttc and audio-visual contrastive loss Lavc are defined as: log exp(sim(ˆt I i ,ˆt A i )/τ2) PB j=1 exp(sim(ˆt I i ,ˆt A j )/τ2) + log exp(sim(ˆt A i ,ˆt I i )/τ2) PB j=1 exp(sim(ˆt A i ,ˆt I j)/τ2) log exp(sim(ˆv I i , ˆa A i )/τ3) PB j=1 exp(sim(ˆv I i , ˆa A j )/τ3) + log exp(sim(ˆa A i , ˆv I i )/τ3) PB j=1 exp(sim(ˆa A i , ˆv I j)/τ3) B corresponds to the batch, and τ2, τ3 are the temperature hyperparameters. The inter-MCR alignment loss Linter is the combination of the two contrastive losses: Linter = Lttc + Lavc (7) The Lttc and Lavc are complementary to each other. The semantics between (t I,t A) are highly consistent, thus the connection learned from them is much more robust, but their alignment is indirect for audio-visual representation. On the other hand, (v I,a A) pairs are directly beneficial to audiovisual representation learning, but their semantic coherence is less reliable. Note that since the semantic consistency in (v I,a A) is derived from (t I,t A), the connection learned from pseudo pair (v I,a A) can still be considered as being established via overlapping modalities. 3.4 Intra-MCR Alignment As discussed in [16], there exists a phenomenon known as the modality gap in MCR spaces. Although the embeddings from different modalities are semantically aligned in an MCR space, they are distributed in entirely distinct regions of the representation space. This implies that the more stable connection learned from (t I i ,t A i ) may not accommodate the inputs from audio and image. To better maintain the connection, we propose closing the modality gap and guaranteeing that embeddings from different modalities with similar semantics are distributed in the same region of the representation space. The analysis in [16] suggests that the repulsive structure in contrastive loss preserves the modality gap. Inspired by this observation, we derive the intra-MCR alignment loss by removing the repulsive structure in the contrastive loss. As introduced in 3.1, a typical contrastive item can be formulated as: log exp(sim(xi, zi)/τ) PN j=1 exp(sim(xi, zj)/τ) = sim(xi, zi)/τ | {z } pull positive close j=1 exp(sim(xi, zj)/τ) | {z } push negative away We only retain the mechanism of pulling samples closer together and remove the repulsive effect between negative pairs, which helps to close the modality gap in the newly learned MCR space. In the L2-normalized MCR space, there are (xi yi)T (xi yi) = 2(1 x T i yi). After removing the gradient-irrelevant constant terms, our intra-MCR alignment loss Lintra can be expressed as: i=1 ( ˆt I i ˆv I i 2 + ˆt A i ˆa A i 2) (9) By realigning text-guided cross-modal semantically consistent embeddings in each MCR space, i.e., aligning (ˆt I i , ˆv I i ) for CLIP and (ˆt A i , ˆa A i ) for CLAP, the modality gap between embeddings from same MCR can be effectively alleviated in the new space. As a result, the more stable connection provided by Equation 5 can be maintained for audio-visual inputs. 3.5 Training and Inference During training, all pre-trained encoders in CLIP and CLAP are frozen to preserve the semantic correspondences between image-text and audio-text, and only the two projectors are learnable. To make the training more efficient, we pre-extract the text embeddings t I i and t A i . Since the semantic enhancements are training-free, the inter-modality semantic consistency strategy can also be precomputed, and the semantically consistent image v I i and audio embedding a A i are stored offline. We apply a combination of interand intra-MCR alignment loss to optimize the two projectors for establishing a stable connection between CLIP and CLAP representation spaces, formulated as: L = Linter + λLintra (10) λ is the hyper-parameter to balance the two terms. During inference, as shown in Figure 1 (b), the image embedding in CLIP and the audio embedding in CLAP can be mapped into a shared space through corresponding projectors. The cosine scores in this space reflect the semantic similarity between images and audio. 4 Experiments 4.1 Details of Connecting CLAP and CLIP Text Datasets. We collected texts from three sources: image-text datasets (COCO [58] and CC3M [59]), video-text datasets (MSRVTT [60], MAD [61]), and audio-text datasets (Audio Cap [62], Clotho [63]), to ensure that the texts contain sufficient visual, action, and audio information. To avoid overfitting visual information, we randomly selected one million descriptions from CC3M. In summary, the texts from image-text, video-text, and audio-text are 1.66M, 0.58M, and 77K, respectively, and there are 2.33M texts in total. Audio/Image Memory. Audio Set [21] provides a vast collection of audio snippets from You Tube videos. All 1.8M audio data in the training set are encoded by the CLAP audio encoder to serve as the audio memory. Image Net1K [64] is a large-scale image recognition dataset. We encoded all the 1.3M images in the train set of Image Net1K using the CLIP image encoder to construct the image memory. It is worth noting that no annotations related to the audio and images are used. Implementation Details. We employ a frozen pre-trained CLIP Vi T-B/32 model [1] and CLAP model [13]. We adopt simple multi-layer perceptrons as our projectors f1( ) and f2( ). The τ1, τ2 and τ3 in Euqation 2, 5 and 6 are all set to 1/100. The variance σ2 of the noises in Equation 3 is set as 0.004. The hyper-parameter λ in Equation 10 is set to 0.1. We train our projectors for 36 epochs using a batch size of 10240. We use the Adam W optimizer with the initial learning rate 1e 3 and the cosine learning rate decay strategy. 4.2 Details of Connecting ULIP and CLIP Image Datasets. The image dataset used for connecting ULIP and CLIP is Image Net1K [64], total 1.3M images without any annotations. Text/3D Memory. We use the same 2.33M text dataset as described in Section 4.1, to construct the corresponding text memory. The 3D object point clouds from the training set of Objaverse [48] are utilized to construct 3D memory, 0.8M samples in total. Implementation Details. We employ a frozen pre-trained CLIP Vi T-B/32 model [1], and ULIP-2 Point BERT model [65, 13] pre-trained on ULIP-Objaverse triplets. The structure of the projector and the temperature parameters remain the same in Section 4.1. The variance σ2 of the noises in Equation 3 is set as 0.002. The hyper-parameter λ in Equation 10 is set to 0.4. We train our projectors for 24 epochs using a batch size of 8192. We also use the Adam W optimizer with the initial learning rate 5e 3 and the cosine learning rate decay strategy. 4.3 Evaluation of Audio-Visual Representations 4.3.1 Downstream Audio-Visual Tasks We assess the quality of audio-visual representations on three downstream audio-visual tasks in a zero-shot manner. More details about the datasets and implementation details are in Appendix. Audio-Image Retrieval. It contains two subtasks: image-to-audio retrieval (I2A) and audio-toimage retrieval (A2I). We assess the zero-shot image-audio retrieval on the AVE [66] and Flickr Sound Net [32]. Due to the small size of the test sets in both datasets, we utilized all available data in the train, eval, and test sets for evaluation, resulting in 4095 samples for AVE and 5000 samples for Flickr-Sound Net. For zero-shot inference, we encode all audio and images into our newly learned audio-visual MCR space and computed the cosine similarity for all audio-image pairs. The m AP, Top-1, and Top-5 metrics are used to evaluate retrieval accuracy. Audio-Visual Source Localization. Audio-visual source localization aims to localize the visual sound sources in an image. The test sets of widely-used VGGSS [67] and MUSIC [68] benchmarks are employed for evaluation. To enable zero-shot inference, we first use a pre-trained object detector [3] to extract object proposals from the images and calculate the cosine similarity between each proposal and audio in our representations space. The proposal with the highest similarity score is token as the final prediction. We adopt Consensus Intersection over Union (c Io U) and Area Under Curve(AUC) metrics following [28, 69]. Counterfactual Audio-Image Recognition. For the non-visible sounds and images with silent objects, this task requires a model to distinguish the semantically unpaired audio-image from audioimage pairs. During the zero-shot inference phase, we employ an object detector [3] to extract object proposals from the image. Subsequently, for each image, the proposal with the highest matching score is considered the predicted object, and the matching score is regarded as the confidence score for this prediction. Experiments are conducted on the Extended VGGSS (Ex-VGGSS) [69] and Extended Flickr-Sound Net (Ex-Flickr Net) [69], and the comparison is based on the Average Precision (AP) and maximum F1 (Max-F1) metrics following [69]. In summary, these three tasks can evaluate a group of audio-visual contrastive representations from various perspectives. Audio-image retrieval is employed to assess the ability to match coarse-grained images and audio, audio-visual source localization is used to evaluate the ability to match fine-grained objects and audio, and counterfactual audio-image recognition is used to evaluate the understanding and reasoning ability of audio and visual inputs. 4.3.2 Analysis on Zero-shot Image-Audio Retrieval We compared our model with Audio CLIP [19] and WAV2CLIP [20] which are contrastively pretrained on image-audio pairs from Audio Set [21] and VGGSound [22], respectively. Results in Table 1 demonstrate that C-MCR achieves state-of-the-art zero-shot retrieval performance. Besides, the generalization ability of Audio CLIP and WAV2CLIP is not stable. For instance, WAV2CLIP performs well on AVE but poorly on Flickr-Sound Net, while Audio CLIP achieves good results on Flickr-Sound Net but poor accuracy on AVE. Similar situations can also be observed in Table 2. In contrast, our C-MCR exhibits stronger and more stable generalization ability. Moreover, since Table 1: Zero-shot audio-image retrieval results on AVE and Flickr-Sound Net. A-V pairs stands for whether training on paired audio-visual data; Tr. Param for Trainable parameters number. Method A-V Pairs Tr. Param AVE Flickr-Sound Net A2I I2A A2I I2A m AP R@1 R@5 m AP R@1 R@5 m AP R@1 R@5 m AP R@1 R@5 Random - - 0.25 0.02 0.12 0.25 0.02 0.12 0.17 0.02 0.06 0.17 0.02 0.06 WAV2CLIP 11.7M 2.80 0.76 3.08 4.01 1.14 4.42 2.52 0.58 3.16 3.47 1.12 4.34 Audio CLIP 134.1M 0.98 0.22 0.85 2.50 1.00 2.83 3.10 1.00 4.02 4.43 1.58 5.92 C-MCR 2.1M 4.11 1.25 4.54 4.13 1.25 4.44 4.57 1.38 5.40 4.92 1.58 5.98 Figure 2: Visualization of audio-to-image retrieval on AVE and Flickr-Sound Net. C-MCR does not utilize any paired data and has much fewer learnable parameters, Audio CLIP and WAV2CLIP are not truly fair baselines to compare C-MCR with. Nevertheless, C-MCR still demonstrates superior performance compared to these pre-trained audio-visual models. Figure 2 provides a few visualizations of audio-to-image retrieval. 4.3.3 Analysis on Zero-shot Audio-Visual Source Localization Table 2 presents the zero-shot audio-visual source localization performance on MUSIC-Solo and VGGSS datasets and the comparison with previous audio-visual source localization methods. Remarkably, despite not using any audio-visual paired data and any fine-tuning, C-MCR demonstrates state-of-the-art performance, achieving a relative improvement of around 25% over the previous leading methods. Additionally, to demonstrate that the improvements are not by introducing the powerful object detector, we also performed the same zero-shot inference using audio-visual representations in Audio CLIP and WAV2CLIP. These methods exhibit unstable generalization performance on the two datasets, and our C-MCR demonstrates a significantly better overall performance than both. These results show that the state-of-the-art performances in audio-visual source localization mainly benefit from the stronger and more robust fine-grained audio-visual matching capability. 4.3.4 Analysis on Zero-shot Counterfactual Audio-Image Recognition Table 3 shows the comparisons on counterfactual audio-image recognition. Our C-MCR significantly outperforms previous methods that trained on the training set and exhibits overall improvement compared to other audio-visual representation models. The state-of-the-art performance on zeroshot counterfactual audio-image recognition further demonstrates the superiority of our method in understanding the deep semantic relationship between audio and visual modalities. Table 2: Zero-shot audio-visual source localization on MUSIC-Solo and VGGSS. Method MUSIC-Solo VGGSS c Io U AUC c Io U AUC Attention [32] 37.20 38.70 17.10 28.70 DMC [31] 29.10 38.00 23.90 - DSOL [33] 51.40 43.60 29.91 - TURN [28] 33.70 45.20 34.60 39.10 EZ-VSL [36] - - 38.85 39.54 SLAVC [69] - - 39.80 - WAV2CLIP [20] 47.49 53.80 36.91 39.58 Audio CLIP [19] 30.56 37.16 43.93 45.96 C-MCR(Ours) 53.78 56.09 48.08 48.69 Table 3: Zero-shot counterfactual audio-image recognition on Ex-VGGSS and Ex-Flickr Net. Method Ex-VGGSS Ex-Flickr Net AP Max-F1 AP Max-F1 Attention [32] 6.70 13.10 15.98 24.00 DMC [31] 11.53 20.30 25.56 41.80 DSOL [33] 16.84 25.60 38.32 49.40 OGL [36] 18.73 30.90 40.20 55.70 EZ-VSL [36] 27.71 34.60 48.75 56.80 SLAVC [69] 34.46 41.50 52.15 60.10 WAV2CLIP [20] 33.86 47.69 60.54 66.20 Audio CLIP [19] 42.59 55.43 72.78 71.98 C-MCR(Ours) 50.91 58.98 73.67 74.02 4.4 Evaluation of 3D-language Representations Table 4: Zero-shot 3D point cloud classification results on Model Net40. Method Top1 Top3 Top5 Re Con [70] 61.2 73.9 78.1 CG3D [15] 48.7 60.7 66.5 ULIP [17] 60.4 79.0 84.4 ULIP-2 [18] 74.0 86.5 90.0 C-MCR 64.9 87.0 92.8 In order to verify the performance of the 3D-language representation obtained by connecting ULIP-2 and CLIP, we evaluate the zero-shot 3D point cloud classification accuracy on Model Net40, and the results are shown in Table 4. Our C-MCR achieves state-of-the-art zero-shot classification results compared with methods trained on 3D-language data. The advanced performance in the 3D-language field further demonstrates the great potential of C-MCR to learn contrastive representations for modalities lacking paired data. 4.5 Ablation Studies We conduct ablation studies on audio-image retrieval over AVE and Flickr-Sound Net to examine the effectiveness of our method. All the results are presented in Table 5 and Figure 3. Semantic Consistency. We use a softmax function to softly aggregate embeddings in memory and produce the semantic consistent embedding in Equation 2. For comparison, Row I selects the embedding in the memory with the highest similarity as generated embedding, while Row J randomly selects an embedding in the memory. Compared to Row I, the significantly better results in Rows J and K highlight the necessity of inter-modality semantic consistency. Compared to the approach of hardly selecting one embedding, as in Row I, our soft clustering of memories slightly improves the performance by generating more diverse embeddings. Semantic Completion. By comparing H and K, we can find that semantic bias in the MCR space indeed dramatically affects the learning of connections, and adding noise to embeddings can effectively alleviate this issue by enhancing semantic completeness and robustness. The results in G and K demonstrate that our re-normalized operator in Equation 3 is beneficial for learning alignment on the unit sphere. Moreover, in Figure 3, we vary the variance σ2 of noises and report its effect. Generally, the performance is not sensitive to changes in σ2. Inter-MCR alignment. Comparison between D, E, and K demonstrates that the connection learned from the native text-text pairs is much more crucial than from the pseudo-consistent audio and visual embeddings, and using both native text-text pairs and pseudo audio-visual pairs produces the best performance. Furthermore, if no connections are made, as in Row F, the image and audio embeddings would have no semantic relationship since the original CLIP and CLAP spaces are isolated. Intra-MCR alignment. Results in A, B, and K indicate that alignment within either CLIP or CLAP can provide a relatively reliable connection, and aligning both leads to even better results. Conversely, not aligning CLIP and CLAP as in C results in a connection not well adapted to audio-visual input. More importantly, the results in C are similar to those in Row D, where connections are not learned from text-text pairs. This observation indicates that the primary function of intra-MCR alignment is to alleviate the modality gap, thereby enabling the more stable connections learned from text-text pairs to adapt to audio-visual inputs. Table 5: Ablation studies in AVE and Flickr-Sound Net retrieval. We report the m AP" metric on both A2I and I2A subtasks. Re-norm stands for the re-normalized operator in Equation 3; CLIP for intra-CLIP alignment item in Equation 9; CLAP for intra-CLAP alignment item in Equation 9; Flickr Net for Flickr-Sound Net dataset. Rows Consistency Completion Inter-MCR Intra-MCR AVE Flickr Net Re-norm Noise Lttc Lavc CLAP CLIP A2I I2A A2I I2A A softmax 4.09 4.11 4.52 4.71 B softmax 3.97 4.08 4.44 4.79 C softmax 3.14 3.22 3.63 3.51 D softmax 3.28 3.30 3.69 3.50 E softmax 4.09 4.10 4.42 4.54 F softmax 0.22 0.23 0.18 0.19 G softmax 3.70 3.88 4.57 4.62 H softmax 2.77 2.37 2.72 2.57 I argmax 4.01 3.99 4.49 4.52 J 2.62 2.84 2.52 2.76 K softmax 4.11 4.13 4.57 4.92 Figure 3: Effect of different variance σ2 of the noises in Equation 3 on AVE and Flickr-Sound Net retrieval. The average m AP is the mean value of the m AP in I2A and A2I subtasks. 5 Conclusion This paper proposes Connecting Multi-modal Contrastive Representation (C-MCR), a new flexible and training-efficient method for learning multi-modal contrastive representation. C-MCR eliminates the need for large-scale, high-quality data pairs and instead extends the acquired multi-modal alignment knowledge in existing MCRs. By connecting existing MCRs via overlapping modality, we are able to discover more generalized contrastive representations across a broader range of modalities. Experimentally, we learn state-of-the-art audio-visual contrastive representations by connecting CLIP and CLAP through texts, and advanced 3D-language representations by connecting CLIP and ULIP via images. Despite not utilizing any paired data, the representations obtained by C-MCR significantly outperform previous representations learned from data pairs on different downstream tasks. Acknowledgments This work was supported in part by National Key R&D Program of China under Grant No.2022ZD0162000, National Natural Science Foundation of China under Grant No.62222211, No.62072397 and No.61836002. [1] Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International conference on machine learning, pages 8748 8763. PMLR, 2021. [2] Chao Jia, Yinfei Yang, Ye Xia, Yi-Ting Chen, Zarana Parekh, Hieu Pham, Quoc Le, Yun Hsuan Sung, Zhen Li, and Tom Duerig. Scaling up visual and vision-language representation learning with noisy text supervision. In International Conference on Machine Learning, pages 4904 4916. PMLR, 2021. [3] Liunian Harold Li, Pengchuan Zhang, Haotian Zhang, Jianwei Yang, Chunyuan Li, Yiwu Zhong, Lijuan Wang, Lu Yuan, Lei Zhang, Jenq-Neng Hwang, et al. Grounded language-image pre-training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 10965 10975, 2022. [4] Jiarui Xu, Shalini De Mello, Sifei Liu, Wonmin Byeon, Thomas Breuel, Jan Kautz, and Xiaolong Wang. Groupvit: Semantic segmentation emerges from text supervision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18134 18144, 2022. [5] Prafulla Dhariwal and Alexander Nichol. Diffusion models beat gans on image synthesis. Advances in Neural Information Processing Systems, 34:8780 8794, 2021. [6] Aditya Ramesh, Prafulla Dhariwal, Alex Nichol, Casey Chu, and Mark Chen. Hierarchical text-conditional image generation with clip latents. ar Xiv preprint ar Xiv:2204.06125, 2022. [7] Huaishao Luo, Lei Ji, Ming Zhong, Yang Chen, Wen Lei, Nan Duan, and Tianrui Li. Clip4clip: An empirical study of clip for end to end video clip retrieval and captioning. Neurocomputing, 508:293 304, 2022. [8] Yongming Rao, Wenliang Zhao, Guangyi Chen, Yansong Tang, Zheng Zhu, Guan Huang, Jie Zhou, and Jiwen Lu. Denseclip: Language-guided dense prediction with context-aware prompting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18082 18091, 2022. [9] Medhini Narasimhan, Anna Rohrbach, and Trevor Darrell. Clip-it! language-guided video summarization. Advances in Neural Information Processing Systems, 34:13988 14000, 2021. [10] Renrui Zhang, Ziyu Guo, Wei Zhang, Kunchang Li, Xupeng Miao, Bin Cui, Yu Qiao, Peng Gao, and Hongsheng Li. Pointclip: Point cloud understanding by clip. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8552 8562, 2022. [11] Yuhao Zhang, Hang Jiang, Yasuhide Miura, Christopher D Manning, and Curtis P Langlotz. Contrastive learning of medical visual representations from paired images and text. In Machine Learning for Healthcare Conference, pages 2 25. PMLR, 2022. [12] Benjamin Elizalde, Soham Deshmukh, Mahmoud Al Ismail, and Huaming Wang. Clap learning audio concepts from natural language supervision. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1 5. IEEE, 2023. [13] Yusong Wu, Ke Chen, Tianyu Zhang, Yuchen Hui, Taylor Berg-Kirkpatrick, and Shlomo Dubnov. Large-scale contrastive language-audio pretraining with feature fusion and keywordto-caption augmentation. ar Xiv preprint ar Xiv:2211.06687, 2022. [14] Hu Xu, Gargi Ghosh, Po-Yao Huang, Dmytro Okhonko, Armen Aghajanyan, Florian Metze, Luke Zettlemoyer, and Christoph Feichtenhofer. Videoclip: Contrastive pre-training for zeroshot video-text understanding. ar Xiv preprint ar Xiv:2109.14084, 2021. [15] Deepti Hegde, Jeya Maria Jose Valanarasu, and Vishal Patel. Clip goes 3d: Leveraging prompt tuning for language grounded 3d recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2028 2038, 2023. [16] Victor Weixin Liang, Yuhui Zhang, Yongchan Kwon, Serena Yeung, and James Y Zou. Mind the gap: Understanding the modality gap in multi-modal contrastive representation learning. Advances in Neural Information Processing Systems, 35:17612 17625, 2022. [17] Le Xue, Mingfei Gao, Chen Xing, Roberto Martín-Martín, Jiajun Wu, Caiming Xiong, Ran Xu, Juan Carlos Niebles, and Silvio Savarese. Ulip: Learning a unified representation of language, images, and point clouds for 3d understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 1179 1189, 2023. [18] Le Xue, Ning Yu, Shu Zhang, Junnan Li, Roberto Martín-Martín, Jiajun Wu, Caiming Xiong, Ran Xu, Juan Carlos Niebles, and Silvio Savarese. Ulip-2: Towards scalable multimodal pre-training for 3d understanding. ar Xiv preprint ar Xiv:2305.08275, 2023. [19] Andrey Guzhov, Federico Raue, Jörn Hees, and Andreas Dengel. Audioclip: Extending clip to image, text and audio. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 976 980. IEEE, 2022. [20] Ho-Hsiang Wu, Prem Seetharaman, Kundan Kumar, and Juan Pablo Bello. Wav2clip: Learning robust audio representations from clip. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4563 4567. IEEE, 2022. [21] Jort F Gemmeke, Daniel PW Ellis, Dylan Freedman, Aren Jansen, Wade Lawrence, R Channing Moore, Manoj Plakal, and Marvin Ritter. Audio set: An ontology and human-labeled dataset for audio events. In 2017 IEEE international conference on acoustics, speech and signal processing (ICASSP), pages 776 780. IEEE, 2017. [22] Honglie Chen, Weidi Xie, Andrea Vedaldi, and Andrew Zisserman. Vggsound: A large-scale audio-visual dataset. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 721 725. IEEE, 2020. [23] Hao Zhu, Man-Di Luo, Rui Wang, Ai-Hua Zheng, and Ran He. Deep audio-visual learning: A survey. International Journal of Automation and Computing, 18:351 376, 2021. [24] Donghuo Zeng, Yanan Wang, Jianming Wu, and Kazushi Ikeda. Complete cross-triplet loss in label space for audio-visual cross-modal retrieval. ar Xiv preprint ar Xiv:2211.03434, 2022. [25] Donghuo Zeng, Yi Yu, and Keizo Oyama. Deep triplet neural networks with cluster-cca for audio-visual cross-modal retrieval. ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM), 16(3):1 23, 2020. [26] Luís Vilaça, Yi Yu, and Paula Viana. Recent advances and challenges in deep audio-visual correlation learning. ar Xiv preprint ar Xiv:2202.13673, 2022. [27] Yan-Bo Lin, Yi-Lin Sung, Jie Lei, Mohit Bansal, and Gedas Bertasius. Vision transformers are parameter-efficient audio-visual learners. ar Xiv preprint ar Xiv:2212.07983, 2022. [28] Yang Zhao, Chen Zhang, Haifeng Huang, Haoyuan Li, and Zhou Zhao. Towards effective multi-modal interchanges in zero-resource sounding object localization. Advances in Neural Information Processing Systems, 35:38089 38102, 2022. [29] Rui Qian, Di Hu, Heinrich Dinkel, Mengyue Wu, Ning Xu, and Weiyao Lin. Multiple sound sources localization from coarse to fine. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XX 16, pages 292 308. Springer, 2020. [30] Triantafyllos Afouras, Andrew Owens, Joon Son Chung, and Andrew Zisserman. Selfsupervised learning of audio-visual objects from video. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part XVIII 16, pages 208 224. Springer, 2020. [31] Di Hu, Feiping Nie, and Xuelong Li. Deep multimodal clustering for unsupervised audiovisual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9248 9257, 2019. [32] Arda Senocak, Tae-Hyun Oh, Junsik Kim, Ming-Hsuan Yang, and In So Kweon. Learning to localize sound source in visual scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4358 4366, 2018. [33] Di Hu, Rui Qian, Minyue Jiang, Xiao Tan, Shilei Wen, Errui Ding, Weiyao Lin, and Dejing Dou. Discriminative sounding objects localization via self-supervised audiovisual matching. Advances in Neural Information Processing Systems, 33:10077 10087, 2020. [34] Xian Liu, Rui Qian, Hang Zhou, Di Hu, Weiyao Lin, Ziwei Liu, Bolei Zhou, and Xiaowei Zhou. Visual sound localization in the wild by cross-modal interference erasing. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 1801 1809, 2022. [35] Arda Senocak, Hyeonggon Ryu, Junsik Kim, and In So Kweon. Learning sound localization better from semantically similar samples. In ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 4863 4867. IEEE, 2022. [36] Shentong Mo and Pedro Morgado. Localizing visual sounds the easy way. In Computer Vision ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23 27, 2022, Proceedings, Part XXXVII, pages 218 234. Springer, 2022. [37] Kim Sung-Bin, Arda Senocak, Hyunwoo Ha, Andrew Owens, and Tae-Hyun Oh. Sound to visual scene generation by audio-to-visual latent alignment. ar Xiv preprint ar Xiv:2303.17490, 2023. [38] Ludan Ruan, Yiyang Ma, Huan Yang, Huiguo He, Bei Liu, Jianlong Fu, Nicholas Jing Yuan, Qin Jin, and Baining Guo. Mm-diffusion: Learning multi-modal diffusion models for joint audio and video generation. ar Xiv preprint ar Xiv:2212.09478, 2022. [39] Zhaofeng Shi. A survey on audio synthesis and audio-visual multimodal processing. ar Xiv preprint ar Xiv:2108.00443, 2021. [40] Roy Sheffer and Yossi Adi. I hear your true colors: Image guided audio generation. ar Xiv preprint ar Xiv:2211.03089, 2022. [41] Ye Zhu, Yu Wu, Kyle Olszewski, Jian Ren, Sergey Tulyakov, and Yan Yan. Discrete contrastive diffusion for cross-modal music and image generation. In The Eleventh International Conference on Learning Representations. [42] Kun Su, Xiulong Liu, and Eli Shlizerman. Audeo: Audio generation for a silent performance video. Advances in Neural Information Processing Systems, 33:3325 3337, 2020. [43] Kun Su, Xiulong Liu, and Eli Shlizerman. How does it sound? Advances in Neural Information Processing Systems, 34:29258 29273, 2021. [44] Yipin Zhou, Zhaowen Wang, Chen Fang, Trung Bui, and Tamara L Berg. Visual to sound: Generating natural sound for videos in the wild. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3550 3558, 2018. [45] Angel X Chang, Thomas Funkhouser, Leonidas Guibas, Pat Hanrahan, Qixing Huang, Zimo Li, Silvio Savarese, Manolis Savva, Shuran Song, Hao Su, et al. Shapenet: An information-rich 3d model repository. ar Xiv preprint ar Xiv:1512.03012, 2015. [46] Angela Dai, Angel X Chang, Manolis Savva, Maciej Halber, Thomas Funkhouser, and Matthias Nießner. Scannet: Richly-annotated 3d reconstructions of indoor scenes. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5828 5839, 2017. [47] Zhirong Wu, Shuran Song, Aditya Khosla, Fisher Yu, Linguang Zhang, Xiaoou Tang, and Jianxiong Xiao. 3d shapenets: A deep representation for volumetric shapes. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 1912 1920, 2015. [48] Matt Deitke, Dustin Schwenk, Jordi Salvador, Luca Weihs, Oscar Michel, Eli Vander Bilt, Ludwig Schmidt, Kiana Ehsani, Aniruddha Kembhavi, and Ali Farhadi. Objaverse: A universe of annotated 3d objects. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 13142 13153, 2023. [49] Dave Zhenyu Chen, Angel X Chang, and Matthias Nießner. Scanrefer: 3d object localization in rgb-d scans using natural language. In European conference on computer vision, pages 202 221. Springer, 2020. [50] Panos Achlioptas, Ahmed Abdelreheem, Fei Xia, Mohamed Elhoseiny, and Leonidas Guibas. Referit3d: Neural listeners for fine-grained 3d object identification in real-world scenes. In Computer Vision ECCV 2020: 16th European Conference, Glasgow, UK, August 23 28, 2020, Proceedings, Part I 16, pages 422 440. Springer, 2020. [51] Lichen Zhao, Daigang Cai, Lu Sheng, and Dong Xu. 3dvg-transformer: Relation modeling for visual grounding on point clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2928 2937, 2021. [52] Zehan Wang, Haifeng Huang, Yang Zhao, Linjun Li, Xize Cheng, Yichen Zhu, Aoxiong Yin, and Zhou Zhao. 3drp-net: 3d relative position-aware network for 3d visual grounding. ar Xiv preprint ar Xiv:2307.13363, 2023. [53] Zehan Wang, Haifeng Huang, Yang Zhao, Linjun Li, Xize Cheng, Yichen Zhu, Aoxiong Yin, and Zhou Zhao. Distilling coarse-to-fine semantic matching knowledge for weakly supervised 3d visual grounding. ICCV 2023, 2023. [54] Daichi Azuma, Taiki Miyanishi, Shuhei Kurita, and Motoaki Kawanabe. Scanqa: 3d question answering for spatial scene understanding. In proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 19129 19139, 2022. [55] Xiaojian Ma, Silong Yong, Zilong Zheng, Qing Li, Yitao Liang, Song-Chun Zhu, and Siyuan Huang. Sqa3d: Situated question answering in 3d scenes. ar Xiv preprint ar Xiv:2210.07474, 2022. [56] Zehan Wang, Haifeng Huang, Yang Zhao, Ziang Zhang, and Zhou Zhao. Chat-3d: Dataefficiently tuning large language model for universal dialogue of 3d scenes. ar Xiv preprint ar Xiv:2308.08769, 2023. [57] Yining Hong, Haoyu Zhen, Peihao Chen, Shuhong Zheng, Yilun Du, Zhenfang Chen, and Chuang Gan. 3d-llm: Injecting the 3d world into large language models. ar Xiv preprint ar Xiv:2307.12981, 2023. [58] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft coco: Common objects in context. In Computer Vision ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pages 740 755. Springer, 2014. [59] Piyush Sharma, Nan Ding, Sebastian Goodman, and Radu Soricut. Conceptual captions: A cleaned, hypernymed, image alt-text dataset for automatic image captioning. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2556 2565, 2018. [60] Jun Xu, Tao Mei, Ting Yao, and Yong Rui. Msr-vtt: A large video description dataset for bridging video and language. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 5288 5296, 2016. [61] Mattia Soldan, Alejandro Pardo, Juan León Alcázar, Fabian Caba, Chen Zhao, Silvio Giancola, and Bernard Ghanem. Mad: A scalable dataset for language grounding in videos from movie audio descriptions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5026 5035, 2022. [62] Chris Dongjoo Kim, Byeongchang Kim, Hyunmin Lee, and Gunhee Kim. Audiocaps: Generating captions for audios in the wild. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 119 132, 2019. [63] Konstantinos Drossos, Samuel Lipping, and Tuomas Virtanen. Clotho: An audio captioning dataset. In ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 736 740. IEEE, 2020. [64] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A largescale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pages 248 255. Ieee, 2009. [65] Xumin Yu, Lulu Tang, Yongming Rao, Tiejun Huang, Jie Zhou, and Jiwen Lu. Point-bert: Pre-training 3d point cloud transformers with masked point modeling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 19313 19322, 2022. [66] Yapeng Tian, Jing Shi, Bochen Li, Zhiyao Duan, and Chenliang Xu. Audio-visual event localization in unconstrained videos. In Proceedings of the European Conference on Computer Vision (ECCV), pages 247 263, 2018. [67] Honglie Chen, Weidi Xie, Triantafyllos Afouras, Arsha Nagrani, Andrea Vedaldi, and Andrew Zisserman. Localizing visual sounds the hard way. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16867 16876, 2021. [68] Hang Zhao, Chuang Gan, Andrew Rouditchenko, Carl Vondrick, Josh Mc Dermott, and Antonio Torralba. The sound of pixels. In Proceedings of the European conference on computer vision (ECCV), pages 570 586, 2018. [69] Shentong Mo and Pedro Morgado. A closer look at weakly-supervised audio-visual source localization. ar Xiv preprint ar Xiv:2209.09634, 2022. [70] Zekun Qi, Runpei Dong, Guofan Fan, Zheng Ge, Xiangyu Zhang, Kaisheng Ma, and Li Yi. Contrast with reconstruct: Contrastive 3d representation learning guided by generative pretraining. ar Xiv preprint ar Xiv:2302.02318, 2023. A Ablation Study about Text Dataset. We conduct more experiments on audio-image retrieval with training texts from different sources. Furthermore, we provide insights about selecting training data when employing our C-MCR to connect other MCRs. As discussed in Sec 4.1, our training texts are collected from three sources: image-text datasets (COCO [58] and CC3M [59]), video-text datasets (MSRVTT [60] and MAD [61]), and audio-text datasets (Audio Cap [62] and Clotho [63]). Figure 4: Ablation studies about text datasets for training. Dataset AVE Flickr A2I I2A A2I I2A Full 4.11 4.13 4.57 4.92 w/o image-text 3.83 3.76 3.97 3.91 w/o video-text 4.04 4.05 4.41 4.78 w/o audio-text 4.07 3.88 4.31 4.59 In Table 4, we exclude the text data from imagetext, video-text, and audio-text datasets, respectively. The results demonstrate that combining data from all three sources achieves the best performance. Furthermore, our findings suggest that the information in video-text datasets is relatively less important than in image-text and audio-text datasets. When applying our C-MCR to connect other MCRs, it is critical to collect overlapping modality data associated with information from non-overlapping modalities to ensure robust connections. The data from the pre-training datasets used by MCR could serve as an appropriate starting point. Combining the overlapping modality data from these sources ensures that the data used for constructing connections contains sufficient information from non-overlapping modalities. Additionally, this data is easily accessible and scalable, which greatly enhances the practicality of our C-MCR. B Downstream Task Details. B.1 Audio-Image Retrieval. We consider two datasets for this task: AVE [66] and Flickr-Sound Net [32], both of which consist of semantically matched image and audio pairs that were manually curated. To more comprehensively and stably reflect the retrieval capability of the model, we use all available data in these two datasets for evaluation, resulting in 4,095 samples for AVE and 5,000 samples for Flickr-Sound Net. B.2 Audio-Visual Source Localization. We conduct experiments on the VGGSS [67] and MUSIC [68] datasets. VGGSS is derived from VGGSound, and its test set comprises 5,158 audio-image pairs. MUSIC consists of 489 untrimmed videos of musical solos spanning 11 instrument categories for testing. It is worth noting that we use the category names from the COCO dataset as prompts to enable the open-vocabulary object detector GLIP [3] to extract object proposals. B.3 Counterfactual Audio-Image Recognition. The Extended Flickr-Sound Net [69] and Extended VGGSS [69] are constructed by adding 250 and 5,158 negative samples to the test sets of the original Flickr-Sound Net and VGGSS datasets, respectively. The prompts used for the object detector GLIP [3] are also the category names from the COCO dataset. We evaluate the counterfactual Audio-Image Recognition performance using the Maximum F1 (Max-F1) and Average Precision (AP) metrics, following [69]. During inference, for the i-th image-audio pair, the proposal with the highest matching score with the audio is considered the predicted object, and its matching score is considered the confidence score ci. The CIo U of the predicted object is denoted as Io Ui. The ground-truth map is denoted as Gi, and the ground-truth maps of negative samples are . Under these definitions, the true positives T P, false positives FP, and false negatives FN are computed as: T P(γ, δ) = {i|Gi = , Io Ui > γ, ci > δ} FP(γ, δ) = {i|Gi = , Io Ui γ, ci > δ} {i|Gi = , ci > δ} FN(γ, δ) = {i|Gi = , ci δ} (11) where γ is the threshold of Io U and δ is the threshold of confidence score. Following previous work, the γ is set as 0.5. The F1 score can be represented as: F1(γ, δ) = 2 Precision(γ, δ) Recall(γ, δ) Precision(γ, δ) + Recall(γ, δ) (12) Precision(γ, δ) = |T P(γ, δ)| |T P(γ, δ)| + |FP(γ, δ)|; Recall(γ, δ) = |T P(γ, δ)| |T P(γ, δ)| + |FN(γ, δ)| (13) In accordance with [69], we calculate F1 scores for all values of δ and report the maximum F1 score (Max-F1). Average Precision (AP) is another commonly used metric in object detection, its computation is detailed in [58, 69]. C Model Configurations. Table 6: Model configurations of projectors. Module Block Cin Cout Linear 512 1024 Batch Norm1D 1024 1024 Relu - - Linear 1024 512 Batch Norm1D 512 512 Relu - - Linear 512 1024 Batch Norm1D 1024 1024 Relu - - Linear 1024 512 Batch Norm1D 512 512 Relu - - The model configurations of our projectors are shown in Table 6. D Limitations and Future Work. While C-MCR offers an efficient and effective contrastive representation learning method for modalities that lack high-quality, large-scale paired data, it still necessitates an intermediate modality to associate these modalities. Exploring ways to reduce data requirements further while maintaining representation performance is an intriguing direction for future research. E Social Impacts. Although C-MCR achieves outstanding performance in audio-visual learning by connecting CLIP and CLAP, further analysis of the capability boundary of this representation is necessary before applying it to additional modalities or deploying it in practice. C-MCR only requires unpaired unimodal data during training, significantly reducing the data requirements for learning a generalizable representation. However, this also means that unsuitable and harmful data in each modality are more likely to be used for training.