# crossmodal_learning_with_adversarial_samples__dc35fe90.pdf Cross-Modal Learning with Adversarial Samples Chao Li1,2 Cheng Deng1, Shangqian Gao2 De Xie1 Wei Liu3, 1School of Electronic Engineering, Xidian University, Xi an, Shaanxi, China 2Electrical and Computer Engineering, University of Pittsburgh, Pittsburgh, PA, USA 3Tencent AI Lab, China {chaolee.xd, chdeng.xd, xiede.xd}@gmail.com, shg84@pitt.edu, wl2223@columbia.edu With the rapid developments of deep neural networks, numerous deep cross-modal analysis methods have been presented and are being applied in widespread realworld applications, including healthcare and safety-critical environments. However, the recent studies on robustness and stability of deep neural networks show that a microscopic modification, known as adversarial sample, which is even imperceptible to humans, can easily fool a well-performed deep neural network and brings a new obstacle to deep cross-modal correlation exploring. In this paper, we propose a novel Cross-Modal correlation Learning with Adversarial samples, namely CMLA, which for the first time presents the existence of adversarial samples in cross-modal data. Moreover, we provide a simple yet effective adversarial sample learning method, where interand intramodality similarity regularizations across different modalities are simultaneously integrated into the learning of adversarial samples. Finally, our proposed CMLA is demonstrated to be highly effective in cross-modal hashing based retrieval. Extensive experiments on two cross-modal benchmark datasets show that the adversarial examples produced by our CMLA are efficient in fooling a target deep cross-modal hashing network. On the other hand, such adversarial examples can significantly strengthen the robustness of the target network by conducting an adversarial training. 1 Introduction Cross-modal learning, such as cross-modal retrieval, enables a user to achieve what he/she prefers in one modality (e.g., image) that is relevant to a given query in another (e.g., text). However, due to the drastic growth of multimedia, learning in such tremendous amounts of multimedia data has been a new challenge. The recent success of deep learning and its role in cross-modal learning seem to obviate concerns about the performance in both accuracy and speed: exploiting deep neural networks to map data samples of different modalities into compact hash codes and using fast bitwise XOR operations to perform retrieval. Extensive efforts [31, 32, 30, 29, 39, 28, 44, 33, 23, 47, 5, 18, 21, 24, 27, 40, 12, 13, 49] have been made and achieved remarkable retrieval accuracy. The current studies [17, 20, 37] show that deep networks are vulnerable against purposeful input samples, namely adversarial samples. These samples can easily fool a well-performed deep learning model by only adding a little perturbation which is even imperceptible to humans. Besides, adversarial samples have been observed in wide areas, such as image classification [36], object recognition [9], object detection [45], speech recognition [2], etc. However, the potential risks of deep neural networks being vulnerable to adversarial samples in cross-modal learning have not been delineated. In this paper, we take cross-modal hashing retrieval as a representative case of cross-modal learning, where search space can roughly be divided into four parts: T2T, I2I, I2T/T2I, and NR, as shown in Corresponding authors. 33rd Conference on Neural Information Processing Systems (Neur IPS 2019), Vancouver, Canada. (a) Cross-modal Search Space (b) Adversarial Sample Learning (c) Adversarial Training T2I/I2T I2I T2T T2I/I2T I2I T2T Figure 1: (a) The search space in cross-modal retrieval, which can be divided into four parts: I2I (query is image, database is image), T2T (query is text, database is text), I2T (query is image, database is text)/T2I (query is text, database is image), and NR (not relevant samples in two modalities). (b) Adversarial sample learning. (c) Adversarial training. Fig. 1. NR means that the returned samples are not relevant to the query. In the T2T space, taking a text as a query, only semantically similar texts can be successfully retrieved, and vice versa in I2I. Different from T2T and I2I, I2T/T2I is the main concern in the cross-modal retrieval community, which focuses on information retrieval across different modalities. Existing deep cross-modal hashing works uniformly build the relationships across different modalities by constructing a multi-layer neural network, and learn hash codes via an objective of similarity metric. However, due to the lack of sufficient training data points and neglecting the robustness of their hashing networks, the search space of these methods cannot be sufficiently explored, so the retrieval performance can easily be compromised. Inspired by adversarial training [42] which successfully increases the robustness by augmenting training data with adversarial samples, we propose to explore and utilize such adversarial samples to construct a more robust search space in I2T/T2I. Different from other tasks mentioned above, for cross-modal retrieval, adversarial samples generated from four different spaces, i.e., T2T, I2I, T2I/I2T, and NR, have different capacities in attacking. However, a more ideal and deceptive adversarial sample for cross-modal not only can make an effective attack to cross-modal retrieval, but also should keep the non-decreasing retrieval performance compared with a clean sample when executing single-modal retrieval. To be specific, given a cross-modal system, adversarial samples with errors in both single-modal and cross-modal are suspicious and can be easily detected. On the contrary, adversarial samples with errors merely in cross-modal but being correct in single-modal are much harder to be discovered, which are more deceptive adversarial samples. In this paper, we propose a novel Cross-Modal Learning with Adversarial samples, namely CMLA, which can improve the robustness of cross-modal learning, such as a deep cross-modal hashing network. As such, accurate connections across different modalities can be bridged and a more robust cross-modal retrieval system can be established. The highlights of our work can be summarized as follows: We propose a simple yet effective cross-modal learning method by exploring cross-modal adversarial samples, where adversarial sample is defined in two aspects: two perturbations for different modalities are learned to fool a deep cross-modal network; the perturbations on each modality will not impact the performance within its modality. A novel cross-modal adversarial sample learning algorithm is presented. To learn crossmodal adversarial samples with high attacking capability, we propose to decrease the intermodality similarity and simultaneously keep intra-modality similarity by one optimization. We additionally apply the proposed CMLA to cross-modal hash learning. Experiments on two widely used cross-modal retrieval benchmarks show the effectiveness of our CMLA in attacking a target retrieval system and further improving its robustness. 2 Related Works Cross-modal hashing methods focus on building the correlation between different modalities and learning reliable hash codes, which can be mainly categorized into two settings: data-independent hashing methods and data-dependent ones. In data-independent hashing methods, hash codes are learned based on random projections, e.g., locality-sensitive hashing (LSH) [16]. Compared with 0 1 0 0 1 0 1 1 0 1 0 1 1 0 1 0 1 0 1 1 1 1 0 0 1 0 1 0 0 1 0 1 0 0 0 0 1 1 0 1 0 1 1 0 1 1 0 1 0 0 1 0 1 0 0 1 -1 1 + -1 1 + Intra-Modal Similarity -1 1 + -1 1 + -1 -1 1 + -1 1 + -1 1 + -1 1 + -1 CNN-Based Network Multilayer Perceptron Inter-Modal Similarity 1 + Hash Layer Figure 2: The pipeline of our proposed CMLA for cross-modal hash learning. data-dependent methods, data-independent methods always require long bits to encode multi-modal data. Thus, most currently proposed methods are data-dependent ones, which can learn compact hash codes using partial data as a training set. Collective Matrix Factorization Hashing (CMFH) [14] learns hash codes for different modalities by executing collective matrix factorization of different views. Brostein et al. [3] presented a cross-modal hashing by preserving intra-class similarity via eigen-decomposition and boosting. Semantics-preserving hashing (Se PH) [28] produces unified binary codes by modeling an affinity matrix in a probability distribution while minimizing the Kullback-Leibler divergence. Most of these methods, which depend on hand-crafted features, lack the capacity of fully exploiting heterogeneous relations across modalities. Recently, models based on deep neural networks can easily access more discriminative features, leading to a boost in the development of deep cross-modal hashing techniques [6, 15, 23, 43, 26, 11]. Deep cross-modal hashing (DCMH) [23] utilizes a deep neural network to perform feature extraction and hash code learning from scratch. Pairwise relationship guided deep hashing (PRDH) [48] integrates different pairwise constraints to purify the similarities of hash codes from inter-modality and intra-modality. Self-supervised adversarial hashing (SSAH) [25] introduces a label network into the hash learning process, which facilitates hash code generation. Despite the outstanding performance and remarkable achievements, the DNN systems have recently been shown to be vulnerable to adversarial attacks. Szegedy et al. [42] first showed that a welldesigned small perturbation on images can fool state-of-the-art deep neural networks with high probability. In the following, a series of research efforts on attack [41, 46, 34] and defense [7, 35, 38] have been presented. Fast gradient sign method (FGSM) [17], which uses a sign function on gradients for the inputs to learn adversarial examples, is one of the representative efficient attack algorithms. For the defense methods, adversarial training [42, 17] is adopted to augment their training data of the classifier by learning adversarial samples. Distillation technique [38] was presented to reduce the magnitude of the gradients used for adversarial sample creation and also to increase the minimum feature numbers to be modified. However, concerning large-scale cross-modal retrieval, although plenty of deep cross-modal hashing networks have been constructed, there is no attempt to focus on the security of DNNs in deep cross-modal hashing models. 3 Cross-Modal Learning with Adversarial Samples 3.1 Problem Definition Given a cross-modal benchmark O = {oi}N i=1 with N data points, oi = (ov i , ot i, ol i), where ov i and ot i are collected in pair, respectively, denoting image and textual description for the ith data point, ol i is a multi-label annotation assigned to oi. Let S denote a pairwise similarity matrix which describes semantic similarity between each pair of data points, where Sij = 1 means that oi and oj are semantically similar, otherwise Sij = 0. In a multi-label setting, when oi and oj share at least one label, Sij = 1; otherwise Sij = 0. The main task of cross-modal hashing is to learn two hash functions H , {v, t}, which build cross-modal correlations and generate hash codes B { 1, 1}K for cross-modal data, where K is code length, and H are usually learned by deep neural networks in deep cross-modal hashing. We additionally define the outputs of hash layers as Hv and Ht for image and text, respectively. Binary hash codes B are generated by applying a sign function to H : B = sign(H ), {v, t}. (1) For deep hashing networks H (o , θ ), let θ be network parameters and J(θv, θt, ov, ot) be the loss function, respectively. The aim of a cross-modal adversarial attack is to find the minimum perturbations δv and δt that cause the change of retrieval accuracy. Formally, (o , H ) := min δ δ p, s.t. max δ D (H (o + δ ; θ ) , H (o ; θ )) , δ p ϵ, {v, t}, (2) where hash codes H are generated from hash layer H by learning a deep network θ , and D( , ) is a distance measure. p , p = {1, 2, } which denotes Lp norm, denotes the distance between the learned adversarial sample and the original sample. 3.2 Proposed CMLA The overall flowchart of the proposed CMLA model is illustrated in Fig. 2. For a target deep crossmodal hashing network that consists of CNN-based Network and Multilayer Perceptron Network, following which are two hash layers to output hash codes for different modalities. In addition, two regularizations, namely inter-modal similarity regularization and intra-modal similarity regularization, are combined to optimize the learned adversarial samples. For a better understanding, taking an image data point ov for example, we intend to generate an adversarial sample ˆov by learning a small perturbation δv, where ˆov = ov + δv. In this way, feeding ˆo into the target deep cross-modal hashing network, semantically irrelevant results in the text modality should be returned. To achieve this goal, original text information is first fed into the deep hashing network to generate regular hash codes Ht = Ht(ot, θt). The correlation between the two modalities built during cross-modal hash codes generation is treated as a supervision signal to learn the optimal perturbation for each modality. With Ht, adversarial sample learning can be transferred to a problem of maximizing the Hamming distance between cross-modal hash codes. This can be solved by introducing an inter-modal similarity regularization, with which δv will be optimized by maximizing the Hamming distance between ˆHv and Ht. We formulate this inter-modal similarity loss function as: min δv J v inter = log 1 + e Γij + SijΓij i=1 ˆov i ov i p , Γij = 1 2( ˆHv i )(Ht j) . Moreover, compared with single-modal retrieval, a notable difference existing in cross-modal learning is that the latter not only can build correlations across modalities but also can remain the correlations within each modality. Thus, during adversarial sample learning for cross-modal data, the modality correlation within an individual modality should be kept, which means that the learned cross-modal perturbation cannot change the intra-modal similarity relationship. To learn this perturbation, an additional intra-modal similarity regularization function is adopted in our CMLA, which can be written as: min δv J v intra = log 1 + eΘij SijΘij , Θij = 1 2( ˆHv i )(Hv j ) . (4) Algorithm 1 Cross-Modal correlation Learning with Adversarial samples (CMLA). Input: target deep cross-modal hashing networks: H(o , θ ), {v, t}, and a cross-modal dataset with N data points: {image, text, and label}; Output: optimal perturbations: δv, δt; 1 Maximum iteration = Tmax, Batch_Size = 128, n = N/128 ; 2 for j = 1, j n do 3 initialize iter = 0; while iter Tmax do 4 compute Hv = Hv(ov, θv), Ht = Ht(ot, θt); 5 if not converged then 6 update δv and δt: δv = arg minθv Jv(ov, θv, Hv, Ht); δt = arg minθt Jt(ot, θt, Hv, Ht); 10 return δv and δt. Therefore, the objective function of our CMLA for image modality adversarial sample learning is formulated as: min δv J v = α log 1 + e Γij + SijΓij + β log 1 + eΘij SijΘij i ˆov i ov i p , where Γij = 1 2( ˆHv i )(Ht j) , Θij = 1 2( ˆHv i )(Hv j ) , and α, β, and γ are hyper-parameters. In a similar way, adversarial samples for the text modality can be learned, where the objective function is written as: min δt J t = λ log 1 + e Υij + SijΥij + ξ log 1 + eΨij SijΨij ˆot i ot i p , where Υij = 1 2( ˆHt i )(Hv j ) , Ψij = 1 2( ˆHt i )(Ht j) , and λ, ξ, and η are hyper-parameters. To solve the problems in Eqs.(5)(6), we fix the network parameters and optimize δv and δt. Considering the structure difference between perturbations of image and text, CMLA learns different perturbations for two modalities, with δv and δt being updated iteratively. Algorithm 1 summarizes the learning procedure of the proposed CMLA. 4 Experiments 4.1 Experimental Setup Extensive experiments on two benchmarks: MIRFlickr-25K [22] and NUS-WIDE [10] are conducted to evaluate the performances of our proposed CMLA and two state-of-the-art deep cross-modal hashing networks as well as their variations. MIRFlickr-25K [22] is collected from Flickr, which contains 25,000 images. Each image is labeled with an associated text description. 20,015 image-text pairs are selected in our experiments, and each image-text pair is annotated with at least one of 24 unique labels. For the text modality, each text is represented by a 1,386-dimensional bag-of-words vector. NUS-WIDE [10] is a public web image dataset containing 269,648 web images. 81 ground-truth concepts have been annotated for retrieval evaluation. After pruning the data point that has no label Table 1: Comparison in terms of MAP scores of two retrieval tasks on MIRFlickr-25K and NUSWIDE datasets with different lengths of hash codes. Task Method MIRFlickr-25K NUS-WIDE 16 32 48 64 16 32 48 64 Image Query v.s. Text Database DCMH 0.736 0.749 0.756 0.761 0.595 0.607 0.620 0.641 DCMH+ 0.805 0.816 0.825 0.828 0.658 0.679 0.686 0.683 SSAH 0.797 0.805 0.807 0.807 0.645 0.660 0.670 0.672 SSAH+ 0.804 0.815 0.826 0.829 0.660 0.675 0.690 0.694 Text Query v.s. Image Database DCMH 0.796 0.797 0.804 0.806 0.601 0.614 0.623 0.645 DCMH+ 0.810 0.820 0.820 0.819 0.679 0.691 0.693 0.690 SSAH 0.798 0.805 0.807 0.804 0.661 0.677 0.681 0.684 SSAH+ 0.808 0.809 0.814 0.815 0.671 0.685 0.693 0.697 or text information, a subset of 190,421 image-text pairs that belong to the 21 most-frequent concepts are selected as a dataset in our experiments. We use a 1,000-dimensional bag-of-words vector to represent each text data point. Evaluations. In order to evaluate the performance of the proposed CMLA, we follow previous works [28, 4, 5] and adopt three commonly used evaluation criteria in cross-modal retrieval: Mean Average Precision (MAP) which is used to measure the accuracy of the Hamming distances, precisionrecall curve (PR curve) which is used to measure the accuracy of hash lookups, and Precision@1000 curve which is used to evaluate the precision with respect to top 1,000 retrieved results. The distortion D between the original modality data o and distorted modality data ˆo is measured by D = q P (ˆo o )2 M , {v, t}. M is set as 150,528 (224 224 3) for the image modality, while for the text modality, we set M as 1,380 and 1,000 for MIRFlickr-25K and NUS-WIDE, respectively, depending on their dimensions. Baselines. DCMH [23] and SSAH [25], which are two representative deep cross-modal hashing networks, are selected as the targeted cross-modal hashing models in [23] [25]. Following the setting in [23] [25], we retrain DCMH and SSAH. Moreover, in order to evaluate the cross-network transfer, we additionally construct two improved versions DCMH+ and SSAH+. DCMH+ and SSAH+ are built by replacing the vgg-f [8] network with the Res Net50 [19] network. Implementation Details. Our proposed CMLA is implemented via Tensor Flow [1] and is run on a server with two NVIDIA Tesla P40 GPUs holding a graphics memory capacity of 24GB for each one. All images are resized to 224 224 3 before being used as the inputs. In adversarial sample learning, we use the Adam optimizer respectively with initial learning rates 0.5 and 0.002 for the image and text modalities, and train each sample for Tmax iterations. All hyper-parameters α, β, λ, ξ, γ, and η are set as 1 empirically. The mini-batch size is fixed at 128. ϵv is set as 8 for the image modality, and ϵt is set as 0.01 for the text modality. After the adversarial sample is generated, we clip the image into 0 255 and clip text into 0 1, respectively. The results reported in our experiments are all average results after a run for 10 times. 4.2 Results For MIRFlickr-25K, 2,000 data points are randomly selected as a query set, 10,000 data points are used as a training set to train the target retrieval network model, and the remainder is kept as a retrieval database. 5,000 data points from the training set are further sampled to learn adversarial samples. For NUS-WIDE, we randomly sample 2,100 data points as a query set and 10,500 data points as a training set. Similarly, 5,000 data points from the training set are sampled to learn adversarial samples. The source codes of DCMH and SSAH are provided by the authors. Moreover, two variations DCMH+ and SSAH+ are constructed by replacing the vgg-f network with the Res Net50 network. All models are retrained from scratch and their performances are shown in Table 1. It can be seen that the target networks DCMH and SSAH achieve similar performances to their original papers and can achieve more promising results after equipped with Res Net50 which has more layers than vgg-f. Given a target deep cross-modal hashing network, to evaluate the attacking performance of the proposed cross-modal adversarial samples learning method, we first fix the network parameters and Table 2: Comparison in terms of MAP scores and distortions (D) of two retrieval tasks on MIRFlickr25K and NUS-WIDE datasets with 32-bit code length. Task Iteration MIRFlickr-25K NUS-WIDE DCMH DCMH+ SSAH SSAH+ DCMH DCMH+ SSAH SSAH+ Image Query v.s. Text Database 100 MAP 0.579 0.631 0.679 0.681 0.526 0.609 0.587 0.591 D 0.039 0.041 0.034 0.038 0.031 0.033 0.032 0.025 200 MAP 0.563 0.599 0.671 0.699 0.499 0.583 0.534 0.543 D 0.023 0.038 0.028 0.032 0.026 0.031 0.029 0.026 500 MAP 0.521 0.554 0.665 0.674 0.457 0.578 0.460 0.502 D 0.019 0.029 0.020 0.023 0.025 0.028 0.026 0.024 Text Query v.s. Image Database 100 MAP 0.615 0.619 0.603 0.611 0.523 0.628 0.501 0.523 D 0.048 0.037 0.031 0.021 0.037 0.035 0.042 0.025 200 MAP 0.587 0.577 0.595 0.605 0.447 0.549 0.454 0.474 D 0.027 0.033 0.025 0.019 0.035 0.031 0.035 0.023 500 MAP 0.561 0.564 0.589 0.593 0.371 0.533 0.351 0.427 D 0.019 0.021 0.023 0.017 0.030 0.027 0.017 0.019 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Recall Image-query-Text on MIRFlickr-25K DCMH-C DCMH-A SSAH-C SSAH-A 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Recall Text-query-Image on MIRFlickr-25K DCMH-C DCMH-A SSAH-C SSAH-A 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 R(*1000) Image-query-Text on MIRFlickr-25K DCMH-C DCMH-A SSAH-C SSAH-A 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 R(*1000) Text-query-Image on MIRFlickr-25K DCMH-C DCMH-A SSAH-C SSAH-A 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Recall Image-query-Text on NUS-WIDE DCMH-C DCMH-A SSAH-C SSAH-A 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Recall Text-query-Image on NUS-WIDE DCMH-C DCMH-A SSAH-C SSAH-A 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 R(*1000) Image-query-Text on NUS-WIDE DCMH-C DCMH-A SSAH-C SSAH-A 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 R(*1000) Text-query-Image on NUS-WIDE DCMH-C DCMH-A SSAH-C SSAH-A Figure 3: PR and Precision@1000 evaluated on MIRFlickr-25K and NUS-WID datasets with 32 bits. execute CMLA to generate the adversarial sample for each query data point. Then, we respectively use each adversarial query sample to retrieve data points across different modalities, where the evaluation settings are the same as those in Table 1. Taking hash code learning with the code length of 32-bit as an example, we train each sample with different iterations from 100 to 500. The results are shown in Table 2, where we provide MAP values and distortions (D) to show the relationship between retrieval performance and distortion with the growth of training iterations. Compared with the results reported in Table 1, it is obvious that: (1) The performances of both DCMH and SSAH are severely decreased by only adding a small distortion to original modality data; (2) With the growth of training iterations, CMLA can simultaneously maintain a high attacking performance and continuously reduce the magnitude of learned disturbance. The same results are shown in Fig. 3, where PR-curves and Precision@1000 curves are provided to show the effectiveness of our proposed CMLA. Some adversarial samples and corresponding original data points are also given in Fig. 4. The images listed above are original images, where their corresponding adversarial samples are shown below them. It is nearly imperceptible to humans eyes. For the text modality, we show the learned adversarial sample, which is constructed by mixing learned distortions and the original bag-of-words vector. Compared with the image modality, the text adversarial sample has relatively large distortions due to its nature of discrete representation. 0.8 0.8 0.8 0.0 0.0 0.0 200 400 600 800 1000 1200 1400 200 400 600 800 1000 1200 1400 200 400 600 800 1000 1200 1400 1.0 1.0 1.0 0.8 0.8 0.8 0.6 0.6 0.6 00 0.0 0.0 200 400 600 800 1000 1200 1400 200 400 600 800 1000 1200 1400 200 400 600 800 1000 1200 1400 Figure 4: Adversarial samples of different modalities learned by the proposed CMLA. With Intra-Modal Loss Without Intra-Modal loss Ablation Study of Intra-Modal Similarity on MIRFlickr-25K (a) Adversarial Image With Intra-Modal Loss Without Intra-Modal loss Ablation Study of Intra-Modal Similarity on MIRFlickr-25K (b) Adversarial Text With Intra-Modal Loss Without Intra-Modal loss Ablation Study of Intra-Modal Similarity on NUS-WIDE (c) Adversarial Image With Intra-Modal Loss Without Intra-Modal loss Ablation Study of Intra-Modal Similarity on NUS-WIDE (d) Adversarial Text Figure 5: Ablation studies of intra-modal similarity evaluated on MIRFlickr-25K and NUS-WIDE datasets with 32 bits. 4.3 Further Analysis As mentioned above, there is a main difference between cross-modal retrieval and single-modal retrieval that only retrieves data points within an identical modality. A well-designed cross-modal retrieval system can achieve both sameand differentmodality data points in high accuracy. Therefore, a more deceptive cross-modal adversarial sample not only can fool a cross-modal retrieval system but also can simultaneously maintain high performance for single-modal retrieval. For a detailed elaboration, an ablation study is conducted by cutting off the intra-modal loss in our CMLA. We take SSAH as an example. Two single-modal retrieval tasks I2I and T2T are respectively executed on MIRFlickr-25K and NUS-WIDE. Fig. 5 shows the retrieval results, from which it is obvious that: The adversarial sample learned without using an intra-modal similarity constraint on MIRFlickr-25K achieves 15.7% and 9.3% performance drops on I2T and T2I compared with that learned equipped with this constraint, but the retrieval performances in single modality also obviously decrease from 0.674 (0.549) to 0.420 (0.426) on I2I (T2T). These adversarial samples, which have a lower performance in single-modal retrieval, can be easily detected by a single-modal retrieval verification before feeding into a cross-modal retrieval system. While equipped with our intra-modal similarity constraint, the variance of CMLA is further constrained and forced to learn more deceptive adversarial samples. In this way, CMLA can fool a cross-modal retrieval system and will not be detected by a single-modal retrieval verification. The main goal of our adversarial sample learning is to improve the robustness of a deep cross-modal hashing network. Thus, we additionally learn adversarial samples from the training set and then conduct an adversarial training by combining the adversarial samples with the training set together. The retrieval performances of the retrained deep cross-modal hashing network under identical attacks Table 3: Comparison in terms of MAP scores on MIRFlickr-25K and NUS-WIDE datasets with different lengths of hash codes. All networks are evaluated after adversarial training. Task Method MIRFlickr-25K NUS-WIDE 16 32 48 64 16 32 48 64 Image Query v.s. Text Database DCMH 0.711 0.723 0.735 0.759 0.578 0.587 0.611 0.628 DCMH+ 0.779 0.781 0.803 0.801 0.647 0.649 0.666 0.677 SSAH 0.783 0.784 0.788 0.785 0.615 0.640 0.658 0.661 SSAH+ 0.784 0.788 0.787 0.789 0.621 0.635 0.662 0.670 Text Query v.s. Image Database DCMH 0.771 0.775 0.782 0.786 0.610 0.603 0.611 0.620 DCMH+ 0.793 0.801 0.803 0.799 0.655 0.673 0.675 0.676 SSAH 0.790 0.788 0.792 0.791 0.638 0.659 0.660 0.664 SSAH+ 0.789 0.789 0.794 0.793 0.641 0.664 0.668 0.667 of adversarial query samples are shown in Table 3. Comparing Table 1, Table 2, and Table 3, we can see that each deep cross-modal hashing network achieves a significant performance increase after adversarial training. Thus, the proposed CMLA can effectively learn adversarial samples, and in turn the learned adversarial samples can also be leveraged to improve the robustness of the targeted deep cross-modal hashing network. 5 Conclusions This paper presents a novel Cross-Modal Learning method with Adversarial samples, dubbed CMLA. First, we made an observation of the existence of the adversarial samples across two different modalities. Second, by simultaneously maximizing inter-modality similarity and minimizing intramodality similarity, an effective adversarial sample learning method was proposed. Moreover, a task on cross-modal hashing retrieval was conducted to verify our proposed CMLA, where extensive results were shown in experiments. As our main purpose is to build an accurate relationship across modalities and to improve the robustness of a target retrieval system, additional adversarial training was enforced for the targeted cross-modal hashing network. The experiments on two widely-used cross-modal retrieval datasets show the high attacking efficiency of our proposed adversarial sample learning method. Besides, these adversarial samples, in turn, can further improve the robustness of existing deep cross-modal hashing networks and achieve state-of-the-art performances on cross-modal retrieval tasks. Acknowledgments This work was partially supported by the National Natural Science Foundation of China 61572388, the National Key Research and Development Program of China (2017YFE0104100), and the Key R&D Program-The Key Industry Innovation Chain of Shaanxi under Grant 2018ZDXM-GY-176. [1] Martín Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay Ghemawat, Geoffrey Irving, Michael Isard, et al. Tensorflow: A system for large-scale machine learning. In OSDI, pages 265 283, 2016. [2] Giuseppe Ateniese, Giovanni Felici, Luigi V Mancini, Angelo Spognardi, Antonio Villani, and Domenico Vitali. Hacking smart machines with smarter ones: How to extract meaningful data from machine learning classifiers. ar Xiv preprint ar Xiv:1306.4447, 2013. [3] Michael M. Bronstein, Alexander M. Bronstein, Fabrice Michel, and Nikos Paragios. Data fusion through cross-modality metric learning using similarity-sensitive hashing. In CVPR, pages 3594 3601, 2010. [4] Juan C Caicedo and Svetlana Lazebnik. Active object localization with deep reinforcement learning. In CVPR, pages 2488 2496, 2015. [5] Yue Cao, Bin Liu, Mingsheng Long, and Jianmin Wang. Cross-modal hamming hashing. In ECCV, pages 207 223, 2018. [6] Yue Cao, Mingsheng Long, Jianmin Wang, Qiang Yang, and Philip S. Yu. Deep visual-semantic hashing for cross-modal retrieval. In KDD, pages 1445 1454, 2016. [7] Nicholas Carlini and David Wagner. Adversarial examples are not easily detected: Bypassing ten detection methods. In ACM AISEC, pages 3 14. ACM, 2017. [8] Ken Chatfield, Karen Simonyan, Andrea Vedaldi, and Andrew Zisserman. Return of the devil in the details: Delving deep into convolutional nets. In ar Xiv preprint ar Xiv:1405.3531, 2014. [9] Shang-Tse Chen, Cory Cornelius, Jason Martin, and Duen Horng Polo Chau. Shapeshifter: Robust physical adversarial attack on faster r-cnn object detector. In ECML, pages 52 68. Springer, 2018. [10] Tat-Seng Chua, Jinhui Tang, Richang Hong, Haojie Li, Zhiping Luo, and Yantao Zheng. Nus-wide: a real-world web image database from national university of singapore. In CIVR, page 48, 2009. [11] Cheng Deng, Zhaojia Chen, Xianglong Liu, Xinbo Gao, and Dacheng Tao. Triplet-based deep hashing network for cross-modal retrieval. IEEE Transactions on Image Processing, 27(8):3893 3903, 2018. [12] Cheng Deng, Erkun Yang, Tongliang Liu, Jie Li, Wei Liu, and Dacheng Tao. Unsupervised semanticpreserving adversarial hashing for image search. IEEE Transactions on Image Processing, 28(8):4032 4044, 2019. [13] Cheng Deng, Erkun Yang, Tongliang Liu, and Dacheng Tao. Two-stream deep hashing with class-specific centers for supervised image search. IEEE Transactions on Neural Networks and Learning Systems, 2019. [14] Guiguang Ding, Yuchen Guo, and Jile Zhou. Collective matrix factorization hashing for multimodal data. In CVPR, pages 2083 2090, 2014. [15] Venice Erin Liong, Jiwen Lu, Yap-Peng Tan, and Jie Zhou. Cross-modal deep variational hashing. In ICCV, pages 4077 4085, 2017. [16] Aristides Gionis, Piotr Indyk, Rajeev Motwani, et al. Similarity search in high dimensions via hashing. In Vldb, volume 99, pages 518 529, 1999. [17] Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining and harnessing adversarial examples. ar Xiv preprint ar Xiv:1412.6572, 2014. [18] Jiuxiang Gu, Jianfei Cai, Shafiq R Joty, Li Niu, and Gang Wang. Look, imagine and match: Improving textual-visual cross-modal retrieval with generative models. In CVPR, pages 7181 7189, 2018. [19] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In CVPR, pages 770 778, 2016. [20] Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. ar Xiv preprint ar Xiv:1503.02531, 2015. [21] Yan Huang, Qi Wu, Chunfeng Song, and Liang Wang. Learning semantic concepts and order for image and sentence matching. In CVPR, pages 6163 6171, 2018. [22] Mark J Huiskes and Michael S Lew. The mir flickr retrieval evaluation. In MIPR, pages 39 43, 2008. [23] Qing-Yuan Jiang and Wu-Jun Li. Deep cross-modal hashing. In CVPR, pages 3232 3240, 2017. [24] Kuang-Huei Lee, Xi Chen, Gang Hua, Houdong Hu, and Xiaodong He. Stacked cross attention for image-text matching. In ECCV, pages 201 216, 2018. [25] Chao Li, Cheng Deng, Ning Li, Wei Liu, Xinbo Gao, and Dacheng Tao. Self-supervised adversarial hashing networks for cross-modal retrieval. In CVPR, pages 4242 4251, 2018. [26] Chao Li, Cheng Deng, Lei Wang, De Xie, and Xianglong Liu. Coupled cyclegan: Unsupervised hashing network for cross-modal retrieval. In AAAI, 2019. [27] Yeqing Li, Wei Liu, and Junzhou Huang. Sub-selective quantization for learning binary codes in large-scale image search. IEEE transactions on pattern analysis and machine intelligence, 40(6):1526 1532, 2018. [28] Zijia Lin, Guiguang Ding, Mingqing Hu, and Jianmin Wang. Semantics-preserving hashing for cross-view retrieval. In CVPR, pages 3864 3872, 2015. [29] Wei Liu, Cun Mu, Sanjiv Kumar, and Shih-Fu Chang. Discrete graph hashing. In NIPS, pages 3419 3427, 2014. [30] Wei Liu, Jun Wang, Rongrong Ji, Yu-Gang Jiang, and Shih-Fu Chang. Supervised hashing with kernels. In CVPR, pages 2074 2081. IEEE, 2012. [31] Wei Liu, Jun Wang, Sanjiv Kumar, and Shih-Fu Chang. Hashing with graphs. In ICML, pages 1 8, 2011. [32] Wei Liu, Jun Wang, Yadong Mu, Sanjiv Kumar, and Shih-Fu Chang. Compact hyperplane hashing with bilinear functions. In ICML, pages 467 474, 2012. [33] Wei Liu and Tongtao Zhang. Multimedia hashing and networking. IEEE Multi Media, 23(3):75 79, 2016. [34] Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. ar Xiv preprint ar Xiv:1706.06083, 2017. [35] Jan Hendrik Metzen, Tim Genewein, Volker Fischer, and Bastian Bischoff. On detecting adversarial perturbations. ar Xiv preprint ar Xiv:1702.04267, 2017. [36] Seyed-Mohsen Moosavi-Dezfooli, Alhussein Fawzi, and Pascal Frossard. Deepfool: a simple and accurate method to fool deep neural networks. In CVPR, pages 2574 2582, 2016. [37] Nicolas Papernot, Patrick Mc Daniel, and Ian Goodfellow. Transferability in machine learning: from phenomena to black-box attacks using adversarial samples. ar Xiv preprint ar Xiv:1605.07277, 2016. [38] Nicolas Papernot, Patrick Mc Daniel, Xi Wu, Somesh Jha, and Ananthram Swami. Distillation as a defense to adversarial perturbations against deep neural networks. In SP, pages 582 597. IEEE, 2016. [39] Fumin Shen, Chunhua Shen, Wei Liu, and Heng Tao Shen. Supervised discrete hashing. In CVPR, pages 37 45, 2015. [40] Yuming Shen, Li Liu, and Ling Shao. Unsupervised binary representation learning with deep variational networks. International Journal of Computer Vision, pages 1 15, 2019. [41] Mengying Sun, Fengyi Tang, Jinfeng Yi, Fei Wang, and Jiayu Zhou. Identify susceptible locations in medical records via adversarial attacks on deep predictive models. In ACM SIGKDD, pages 793 801. ACM, 2018. [42] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. Intriguing properties of neural networks. ar Xiv preprint ar Xiv:1312.6199, 2013. [43] Bokun Wang, Yang Yang, Xing Xu, Alan Hanjalic, and Heng Tao Shen. Adversarial cross-modal retrieval. In ACMMM, pages 154 162, 2017. [44] Jun Wang, Wei Liu, Sanjiv Kumar, and Shih-Fu Chang. Learning to hash for indexing big data a survey. Proceedings of the IEEE, 104(1):34 57, 2015. [45] Cihang Xie, Jianyu Wang, Zhishuai Zhang, Yuyin Zhou, Lingxi Xie, and Alan Yuille. Adversarial examples for semantic segmentation and object detection. In ICCV, pages 1369 1378, 2017. [46] Xiaojun Xu, Xinyun Chen, Chang Liu, Anna Rohrbach, Trevor Darell, and Dawn Song. Can you fool ai with adversarial examples on a visual turing test. ar Xiv preprint ar Xiv:1709.08693, 2017. [47] Erkun Yang, Cheng Deng, Chao Li, Wei Liu, Jie Li, and Dacheng Tao. Shared predictive cross-modal deep quantization. IEEE Transactions on Neural Networks and Learning Systems, (99):1 12, 2018. [48] Erkun Yang, Cheng Deng, Wei Liu, Xianglong Liu, Dacheng Tao, and Xinbo Gao. Pairwise relationship guided deep hashing for cross-modal retrieval. In AAAI, pages 1618 1625, 2017. [49] Erkun Yang, Tongliang Liu, Cheng Deng, Wei Liu, and Dacheng Tao. Distillhash: Unsupervised deep hashing by distilling data pairs. In CVPR, pages 2946 2955, 2019.