# bayesian_attention_modules__77702186.pdf Bayesian Attention Modules Xinjie Fan ,1, Shujian Zhang ,1, Bo Chen2, and Mingyuan Zhou1 1The University of Texas at Austin and 2Xidian University xfan@utexas.edu, szhang19@utexas.edu, bchen@mail.xidian.edu.cn, mingyuan.zhou@mccombs.utexas.edu Attention modules, as simple and effective tools, have not only enabled deep neural networks to achieve state-of-the-art results in many domains, but also enhanced their interpretability. Most current models use deterministic attention modules due to their simplicity and ease of optimization. Stochastic counterparts, on the other hand, are less popular despite their potential benefits. The main reason is that stochastic attention often introduces optimization issues or requires significant model changes. In this paper, we propose a scalable stochastic version of attention that is easy to implement and optimize. We construct simplex-constrained attention distributions by normalizing reparameterizable distributions, making the training process differentiable. We learn their parameters in a Bayesian framework where a data-dependent prior is introduced for regularization. We apply the proposed stochastic attention modules to various attention-based models, with applications to graph node classification, visual question answering, image captioning, machine translation, and language understanding. Our experiments show the proposed method brings consistent improvements over the corresponding baselines. 1 Introduction Attention modules, aggregating features with weights obtained by aligning latent states, have become critical components for state-of-the-art neural network models in various applications, such as natural language processing [1 4], computer vision [5, 6], graph analysis [7, 8], and multi-modal learning [9 12]. They have been proven to be effective in not only being combined with other types of neural network components, such as recurrent [1, 9] and convolutional units [5, 13], but also being used to build a stand-alone architecture [2, 6, 14]. Besides boosting the performance, using them also often helps aid model visualization and enhance interpretability [1, 9]. While the attention mechanism provides useful inductive bias, attention weights are often treated as deterministic rather than random variables. Consequently, the only source of randomness lies at the model output layer. For example, in classification models, the randomness is in the final logistic regression layer, while in discrete sequence generation, it is in the conditional categorical output layer. However, a single stochastic output layer is often insufficient in modeling complex dependencies [15]. The idea of augmenting deterministic neural networks with latent random variables has achieved success in many fields to model highly structured data, such as texts [16, 17], speeches [15, 18, 19], natural language sequences [20 23], and images [24 26]. Such modification may not only boost the performance, but also provide better uncertainty estimation [27, 28]. As attention weights can be interpreted as alignment weights, it is intuitive to connect the attention module with latent alignment models [9, 29, 30], where latent alignment variables are stochastic and the objective becomes a lower bound of the log marginal likelihood. Making the attention weights stochastic and learning the alignment distribution in a probabilistic manner brings several potential Equal contribution. Corresponding to: mingyuan.zhou@mccombs.utexas.edu 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. advantages. First, adding latent random variable enhances the model s ability to capture complicated dependencies in the target data distribution. Second, we are able to adopt Bayesian inference, where we may build our prior knowledge into prior regularization on the attention weights and utilize posterior inference to provide a better basis for model analysis and uncertainty estimation [28, 29]. Most current work on stochastic attention focus on hard attention [9, 29, 31], where the attention weights are discrete random variables sampled from categorical distributions. However, standard backpropagation no longer applies to the training process of such models and one often resorts to a REINFORCE gradient estimator [32], which has large variance. Such models generally underperform their deterministic counterparts, with a few exceptions, where a careful design of baselines and curriculum learning are required [9, 29, 30]. While attending to multiple positions at one time is intuitively more preferable, probabilistic soft attention is less explored. Bahuleyan et al. [33] propose to use the normal distribution to generate the attention weights, which are hence possibly negative and do not sum to one. Deng et al. [29] consider sampling attention weights from the Dirichlet distribution, which is not reparameterizable and hence not amenable to gradient descent based optimization. In this paper, we propose Bayesian attention modules where the attention weights are treated as latent random variables, whose distribution parameters are obtained by aligning keys and queries. We satisfy the simplex constraint on the attention weights, by normalizing the random variables drawn from either the Lognormal or Weibull distributions. Both distributions generate non-negative random numbers that are reparameterizable. In this way, the whole training process can be made differentiable via the reparameterization trick. We introduce a contextual prior distribution whose parameters are functions of keys to impose a Kullback Leibler (KL) divergence based regularization. To reduce the variance of gradient estimation, we pick the prior distribution such that the KL term can be rewritten in a semi-analytic form, i.e., an expectation of analytic functions. Compared to previous stochastic attentions, our method is much simpler to implement, requires only a few modifications to standard deterministic attention, is stable to train, and maintains good scalability, thereby making it attractive for large-scale deep learning applications. We evaluate the proposed stochastic attention module on a broad range of tasks, including graph node classification, visual question answering, image captioning, machine translation, and language understanding, where attention plays an important role. We show that the proposed method consistently outperforms baseline attention modules and provides better uncertainty estimation. Further, we conduct a number of ablation studies to reason the effectiveness of the proposed model. 2 Preliminaries on attention modules In this section, we briefly review the standard deterministic soft attention modules that have been widely used in various neural networks. Basic module: Consider n key-value pairs, packed into a key matrix K Rn dk and a value matrix V Rn dv, and m queries packed into Q Rm dk, where the dimensions of queries and keys are both equal to dk. Depending on the applications, key, value, and query may have different meanings. For example, in self-attention layers [2], key, value, and query are all from the same source, i.e., the output of the previous layer and in this case m equals to n. In encoder-decoder attention layers, the queries come from the decoder layer, while the keys and values come from the output of the encoder [1, 2, 9]. When attention is used for multi-modal cases, the queries often come from one modality while the keys and values come from the other [11]. Attention modules make use of keys and queries to obtain deterministic attention weights W, which are used to aggregate values V into output features O = WV Rm dv. Specifically, W is obtained through a softmax function across the key dimension as W = softmax(f(Q, K)) Rm n, so that it is a non-negative matrix with each row summing to one. Thus if we denote Φ = f(Q, K), then Wi,j = exp(Φi,j) Pn j =1 exp(Φi,j ). (1) Intuitively, the scale of element Wi,j represents the importance of the jth key to the ith query, and the neural network should learn Q and K such that W gives higher weights to more important features. There are many choices of the alignment score function f, including scaled dot-product [2, 3], additive attention [1, 9, 10], and several other variations [13, 34, 35]. Multi-head and multi-layer attention: Multi-head attention is proposed to attend to information from different representation subspaces [2], where queries, keys, and values are linearly projected H times by H different projection matrices, producing H output values that are concatenated as the final attention layer output. One may then stack attention layers by placing one on top of another, leading to deep attention modules [2, 11]. For a deep attention module with L attention layers and H heads for each layer, the output of the lth layer would be Ol = [W l,1V l,1, ..., W l,HV l,H], where W l,h = softmax(f(Ql,h, Kl,h)), Ql,h = Ql M l,h Q , Kl,h = Kl M l,h K , and V l,h = V l M l,h V for h = 1, ..., H, and the M s are parametric matrices that the neural network needs to learn. Then the output of the lth attention layer, Ol, is fed into the next attention layer (possibly after some transformations) and the queries, keys, and values of the (l + 1)th layer would be functions of Ol. 3 Bayesian attention modules: a general recipe for stochastic attention We suggest a general recipe for stochastic attention: 1) treat attention weights as data-dependent local random variables and learn their distributions in a Bayesian framework, 2) use normalized reparametrizable distributions to construct attention distributions over simplex, and 3) use a key-based contextual prior as regularization. 3.1 Learning attention distributions in a Bayesian way Consider a supervised learning problem with training data D := {xi, yi}N i=1, where we model the conditional probability pθ(yi | xi) using a neural network parameterized by θ, which includes the attention projections M s. For notational convenience, below we drop the data index i. Using vanilla attention modules, the mapping from x to the likelihood pθ(y|x) is deterministic, so the whole model is differentiable meaning that it is tractable to directly maximize the likelihood. Now, we turn the mapping from queries and keys to attention weights W stochastic. Instead of using deterministic weights obtained from queries and keys to aggregate values, we treat W = {W l,h}l=1:L,h=1:H as a set of data-dependent local latent variables sampled from qφ, which can be parameterized by some functions of queries and keys. Intuitively, we argue that this distribution can be viewed as a variational distribution approximating the posterior of local attention weights W, under a Bayesian model, given the data x, y. Therefore, we can learn qφ with amortized variational inference [24]. Note that, unlike Deng et al. [29] and Lawson et al. [30], we do not enforce qφ to be dependent on y, which might not be available during testing. Instead, we use the queries and keys in standard attention modules to construct qφ, so qφ depends on x only or both x and the part of y that has already been observed or generated by the model. For example, in visual question answering or graph node classification, qφ only depends on input x. While in sequence generation, like image captioning or machine translation, qφ could be dependent on the observed part of y as the queries come from y. Constructing variational distribution in such a way has several advantages. First, as we will show in the next section, by utilizing keys and values, transforming a set of deterministic attention weights into an attention distribution becomes straightforward and requires minimal changes to standard attention models. We can even easily adapt pretrained standard attention models for variational finetuning (shown in Section 4.5). Otherwise, building an efficient variational distribution often requires domain knowledge [29] and case by case consideration. Second, due to a similar structure as standard attention modules, qφ introduces little additional memory and computational cost, for which we provide a complexity analysis in Section 3.4. Third, as keys and values are available for both training and testing, we can use the variational distribution qφ during testing. By contrast, previous works [29, 30] enforce qφ to include information not available during testing, restricting its usage at the testing time. Further, as keys and queries depend on the realization of attention weights in previous layers, this structure naturally allows cross-layer dependency between attention weights in different layers so that it is capable of modeling complex distributions. Consider a Bayesian model, where we have prior pη(W) and likelihood pθ(y | x, W) that share a common structure with vanilla deterministic soft attention. We learn the distribution qφ by minimizing KL(qφ(W)||p(W | x, y)), the KL divergence from the posterior distribution of W given x and y to qφ. With amortized variational inference, it is equivalent to maximizing LD = P (x,y) D L(x, y), an evidence lower bound (ELBO) [29, 36, 37] of the intractable log marginal likelihood P (x,y) D log p(y | x) = P (x,y) D log R pθ(y | x, W)pη(W)d W, where L(x, y) := Eqφ(W ) [log pθ(y | x, W)] KL(qφ(W)||pη(W)) = Eqφ(W ) h log pθ(y | x,W )pη(W ) Learning attention distribution qφ via amortized variational inference provides a natural regularization for qφ from prior pη, where we can inject our prior beliefs on attention distributions. We will show we can parameterize the prior distribution with keys, so that the prior distribution can be data-dependent and encode the importance information of each keys. Meanwhile, we can update θ and η to maximize the ELBO. As qφ becomes closer to the posterior, the ELBO becomes a tighter lower bound. 3.2 Reparameterizable attention distributions A challenge of using stochastic attention weights is to optimize their distribution parameters. Existing methods [9, 29, 31] construct attention distributions in a way that standard backpropagation based training no longer applies. Without carefully customizing a training procedure for each specific task, it is generally hard to learn such distributions. Below we introduce reparameterizable soft stochastic attentions that allow effectively optimizing the distribution parameters in a simple and general way. Our goal is to construct a reparameterizable attention distribution qφ over the simplex, i.e., W l,h i,j 0 and P j W l,h i,j = 1. While the Dirichlet distribution, satisfying the simplex constraint and encouraging sparsity, appears to be a natural choice, it is not reparameterizable and hence not amenable to gradient descent based optimization. Here, we consider satisfying the simplex constraint by normalizing random variables drawn from non-negative reparameterizable distributions. In particular, we consider the Weibull and Lognormal distributions. We choose them mainly because they both lead to optimization objectives that are simple to optimize, as described below. Weibull distribution: The Weibull distribution S Weibull(k, λ) has probability density function (PDF) p(S | k, λ) = k λk Sk 1e (S/λ)k, where S R+. Its expectation is λΓ(1 + 1/k) and variance is λ2 Γ (1 + 2/k) (Γ (1 + 1/k))2 . It is reparameterizable as drawing S Weibull(k, λ) is equivalent to letting S = g(ϵ) := λ( log(1 ϵ))1/k, ϵ Uniform(0, 1). It resembles the gamma distribution, and with γ denoted as the Euler Mascheroni constant, the KL divergence from the gamma to Weibull distributions has an analytic expression [17] as KL(Weibull(k, λ)||Gamma(α, β)) = γα k α log λ+log k+βλΓ(1+ 1 k) γ 1 α log β+log Γ(α). Lognormal distribution: The Lognormal distribution S Lognormal(µ, σ2) has PDF p(S | µ, σ) = 1 Sσ 2π exp h (log S µ)2 2σ2 i , where S R+. Its expectation is exp(µ + σ2/2) and vari- ance is [exp(σ2) 1] exp(2µ + σ2). It is also reparameterizable as drawing S Lognormal(µ, σ2) is equivalent to letting S = g(ϵ) = exp(ϵσ + µ), ϵ N(0, 1). The KL divergence is analytic as KL(Lognormal(µ1, σ2 1)||Lognormal(µ2, σ2 2)) = log σ2 σ1 + σ2 1+(µ1 µ2)2 Sampling Sl,h i,j from either the Weibull or Lognormal distribution, we obtain the simplex-constrained random attention weights W by applying a normalization function g over S as W l,h i = g(Sl,h i ) := Sl,h i / P j Sl,h i,j . Note W is reparameterizable but often does not have an analytic PDF. Parameterizing variational attention distributions: To change the deterministic mapping from the queries and keys to attention weights to a stochastic one, we use queries and keys to obtain the distribution parameters of unnormalized attention weights S, which further define the variational distribution of the normalized attention weights W. With the Weibull distribution, we treat k as a global hyperparameter and let λl,h i,j = exp(Φl,h i,j )/Γ(1 + 1/k), and like before Φl,h = f(Ql,h, Kl,h). Then, we sample Sl,h i,j Weibull(k, λl,h i,j ), which is the same as letting Sl,h i,j = exp(Φl,h i,j ) ( log(1 ϵh,l i,j ))1/k Γ(1+1/k) , ϵh,l i,j Uniform(0, 1) . With the Lognor- mal distribution, we treat σ as a global hyperparameter and let µl,h i,j = Φl,h i,j σ2/2. Then, we sample Sl,h i,j Lognormal(µl,h i,j , σ2), which is the same as letting Sl,h i,j = exp(Φl,h i,j ) exp(ϵh,l i,j σ σ2/2), ϵh,l i,j N(0, 1). Note our parameterizations ensure that E[Sl,h i,j ] = exp(Φl,h i,j ). Therefore, if, instead of sampling Sl,h i,j from either distribution, we use its expectation as a substitute, then the mapping becomes equivalent to that of vanilla soft attention, whose weights are defined as in (1). In other words, if we let k of the Weibull distribution go to infinity, or σ of the Lognormal distribution go to zero, which means the variance of Sl,h i,j goes to zero and the distribution becomes a point mass concentrated at the expectation, then the proposed stochastic soft attention reduces to deterministic soft attention. Therefore, the proposed stochastic soft attention can be viewed as a generalization of vanilla deterministic soft attention. We have now constructed qφ to be a reparameterizable distribution W = gφ(ϵ) := g( gφ(ϵ)), where ϵ is a collection of i.i.d. random noises with the same size as W. To estimate the gradient of the ELBO, however, we need either the analytic forms of both pη and qφ, or the analytic form of the KL term, neither of which are available. In the next section, we show how to work around this issue by imposing the KL regularization for latent variable S before normalization and decomposing the joint distribution into a sequence of conditionals. 3.3 Contextual prior: key-dependent Bayesian regularization In the ELBO objective, there is a built-in regularization term, i.e., the KL divergence from the prior distribution pη to variational distribution qφ. To estimate the gradients, we need to evaluate qφ(W) and pη(W) for given attention weights W. We note that even though the analytic form of qφ(W) is not available, qφ(S) = QL l=1 qφ(Sl | S1:l 1) is a product of analytic PDFs (Weibull or Lognormal) for unnormalized weights S, so we rewrite the ELBO in terms of S (we keep using qφ, pη for S as the distribution of W is defined by S), L(x, y) := Eqφ(S) [log pθ(y | x, S)] KL(qφ(S)||pη(S)) (2) Note the KL divergence, as shown in Section 3.2, can be made analytic and hence it is natural to use either the gamma or Lognormal distribution to construct pη(S). Regardless of whether the gamma or Lognormal is used to construct pη(S), due to the dependencies between different stochastic attention layers, we do not have analytic expressions for KL(qφ(S)||pη(S)). Fortunately, as shown in Lemma 1, by decomposing the joint into a sequence of conditionals and exploiting these analytic KL divergence expressions, we can express each KL term in a semi-analytic form. According to the Rao-Blackwellization theorem [38], we can reduce the Monte Carlo estimation variance by plugging in the analytic part. Lemma 1. The KL divergence from the prior to variational distributions is semi-analytic as KL(qφ(S)||pη(S)) = XL l=1 Eqφ(S1:l 1) KL(qφ(Sl|S1:l 1)||pη(Sl|S1:l 1)) | {z } analytic (3) KL(qφ(S)||pη(S)) =Eqφ(S) l=1 (log qφ(Sl|S1:l 1) log pη(Sl|S1:l 1)) l=1 Eqφ(S) [log qφ(Sl|S1:l 1) log pη(Sl|S1:l 1)] l=1 Eqφ(S1:l 1)Eqφ(Sl|S1:l 1) [log qφ(Sl|S1:l 1) log pη(Sl|S1:l 1)] . Key-based contextual prior: Instead of treating the prior as a fixed distribution independent of the input x, here we make the prior depend on the input through keys. The motivation comes from our application in image captioning. Intuitively, given an image (keys), there should be a global prior attention distribution over the image, indicating the importance of each part of the image even before the caption generation process. Based on the prior distribution, the attention distribution can be updated locally using the current state of generation (queries) at the each step (see Figure 1). This intuition can be extended to the general attention framework, where the prior distribution encodes the global importance of each keys shared by all queries, while the posterior encodes the local importance Figure 1: Visualization of attention weight samples from contextual prior distribution and variational distributions at each step for image captioning. Given the image, prior attention distribution over the image areas encodes the importance of each part before the caption generation process. Based on the prior distribution, the attention distribution can be updated at the each step using the current state of generation. of each keys for each query. To obtain the prior parameters, we take a nonlinear transformation of the key features, followed by a softmax to obtain positive values and enable the interactions between keys. Formally, let Ψl,h = softmax(F2(FNL(F1(Kl,h)))) Rn 1, where F1 is linear mapping from Rdk to a hidden dimension Rdmid, F2 is linear mapping from Rdmid to R, and FNL denotes a nonlinear activation function, such as Re LU [39]. With the gamma prior, we treat β as a hyperparameter and let αl,h i,j = Ψl,h i,1. With the Lognormal, we treat σ as a hyperparameter and let µl,h i,j = Ψl,h i,1. Following previous work [40], we add a weight λ to the KL term and anneal it from a small value to one. 3.4 Putting it all together Combining (2) and (3) and using reparameterization, we have Lλ(x, y) = Eϵ[Lλ(x, y, ϵ)], where Lλ(x, y, ϵ) = log pθ(y | x, gφ(ϵ)) λ XL l=1 KL(qφ(Sl | gφ(ϵ1:l 1))||pη(Sl | gφ(ϵ1:l 1))) | {z } analytic . (5) To estimate the gradient of Lλ(x, y) with respect to φ, θ, η, we compute the gradient of Lλ(x, y, ϵ), which is a Monte Carlo estimator with one sample of ϵ. This way provides unbiased and low-variance gradient estimates (see the pseudo code in Algorithm 1 in Appendix). At the testing stage, to obtain point estimates, we adopt the common practice of approximating the posterior means of prediction probabilities by substituting the latent variables by their posterior expectations [41]. To calibrate estimation uncertainties, we draw multiple posterior samples, each of which produces one posterior prediction probability sample. Complexity analysis: Our framework is computationally and memory efficient due to parameter sharing between the variational, prior, and likelihood networks. Extra memory cost comes from the contextual prior network which, for a single layer and single head attention, is of scale O(dkdmid). This is insignificant compared to the memory scale of M s, O(dkdv + dvdv), as dmid is as small as 10 dv. Meanwhile, the additional computations involve the sampling process and computing the KL term which is of scale O(mn). Computing the contextual prior is of scale O(ndkdmid). All above is inconsiderable compared to the computational scale of deterministic attentions, O(mndkdv). 4 Experiments Our method can be straightforwardly implemented in any attention based models. To test the general applicability of our method, we conduct experiments on a wide range of tasks where attention is essential, covering graphs (node classification), multi-modal domains (visual question answering, image captioning), and natural language processing (machine translation, language understanding). A variety of attention types appear in these domains, including self, encoder-decoder, and guided attentions. In this section, we summarize the main experimental settings and results, and include the details in Appendix B. All experiments are conducted on a single Nvidia Tesla V100 GPU with 16 GB memory. Python code is available at https://github.com/zhougroup/BAM 4.1 Attention in graph neural networks We first adapt our method to graph attention networks (GAT) [7], which leverages deterministic self-attention layers to process node-features for graph node classification. The graph structure is encoded in the attention masks in a way that nodes can only attend to their neighborhoods features in the graph. GAT is computationally efficient, capable of processing graphs of different sizes, and achieves state-of-the-art results on benchmark graphs. We use the same model and experimental setup as in GAT [7], as summarized in Appendix B.1. We experiment with three benchmark graphs, including Cora, Citeseer, and Pubmed, for node classification in a transductive setting, meaning that training and testing are performed on different nodes of the same graph [42]. We include a summary of these datasets in Table 6 in Appendix. For large and sparse graph datasets like Pubmed, following GAT [7], we implement a sparse version of the proposed method, where sparse tensor operations are leveraged to limit the memory complexity to be linear in the number of edges. Table 1: Classification accuracy for graphs. Attention Cora Citeseer Pub Med GAT 83.00 72.50 77.26 BAM (NO KL) 83.39 72.91 78.50 BAM-LF 83.24 72.86 78.30 BAM-LC 83.34 73.04 78.76 BAM-WF 83.48 0.2 73.18 0.3 78.50 0.3 BAM-WC 83.81 0.3 73.52 0.4 78.82 0.3 Results. The results are summarized in Table 1. We report the results of soft attention (GAT), and 5 versions of Bayesian Attention Modules (BAM): no KL regularization (NO KL), Lognormal with fixed prior (LF), Lognormal with contextual prior (LC), Weibull and fixed prior (WF), and Weibull and contextual prior (WC). We report the mean classification accuracies on test nodes over 5 random runs, and the standard deviations of BAM-WC. Note the results of GAT are reproduced by running the code provided by the authors (https://github.com/Petar V-/GAT). Our results demonstrate that adapting the deterministic soft attention module in GAT to Bayesian attention consistently improves the performance. Weibull distribution performs better that Lognormal, and contextual prior outperforms fixed prior and no prior. 4.2 Attention in visual question answering models We consider a multi-modal learning task, visual question answering (VQA), where a model predicts an answer to a question relevant to the content of a given image. The recently proposed MCAN [11] uses self-attention to learn the fine-grained semantic meaning of both the image and question, and guided-attention to learn the reasoning between these two modalities. We apply BAM to MCAN and conduct experiments on the VQA-v2 dataset [43], consisting of human-annotated question-answer pairs for images from the MS-COCO dataset [44] (see detailed experiment settings in Appendix B.2). To investigate the model s robustness to noise, we also perturb the input by adding Gaussian noise to the image features [45]. For evaluation, we consider both accuracy and uncertainty, which is necessary here as some questions are so challenging that even human annotators might have different answers. We use a hypothesis testing based Patch Accuracy vs Patch Uncertainty (PAv PU) [46] to evaluate the quality of uncertainty estimation, which reflects whether the model is uncertain about its mistakes. We defer the details of this metric to Appendix B.2.1. Table 2: Results of different attentions on VQA. METRIC ACCURACY PAVPU DATA ORIGINAL NOISY ORIGINAL NOISY SOFT 66.95 61.25 70.04 65.34 BAM-LF 66.89 61.43 69.92 65.48 BAM-LC 66.93 61.58 70.14 65.60 BAM-WF 66.93 61.60 70.09 65.62 BAM-WC 67.02 0.04 62.89 0.06 71.21 0.06 66.75 0.08 Results. In Table 2, we report the accuracy and uncertainty for both original and noisy data (see complete results in Table 7). In terms of accuracy, BAM performs similarly as soft attention on the original dataset, but clearly outperforms it on the more challenging noisy dataset, showing that stochastic soft attention is more robust to noise than deterministic ones. For uncertainty, as soft attention is deterministic we use the dropout in the model to obtain uncertainty, while for BAM we use both dropout and stochastic attention to obtain uncertainty. We observe that on both original and noisy datasets, BAM has better uncertainty estimations, meaning in general it is more uncertain on its mistakes and more certain on its correct predictions. We provide qualitative analysis for uncertainty estimation by visualizing the predictions and uncertainties of three VQA examples in Figure 2 in Appendix. We note that we again observe the improvement from using contextual prior and that Weibull also performs better than Lognormal. 4.3 Attention in image captioning models We further experiment a multi-modal sequence generation task, image captioning, where probabilistic attention (hard attention) was found to outperform deterministic ones [9]. Image captioning models map an image x to a sentence y = (y1, . . . , y T ) that summarizes the image information. Encoderdecoder attention is commonly adopted in state-of-the-art models. During encoding, bottom-up bounding box features [47] are extracted from images by a pretrained Faster R-CNN [48]. At each step of decoding, a weighted sum of bounding box features is injected into the hidden states of an LSTM-based RNN [49, 50] to generate words. The weights are computed by aligning the bounding box features (keys) and hidden states from the last step (queries). We conduct our experiments on MS-COCO [44], following the setup of Luo et al. [51]. For the model architecture, we employ an attention-based model (Att2in) of Rennie et al. [10], which we implement based on the code by Luo et al. [51] and replace the Res Net-encoded features by bounding box features (see details in Appendix B.3). In all experiments, we use maximum likelihood estimation (MLE) for training; we do not consider reinforcement learning based fine-tuning [10, 52, 53], which is beyond the scope of this paper and we leave it as future work. We report four widely used evaluation metrics, including BLEU [54], CIDEr [55], ROUGE [56], and METEOR [57]. Table 3: Comparing different attention modules on image captioning. ATTENTION BLEU-4 BLEU-3 BLEU-2 BLEU-1 CIDER ROUGE METEOR SOFT[9] 24.3 34.4 49.2 70.7 - - 23.9 HARD[9] 25.0 35.7 50.4 71.8 - - 23.0 SOFT (OURS) 32.2 43.6 58.3 74.9 104.0 54.7 26.1 HARD (OURS) 26.5 37.2 51.9 69.8 84.4 50.7 23.3 BAM-LC 32.7 44.0 58.7 75.1 105.0 54.8 26.3 BAM-WC 32.8 0.1 44.1 0.1 58.8 0.1 75.3 0.1 104.5 0.1 54.9 0.1 26.2 0.1 Results. We incorporate the results of both deterministic soft attention and probabilistic hard attention from Xu et al. [9]. We also report those results based on an improved network architecture used by BAM. Results in Table 3 show that the proposed probabilistic soft attention module (BAM) consistently outperforms the deterministic ones. In our implementation, we observe that it is difficult to make hard attention work well due to the high variance of gradients. Moreover, we experiment on modeling the attention weights as Gaussian distribution directly as in Bahuleyan et al. [33]. Our experiment shows that naively modeling attention weights with Gaussian distribution would easily lead to NAN results, as it allows the attention weights to be negative and not sum to 1. Therefore, it is desirable to model attention weights with simplex constrained distributions. We also experiment with BAM where the KL term is completely sampled and observe that the training becomes very unstable and often lead to NAN results. Therefore, constructing prior and posterior in a way that the KL term is semi-analytic does bring clear advantages to our model. In Figure 1, we visualize the prior attention weights and posterior attention weights at each step of generation, where we can visually see how the posterior attention adapts from prior across steps. Further, we again observe that BAM-WC ourperforms BAM-LC and hence for the following tasks, we focus on using BAM-WC. 4.4 Attention in neural machine translation We also experiment with neural machine translation, where we compare with the variational attention method proposed by Deng et al. [29]. We follow them to set the base model structure and experimental setting (see in Appendix B.4), and adapt their deterministic attention to BAM-WC. We compare the BLEU score [54] of BAM-WC and several variants of variational attention in Deng et al. [29]. Results. As shown in Table 4, BAM-WC outperforms deterministic soft attention significantly in BLEU score with same-level computational cost. Moreover, compared to all variants of variational attention [29], BAM-WC achieves better performance with much less training cost, as BAM does not require the training of a completely separate variational network. In Table 8 in Appendix, we compare the run time and number of parameters of BAM and variational attention [26], where we show that BAM achieves better results while being more efficient in both time and memory. It is Table 4: Results on IWSLT. Soft Attention 32.77 Variational Relaxed Attention 30.05 Variational Attention + Enum 33.68 Variational Attention + Sample 33.30 BAM-WC (Ours) 33.81 0.02 interesting to note that in Deng et al. [29] variational relaxed attention (probabilistic soft attention) underperforms variational hard attention, while BAM-WC, which is also probabilistic soft attention, can achieve better results. One of the main reason is that BAM-WC is reparameterizable and has stable gradients, while Deng et al. [29] use the Dirichlet distribution which is not reparameterizable so the gradient estimations still have high variances despite the use of a rejection based sampling and implicit differentiation [58]. Also, we note that, compared to Deng et al. [29], our method is much more general because we do not need to construct the variational distribution on a case-by-case basis. 4.5 Attention in pretrained language models Finally, we adapt the proposed method to finetune deterministic self-attention [2] based language models pretrained on large corpora. Our variational distribution parameters use the pretrained parameters from the deterministic models, and we randomly initialize the parameters for contextual prior. Then we finetune BAM for downstream tasks. We conduct experiments on 8 benckmark datasets from General Language Understanding Evaluation (GLUE) [59] and two versions of Stanford Question Answering Datasets (SQu AD) [60, 61]. We leverage the state-of-the-art pretrained model, ALBERT [4], which is a memory-efficient version of BERT [3] with parameter sharing and embedding factorization. Our implementation is based on Huggingface Py Torch Transformer [62] and we use the base version of ALBERT following the same setting [4] (summarized in Appendix B.5). Table 5: Performance of BAM on GLUE and SQu AD benchmarks. MRPC COLA RTE MNLI QNLI QQP SST STS SQUAD 1.1 SQUAD 2.0 ALBERT-BASE 86.5 54.5 75.8 85.1 90.9 90.8 92.4 90.3 80.86/88.70 78.80/82.07 ALBERT-BASE+BAM-WC 88.5 55.8 76.2 85.6 91.5 90.7 92.7 91.1 81.40/88.82 78.97/82.23 Results. In Table 5, we compare the results of ALBERT, which uses deterministic soft attention finetuned on each dataset with those finetuned with BAM-WC, resuming from the same checkpoints. We observe consistent improvements from using BAM-WC in both GLUE and SQu AD datasets even by only using BAM at the finetuning stage. We leave as future work using BAM at the pretrain stage. 5 Conclusion We have proposed a simple and scalable Bayesian attention module (BAM) that achieves strong performance on a broad range of tasks but requires surprisingly few modifications to standard deterministic attention. The attention weights are obtained by normalizing reparameterizable distributions parameterized by functions of keys and queries. We learn the distributions in a Bayesian framework, introducing a key-dependent contextual prior such that the KL term used for regularization is semianalytic. Our experiments on a variety of tasks, including graph node classification, visual question answering, image captioning, and machine translation, show that BAM consistently outperforms corresponding baselines and provides better uncertainty estimation at the expense of only slightly increased computational and memory cost. Further, on language understanding benchmarks, we show it is possible to finetune a pretrained deterministic attention with BAM and achieve better performance than finetuning with the original deterministic soft attention. With extensive experiments and ablation studies, we demonstrate the effectiveness of each component of the proposed architecture, and show that BAM can serve as an efficient alternative to deterministic attention in the versatile tool box of attention modules. Broader Impact Attention modules have become critical components for state-of-the-art neural network models in various applications, including computer vision, natural language processing, graph analysis, and multi-modal tasks, to name a few. While we show improvements brought by our work on five representative tasks from a broad range of domains, our framework is general enough that it could be used to improve potentially any attention based models. Also, our framework solves two main issues of previously proposed probabilistic attentions that restrict their popularity, i.e., optimization difficulty and complicated model design. We hope that our work will encourage the community to pay more attention to stochastic attention and study from a probabilistic perspective. Considering that attention models have been adopted in many machine learning systems, our work could have an important impact on those systems, such as self-driving [63], healthcare [64], and recommender systems [65]. However, there are potential risks of applying such systems in real-life scenario, because the data we encounter in real-life is biased and long-tailed, and also the discrepancy between training data and testing data might be large. Therefore, an undue trust in deep learning models, incautious usage or imprecise interpretation of model output by inexperienced practitioners might lead to unexpected false reaction in real-life and unexpected consequences. However, we see opportunities that our work can help mitigate the risks with uncertainty estimation. Knowing when mistakes happen would enable us to know when to ask for human-aid if needed for real-life applications [66]. Acknowledgements X. Fan, S. Zhang, and M. Zhou acknowledge the support of Grants IIS-1812699 and ECCS-1952193 from the U.S. National Science Foundation, the support of NVIDIA Corporation with the donation of the Titan Xp GPU used for this research, and the Texas Advanced Computing Center (TACC) at The University of Texas at Austin for providing HPC resources that have contributed to the research results reported within this paper (URL: http://www.tacc.utexas.edu). B. Chen acknowledges the support of the Program for Young Thousand Talent by Chinese Central Government, the 111 Project (No. B18039), NSFC (61771361), Shanxi Innovation Team Project, and the Innovation Fund of Xidian University. [1] Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Bengio. Neural machine translation by jointly learning to align and translate. ar Xiv preprint ar Xiv:1409.0473, 2014. [2] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in neural information processing systems, pages 5998 6008, 2017. [3] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. ar Xiv preprint ar Xiv:1810.04805, 2018. [4] Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. ALBERT: A lite BERT for self-supervised learning of language representations. ar Xiv preprint ar Xiv:1909.11942, 2019. [5] Irwan Bello, Barret Zoph, Ashish Vaswani, Jonathon Shlens, and Quoc V Le. Attention augmented convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, pages 3286 3295, 2019. [6] Prajit Ramachandran, Niki Parmar, Ashish Vaswani, Irwan Bello, Anselm Levskaya, and Jonathon Shlens. Stand-alone self-attention in vision models. ar Xiv preprint ar Xiv:1906.05909, 2019. [7] Petar Veliˇckovi c, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Lio, and Yoshua Bengio. Graph attention networks. ar Xiv preprint ar Xiv:1710.10903, 2017. [8] John Boaz Lee, Ryan A Rossi, Sungchul Kim, Nesreen K Ahmed, and Eunyee Koh. Attention models in graphs: A survey. ACM Transactions on Knowledge Discovery from Data (TKDD), 13(6):1 25, 2019. [9] Kelvin Xu, Jimmy Ba, Ryan Kiros, Kyunghyun Cho, Aaron Courville, Ruslan Salakhudinov, Rich Zemel, and Yoshua Bengio. Show, attend and tell: Neural image caption generation with visual attention. In International conference on machine learning, pages 2048 2057, 2015. [10] Steven J Rennie, Etienne Marcheret, Youssef Mroueh, Jerret Ross, and Vaibhava Goel. Selfcritical sequence training for image captioning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 7008 7024, 2017. [11] Zhou Yu, Jun Yu, Yuhao Cui, Dacheng Tao, and Qi Tian. Deep modular co-attention networks for visual question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 6281 6290, 2019. [12] Xin Wang, Qiuyuan Huang, Asli Celikyilmaz, Jianfeng Gao, Dinghan Shen, Yuan-Fang Wang, William Yang Wang, and Lei Zhang. Reinforced cross-modal matching and self-supervised imitation learning for vision-language navigation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 6629 6638, 2019. [13] Xiaolong Wang, Ross Girshick, Abhinav Gupta, and Kaiming He. Non-local neural networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7794 7803, 2018. [14] Niki Parmar, Ashish Vaswani, Jakob Uszkoreit, Łukasz Kaiser, Noam Shazeer, Alexander Ku, and Dustin Tran. Image Transformer. ar Xiv preprint ar Xiv:1802.05751, 2018. [15] Junyoung Chung, Kyle Kastner, Laurent Dinh, Kratarth Goel, Aaron C Courville, and Yoshua Bengio. A recurrent latent variable model for sequential data. In Advances in neural information processing systems, pages 2980 2988, 2015. [16] Mingyuan Zhou, Yulai Cong, and Bo Chen. Augmentable gamma belief networks. Journal of Machine Learning Research, 17(163):1 44, 2016. [17] Hao Zhang, Bo Chen, Dandan Guo, and Mingyuan Zhou. WHAI: Weibull hybrid autoencoding inference for deep topic modeling. In International Conference on Learning Representations, 2018. [18] Marco Fraccaro, Søren Kaae Sønderby, Ulrich Paquet, and Ole Winther. Sequential neural models with stochastic layers. In Advances in neural information processing systems, pages 2199 2207, 2016. [19] Justin Bayer and Christian Osendorfer. Learning stochastic recurrent networks. ar Xiv preprint ar Xiv:1411.7610, 2014. [20] Samuel R Bowman, Luke Vilnis, Oriol Vinyals, Andrew M Dai, Rafal Jozefowicz, and Samy Bengio. Generating sentences from a continuous space. ar Xiv preprint ar Xiv:1511.06349, 2015. [21] Xuanli He, Gholamreza Haffari, and Mohammad Norouzi. Sequence to sequence mixture model for diverse machine translation. ar Xiv preprint ar Xiv:1810.07391, 2018. [22] Biao Zhang, Deyi Xiong, Jinsong Su, Hong Duan, and Min Zhang. Variational neural machine translation. ar Xiv preprint ar Xiv:1605.07869, 2016. [23] Dandan Guo, Bo Chen, Ruiying Lu, and Mingyuan Zhou. Recurrent hierarchical topic-guided neural language models. In International Conference on Machine Learning, 2020. [24] Diederik P Kingma and Max Welling. Auto-encoding variational Bayes. ar Xiv preprint ar Xiv:1312.6114, 2013. [25] Irina Higgins, Loic Matthey, Arka Pal, Christopher Burgess, Xavier Glorot, Matthew Botvinick, Shakir Mohamed, and Alexander Lerchner. beta-VAE: Learning basic visual concepts with a constrained variational framework. [26] Charles Blundell, Julien Cornebise, Koray Kavukcuoglu, and Daan Wierstra. Weight uncertainty in neural networks. ar Xiv preprint ar Xiv:1505.05424, 2015. [27] Yarin Gal and Zoubin Ghahramani. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In international conference on machine learning, pages 1050 1059, 2016. [28] Yarin Gal, Jiri Hron, and Alex Kendall. Concrete dropout. In Advances in Neural Information Processing Systems, pages 3581 3590, 2017. [29] Yuntian Deng, Yoon Kim, Justin Chiu, Demi Guo, and Alexander Rush. Latent alignment and variational attention. In Advances in Neural Information Processing Systems, pages 9712 9724, 2018. [30] Dieterich Lawson, Chung-Cheng Chiu, George Tucker, Colin Raffel, Kevin Swersky, and Navdeep Jaitly. Learning hard alignments with variational inference. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 5799 5803. IEEE, 2018. [31] Shiv Shankar and Sunita Sarawagi. Posterior attention models for sequence to sequence learning. 2018. [32] Ronald J Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. In Reinforcement Learning, pages 5 32. Springer, 1992. [33] Hareesh Bahuleyan, Lili Mou, Olga Vechtomova, and Pascal Poupart. Variational attention for sequence-to-sequence models. ar Xiv preprint ar Xiv:1712.08207, 2017. [34] Minh-Thang Luong, Hieu Pham, and Christopher D Manning. Effective approaches to attentionbased neural machine translation. ar Xiv preprint ar Xiv:1508.04025, 2015. [35] Alex Graves, Greg Wayne, and Ivo Danihelka. Neural turing machines. ar Xiv preprint ar Xiv:1410.5401, 2014. [36] Matthew D Hoffman, David M Blei, Chong Wang, and John Paisley. Stochastic variational inference. The Journal of Machine Learning Research, 14(1):1303 1347, 2013. [37] David M Blei, Alp Kucukelbir, and Jon D Mc Auliffe. Variational inference: A review for statisticians. Journal of the American Statistical Association, 112(518):859 877, 2017. [38] Art B. Owen. Monte Carlo Theory, Methods and Examples, chapter 8 Variance Reduction. 2013. [39] Vinod Nair and Geoffrey E Hinton. Rectified linear units improve restricted Boltzmann machines. In Proceedings of the 27th international conference on machine learning (ICML-10), pages 807 814, 2010. [40] Samuel R Bowman, Luke Vilnis, Oriol Vinyals, Andrew Dai, Rafal Jozefowicz, and Samy Bengio. Generating sentences from a continuous space. In Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning, pages 10 21, 2016. [41] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: A simple way to prevent neural networks from overfitting. The Journal of Machine Learning Research, 15(1):1929 1958, 2014. [42] Zhilin Yang, William W Cohen, and Ruslan Salakhutdinov. Revisiting semi-supervised learning with graph embeddings. ar Xiv preprint ar Xiv:1603.08861, 2016. [43] Yash Goyal, Tejas Khot, Douglas Summers-Stay, Dhruv Batra, and Devi Parikh. Making the V in VQA matter: Elevating the role of image understanding in visual question answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 6904 6913, 2017. [44] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays, Pietro Perona, Deva Ramanan, Piotr Dollár, and C Lawrence Zitnick. Microsoft COCO: Common objects in context. In European conference on computer vision, pages 740 755. Springer, 2014. [45] Hugo Larochelle, Dumitru Erhan, Aaron Courville, James Bergstra, and Yoshua Bengio. An empirical evaluation of deep architectures on problems with many factors of variation. In Proceedings of the 24th international conference on Machine learning, pages 473 480, 2007. [46] Jishnu Mukhoti and Yarin Gal. Evaluating Bayesian deep learning methods for semantic segmentation. ar Xiv preprint ar Xiv:1811.12709, 2018. [47] Peter Anderson, Xiaodong He, Chris Buehler, Damien Teney, Mark Johnson, Stephen Gould, and Lei Zhang. Bottom-up and top-down attention for image captioning and visual question answering. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 6077 6086, 2018. [48] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun. Faster R-CNN: Towards real-time object detection with region proposal networks. In Advances in neural information processing systems, pages 91 99, 2015. [49] Sepp Hochreiter and Jürgen Schmidhuber. Long short-term memory. Neural computation, 9(8): 1735 1780, 1997. [50] Tomáš Mikolov, Martin Karafiát, Lukáš Burget, Jan ˇCernock y, and Sanjeev Khudanpur. Recurrent neural network based language model. In Eleventh annual conference of the international speech communication association, 2010. [51] Ruotian Luo, Brian Price, Scott Cohen, and Gregory Shakhnarovich. Discriminability objective for training descriptive captions. ar Xiv preprint ar Xiv:1803.04376, 2018. [52] Marc Aurelio Ranzato, Sumit Chopra, Michael Auli, and Wojciech Zaremba. Sequence level training with recurrent neural networks. ar Xiv preprint ar Xiv:1511.06732, 2015. [53] Xinjie Fan, Yizhe Zhang, Zhendong Wang, and Mingyuan Zhou. Adaptive correlated Monte Carlo for contextual categorical sequence generation. In International Conference on Learning Representations, 2020. [54] Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. BLEU: A method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting on association for computational linguistics, pages 311 318. Association for Computational Linguistics, 2002. [55] Ramakrishna Vedantam, C Lawrence Zitnick, and Devi Parikh. CIDEr: Consensus-based image description evaluation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4566 4575, 2015. [56] Chin-Yew Lin. ROUGE: A package for automatic evaluation of summaries. In Text Summarization Branches Out, pages 74 81, Barcelona, Spain, July 2004. Association for Computational Linguistics. URL https://www.aclweb.org/anthology/W04-1013. [57] Satanjeev Banerjee and Alon Lavie. METEOR: An automatic metric for MT evaluation with improved correlation with human judgments. In Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization, pages 65 72, Ann Arbor, Michigan, June 2005. Association for Computational Linguistics. URL https://www.aclweb.org/anthology/W05-0909. [58] Martin Jankowiak and Fritz Obermeyer. Pathwise derivatives beyond the reparameterization trick. ar Xiv preprint ar Xiv:1806.01851, 2018. [59] Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R Bowman. GLUE: A multi-task benchmark and analysis platform for natural language understanding. ar Xiv preprint ar Xiv:1804.07461, 2018. [60] Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. SQu AD: 100,000+ questions for machine comprehension of text. ar Xiv preprint ar Xiv:1606.05250, 2016. [61] Pranav Rajpurkar, Robin Jia, and Percy Liang. Know what you don t know: Unanswerable questions for SQu AD. ar Xiv preprint ar Xiv:1806.03822, 2018. [62] Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, et al. Transformers: State-ofthe-art natural language processing. ar Xiv preprint ar Xiv:1910.03771, 2019. [63] Jinkyu Kim and John Canny. Interpretable learning for self-driving cars by visualizing causal attention. In Proceedings of the IEEE international conference on computer vision, pages 2942 2950, 2017. [64] Edward Choi, Mohammad Taha Bahadori, Jimeng Sun, Joshua Kulas, Andy Schuetz, and Walter Stewart. Retain: An interpretable predictive model for healthcare using reverse time attention mechanism. In Advances in Neural Information Processing Systems, pages 3504 3512, 2016. [65] Yi Tay, Anh Tuan Luu, and Siu Cheung Hui. Multi-pointer co-attention networks for recommendation. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 2309 2318, 2018. [66] Yaniv Ovadia, Emily Fertig, Jie Ren, Zachary Nado, D Sculley, Sebastian Nowozin, Joshua V Dillon, Balaji Lakshminarayanan, and Jasper Snoek. Can you trust your model s uncertainty? evaluating predictive uncertainty under dataset shift. ar Xiv preprint ar Xiv:1906.02530, 2019. [67] Djork-Arné Clevert, Thomas Unterthiner, and Sepp Hochreiter. Fast and accurate deep network learning by exponential linear units (elus). ar Xiv preprint ar Xiv:1511.07289, 2015. [68] Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the thirteenth international conference on artificial intelligence and statistics, pages 249 256, 2010. [69] Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980, 2014. [70] Damien Teney, Peter Anderson, Xiaodong He, and Anton Van Den Hengel. Tips and tricks for visual question answering: Learnings from the 2017 challenge. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4223 4232, 2018. [71] Andrej Karpathy and Li Fei-Fei. Deep visual-semantic alignments for generating image descriptions. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 3128 3137, 2015. [72] Mauro Cettolo, Jan Niehues, Sebastian Stüker, Luisa Bentivogli, and Marcello Federico. Report on the 11th iwslt evaluation campaign, iwslt 2014. [73] Rico Sennrich, Barry Haddow, and Alexandra Birch. Neural machine translation of rare words with subword units. ar Xiv preprint ar Xiv:1508.07909, 2015. [74] Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V Le, Mohammad Norouzi, Wolfgang Macherey, Maxim Krikun, Yuan Cao, Qin Gao, Klaus Macherey, et al. Google s neural machine translation system: Bridging the gap between human and machine translation. ar Xiv preprint ar Xiv:1609.08144, 2016. [75] Alex Warstadt, Amanpreet Singh, and Samuel R Bowman. Neural network acceptability judgments. Transactions of the Association for Computational Linguistics, 7:625 641, 2019. [76] Richard Socher, Alex Perelygin, Jean Wu, Jason Chuang, Christopher D Manning, Andrew Y Ng, and Christopher Potts. Recursive deep models for semantic compositionality over a sentiment treebank. In Proceedings of the 2013 conference on empirical methods in natural language processing, pages 1631 1642, 2013. [77] William B Dolan and Chris Brockett. Automatically constructing a corpus of sentential paraphrases. In Proceedings of the Third International Workshop on Paraphrasing (IWP2005), 2005. [78] Daniel Cer, Mona Diab, Eneko Agirre, Inigo Lopez-Gazpio, and Lucia Specia. Semeval-2017 task 1: Semantic textual similarity-multilingual and cross-lingual focused evaluation. ar Xiv preprint ar Xiv:1708.00055, 2017. [79] Shankar Iyer, Nikhil Dandekar, and Kornél Csernai. First quora dataset release: Question pairs. data. quora. com, 2017. [80] Adina Williams, Nikita Nangia, and Samuel R Bowman. A broad-coverage challenge corpus for sentence understanding through inference. ar Xiv preprint ar Xiv:1704.05426, 2017. [81] Ido Dagan, Oren Glickman, and Bernardo Magnini. The pascal recognising textual entailment challenge. In Machine Learning Challenges Workshop, pages 177 190. Springer, 2005.