# massively_multilingual_sparse_word_representations__335a0c7b.pdf Published as a conference paper at ICLR 2020 MASSIVELY MULTILINGUAL SPARSE WORD REPRESENTATIONS G abor Berend 1,2 1 University of Szeged, Institute of Informatics Szeged, Hungary 2 MTA-SZTE RGAI, Szeged Szeged, Hungary berendg@inf.u-szeged In this paper, we introduce MAMUS for constructing multilingual sparse word representations. Our algorithm operates by determining a shared set of semantic units which get reutilized across languages, providing it a competitive edge both in terms of speed and evaluation performance. We demonstrate that our proposed algorithm behaves competitively to strong baselines through a series of rigorous experiments performed towards downstream applications spanning over dependency parsing, document classification and natural language inference. Additionally, our experiments relying on the QVEC-CCA evaluation score suggests that the proposed sparse word representations convey an increased interpretability as opposed to alternative approaches. Finally, we are releasing our multilingual sparse word representations for the 27 typologically diverse set of languages that we conducted our various experiments on. 1 INTRODUCTION Cross-lingual transferability of natural language processing models is of paramount importance for low-resource languages which often lack a sufficient amount of training data for various NLP tasks. A series of attempts have been made to remedy the shortage of labeled training data. Both part-ofspeech tagging (Fang & Cohn, 2017; Zhang et al., 2016; Haghighi & Klein, 2006; Gouws & Søgaard, 2015; Kim et al., 2015; Agi c et al., 2015) and dependency parsing (Guo et al., 2015; Agi c et al., 2016) have been investigated from that perspective. The mapping of distributed word representations of low-resource languages to the embedding space of a resource-rich language makes NLP models trained on the source language directly applicable for texts in resource-scarce languages. Overcomplete word representations aim at expressing low dimensional distributed word representations as a sparse linear combination of an overcomplete set of basis vectors. Using sparse word representations has not only been reported to give better performance compared to dense representations for certain problems (Yogatama et al., 2015; Faruqui et al., 2015; Sun et al., 2016; Berend, 2017), but it is also argued to provide increased interpretability (Murphy et al., 2012; Vyas & Carpuat, 2016; Park et al., 2017; Subramanian et al., 2018). Such bilingual sparse representations have straightforward benefits as any machine learning model that is trained on the labeled training data using the sparse representation of some resource-rich source language can be used directly and interchangeably for texts written in some target language. This way we can enjoy the benefits of sparsity such as smaller and more interpretable models. In this work, we propose a new algorithm for determining multilingual sparse word representations in which words with similar meaning across different languages are described with similar sparse vectors. Our work has multiple advantageous properties compared to previous similar attempts (Vyas & Carpuat, 2016; Upadhyay et al., 2018). Firstly, our algorithm naturally fits multilingual setting, i.e. it is not limited to pairs of languages, but is capable of determining cross-lingual sparse embeddings for multiple languages. Our algorithm differs from previous approaches in that we obtain target language sparse word representations via a series of convex optimization problems with a substantially reduced parameter set, Published as a conference paper at ICLR 2020 whereas previous solutions perform non-convex optimization for orders of more parameters. The model formulation we propose not only speeds up computation drastically, but our empirical results also suggest that the proposed sparse word vectors perform better when employed in various extrinsic tasks, i.e. dependency parsing, document classification and natural language inference. We make our sparse embeddings for 27 languages and the source code that we used to obtain them publicly available at https://github.com/begab/mamus. 2 MAMUS THE PROPOSED ALGORITHM In order to facilitate the introduction of the MAssively MUltilingual Sparse (dubbed as MAMUS) word representations, we first introduce our notation. In what follows, we denote dense word embedding matrices of the source and target languages as S Rm |Vs| and T Rn |Vt|, respectively, with Vs and Vt indicating the vocabulary of the source and target languages. The m = n condition does not have to hold in general, however, it is going to be the case throughout the paper. We shall denote some symbolic word form as x and its corresponding distributed vectorial representation in boldface, i.e. x. 2.1 PREPROCESSING OF INPUT WORD EMBEDDINGS Our algorithm takes as input a pair of traditionally trained distributed dense word embeddings such as Glove (Pennington et al., 2014), word2vec (Mikolov et al., 2013a) or fasttext (Bojanowski et al., 2017). It is important to note that we do not assume the existence of parallel text, meaning that we can train the input word embeddings in total isolation from each other. We transform the dense input embedding matrices so that the word vectors comprising it have unit norm, as it is frequently met in the literature (Xing et al., 2015; Artetxe et al., 2016; 2017). This preprocessing step ensures that the dot product of word embeddings equals their cosine similarity. Unit normalization of word embeddings also makes cross-lingual comparison more natural, since all embeddings have identical length, irrespective of the language they belong to. 2.2 MAPPING OF WORD EMBEDDINGS In order to align independently trained word representations, we learn a linear mapping W. The mapping is expected to bring target language word vectors close to their semantically equivalent counterparts in the source embedding space. As proposed by Mikolov et al. (2013b) such W can be simply defined by minimizing l P i=1 si Wti , with {(si, ti)}l i=1 being a set of word pairs which are cross-lingual equivalents of each other. If we construct matrices So = [si]l i=1 and To = [ti]l i=1 by stacking the embeddings of translation pairs, we can express the solution as So T + o , with T + o denoting the Moore-Penrose pseudoinverse of To. Multiple studies have argued that ensuring W to be orthonormal can significantly improve the quality of the mapping of word embeddings across languages (Smith et al., 2017; Xing et al., 2015; Artetxe et al., 2016; Hamilton et al., 2016). Finding the optimal orthonormal W can be viewed as an instance of the orthogonal Procrustes problem (Sch onemann, 1966) which can be solved by W = UV , with U and V coming from the singular value decomposition of the matrix product S o To. By applying an isometric transformation, we ensure that transformed embeddings preserve their norm. 2.3 SPARSE CODING OF THE EMBEDDING SPACES Sparse coding decomposes a X Rm v matrix of signals (word embeddings in our case) into the product of a dictionary matrix D Rm k and a matrix of sparse coefficients α Rk v, where k denotes the number of basis vectors to be employed. The columns of D form an overcomplete set of basis vectors and the sparse nonzero coefficients in the ith column of α indicate which column Published as a conference paper at ICLR 2020 vectors from D should be incorporated in the reconstruction of xi. We minimize the objective min D C,α Rk |v| 0 1 2 X Dα 2 F + λ α 1, (1) where C denotes the convex set of matrices with column norm at most 1 and the sparse coefficients in α are required to be non-negative. Ensuring that all the coefficients in α are non-negative makes their cross-lingual comparison more natural, as the signs of sparse word coefficients cannot mismatch. Additionally, non-negativity has been reported to provide increased interpretability (Murphy et al., 2012). We used the SPAMS library (Mairal et al., 2009) for calculating D and α. We can perform sparse coding as defined in (1) for the independently created source and target embedding matrices S and T, obtaining decompositions S Dsαs and T Dtαt. If we do so, however, sparse word representations extractable from matrices αs and αt are not comparable to any extent due to the fact that S and T are decomposed independently of each other. We propose to first solve a single instance of the non-convex dictionary learning problem as defined in (1) for the source language alone as if our goal was to obtain monolingual sparse word representations. After determining Ds, we can apply this dictionary matrix to find sparse coefficient matrices for the isometrically transformed embeddings of all the target languages, hence solving a much simpler complex optimization problem with reasonably fewer parameters in the case of target languages of the form min α Rk |v| 0 1 2 W X Dsα 2 F + λ α 1, (2) with W denoting the isometric transformation specific to a given source-target language pair. We summarize the working mechanisms of MAMUS in Algorithm 1. Algorithm 1 Pseudocode of MAMUS Require: source and target embeddings S Rm |Vs|, T1 Rm |Vt1|, . . . , TN Rm |Vt N | semantically equivalent word pairs {(s(i) 1 , t(i) 1 )}l1 i=1, . . . , {(s(i) N , t(i) N )}l N i=1 Ensure: Sparse representation matrices (αs, αt1, . . . , αt N ) procedure MAMUS(S, T) S UNITNORMALIZE(S) D s, α s arg min α,D S Dα F + λ α 1 for k 1 to N do Tk UNITNORMALIZE(Tk) W arg min W W =I i=1 s(i) k Wt(i) k αtk arg min α R 0 W Tk D sα F + λ α 1 end for return αs, αt1, . . . , αt N end procedure 2.4 DECISIVE DIFFERENCES TO BISPARSE The goal of the BISPARSE (Vyas & Carpuat, 2016) is to determine such sparse word representations that behave similarly for a pair of languages. We subsequently juxtapose the conceptual differences between BISPARSE and MAMUS both from theoretical and practical perspectives. BISPARSE (Vyas & Carpuat, 2016) determines sparse bilingual word representations by solving min Ds,Dt Rm k, αs Rk |Vs| 0 , αt Rk |Vt| 0 Xs Dsαs 2 F + Xt Dtαt 2 F +λx j=1 Mij α(i) s α(j) t 2 2 +λt αt 1+λs αs 1, Published as a conference paper at ICLR 2020 where M is a pairwise similarity matrix between the source and target language word pairs, λs, λt and λx denotes the regularization coefficients for the source, target language and the cross-lingual loss, respectively. Finally, α(i) s and α(j) t refers to the ith and jth columns of matrices αs and αt, respectively. As a first notable difference, BISPARSE operates by employing an explicit |Vs| |Vt| pairwise similarity matrix, whereas MAMUS relies on a list of word translation pairs, which is used to rotate all the target language embeddings into the embedding space of the source language. BISPARSE hence requires an additional hyperparameter which controls for the strength of the cross-lingual regularization (cf. λx). According to our experiments choosing this hyperparameter is a crucial factor in employing BISPARSE. The objective in (3) reveals that the original BISPARSE model optimizes the source and target word representations jointly, i.e. a separate source language representation is created whenever applied to a new target. This behavior of BISPARSE limits its application for a single pair of languages, instead of the more general multilingual setting, when our goal is to represent more than two languages at a time. As the default behavior of BISPARSE hampers truly multilingual application, we adopt it similar to MAMUS, i.e. we first train sparse word representations for the source language by solving min Ds Rm k,αs Rk |Vs| 0 1 2 Xs Dsαs 2 F + λs αs 1, (4) with the source language being chosen as English. Once the dictionary matrix of semantic atoms (D s) and the sparse coefficients (α s) are calculated for the source language, we next solve min Dt Rm k,αt Rk |Vt| 0 1 2 Xt Dtαt 2 F + λt αt 1 + 1 j=1 Mij α (i) s α(j) t 2 2. (5) Identical to BISPARSE (Vyas & Carpuat, 2016), we solved the modified optimization problems in (4) and (5) relying on the efficient Forward-Backward Splitting Solver implementation of the FASTA framework (Goldstein et al., 2014; 2015). To summarize, the way BISPARSE works in its original formulation is that it creates sparse word representations for pairs of languages, both for the source and target language. One compelling characteristic of MAMUS, however, is that it learns the Ds and αs parameters independent from the target embedding space, thus the same decomposition of source language embeddings can be utilized for multiple target languages. This makes our algorithm conveniently applicable to an arbitrary number of languages. Furthermore, by using the dictionary matrix D s for all the target languages, MAMUS not only involves optimization over a substantially reduced set of parameters, it additionally enjoys the ease of solving a series a convex optimization problems. The model formulation of BISPARSE on the other hand is such, that it requires to solve a series of highly non-convex optimization problems for each of the target languages, since it requires the optimization over Dt as well. This property is also true for the modified version of BISPARSE which is adapted for multilingual usage. We will refer to this adapted version of BISPARSE as MULTISPARSE for the rest of the paper. 3 EXPERIMENTS Our primary source for evaluating our proposed representations is the massively multilingual evaluation framework from (Ammar et al., 2016b), which also includes recommended corpora to be used for training word representations for more than 70 languages. All the embeddings used in our experiments were trained over these recommended resources, which is a combination of the Leipzig Corpora Collection (Goldhahn et al., 2012) and Europarl (Koehn, 2005). For 11 languages (bg, cs, da, de, el, es, fi, fr, hu, it, sv) with ample parallel text to English bilingual dictionaries are also released as part of the evaluation framework. For the remaining languages, Ammar et al. (2016b) released dictionaries that were obtained by translating the 20k most common English words with Google Translate. All the word representations involved in our Published as a conference paper at ICLR 2020 Table 1: QVEC-CCA results on the English dev set from (Ammar et al., 2016b) as a function of λ. λ 0.1 0.2 0.3 0.4 0.5 English QVEC-CCA (development set) 0.811 0.806 0.804 0.803 0.799 Avg. nonzero coefficients per word form 22.9 9.1 5.2 3.4 2.3 experiments were trained over the previously introduced resources from (Ammar et al., 2016b) in order to ensure the comparability of the different approaches. We trained fasttext-CBOW (Bojanowski et al., 2017) dense embeddings as inputs to our experiments. We simply used the default settings of fasttext for training, meaning that the original dense word representations were 100 dimensional. We conducted additional experiments with different input embeddings (word2vec and Glove) and also with different dimensionality (300), however, we omit these results for brevity as the general trends we observed were very similar. We set the number of semantic atoms in the dictionary matrix D consistently as k = 1200 throughout all our experiments. 3.1 MONOLINGUAL EXPERIMENTS We first performed monolingual experiments in order to investigate the effects of choosing different regularization coefficients λ, controlling the sparsity and the quality of sparse word representations. Table 1 reports the quality of our monolingual representations evaluated by the QVEC-CCA (Tsvetkov et al., 2016) correlation-based intrinsic evaluation metric, which was proposed as an improvement over the QVEC evaluation technique (Tsvetkov et al., 2015). The goal for both QVEC and QVEC-CCA is to quantify the extent to which embedding dimensions can be aligned to humaninterpretable concepts, such as word supersenses. The evaluation environment we utilize includes supersense tag annotations for Danish (Alonso et al., 2015; Martinez Alonso et al., 2016), English (Miller et al., 1993) and Italian (Montemagni et al., 2003). During the monolingual experiments, we were solely focusing on the development set for English to set the hyperparameter controlling the sparsity of the representations. Table 1 additionally contains the average number of nonzero coefficients yielded by the different regularization coefficients. Based on our monolingual evaluation results from Table 1, we decided to fix the regularization coefficient for MAMUS at λ = 0.1 for all of our upcoming mutlilingual experiments. As our choice for the value of λ was based on the performance achieved by MAMUS on the development set of a single language and a single evaluation criterion, we chose this hyperparameter without the risk of overfitting to the test data regarding any of the multilingual evaluation scenarios. In the case of MULTISPARSE, we managed to obtain a similar performance and average sparseness (55.2 nonzero coefficient per word form) when utilizing λs = 2. During our multilingual experiments, we set the target language regularization coefficient identically, i.e. λt = 2. For the cross-lingual regularization term of MULTISPARSE, we chose λx = 5 based on our preliminary investigation based on the remaining two languages we had supersense tag annotations for, i.e. Danish and Italian. The above set of hyperparameters resulted in 41.0 and 44.3 nonzero coefficients per word form on average, respectively. 3.1.1 MONOLINGUAL EVALUATION OF BISPARSE AND MULTISPARSE REPRESENTATIONS We conducted an additional monolingual QVEC-CCA evaluation on Danish, English and Italian. In this experiment we compared the results of sparse word representations obtained by the vanilla BISPARSE formulation and its multilingual extension MULTISPARSE. This way we can assess the effects of multilingual training as opposed to bilingual one. The QVEC-CCA scores we obtained for the standard BISPARSE representations were 0.602, 0.789/0.786 and 0.585 for Danish, English and Italian, respectively. We report two English results (0.789/0.786) as BISPARSE provided us with two distinct representations for English depending on the target language it was jointly trained with. MULTISPARSE representations on the other hand obtained QVEC-CCA scores of 0.612, 0.808 and 0.596 for Danish, English and Italian, respectively. Published as a conference paper at ICLR 2020 Our monolingual QVEC-CCA results suggest that the multilingual training objective performs comparably (or even better) to the bilingual one employed in BISPARSE. A possible explanation for this is that BISPARSE tries to optimize a more difficult objective as it involves the joint optimization of two non-convex problems with an increased number of parameters hence finding a good solution is more difficult for BISPARSE. 3.2 MULTILINGUAL EXPERIMENTS In the recent survey, Ruder et al. (2019) recommend to evaluate cross-lingual word embeddings on an intrinsic task and at least one downstream task besides document classification. Following this suggestion, we measure the quality of word representations based on QVEC-CCA similar to our monolingual evaluation and also report their performance scores in cross-lingual document classification (CLDC), cross-lingual dependency parsing and cross-lingual natural language inference (XNLI). Results from this point on are all obtained with the exact same hyperparameters as introduced before and on the test sets of the various tasks, unless stated otherwise. 3.2.1 BASELINE DENSE CROSS-LINGUAL REPRESENTATIONS Before delving into the multilingual experimental results, we introduce those additional dense distributed cross-lingual representations that we used in our evaluation. These representations were trained over the same corpora and translation lists introduced earlier. multi Cluster (Ammar et al., 2016b) builds a graph from word forms across multiple languages with edges going between translated word pairs. Clusters are then formed based on the connected components of the graph. Word forms are finally replaced by the random identifier of the cluster they belong to and a skip-gram model is trained for the corpus obtained this way. multi CCA (Ammar et al., 2016b) is an extension of the approach introduced by Faruqui & Dyer (2014). It seeks a linear operator which projects pre trained word embeddings of some language to the embedding space of English such that the correlation between the representation of translated word pairs is maximized. Even though we trained our representations and further baselines on the same corpora and word translation lists as the multi Cluster and multi CCA were built on, minor differences in the learned vocabularies exists (even between multi Cluster and multi CCA) which are likely due to small differences in preprocessing. For this reason, pre trained multi Cluster and multi CCA allow for a less stringent comparison with the remaining representations, which were trained by ensuring to have identical vocabularies to the one employed in our approach. With this caution in mind, we still think it is useful to report evaluation results of these pre-trained embeddings as well. MUSE (Lample et al., 2018) is capable of training dense cross-lingual word embeddings in both unsupervised and supervised manner. We experiment with the supervised variant of MUSE which requires translated word pairs as input similar to MAMUS. MUSE also incorporates a method for iteratively refining the list of word translation pairs that we relied on using its default parametrization. 3.2.2 MULTILINGUAL EXPERIMENTAL RESULTS QVEC-CCA We next assess the quality of the different multilingual word representations according to the QVEC-CCA evaluation score. As mentioned earlier, the evaluation framework introduced by Ammar et al. (2016b) provides supersense tagging for Danish, English and Italian. In Table 2a, we report the individual evaluation scores over the three subspaces and their combination, from which we can conclude that MAMUS has a clear advantage over all the alternative word representation regarding this interpretability oriented evaluation metric. Downstream evaluations from Ammar et al. (2016b) We also performed a 4-class cross-lingual document classification (CLDC) on newswire texts originating from the RCV corpus (Lewis et al., 2004) over 7 languages (da, de, en, es, fr, it, sv). The model was simultaneously tuned on the training sections of the different languages as implemented in (Ammar et al., 2016b). Table 2b lists classification accuracies of the different methods. Published as a conference paper at ICLR 2020 Table 2: Intrinsic (a) and extrinsic (b) evaluations from the test bed from (Ammar et al., 2016b). (a) Intrinsic QVEC-CCA evaluation scores of various crosslingual word representations. da en it {da,en,it} multi Cluster 0.475 0.539 0.468 0.372 multi CCA 0.501 0.635 0.510 0.415 MUSE 0.343 0.393 0.338 0.294 MULTISPARSE 0.612 0.808 0.596 0.480 MAMUS 0.632 0.819 0.620 0.503 (b) Downstream evaluation results as average accuracy for CLDC and UAS for dependency parsing. CLDC dependency parsing 90.79 61.39 92.18 62.82 87.34 64.47 86.45 59.20 91.84 63.53 Regarding our evaluation towards dependency parsing, we evaluated the transition-based stack LSTM parser from (Ammar et al., 2016a) over the Universal Dependencies v1.1 treebanks (Agi c et al., 2015) covering 18 languages (bg, cs, da, de, el, en, es, eu, fa, fi, fr, ga, he, hr, hu, id, it, sv). Even though the parser is capable of incorporating additional features besides word embeddings, this capability of the parser was disabled, so that the effects of employing different word representations can be assessed on their own. The performance metric for parsing is reported in Table 2b as unlabeled attachment score (UAS). Table 2b illustrates that the performance obtained by dense representations varies largely between downstream tasks. The performance of MUSE, for instance, is the best when evaluated on dependency parsing, however, it has the second lowest overall accuracy for CLDC. Evaluation scores for those models that are based on MAMUS representations rank second on both downstream evaluation tasks with a minor performance gap to the best results obtained by different dense representations on the two tasks. Natural Language Inference In order to assess the capabilities of the different representations towards natural language inference (NLI) in a multilingual setting, we also performed evaluation towards the XNLI dataset (Conneau et al., 2018). XNLI covers 15 languages and it can be viewed as a multilingual extension of the multi NLI (Williams et al., 2018) dataset. The task in XNLI and multi NLI is to categorize sentence pairs comprising of a premise (p) and a hypothesis (h) sentence whether the relation between p and h is entailing, contradictory or neutral. We implemented a simple multilayer perceptron in Py Torch v1.1 (Paszke et al., 2017) with two hidden layers employing Re LU nonlinearity. The MLP uses the categorical cross-entropy for loss function, which was optimized by Adam (Kingma & Ba, 2014). Based on the differently constructed sparse and dense word representations, we train five different NLI models based on the English multi NLI dataset and report average classification accuracies as the performance score. The five models only differed in the random initialization of the parameters to account for the potential variability in model performances. For a pair of premise and hypothesis sentence representation pair (p, h), the input to the MLP gets determined according to [p; h; p h; p h], similar to (Williams et al., 2018) with ; standing for vector concatenation and denoting elementwise multiplication. The vectorial representations for the individual sentences were also obtained in an identical way to the approach (dubbed as X-CBOW) in (Conneau et al., 2018), i.e. we took the mean of the vectorial representations of the words comprising a sentence. We report the average accuracy achieved by the different word representations for multi NLI (Table 3a) and the 15 languages of XNLI (Table 3b). Although the training corpora compiled by Ammar et al. (2016b) contains datasets for all the XNLI languages, the authors did not release any pre-trained multi Cluster and multi CCA embeddings for Hindi (hi), Thai (th), Urdu (ur) and Vietnamese (vi). As a result, we do not report multi Cluster and multi CCA results for these four languages. We can see in Table 3a that according to the evaluation conducted on the development set of multi NLI multi CCA performed the best (similar to CLDC), and the English subspace of the MAMUS representations achieved the second best results. Table 3b represents the average performance Published as a conference paper at ICLR 2020 Table 3: Averaged multi NLI and XNLI performances of 5 independent MLPs. (a) Average multi NLI matched dev set accuracy of the five models used during XNLI evaluation. multi Cluster multi CCA MUSE MULTISPARSE MAMUS 57.74 63.43 57.10 60.22 61.43 (b) XNLI results obtained. Results in bold are the best for a given language, underlined scores indicate the second best resutls. ar bg de el en es fr ru sw tr zh hi th ur vi multi Cluster 33.36 49.53 50.00 49.52 58.84 48.86 48.46 39.47 35.91 36.97 36.05 multi CCA 38.85 40.59 34.50 42.70 65.43 36.16 33.83 41.25 38.31 38.82 45.36 MUSE 42.83 45.64 41.60 46.24 58.39 44.53 43.64 44.70 37.33 42.21 44.57 38.98 33.50 38.57 35.92 MULTISPARSE 33.10 35.04 38.17 33.33 60.10 40.48 40.92 33.34 36.20 37.78 33.23 35.40 33.06 34.18 34.63 MAMUS 42.67 47.88 44.27 49.58 61.53 51.19 48.93 43.89 37.09 47.24 44.73 41.78 34.59 42.00 34.94 ar bg de el en es fr ru sw tr zh hi th ur vi 0.0 Mamus m BERT Figure 1: Comparison of the accuracies over the 15 languages of the XNLI datasets. of five models that are purely trained on the English multi NLI training data relying on the various multilingual word representations upon evaluation. It can be seen in Table 3b that other than for Swahili the models trained on the English MAMUS representations perform either the best or the second best 14 out of the 15 languages. In order to account for the recent dominance of contextualized word representations, we additionally compared the application of MAMUS to that of multilingual BERT (Devlin et al., 2019) when utilized for XNLI. We employed multilingual BERT (m BERT) representations from Wolf et al. (2019) identically to how we used the non-contextualized word representations previously. That is, we took the average of the contextualized representations that we derived from the last layer of m BERT and trained 5 MLPs with different weight initializations. Similar to our previous experiments, we conducted evaluation in the zero-shot setting for the 15 XNLI languages, i.e. by solely using multi NLI for training our models. The comparative results of the trained MLPs can be seen in Figure 1. Figure 1 reveals that the MLPs trained on top of MAMUS representations are capable of achieving more than 90% relative performance on average to those MLPs which rely on m BERT. Indeed, the average performance of the MLPs over all the languages and random initializations is 49.3 for m BERT, while MAMUS-based MLPs had an average accuracy of 44.8. We thus think that MAMUS can be a viable alternative to m BERT in multilingual settings when a small fraction of accuracy can be sacrificed for obtaining a model in which the calculation of sentence representations is substantially cheaper (cf. the computational requirements of averaging static word embeddings and contextual embeddings which require a forward pass in a transformer architecture). Interestingly, when inspecting results for English in Figure 1, we can see that MLPs taking MAMUS representations as input actually perform better than those utilizing m BERT by a fair margin (61.5 versus 58.3). This could imply that our algorithm has the potential to transform static dense embeddings in a way which makes them as useful as contextualized m BERT representations for certain applications. Note that this hypothesis would require further investigation, that is left for future work. Published as a conference paper at ICLR 2020 0 50 100 150 200 250 300 Frequency bin (1k word per bin) Avg. nnz/embedding method Mamus multisparse 0 20 40 60 80 100 120 140 Frequency bin (1k word per bin) Avg. nnz/embedding method Mamus multisparse 0 50 100 150 200 Frequency bin (1k word per bin) Avg. nnz/embedding method Mamus multisparse Figure 2: The sparsity structure of the learned representations for the languages covered in the multilingual QVEC-CCA evaluation as a function of the frequency of word forms. Plots also include the frequency-binned average norms of the dense input embeddings. The orange curve indicates the norms of the embeddings per frequency bins, the blue are for the number of nonzero coefficients per word forms. 3.2.3 ANALYSIS OF THE SPARSITY STRUCTURE As a final assessment of the sparse word representations, we characterize their number of nonzero coefficients as a function of the frequency of the words they correspond to. From a human cognition point of view, we can argue that frequency can be a good proxy to the specialization in the meaning of a word (Caraballo & Charniak, 1999). Words with high frequency, e.g. car, dog and newspaper tend to refer to easily definable concepts, whereas less frequent words, such as gymnosperms or samizdat have a more complex hence more difficult to describe meaning. In terms of sparse coding, this could be reflected by the fact that words with more complex meaning would rely on more semantic atoms from the dictionary matrix D. Encoding less frequent words with more bits can also be motivated from an information theoretic point of view. In Figure 2, we plot the average number of nonzero elements over the entire vocabulary of the Danish, English and Italian subspaces (grouped into bins of 1,000 words). We omit similar figures for the rest of the languages for space considerations. Figure 2 reveals that the sparsity structures induced by the different approaches differ substantially. MAMUS tends to behave more stable across the languages, i.e. it tends to assign a similar amount of nonzero coefficients for words in different languages that belong to the same frequency bin. The number of nonzero coefficients per a word form determined by MULTISPARSE, however behaves less predictably. Figure 2 also illustrates that the sparsity structure of MULTISPARSE is highly influenced by the norms of the dense input embeddings, which are known to be typically higher for more common word forms (Turian et al., 2010). Representations determined by MAMUS, however, tend to assign more nonzero coefficients to less frequent hence arguably more specialized word forms. 4 RELATED WORK A common technique employed to overcome the absence of labeled training data is to apply cross-lingual projections to port the missing linguistic annotation for low-resource target languages (Yarowsky & Ngai, 2001; Das & Petrov, 2011; T ackstr om et al., 2013; Agi c et al., 2015; Agi c et al., 2016). Such projections are often determined by word alignment algorithms which imply that these kind of approaches inherently require substantial amounts of parallel text. There has been a series of research conducted for handling low-resource languages. A dominant approach is to rely on some form of dictionary between a low-resource and a resource-rich language which can be used to perform canonical correlation analysis (Kim et al., 2015), directly incorporated into the learning procedure of word embeddings (Gouws & Søgaard, 2015) or use it in a post-hoc manner to map independently trained word embeddings (Fang & Cohn, 2017; Zhang et al., 2016). Published as a conference paper at ICLR 2020 Model transfer techniques, in which an initial model is trained on a resource-rich language and adapted for a target language in a semi-supervised or unsupervised manner, are also popular (Fang & Cohn, 2017; Zhang et al., 2016). There is massive research interest in transforming word representations such that they become comparable across languages with little or no supervision (Zhang et al., 2017b;a; Artetxe et al., 2017; Smith et al., 2017; Lample et al., 2018; Joulin et al., 2018). There has been a myriad of further techniques introduced for determining cross-lingual distributed word representations (Klementiev et al., 2012; Hermann & Blunsom, 2014; Faruqui & Dyer, 2014; Huang et al., 2015; Luong et al., 2015; Gouws et al., 2015; Vuli c & Moens, 2015; Ammar et al., 2016b), inter alia. The proposed approaches differ widely in the assumptions they make regarding the amount of available parallel or comparable data for determining bilingual word embeddings. Upadhyay et al. (2016) and Ruder et al. (2019) provide extensive overviews on the available approaches. Contrary to these methods, our algorithm relates to the line of research focusing on sparse word representations. Such sparse word representations have been shown to outperform dense word embeddings in monolingual settings (Murphy et al., 2012; Yogatama et al., 2015; Faruqui et al., 2015; Berend, 2017; Sun et al., 2016; Subramanian et al., 2018). Vyas & Carpuat (2016) proposed BISPARSE for obtaining comparable sparse word representations in a bilingual setting. As mentioned earlier, BISPARSE needs to be modified in order to be applicable for obtaining multilingual representations. Perhaps even more importantly, even when applied in the bilingual case, BISPARSE needs to perform a highly non-convex optimization problem and for a substantially larger number of parameters in contrast to MAMUS. Upadhyay et al. (2018) extends (Vyas & Carpuat, 2016) by additionally incorporating dependency relations into sparse coding. 5 CONCLUSIONS In this paper we introduced MAMUS for determining cross-lingually comparable sparse word representations. Our model formulation allowed us to solve a series of convex optimization problems per each target language, which resulted in a more favorable overall training time (4 hours for MAMUS as opposed to 300 hours when using MULTISPARSE) over the 27 languages we conducted our evaluations on. Finally, we make our multilingual sparse embeddings for 27 languages publicly available at https://github.com/begab/mamus. ACKNOWLEDGEMENTS This research was partly funded by the project Integrated program for training new generation of scientists in the fields of computer science , no EFOP-3.6.3-VEKOP-16-2017-0002, supported by the EU and co-funded by the European Social Fund. This work was in part supported by the National Research, Development and Innovation Office of Hungary through the Artificial Intelligence National Excellence Program (grant no.: 2018-1.2.1-NKP-2018-00008). This research was supported by grant TUDFO/47138-1/2019-ITM of the Ministry for Innovation and Technology, Hungary. ˇZeljko Agi c, Dirk Hovy, and Anders Søgaard. If all you have is a bit of the Bible: Learning POS taggers for truly low-resource languages. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp. 268 272, Beijing, China, July 2015. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/ P15-2044. ˇZeljko Agi c, Maria Jesus Aranzabe, Aitziber Atutxa, Cristina Bosco, Jinho Choi, Marie-Catherine de Marneffe, Timothy Dozat, Rich ard Farkas, Jennifer Foster, Filip Ginter, Iakes Goenaga, Koldo Gojenola, Yoav Goldberg, Jan Hajiˇc, Anders Trærup Johannsen, Jenna Kanerva, Juha Kuokkala, Veronika Laippala, Alessandro Lenci, Krister Lind en, Nikola Ljubeˇsi c, Teresa Lynn, Christopher Manning, H ector Alonso Mart ınez, Ryan Mc Donald, Anna Missil a, Simonetta Montemagni, Joakim Nivre, Hanna Nurmi, Petya Osenova, Slav Petrov, Jussi Piitulainen, Barbara Plank, Published as a conference paper at ICLR 2020 Prokopis Prokopidis, Sampo Pyysalo, Wolfgang Seeker, Mojgan Seraji, Natalia Silveira, Maria Simi, Kiril Simov, Aaron Smith, Reut Tsarfaty, Veronika Vincze, and Daniel Zeman. Universal dependencies 1.1, 2015. URL http://hdl.handle.net/11234/LRT-1478. LINDAT/CLARIN digital library at the Institute of Formal and Applied Linguistics ( UFAL), Faculty of Mathematics and Physics, Charles University. ˇZeljko Agi c, Anders Johannsen, Barbara Plank, H ector Alonso Mart ınez, Natalie Schluter, and Anders Søgaard. Multilingual projection for parsing truly low-resource languages. Transactions of the Association for Computational Linguistics, 4:301 312, 2016. H ector Mart ınez Alonso, Anders Johannsen, Sussi Olsen, Sanni Nimb, Nicolai Hartvig Sørensen, Anna Braasch, Anders Søgaard, and Bolette Sandford Pedersen. Supersense tagging for Danish. In Proceedings of the 20th Nordic Conference of Computational Linguistics, NODALIDA 2015, May 11-13, 2015, Institute of the Lithuanian Language, Vilnius, Lithuania, pp. 21 29, 2015. URL http://aclweb.org/anthology/W/W15/W15-1806.pdf. Waleed Ammar, George Mulcaire, Miguel Ballesteros, Chris Dyer, and Noah Smith. Many languages, one parser. Transactions of the Association for Computational Linguistics, 4:431 444, 2016a. ISSN 2307-387X. URL https://www.transacl.org/ojs/index.php/ tacl/article/view/892. Waleed Ammar, George Mulcaire, Yulia Tsvetkov, Guillaume Lample, Chris Dyer, and Noah A. Smith. Massively multilingual word embeddings. Co RR, abs/1602.01925, 2016b. Mikel Artetxe, Gorka Labaka, and Eneko Agirre. Learning principled bilingual mappings of word embeddings while preserving monolingual invariance. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, pp. 2289 2294. Association for Computational Linguistics, 2016. doi: 10.18653/v1/D16-1250. URL http://www.aclweb.org/ anthology/D16-1250. Mikel Artetxe, Gorka Labaka, and Eneko Agirre. Learning bilingual word embeddings with (almost) no bilingual data. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 451 462. Association for Computational Linguistics, 2017. doi: 10.18653/v1/P17-1042. URL http://aclweb.org/anthology/P17-1042. G abor Berend. Sparse coding of neural word embeddings for multilingual sequence labeling. Transactions of the Association for Computational Linguistics, 5:247 261, 2017. ISSN 2307-387X. URL https://transacl.org/ojs/index.php/tacl/article/view/1063. Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. Enriching word vectors with subword information. Transactions of the Association of Computational Linguistics, 5:135 146, 2017. URL http://www.aclweb.org/anthology/Q17-1010. Sharon A. Caraballo and Eugene Charniak. Determining the specificity of nouns from text. In Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora, EMNLP 1999, College Park, MD, USA, June 21-22, 1999, 1999. URL https:// aclanthology.info/papers/W99-0609/w99-0609. Alexis Conneau, Ruty Rinott, Guillaume Lample, Adina Williams, Samuel Bowman, Holger Schwenk, and Veselin Stoyanov. XNLI: Evaluating cross-lingual sentence representations. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 2475 2485, Brussels, Belgium, October-November 2018. Association for Computational Linguistics. doi: 10.18653/v1/D18-1269. URL https://www.aclweb.org/anthology/ D18-1269. Dipanjan Das and Slav Petrov. Unsupervised part-of-speech tagging with bilingual graph-based projections. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies - Volume 1, HLT 11, pp. 600 609, Stroudsburg, PA, USA, 2011. Association for Computational Linguistics. ISBN 978-1-932432-87-9. URL http://dl.acm.org/citation.cfm?id=2002472.2002549. Published as a conference paper at ICLR 2020 Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp. 4171 4186, Minneapolis, Minnesota, June 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1423. URL https: //www.aclweb.org/anthology/N19-1423. Meng Fang and Trevor Cohn. Model transfer for tagging low-resource languages using a bilingual dictionary. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 587 593. Association for Computational Linguistics, 2017. doi: 10.18653/v1/P17-2093. URL http://www.aclweb.org/anthology/P17-2093. Manaal Faruqui and Chris Dyer. Improving vector space word representations using multilingual correlation. In Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, pp. 462 471, Gothenburg, Sweden, April 2014. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/E14-1049. Manaal Faruqui, Yulia Tsvetkov, Dani Yogatama, Chris Dyer, and Noah A. Smith. Sparse overcomplete word vector representations. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 1491 1500, Beijing, China, July 2015. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/P15-1144. Dirk Goldhahn, Thomas Eckart, and Uwe Quasthoff. Building large monolingual dictionaries at the Leipzig Corpora collection: From 100 to 200 languages. In In Proceedings of the Eight International Conference on Language Resources and Evaluation (LREC 12, 2012. Tom Goldstein, Christoph Studer, and Richard Baraniuk. A field guide to forward-backward splitting with a FASTA implementation. ar Xiv eprint, abs/1411.3406, 2014. URL http://arxiv. org/abs/1411.3406. Tom Goldstein, Christoph Studer, and Richard Baraniuk. FASTA: A generalized implementation of forward-backward splitting, January 2015. http://arxiv.org/abs/1501.04979. Stephan Gouws and Anders Søgaard. Simple task-specific bilingual word embeddings. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1386 1390, Denver, Colorado, May June 2015. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/ N15-1157. Stephan Gouws, Yoshua Bengio, and Greg Corrado. Bil BOWA: Fast bilingual distributed representations without word alignments. In Proceedings of the 32nd International Conference on Machine Learning, 2015. URL http://jmlr.org/proceedings/papers/v37/ gouws15.pdf. Jiang Guo, Wanxiang Che, David Yarowsky, Haifeng Wang, and Ting Liu. Cross-lingual dependency parsing based on distributed representations. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp. 1234 1244, Beijing, China, July 2015. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/ P15-1119. Aria Haghighi and Dan Klein. Prototype-driven learning for sequence models. In Proceedings of the Main Conference on Human Language Technology Conference of the North American Chapter of the Association of Computational Linguistics, HLT-NAACL 06, pp. 320 327, Stroudsburg, PA, USA, 2006. Association for Computational Linguistics. doi: 10.3115/1220835.1220876. URL https://doi.org/10.3115/1220835.1220876. William L. Hamilton, Jure Leskovec, and Dan Jurafsky. Diachronic word embeddings reveal statistical laws of semantic change. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1489 1501, Berlin, Germany, August 2016. Association for Computational Linguistics. URL http://www.aclweb.org/ anthology/P16-1141. Published as a conference paper at ICLR 2020 Karl Moritz Hermann and Phil Blunsom. Multilingual Models for Compositional Distributional Semantics. In Proceedings of ACL, June 2014. URL http://arxiv.org/abs/1404. 4641. Kejun Huang, Matt Gardner, Evangelos Papalexakis, Christos Faloutsos, Nikos Sidiropoulos, Tom Mitchell, Partha P. Talukdar, and Xiao Fu. Translation invariant word embeddings. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1084 1088, Lisbon, Portugal, September 2015. Association for Computational Linguistics. URL http: //aclweb.org/anthology/D15-1127. Armand Joulin, Piotr Bojanowski, Tomas Mikolov, Herv e J egou, and Edouard Grave. Loss in translation: Learning bilingual word mapping with a retrieval criterion. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 2979 2984. Association for Computational Linguistics, 2018. URL http://aclweb.org/anthology/ D18-1330. Young-Bum Kim, Benjamin Snyder, and Ruhi Sarikaya. Part-of-speech taggers for low-resource languages using CCA features. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 1292 1302, Lisbon, Portugal, 2015. Association for Computational Linguistics. URL http://aclweb.org/anthology/D15-1150. Diederik P. Kingma and Jimmy Ba. Adam: A method for stochastic optimization. Co RR, abs/1412.6980, 2014. Alexandre Klementiev, Ivan Titov, and Binod Bhattarai. Inducing crosslingual distributed representations of words. In Proceedings of COLING 2012, pp. 1459 1474. The COLING 2012 Organizing Committee, 2012. URL http://www.aclweb.org/anthology/C12-1089. Philipp Koehn. Europarl: A Parallel Corpus for Statistical Machine Translation. In Conference Proceedings: the tenth Machine Translation Summit, pp. 79 86, Phuket, Thailand, 2005. AAMT, AAMT. URL http://mt-archive.info/MTS-2005-Koehn.pdf. Guillaume Lample, Alexis Conneau, Marc Aurelio Ranzato, Ludovic Denoyer, and Herv e J egou. Word translation without parallel data. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=H196sainb. David D. Lewis, Yiming Yang, Tony G. Rose, and Fan Li. Rcv1: A new benchmark collection for text categorization research. J. Mach. Learn. Res., 5:361 397, December 2004. ISSN 1532-4435. URL http://dl.acm.org/citation.cfm?id=1005332.1005345. Thang Luong, Hieu Pham, and Christopher D. Manning. Bilingual word representations with monolingual quality in mind. In Proceedings of the 1st Workshop on Vector Space Modeling for Natural Language Processing, pp. 151 159, Denver, Colorado, June 2015. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/W15-1521. Julien Mairal, Francis Bach, Jean Ponce, and Guillermo Sapiro. Online dictionary learning for sparse coding. In Proceedings of the 26th Annual International Conference on Machine Learning, ICML 09, pp. 689 696, New York, NY, USA, 2009. ACM. ISBN 978-1-60558-516-1. doi: 10.1145/ 1553374.1553463. URL http://doi.acm.org/10.1145/1553374.1553463. Hector Martinez Alonso, Anders Trærup Johannsen, Sussi Olsen, Sanni Nimb, and Bolette Sandford Pedersen. An empirically grounded expansion of the supersense inventory. In Proceedings of the 8th Global Wordnet Conference, pp. 199 208, 2016. Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. Efficient estimation of word representations in vector space. Co RR, abs/1301.3781, 2013a. URL http://arxiv.org/abs/ 1301.3781. Tomas Mikolov, Quoc V. Le, and Ilya Sutskever. Exploiting similarities among languages for machine translation. Co RR, abs/1309.4168, 2013b. URL http://dblp.uni-trier.de/db/ journals/corr/corr1309.html#Mikolov LS13. Published as a conference paper at ICLR 2020 George A. Miller, Claudia Leacock, Randee Tengi, and Ross T. Bunker. A semantic concordance. In Proceedings of the Workshop on Human Language Technology, HLT 93, pp. 303 308, Stroudsburg, PA, USA, 1993. Association for Computational Linguistics. ISBN 1-55860-324-7. doi: 10.3115/1075671.1075742. URL https://doi.org/10.3115/1075671.1075742. Simonetta Montemagni, Francesco Barsotti, Marco Battista, Nicoletta Calzolari, Ornella Corazzari, Alessandro Lenci, Antonio Zampolli, Francesca Fanciulli, Maria Massetani, Remo Raffaelli, et al. Building the italian syntactic-semantic treebank. In Treebanks, pp. 189 210. Springer Netherlands, 2003. Brian Murphy, Partha Talukdar, and Tom Mitchell. Learning effective and interpretable semantic models using non-negative sparse embedding. In Proceedings of COLING 2012, pp. 1933 1950, Mumbai, India, December 2012. The COLING 2012 Organizing Committee. URL http:// www.aclweb.org/anthology/C12-1118. Sungjoon Park, Jin Yeong Bak, and Alice Oh. Rotated word vector representations and their interpretability. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 401 411. Association for Computational Linguistics, 2017. URL http: //www.aclweb.org/anthology/D17-1041. Adam Paszke, Sam Gross, Soumith Chintala, Gregory Chanan, Edward Yang, Zachary De Vito, Zeming Lin, Alban Desmaison, Luca Antiga, and Adam Lerer. Automatic differentiation in pytorch. In NIPS-W, 2017. Jeffrey Pennington, Richard Socher, and Christopher Manning. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1532 1543, Doha, Qatar, October 2014. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/D14-1162. Sebastian Ruder, Ivan Vuli c, and Anders Søgaard. A survey of cross-lingual word embedding models. J. Artif. Int. Res., 65(1):569 630, May 2019. ISSN 1076-9757. doi: 10.1613/jair.1.11640. URL https://doi.org/10.1613/jair.1.11640. Peter H. Sch onemann. A generalized solution of the orthogonal Procrustes problem. Psychometrika, 31(1), March 1966. doi: 10.1007/BF02289451. URL http://www.springerlink.com/ index/10.1007/BF02289451. Samuel L. Smith, David H. P. Turban, Steven Hamblin, and Nils Y. Hammerla. Offline bilingual word vectors, orthogonal transformations and the inverted softmax. In Proceedings of the 5th International Conference on Learning Representations (ICLR 2017), 2017. Anant Subramanian, Danish Pruthi, Harsh Jhamtani, Taylor Berg-Kirkpatrick, and Eduard H. Hovy. SPINE: sparse interpretable neural embeddings. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, Louisiana, USA, February 2-7, 2018, pp. 4921 4928, 2018. URL https://www.aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/17433. Fei Sun, Jiafeng Guo, Yanyan Lan, Jun Xu, and Xueqi Cheng. Sparse word embeddings using ℓ1 regularized online learning. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, pp. 2915 2921, New York, New York, USA, 9 15 July 2016. AAAI Press / International Joint Conferences on Artificial Intelligence. Oscar T ackstr om, Dipanjan Das, Slav Petrov, Ryan T. Mc Donald, and Joakim Nivre. Token and type constraints for cross-lingual part-of-speech tagging. TACL, 1:1 12, 2013. URL https: //tacl2013.cs.columbia.edu/ojs/index.php/tacl/article/view/44. Yulia Tsvetkov, Manaal Faruqui, Wang Ling, Guillaume Lample, and Chris Dyer. Evaluation of word vector representations by subspace alignment. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, pp. 2049 2054, Lisbon, Portugal, September 2015. Association for Computational Linguistics. doi: 10.18653/v1/D15-1243. URL https://www.aclweb.org/anthology/D15-1243. Published as a conference paper at ICLR 2020 Yulia Tsvetkov, Manaal Faruqui, and Chris Dyer. Correlation-based intrinsic evaluation of word vector representations. In Proceedings of the 1st Workshop on Evaluating Vector-Space Representations for NLP, pp. 111 115, Berlin, Germany, August 2016. Association for Computational Linguistics. doi: 10.18653/v1/W16-2520. URL https://www.aclweb.org/anthology/ W16-2520. Joseph Turian, Lev Ratinov, and Yoshua Bengio. Word representations: A simple and general method for semi-supervised learning. In Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, ACL 10, pp. 384 394, Stroudsburg, PA, USA, 2010. Association for Computational Linguistics. URL http://dl.acm.org/citation.cfm? id=1858681.1858721. Shyam Upadhyay, Manaal Faruqui, Chris Dyer, and Dan Roth. Cross-lingual models of word embeddings: An empirical comparison. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1661 1670, Berlin, Germany, August 2016. Association for Computational Linguistics. URL http://www.aclweb.org/ anthology/P16-1157. Shyam Upadhyay, Yogarshi Vyas, Marine Carpuat, and Dan Roth. Robust cross-lingual hypernymy detection using dependency context. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp. 607 618. Association for Computational Linguistics, 2018. doi: 10.18653/v1/N18-1056. URL http://aclweb.org/anthology/N18-1056. Ivan Vuli c and Marie-Francine Moens. Bilingual word embeddings from non-parallel documentaligned data applied to bilingual lexicon induction. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pp. 719 725. Association for Computational Linguistics, 2015. doi: 10.3115/v1/P15-2118. URL http://www.aclweb.org/ anthology/P15-2118. Yogarshi Vyas and Marine Carpuat. Sparse bilingual word representations for cross-lingual lexical entailment. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1187 1197, San Diego, California, June 2016. Association for Computational Linguistics. URL http: //www.aclweb.org/anthology/N16-1142. Adina Williams, Nikita Nangia, and Samuel Bowman. A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pp. 1112 1122, New Orleans, Louisiana, June 2018. Association for Computational Linguistics. doi: 10.18653/v1/N18-1101. URL https://www.aclweb. org/anthology/N18-1101. Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, Pierric Cistac, Tim Rault, R emi Louf, Morgan Funtowicz, and Jamie Brew. Huggingface s transformers: State-of-the-art natural language processing, 2019. Chao Xing, Dong Wang, Chao Liu, and Yiye Lin. Normalized word embedding and orthogonal transform for bilingual word translation. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1006 1011, Denver, Colorado, May June 2015. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/N15-1104. David Yarowsky and Grace Ngai. Inducing multilingual POS taggers and NP bracketers via robust projection across aligned corpora. In Proceedings of the Second Meeting of the North American Chapter of the Association for Computational Linguistics on Language Technologies, NAACL 01, pp. 1 8, Stroudsburg, PA, USA, 2001. Association for Computational Linguistics. doi: 10. 3115/1073336.1073362. URL https://doi.org/10.3115/1073336.1073362. Dani Yogatama, Manaal Faruqui, Chris Dyer, and Noah A. Smith. Learning word representations with hierarchical sparse coding. In Proceedings of ICML, 2015. Published as a conference paper at ICLR 2020 Meng Zhang, Yang Liu, Huanbo Luan, and Maosong Sun. Earth mover s distance minimization for unsupervised bilingual lexicon induction. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp. 1934 1945, Copenhagen, Denmark, September 2017a. Association for Computational Linguistics. URL https://www.aclweb.org/ anthology/D17-1207. Meng Zhang, Yang Liu, Huanbo Luan, and Maosong Sun. Adversarial training for unsupervised bilingual lexicon induction. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 1959 1970, Vancouver, Canada, July 2017b. Association for Computational Linguistics. URL http://aclweb.org/ anthology/P17-1179. Yuan Zhang, David Gaddy, Regina Barzilay, and Tommi Jaakkola. Ten pairs to tag multilingual POS tagging via coarse mapping between embeddings. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 1307 1317, San Diego, California, June 2016. Association for Computational Linguistics. URL http://www.aclweb.org/anthology/N16-1156.