# reevaluating_word_movers_distance__f5dea79e.pdf Re-evaluating Word Mover s Distance Ryoma Sato 1 2 Makoto Yamada 1 2 Hisashi Kashima 1 2 The word mover s distance (WMD) is a fundamental technique for measuring the similarity of two documents. As the crux of WMD, it can take advantage of the underlying geometry of the word space by employing an optimal transport formulation. The original study on WMD reported that WMD outperforms classical baselines such as bag-of-words (BOW) and TF-IDF by significant margins in various datasets. In this paper, we point out that the evaluation in the original study could be misleading. We re-evaluate the performances of WMD and the classical baselines and find that the classical baselines are competitive with WMD if we employ an appropriate preprocessing, i.e., L1 normalization. In addition, we introduce an analogy between WMD and L1-normalized BOW and find that not only the performance of WMD but also the distance values resemble those of BOW in high dimensional spaces. 1. Introduction The optimal transport (OT) distance is an effective tool for comparing probabilistic distributions. Applications of OT include image processing (Ni et al., 2009; Rabin et al., 2011; De Goes et al., 2012), natural language processing (NLP) (Kusner et al., 2015; Rolet et al., 2016), biology (Schiebinger et al., 2019; Lozupone & Knight, 2005; Evans & Matsen, 2012), and generative models (Arjovsky et al., 2017; Salimans et al., 2018). A prominent application of OT is the word mover s distance (WMD) (Kusner et al., 2015) for document comparison. WMD regards a document as a probabilistic distribution of words, defines the underlying word geometry using pretrained word embeddings, and computes the distance using the optimal transport distance between two word distribu- 1Kyoto University 2RIKEN AIP. Correspondence to: Ryoma Sato . Proceedings of the 39 th International Conference on Machine Learning, Baltimore, Maryland, USA, PMLR 162, 2022. Copyright 2022 by the author(s). Obama speaks to the media in Illinois vs The president greets the press in Chicago concert Japan Obama speaks to the media in Illinois vs The band gave a concert in Japan Figure 1. Neither pair of texts has common words. WMD can choose a similar sentence appropriately, whereas the BOW distance cannot distinguish these cases. tions (i.e., documents). WMD is preferable because it takes the underlying geometry into account. For example, bagof-words (BOW) will conclude that two documents are dissimilar if they have no common words, whereas WMD will determine that they are similar if the words are semantically similar (even if they are not exactly the same), as illustrated in Figure 1. WMD has been widely used in NLP owing to this preferred property. For example, Kusner et al. (2015) and others (Huang et al., 2016; Li et al., 2019) used WMD for document classification, Wu et al. (2018) used WMD for computing document embeddings, Xu et al. (2018) used WMD for topic modeling, Kilickaya et al. (2017) and others (Clark et al., 2019; Zhao et al., 2019; 2020; Wang et al., 2020a; Gao et al., 2020; Lu et al., 2019; Chen et al., 2020b) used WMD for evaluating text generation. Many extensions have been proposed including supervised (Huang et al., 2016; Takezawa et al., 2021) and fast (Le et al., 2019; Backurs et al., 2020; Genevay et al., 2016; Dong et al., 2020; Sato et al., 2020b) variants. WMD is one of the fundamental tools used in NLP, and understanding the deep mechanism of WMD is crucial for further applications. The most fundamental application of WMD is document classification. The original study on WMD (Kusner et al., 2015) conducted extensive experiments using k NN classifiers. Figure 2 shows the classification errors reported in (Kusner et al., 2015). This figure clearly shows that WMD is superior to classical baselines, BOW and TF-IDF1 1It should be noted that (Kusner et al., 2015) used many stronger baselines such as LSI and LDA. We focus on BOW and TF-IDF because (i) BOW was used as the base performance (Figure 4 in (Kusner et al., 2015), Figure 4 in (Yurochkin et al., 2019)), and (ii) BOW is a special case of WMD (Propositoin 4.1). Re-evaluating Word Mover s Distance bbcsport twitter recipe ohsumed classic reuters amazon 20news 0 k NN classification error Figure 2. k NN classification errors reported in the original WMD paper (Figure 3 in (Kusner et al., 2015)). Lower is better. WMD outperformed the naive baselines by significant margins. Figure 2 is surprising in the following senses. First, WMD outperforms the classical baselines by excessively large margins. BOW and TF-IDF have long been recognized as effective tools for document classification. Although it is reasonable for WMD to outperform them, the improvements are surprisingly large. In particular, the performance is ten times better on the classic dataset and five times better on the bbcsport dataset. Such results are excessively impressive. Second, although TF-IDF is known to be more effective than BOW, it performs worse than BOW on the ohsumed, reuters, and amazon datasets. In fact, the number of misclassification doubles on the reuters datasets. In this paper, we point out the possibility that the evaluations conducted in the original WMD study (Kusner et al., 2015) are misleading. Specifically, we found that the main improvements of WMD were due to normalization. Using the same normalization, WMD is comparable to BOW and TF-IDF, or WMD achieves improvements of only two to eight percent at the price of heavy computations. We also confirm that TF-IDF is more effective than raw BOW if we employ adequate normalization. To understand the mechanism of WMD, we introduce an analogy between WMD and L1-normalized BOW. We experimentally find that the distribution of the distances between matched words is not Gaussian-like but bimodal in high dimensional spaces. We then find that not only the performance of WMD but also the distance values resemble those of BOW in high dimensional spaces. The contributions of this paper are summarized as follows. We point out that the performance of WMD is not as high as we previously believed. The performance is comparable to classical baselines in document classification with the same normalization. We introduce an analogy between WMD and L1normalized BOW (Proposition 4.1) and find that WMD resembles BOW in high dimensional spaces (Figure 5). We point out several confusing aspects in the evaluations conducted in the original study on WMD. We suspect that many readers and researchers are unaware of these issues. Clarifying them is crucial for a solid evaluation and analysis in this field Reproducibility. The code is available at https:// github.com/joisino/reeval-wmd. It contains a script to download datasets and pre-computed results, algorithm implementations, and evaluation code. 2. Related Work Word Mover s Distance (WMD) and Optimal Transport (OT) in NLP. WMD (Kusner et al., 2015) is one of the most thriving applications of OT. WMD can take the underlying word geometry into account and inherit many elegant theoretical properties from OT. The success of WMD has facilitated many applications of OT in NLP. Emb Dist (Kobayashi et al., 2015) is a method concurrent with WMD. It also regards a document as a distribution of word embeddings but uses greedy matching instead of OT. Kumar et al. (2017) apply WMD to hidden representations of words instead of raw word embeddings. Yurochkin et al. (2019) consider a document as a distribution of topics and compute the document similarities using OT. They also use the OT distance for computing the ground distance of the topics. Alvarez-Melis et al. (2018) proposed a structured OT and applied it to a document comparison to take the positional consistency into account. Singh et al. (2020) consider a word as a probabilistic distribution of surrounding words and compute the similarity of the words using the OT distance of the distributions. Muzellec & Cuturi (2018) and others (Deudon, 2018; Frogner et al., 2019; Sun et al., 2018) embed words or sentences into distributions instead of vectors and compute the distance between embeddings using OT. Chen et al. (2019) and others (Li et al., 2020; Chen et al., 2020a) regularize the text generation models based on the OT distance between the generated texts and the ground truth texts to improve the generation. Nested Re-evaluating Word Mover s Distance Table 1. Dataset statistics. bbcsport twitter recipe ohsumed classic reuters amazon 20news Number of documents 737 3108 4370 9152 7093 7674 8000 18821 Number of training documents 517 2176 3059 3999 4965 5485 5600 11293 Number of test documents 220 932 1311 5153 2128 2189 2400 7528 Size of the vocabulary 13243 6344 5708 31789 24277 22425 42063 29671 Unique words in a document 116.5 9.9 48.3 60.2 38.7 36.0 44.6 69.3 Number of classes 5 3 15 10 4 8 4 20 Split type five-fold five-fold five-fold one-fold five-fold one-fold five-fold one-fold Duplicate pairs 15 976 48 1873 2588 143 159 59 Duplicate samples 30 474 66 3116 600 197 285 88 Wasserstein (Zhang et al., 2020a) compares the distributions of sequences and is successfully used in imitation learning for text generation. Lei et al. (2019) use WMD to generate paraphrase texts for creating adversarial examples. Zhang et al. (2020b) introduced partial OT to drop meaningless words. Michel et al. (2017) use a Gromov Wasserstein-like distance instead of the standard OT to compare documents. Zhang et al. (2016) and others (Zhang et al., 2017b;a; Grave et al., 2019; Dou & Neubig, 2021) use OT to align word embeddings of different languages. Trapp et al. (2017) use WMD to compare compositional documents by weighting each document. Kilickaya et al. (2017) and others (Clark et al., 2019; Zhao et al., 2019; 2020; Wang et al., 2020a; Gao et al., 2020; Lu et al., 2019; Chen et al., 2020b) used WMD for evaluating text generation. BERTScore (Zhang et al., 2020c) is a relevant method, but it uses greedy matching instead of OT. To summarize, OT and WMD have been used in many NLP tasks. It is important to understand the underlying mechanism of WMD for further advancements in this field. Other Text Retrieval Methods. Many text retrieval methods have been proposed so far. In particular, methods based on neural language models, such as DPR (Karpukhin et al., 2020), Col BERT (Khattab & Zaharia, 2020), and Rocket QA (Qu et al., 2021), have been popular these days. The advantages of WMD over neural network-based methods, namely DPR, are (1) WMD is an unsupervised method, while DPR requires supervision/fine-tuning. The DPR paper reported that it needed about 1000 training examples. The unsupervised nature of WMD is beneficial in particular for low-resource languages and domains, where no or few labeled data are available. (2) WMD is a lightweight method, while DPR requires storing and running the BERT model (Devlin et al., 2019). The compact nature of WMD is beneficial in particular for mobile and Io T devices, where BERT is infeasible. (3) WMD has been one of the basic techniques in NLP, and many methods have been developed based on WMD, such as word mover s embedding (Wu et al., 2018) and word rotator s distance (Yokoi et al., 2020). Clarifying the mechanism of WMD and recognizing the true performance of WMD are important for the future development of this field in its own right. Re-evaluation of Existing Methods. Back in 2009, Armstrong et al. (2009) found that, although many studies have claimed statistically significant improvements against the baselines, most have employed excessively weak baselines, and the performance did not improve from the classical baselines in the information retrieval domain. Dacrema et al. (2019) recently found that many of the deep learning-based recommender systems are extremely difficult to reproduce, and for the methods whose results authors could reproduce, the performance was not as high as people had believed, and the deep approaches were actually comparable to classical baselines with an appropriate hyperparameter tuning. Their paper has had a large impact on the community and was awarded the best paper prize at Rec Sys 2019. Similar observations have also been made in sentence embeddings (Arora et al., 2017; Shen et al., 2018), session-based recommendations (Ludewig et al., 2019), and graph neural networks (Errica et al., 2020) as well. In general, science communication suffers from publication and confirmation biases. The importance of reproducing existing experiments by third-party groups has been widely recognized in science (Lin, 2018; Sculley et al., 2018; Munafò et al., 2017; Collins & Tabak, 2014; Goodman et al., 2016). 3. Backgrounds 3.1. Problem Formulation In this paper, we consider document classification tasks. Each document is represented by a bag-of-word vector x Rm, where m is the number of unique words in the dataset. The i-th component of x represents the number of occurrences of the i-th word in the document. We focus on the k NN classification following the original paper. The k NN classification gathers k samples of the smallest distances (with respect to a certain distance) to a test sample from the training dataset and classifies the sample to the majority class of the gathered labels. The design of the distance function is crucial for the performance of k NN classification. Re-evaluating Word Mover s Distance 3.2. Word Mover s Distance (WMD) WMD provides an effective distance function utilizing pretrained word embeddings. Let zi be the embedding of the i-th word. To utilize OT, WMD first regards a document as a discrete probabilistic distribution of words by normalizing the bag-of-word vector: n L1(x) = x/ X WMD defines the cost matrix C Rm Rm as the distance of the embeddings, i.e., Cij = zi zj 2. The distance between the two documents x and x is the optimal value of the following problem: minimize P Rm m X ij Cij Pij (2) s.t. Pij 0, P 1 = n L1(x), P 1 = n L1(x ), where P denotes the transpose of P , 1 Rm is the vector of ones. Intuitively, Pij represents the amount of word i that is transported to word j. WMD is defined as the minimum total distance to convert one document to another document. Let OT(x, x , C) R be the optimal value of eq. (2). 3.3. Experimental Setups Datasets. We use the same datasets (Greene & Cunningham, 2006; Sanders, 2011; Joachims, 1998; Sebastiani, 2002; Lang, 1995) as in the original paper (Kusner et al., 2015) and use the same train/test splits as in the original paper. Three of the eight datasets have the standard train/test splits (e.g., based on timestamps), and the other datasets do not. Thus, the original paper used five random splits for such datasets. We refer to the former type as one-fold datasets and the latter as five-fold datasets. Table 1 shows the statistics. Here, we point out the first misleading point. Misleading Point 1 Many duplicate samples exist in the datasets. The last two rows in Table 1 report the numbers of duplicate samples. Some of these samples cross the train/test splits, and some of them have different labels despite having the same contents. This causes problems in the evaluations. If the pairs have different labels, it is impossible to classify both of them correctly. If the pairs have the same label, a k NN classifier places more emphasis on this class for no reason. We report this issue in more detail in Appendix A. The datasets released by the WMD paper have been used in many studies (Huang et al., 2016; Yurochkin et al., 2019; Le et al., 2019; Werner & Laber, 2020; Wang et al., 2020b; Wu et al., 2018; Skianis et al., 2020b; Mollaysa et al., 2017; Gupta et al., 2020; Skianis et al., 2020a) with the same protocol. We suspect that many readers and authors were unaware of the duplicate samples, which might have led to a misleading analysis. We release the indices of duplicate documents and a preprocessing script to remove duplicate samples for the following studies. We believe that sharing this fact within the community is important for aiding in a solid evaluation and analysis. In the following, we first use the same dataset as the WMD paper to highlight the essential differences with the original evaluation. We then evaluate using clean datasets to further corroborate the findings. Embeddings. We use the same word embeddings as in the original paper (Kusner et al., 2015). Namely, we use the 300-dimentional word2vec embeddings trained on the Google News corpus. The reasons we use word2vec are that (i) The original paper (Kusner et al., 2015) used the same embeddings. We made the evaluation protocol as identical to the original evaluation as possible to highlight the critical difference. (ii) The choice of word2vec in the original paper has been inherited in many following works (Kilickaya et al., 2017; Wu et al., 2018), and investigating word2vec thus has a broad impact. We found the second misleading point here. Misleading Point 2 The official code of WMD2 normalized the embeddings by the L2 norm, although this was not explicitly stated in the paper. The hint of word normalization is not in the main logic but only in the preprocessed binary file in the WMD s repository. We accidentally found this when we were debugging the code (Appendix C). Note that Yokoi et al. (2020) reported that normalizing the word embeddings improved the performance of WMD, but in fact, the original evaluation of WMD already benefited from this effect without explicitly stating it. We suspect that most readers missed this point. In this paper, we follow the original evaluation and normalize the embeddings using the L2 norm. We provide an ablation study on the cost function in Appendix G. Preprocessing. For WMD, we use the same preprocessing as in the original paper (Kusner et al., 2015), whereas for BOW and TF-IDF, we use a different preprocessing to clearly contrast the results. WMD discards certain words because the word embeddings do not contain all words. In the original study, the authors discard out-of-vocabulary words for WMD but maintain them for BOW and TF-IDF. In this paper, we discard out-of-vocabulary words for BOW and TF-IDF as well and use the same vocabulary for all of WMD, BOW, and TF-IDF. This setting is slightly advantageous for WMD. We show that WMD is comparable to BOW and TF-IDF even under this setting. We include the results with out-of-vocabulary words at the end of Section 4 for completeness. Evaluation Protocol. WMD, BOW, and TF-IDF all have 2https://github.com/mkusner/wmd Re-evaluating Word Mover s Distance bbcsport twitter recipe ohsumed classic reuters amazon 20news 0 k NN classification error 3.9 2.8 5.1 30.028.929.6 40.142.9 44.1 4.1 3.3 2.9 5.7 5.5 4.0 10.4 8.0 7.4 Figure 3. k NN classification errors in our re-evaluation. Lower is better. The shaded bars are the performance without normalization. WMD is comparable to classical baselines with normalization. only one hyperparameter, i.e., size k of the neighborhood in k NN classification. We split the training set into an 80/20 train/validation set uniformly and randomly and select the neighborhood size from {1, 2, , 19} using the validation data. During the process of our re-evaluations, we found that the k NN classifiers were much less capable than we previously thought and that k NN classification may underestimate the performances of distance-based methods. Nevertheless, we adopt k NN evaluations in the main part to clearly contrast our results with the original evaluation. This is justified because we only use k NN classification for all methods. However, this fact can be a misleading point if other classifiers are used for other methods, as in (Wu et al., 2018; Skianis et al., 2020b; Mollaysa et al., 2017; Gupta et al., 2020). We report this issue in detail in Appendix D. Environment. We use a server cluster to compute WMD. Each node has two 2.4GHz Intel Xeon Gold 6148 CPUs. We use a Linux server with Intel Xeon E7-4830 v4 CPUs to evaluate the performances. 4. Normalization is Crucial Importance of Normalization. In the original paper, raw BOW and TF-IDF vectors are used for the nearest neighbor classification. However, this is problematic because the length of these vectors varies based on the length of the documents. Even if two documents share the same topic, the BOW vectors are distant if their lengths differ. We need to normalize these vectors to effectively use them in the nearest neighbor classification. Misleading Point 3 The original evaluation did not normalize the BOW and TF-IDF vectors. To make direct comparisons to WMD, we employ L1 normalization (Eq. (1)) and the L1 distance for BOW and TFIDF vectors, i.e., d BOW/L1/L1(x, x ) = n L1(x) n L1(x ) 1. With this normalization, BOW is a special case of WMD that does not use the underlying geometry. Specifically, let Cunif be the cost matrix of WMD when we use one-hot vectors as word embeddings, i.e., Cunif ij = 0 if i = j and Cunif ij = 2 if i = j. Then, Proposition 4.1. d BOW/L1/L1(x, x ) = OT(x, x , Cunif) The proof is in Appendix H. This proposition shows that the difference in the performances between WMD and L1/L1 BOW indicates the benefit of the underlying geometry. Figure 3 shows the classification errors with normalization. First, we can see that the errors of BOW and TF-IDF drastically decrease. Even BOW performs better than WMD in bbcsport and ohsumed. TF-IDF outperforms WMD in five out of eight datasets. Even in the other datasets where WMD outperforms the baselines, the improvements are far less significant than what was reported in the original evaluation. We can also observe that TF-IDF always performs better than BOW in contrast to the original evaluation. These results make more sense than what was reported in the WMD paper, where BOW outperformed TF-IDF. Comparison with other Normalization. Although normalized BOW and TF-IDF have been employed in other studies (Yurochkin et al., 2019; Li et al., 2019; Werner & Laber, 2020; Wrzalik & Krechel, 2019), it was reported that normalized BOW is still far worse than WMD, which is incompatible with our observations above. We found out that this occurred because of the normalization methods and metrics used to compare the vectors. For example, Yurochkin et al. (2019) used L1 normalization and the L2 metric for BOW, i.e., n L1(x) n L1(x ) 2, and L2 normalization and the L2 metric for TF-IDF, i.e., n L2(x) n L2(x ) 2, where n L2(x) = x/ x 2. Note that the L2/L2 scheme corresponds to the cosine similarity. We investigate the performance with different normalization and metrics using the same protocol as in the previous experiments. Table 2 shows the classification errors. First, it is easy to see in the fifth through eighth rows that the performances degrade without normalization. In addition, even with normalization, the performances are still poor if the normalization and metric use different norms. Although the L2/L2 scheme is better than these schemes, the L1/L1 scheme is the best for all datasets. Re-evaluating Word Mover s Distance Table 2. k NN classification errors. Lower is better. Here, (x/y) uses x as the normalization and y as the metric. The last column reports the average relative performance to the normalized BOW. These values correspond to Figure 4 in the original paper, but we use BOW (L1/L1) as the base performances, while the original paper used BOW (None/L2) as the base performances. Figure 4 in (Yurochkin et al., 2019) also reports the relative performances, but it uses BOW (L1/L2) as the base performances. The standard deviations are reported for five-fold datasets. The first three rows are the same as in Figure 3. For BOW and TF-IDF, a cell is highlighted with bold if the mean score is better than that of WMD. For WMD, a cell is highlighted with bold if the mean score is better than those of both BOW and TF-IDF. The fourth row reports the performances with WMD with TF-IDF weighting. The following rows report the performance with different normalization and metrics. bbcsport twitter recipe ohsumed classic reuters amazon 20news rel. BOW (L1/L1) 3.9 1.1 30.0 1.1 43.4 0.8 44.1 4.1 0.5 5.7 10.4 0.5 29.1 1.000 TF-IDF (L1/L1) 2.8 1.1 28.9 0.8 40.1 0.7 37.8 3.3 0.4 5.5 8.0 0.3 25.9 0.861 WMD 5.1 1.2 29.6 1.5 42.9 0.8 44.5 2.9 0.4 4.0 7.4 0.5 26.8 0.917 WMD-TF-IDF 3.3 0.9 28.3 2.3 39.9 1.1 39.7 2.7 0.3 4.0 6.6 0.2 24.1 0.804 BOW (None/L2)(Kusner et al., 2015) 19.4 3.0 34.2 0.6 60.0 2.3 61.6 35.0 0.9 11.8 28.2 1.0 57.7 3.024 BOW (None/L1) 25.4 1.5 32.7 1.6 65.8 2.5 69.3 52.1 0.5 14.2 31.4 1.2 73.9 3.931 TF-IDF (None/L2)(Kusner et al., 2015) 24.5 1.3 38.2 4.6 65.0 1.9 65.3 38.8 1.0 28.0 41.2 3.2 60.0 3.867 TF-IDF (None/L1) 30.6 1.3 37.8 4.8 70.3 1.3 70.6 52.6 0.2 29.1 41.5 4.9 74.6 4.602 BOW (L1/L2)(Yurochkin et al., 2019) 11.4 3.6 37.0 1.4 50.8 1.1 56.7 17.3 1.5 12.3 35.7 1.3 46.5 2.253 BOW (L2/L1) 15.2 1.5 33.3 1.1 61.1 1.1 65.7 51.1 0.4 16.2 32.2 1.3 77.6 3.622 BOW (L2/L2)(Werner & Laber, 2020) (Wrzalik & Krechel, 2019) 5.5 0.7 31.0 0.8 46.1 0.6 46.2 6.3 0.7 8.8 13.1 0.5 33.2 1.254 TF-IDF (L1/L2) 25.5 11.2 35.7 1.4 54.2 2.7 61.4 22.6 4.2 24.7 41.9 2.0 45.6 3.226 TF-IDF (L2/L1) 27.5 7.2 33.4 1.7 64.9 3.8 69.7 52.0 0.2 19.5 40.8 6.6 78.3 4.245 TF-IDF (L2/L2)(Yurochkin et al., 2019) (Li et al., 2019) 4.0 0.7 29.8 1.5 43.7 1.2 38.4 5.2 0.3 10.5 11.1 0.9 31.6 1.145 Table 3. k NN classification errors with clean data. Lower is better. The same notations as in Table 2. bbcsport twitter recipe ohsumed classic reuters amazon 20news rel. BOW (L1/L1) 3.7 1.0 30.6 1.1 42.9 0.6 39.7 4.2 0.5 5.5 10.6 0.6 29.2 1.000 TF-IDF (L1/L1) 2.3 1.4 30.2 0.7 40.0 1.1 33.4 3.5 0.2 5.9 8.0 0.6 25.9 0.866 WMD 5.5 1.2 30.6 1.2 42.9 0.9 40.6 3.4 0.6 3.8 7.3 0.4 26.9 0.952 WMD-TF-IDF 4.1 1.5 28.8 1.6 40.2 0.9 35.7 2.8 0.3 4.3 6.6 0.3 24.2 0.848 This result is natural if we adopt the stance that a document is a distribution of words, and based on the analogy between the L1/L1 BOW and WMD, i.e., Proposition 4.1. To summarize, the normalization method significantly affects the performance. BOW L1/L1 corresponds to WMD without OT (Proposition 4.1), and BOW None/L1 corresponds BOW L1/L1 without normalization . The performances of these methods are 3.931 (BOW None/L1) introducing normalization 1.000 (BOW L1/L1) introducing OT 0.917 (WMD). This means that WMD owes its performance improvements from the naive method (i.e., BOW None/L1) to normalization (by a factor of 3.9) rather than the OT formulation (by only a factor of 1.09). This ablation study indicates that L1 normalization is the most significant factor of improvement in WMD. We stress that the improvements brought about by L1 normalization are obtained almost for free, whereas the improvements by WMD are at the price of substantial computational costs. We should also point out that many papers have not clarified the normalization scheme. Because the normalization method and metrics used to compare the vectors significantly affect the performance, we recommend clarifying them in the experimental setups. To better illustrate the benefit of WMD, we also use TF-IDF weighting for the marginal measures of WMD. Specifically, let xtfidf Rd denote a TF-IDF vector. We use n L1(xtfidf) instead of n L1(x) in the marginal constraints in Eq. (2). The TF-IDF weighting makes WMD as robust to noise as TF-IDF. The fourth row in Table 2 shows that WMD with TF-IDF weighting performs the best, with approximately a six percent improvement from the normalized TF-IDF, which corresponds to the WMD-TF-IDF with the uniform distance matrix Cunif. Investigating the (BOW, WMD) and (TF-IDF, WMD-TF-IDF) pairs illustrates the benefit of the optimal transport formulation. We can observe six to eight percent improvements in the optimal transport formulation from the naive baselines. This is in contrast to the sixty percent improvement claimed in the original study. Evaluation on Clean Data. We evaluate the methods using clean datasets without duplicate samples. We adopt the same protocol as in the previous experiments except for the removal of duplicate documents. Table 3 reports the classification errors. Although the tendency is the same as that in the previous experiments, the differences between (BOW, WMD) and (TF-IDF, WMD-TF-IDF) become even narrower. Namely, an approximately five percent improvement in WMD and two percent improvement in WMD-TFIDF are found. The results for other normalization and Re-evaluating Word Mover s Distance Table 4. k NN classification errors with all words and the original datasets. Lower is better. The last column reports the average relative performance to the normalized BOW (the first row in Table 2). bbcsport twitter ohsumed classic reuters amazon 20news rel. BOW (L1/L1) 3.0 0.8 29.5 0.7 46.0 3.9 0.4 6.1 12.3 0.5 32.1 1.015 TF-IDF (L1/L1) 2.8 0.8 29.4 0.9 38.7 3.1 0.4 6.8 7.9 0.3 24.8 0.877 Table 5. k NN classification errors with all words and clean data. Lower is better. The last column reports the average relative performance to the normalized BOW (the first row in Table 3). bbcsport twitter ohsumed classic reuters amazon 20news rel. BOW (L1/L1) 2.7 0.6 31.0 2.0 41.0 4.1 0.4 6.3 12.3 0.3 32.1 1.022 TF-IDF (L1/L1) 1.8 0.8 30.3 1.5 34.9 3.4 0.5 6.4 8.1 0.2 24.8 0.849 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 0.4 twitter 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 0.2 ohsumed 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 0.4 reuters 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 bbcsport_5dim 0.0 0.5 1.0 1.5 0.0 0.5 twitter_5dim 0.0 0.5 1.0 1.5 0.0 recipe_5dim 0.0 0.5 1.0 1.5 0.0 0.2 ohsumed_5dim 0.0 0.5 1.0 1.5 Distance classic_5dim 0.0 0.5 1.0 1.5 Distance reuters_5dim 0.0 0.5 1.0 1.5 Distance 0.2 amazon_5dim 0.0 0.5 1.0 1.5 Distance 0.2 20news_5dim Figure 4. Histograms of distances between matched word embeddings. (Top) 300-dimensional embeddings. (Bottom) 5-dimensional embeddings. metrics are reported in Table 7 in the appendix. Evaluation with Out of Vocabulary Words. We had hypothesized that the two to eight percent improvements of WMD were due to the unnecessarily discarded vocabularies of BOW, which made the evaluations slightly advantageous to WMD. We evaluate BOW and TF-IDF, including words not in the word2vec vocabulary. We use all but the recipe dataset because the raw texts of the recipe dataset have not been released by the authors. We include stopwords for the Twitter dataset and remove them for the other datasets following the original paper. Tables 4 and 5 report the classification errors. Against expectations, these results show that the use of all words does not improve the performance. We hypothesize that this is because the word2vec vocabulary implicitly helps the classification by removing noisy words. Summary. We should emphasize that we have not concluded that the improvements brought by WMD are spurious. Based on our careful evaluations, we conclude that the two to eight percent improvements from the naive baselines reported in Tables 2 and 3 are genuine. These improvements are less sensational than those reported in the original paper, i.e., sixty percent improvement. However, we believe that our results indicate the true performance of WMD. These results indicate that if the speed is important, WMD may not be worth trying, whereas if the performance is crucial at any cost, WMD may be a worthwhile candidate over L1/L1 BOW. 5. WMD resembles BOW in High Dimensional Spaces In this section, we experimentally show that not only the performance of WMD but also the distance values of WMD themselves resemble those of L1/L1 BOW. Travel Distance Distribution is Bimodal in High Dimensional Spaces. We had assumed that the distribution of distances of matched word embeddings in WMD was a Gaussian-like unimodal distribution. However, against expectations, we found that the distribution was actually bimodal in high dimensional spaces. The top panels in Figure 4 show the histograms for matched words in nearest neighbor document pairs. The x-axis represents the distances of matched words, and the y-axis represents the frequency. All of the histograms have acute peaks at x = 0, where the source and target words exactly match. Other pairs of matched words are at x 1 and are approximately equally distant. This occurs intuitively because most pairs of embeddings are almost orthogonal in high dimensional spaces. To contrast the results with low dimensional cases, we project the 300-dimensional word2vec to a 5-dimensional space using principal component analysis and compute the same histograms using these low dimensional embeddings. Bottom panels in Figure 4 show the histograms. In contrast Re-evaluating Word Mover s Distance 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) bbcsport ρ = 0.950 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) twitter ρ = 0.886 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) recipe ρ = 0.956 0.5 1.0 1.5 2.0 BOW (L1/L1) ohsumed ρ = 0.882 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) classic ρ = 0.846 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) reuters ρ = 0.979 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) amazon ρ = 0.840 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) 20news ρ = 0.837 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) bbcsport_5dim ρ = 0.482 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) twitter_5dim ρ = 0.391 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) recipe_5dim ρ = 0.611 0.5 1.0 1.5 2.0 BOW (L1/L1) ohsumed_5dim ρ = 0.333 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) classic_5dim ρ = 0.299 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) reuters_5dim ρ = 0.415 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) amazon_5dim ρ = 0.327 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) 20news_5dim ρ = 0.413 Figure 5. Scatter plot of WMD and L1/L1 BOW. (Top) 300-dimensional embeddings. WMD is mostly determined by L1/L1 BOW. (Bottom) 5-dimensional embeddings. to the high dimensional cases, these histograms are almost unimodal. These results indicate that the bimodality is a characteristic phenomenon in high dimensional spaces. WMD Resembles L1/L1 BOW in High Dimensional Spaces. The previous experiments show that the travel distance distribution is bimodal (zero or around one) in high dimensional spaces. In an extreme case, if the distance is exactly zero or one, Proposition 4.1 shows that WMD coincides with L1/L1 BOW. If not exact, we experimentally show that the distance values of WMD themselves resemble those of L1/L1 BOW. Figure 5 shows correlations between WMD and L1/L1 BOW distances and reports the Pearson s correlation coefficients ρ. WMD and L1/L1 BOW are surprisingly similar to each other. To contrast the results with low dimensional cases, we also report the scatter plots using 5-dimensional embeddings in the bottom of Figure 5. In contrast to the high dimensional cases, these scatter plots show that WMD in low dimensional spaces is not similar to L1/L1 BOW. These results indicate that the similarity to L1/L1 BOW is a characteristic phenomenon in high dimensional spaces. For example, the distance between Obama and President is 1.174, and the distance between Obama and band is 1.342 in 300-dimensional word2vec. The distance between speaks and greets is 0.978, and the distance between speaks and gave is 1.309. Although Obama and President seem much more semantically similar than Obama and band, the distances are not much different in high dimensional spaces. In other words, WMD does not identify Obama with President in high dimensional spaces. A Re-evaluating Word Mover s Distance 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 0.4 twitter 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.0 0.5 1.0 1.5 0.0 0.2 ohsumed 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 0.0 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) bbcsport ρ = 0.926 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) twitter ρ = 0.748 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) ohsumed ρ = 0.884 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) classic ρ = 0.715 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) reuters ρ = 0.959 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) amazon ρ = 0.807 0.0 0.5 1.0 1.5 2.0 BOW (L1/L1) 20news ρ = 0.859 Figure 6. Experiments on Glo Ve. Top: Glo Ve counterpart of Figure 4. Bottom: Glo Ve counterpart of 5. The trends are similar to those with word2vec. We omitted the recipe dataset because the original texts have not been released. simple calculation shows that 2WMD( Obama greets , band greets ) = 1.342 + 0 < 2WMD( Obama greets , President speaks ) = 1.174 + 0.978, BOW( Obama greets , band greets ) = 1 < BOW( Obama greets , President speaks ) = 2. This almost equally distant property may not be expected in the two-dimensional illustration (Figure 1). While such an illustration is helpful for understanding the mechanism of WMD, it does not reflect the high dimensional nature of word embeddings. The characteristics of high dimensionality should also be kept in mind to understand the behavior of WMD more precisely. It should be noted that 5-dimensional WMD performs worse than 300-dimensional WMD because 5-dimensional WMD loses much information on the word geometry. On the one hand, by increasing the number of dimensions, the word embeddings become more discriminative, and the performance of WMD increases. On the other hand, by increasing the number of dimensions, the word embeddings become orthogonal, and WMD approaches L1/L1 BOW, i.e., a special case of WMD that completely discriminates each word from the others. Other Embeddings. We used the word2vec embeddings so far. However, the observations above depend on the choice of embeddings, and it is possible that more powerful embeddings overcome the bimodal effect. We confirm if the phenomena can be extended to other embeddings. We run the same experiments with 300 dimensional Glo Ve (Pen- nington et al., 2014). Figure 6 shows that the trends with Glo Ve are similar to those with word2vec. Although the bimodal effect is slightly alleviated in some cases, e.g., on Twitter, the improvement is subtle, and WMD is still suffered from high dimensionality. Overall, WMD with Glo Ve also resembles L1/L1 Bo W (Figure 6 bottom). These results indicate that although there exists the possibility that word embeddings with much higher qualities could alleviate the problem, the trend is robust and tough to overcome. 6. Conclusion In this paper, we pointed out that the major improvement in WMD against classical baselines is not from the inherent feature of WMD but mainly from the normalization. We re-evaluated the performance of WMD and classical baselines and found that classical baselines are competitive with WMD if we normalize the vectors. We also found that WMD resembled BOW much more than we had thought in two-dimensional illustrations owing to the high dimensionality of the word embeddings. In the process of our re-evaluation, we found several confusing aspects in the original evaluations of the original paper on WMD. We pointed them out in this paper to help the following researchers conduct solid evaluations. Acknowledgments This work was supported by the JSPS KAKENHI Grant Number 20H04243, 20H04244 and 21J22490. Re-evaluating Word Mover s Distance Alvarez-Melis, D., Jaakkola, T. S., and Jegelka, S. Structured optimal transport. In Proceedings of the 21st International Conference on Artificial Intelligence and Statistics, AISTATS, volume 84 of Proceedings of Machine Learning Research, pp. 1771 1780. PMLR, 2018. Arjovsky, M., Chintala, S., and Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the 34th International Conference on Machine Learning, ICML, pp. 214 223, 2017. Armstrong, T. G., Moffat, A., Webber, W., and Zobel, J. Improvements that don t add up: ad-hoc retrieval results since 1998. In Proceedings of the 18th ACM Conference on Information and Knowledge Management, CIKM, pp. 601 610. ACM, 2009. Arora, S., Liang, Y., and Ma, T. A simple but tough-to-beat baseline for sentence embeddings. In 5th International Conference on Learning Representations, ICLR. Open Review.net, 2017. Backurs, A., Dong, Y., Indyk, P., Razenshteyn, I. P., and Wagner, T. Scalable nearest neighbor search for optimal transport. In Proceedings of the 37th International Conference on Machine Learning, ICML, pp. 497 506, 2020. Chen, L., Zhang, Y., Zhang, R., Tao, C., Gan, Z., Zhang, H., Li, B., Shen, D., Chen, C., and Carin, L. Improving sequence-to-sequence learning via optimal transport. In 7th International Conference on Learning Representations, ICLR. Open Review.net, 2019. Chen, L., Bai, K., Tao, C., Zhang, Y., Wang, G., Wang, W., Henao, R., and Carin, L. Sequence generation with optimal-transport-enhanced reinforcement learning. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, AAAI, pp. 7512 7520. AAAI Press, 2020a. Chen, Y., Wu, L., and Zaki, M. J. Reinforcement learning based graph-to-sequence model for natural question generation. In 8th International Conference on Learning Representations, ICLR. Open Review.net, 2020b. Clark, E., Celikyilmaz, A., and Smith, N. A. Sentence mover s similarity: Automatic evaluation for multisentence texts. In Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL, pp. 2748 2760. Association for Computational Linguistics, 2019. Collins, F. S. and Tabak, L. A. Policy: NIH plans to enhance reproducibility. Nature News, 505(7485):612, 2014. Cuturi, M. Sinkhorn distances: Lightspeed computation of optimal transport. In Advances in Neural Information Processing Systems 26: Annual Conference on Neural Information Processing Systems 2013, Neur IPS, pp. 2292 2300, 2013. Dacrema, M. F., Cremonesi, P., and Jannach, D. Are we really making much progress? A worrying analysis of recent neural recommendation approaches. In Proceedings of the 13th ACM Conference on Recommender Systems, Rec Sys, pp. 101 109. ACM, 2019. De Goes, F., Breeden, K., Ostromoukhov, V., and Desbrun, M. Blue noise through optimal transport. ACM Transactions on Graphics (TOG), 31(6):1 11, 2012. Deudon, M. Learning semantic similarity in a continuous space. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS, pp. 994 1005, 2018. Devlin, J., Chang, M., Lee, K., and Toutanova, K. BERT: pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT, pp. 4171 4186. ACL, 2019. Dhouib, S., Redko, I., Kerdoncuff, T., Emonet, R., and Sebban, M. A swiss army knife for minimax optimal transport. In Proceedings of the 37th International Conference on Machine Learning, ICML, pp. 2504 2513, 2020. Dong, Y., Gao, Y., Peng, R., Razenshteyn, I. P., and Sawlani, S. A study of performance of optimal transport. ar Xiv, abs/2005.01182, 2020. URL https://arxiv.org/ abs/2005.01182. Dou, Z. and Neubig, G. Word alignment by fine-tuning embeddings on parallel corpora. In Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, EACL, pp. 2112 2128. Association for Computational Linguistics, 2021. Dudley, R. M. The speed of mean glivenko-cantelli convergence. The Annals of Mathematical Statistics, 40(1): 40 50, 1969. Errica, F., Podda, M., Bacciu, D., and Micheli, A. A fair comparison of graph neural networks for graph classification. In 8th International Conference on Learning Representations, ICLR. Open Review.net, 2020. Evans, S. N. and Matsen, F. A. The phylogenetic kantorovich rubinstein metric for environmental sequence samples. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 74(3):569 592, 2012. Re-evaluating Word Mover s Distance Frogner, C., Mirzazadeh, F., and Solomon, J. Learning embeddings into entropic wasserstein spaces. In 7th International Conference on Learning Representations, ICLR. Open Review.net, 2019. Gao, Y., Zhao, W., and Eger, S. SUPERT: towards new frontiers in unsupervised evaluation metrics for multidocument summarization. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL, pp. 1347 1354. Association for Computational Linguistics, 2020. Genevay, A., Cuturi, M., Peyré, G., and Bach, F. R. Stochastic optimization for large-scale optimal transport. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, Neur IPS, pp. 3432 3440, 2016. Genevay, A., Peyré, G., and Cuturi, M. Learning generative models with sinkhorn divergences. In Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, AISTATS, volume 84 of Proceedings of Machine Learning Research, pp. 1608 1617. PMLR, 2018. Genevay, A., Chizat, L., Bach, F. R., Cuturi, M., and Peyré, G. Sample complexity of sinkhorn divergences. In Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics, AISTATS, volume 89 of Proceedings of Machine Learning Research, pp. 1574 1583. PMLR, 2019. Goodman, S. N., Fanelli, D., and Ioannidis, J. P. What does research reproducibility mean? Science translational medicine, 8(341):341ps12 341ps12, 2016. Grave, E., Joulin, A., and Berthet, Q. Unsupervised alignment of embeddings with wasserstein procrustes. In Proceedings of the 22nd International Conference on Artificial Intelligence and Statistics, AISTATS, volume 89 of Proceedings of Machine Learning Research, pp. 1880 1890. PMLR, 2019. Greene, D. and Cunningham, P. Practical solutions to the problem of diagonal dominance in kernel document clustering. In Proceedings of the 23rd International Conference on Machine Learning, ICML, pp. 377 384, 2006. Gupta, V., Saw, A., Nokhiz, P., Netrapalli, P., Rai, P., and Talukdar, P. P. P-SIF: document embeddings using partition averaging. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, AAAI, pp. 7863 7870. AAAI Press, 2020. Huang, G., Guo, C., Kusner, M. J., Sun, Y., Sha, F., and Weinberger, K. Q. Supervised word mover s distance. In Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, Neur IPS, pp. 4862 4870, 2016. Huang, M., Ma, S., and Lai, L. A riemannian block coordinate descent method for computing the projection robust wasserstein distance. In Proceedings of the 38th International Conference on Machine Learning, ICML, volume 139, pp. 4446 4455. PMLR, 2021. Joachims, T. Text categorization with support vector machines: Learning with many relevant features. In Proceedings of the 10th European Conference on Machine Learning, ECML, volume 1398 of Lecture Notes in Computer Science, pp. 137 142. Springer, 1998. Karpukhin, V., Oguz, B., Min, S., Lewis, P. S. H., Wu, L., Edunov, S., Chen, D., and Yih, W. Dense passage retrieval for open-domain question answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 6769 6781, 2020. Khattab, O. and Zaharia, M. Colbert: Efficient and effective passage search via contextualized late interaction over BERT. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR, pp. 39 48. ACM, 2020. Kilickaya, M., Erdem, A., Ikizler-Cinbis, N., and Erdem, E. Re-evaluating automatic metrics for image captioning. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL, pp. 199 209. Association for Computational Linguistics, 2017. Kobayashi, H., Noguchi, M., and Yatsuka, T. Summarization based on embedding distributions. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 1984 1989. The Association for Computational Linguistics, 2015. Kolouri, S., Zou, Y., and Rohde, G. K. Sliced wasserstein kernels for probability distributions. In 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR, pp. 5258 5267, 2016. Kolouri, S., Nadjahi, K., Simsekli, U., Badeau, R., and Rohde, G. K. Generalized sliced wasserstein distances. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS, pp. 261 272, 2019. Kumar, S., Chakrabarti, S., and Roy, S. Earth mover s distance pooling over siamese lstms for automatic short answer grading. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, IJCAI, pp. 2046 2052. ijcai.org, 2017. Re-evaluating Word Mover s Distance Kusner, M. J., Sun, Y., Kolkin, N. I., and Weinberger, K. Q. From word embeddings to document distances. In Proceedings of the 32nd International Conference on Machine Learning, ICML, pp. 957 966, 2015. Lang, K. News Weeder: Learning to filter netnews. In Proceedings of the 12th International Conference on Machine Learning, ICML, pp. 331 339, 1995. Le, T., Yamada, M., Fukumizu, K., and Cuturi, M. Treesliced variants of wasserstein distances. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS, pp. 12283 12294, 2019. Lei, Q., Wu, L., Chen, P., Dimakis, A., Dhillon, I. S., and Witbrock, M. J. Discrete adversarial attacks and submodular optimization with applications to text classification. In Proceedings of Machine Learning and Systems 2019, MLSys. mlsys.org, 2019. Li, C., Ouyang, J., and Li, X. Classifying extremely short texts by exploiting semantic centroids in word mover s distance space. In The World Wide Web Conference, WWW, pp. 939 949. ACM, 2019. Li, J., Li, C., Wang, G., Fu, H., Lin, Y., Chen, L., Zhang, Y., Tao, C., Zhang, R., Wang, W., Shen, D., Yang, Q., and Carin, L. Improving text generation with student-forcing optimal transport. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 9144 9156. Association for Computational Linguistics, 2020. Lin, J. The neural hype and comparisons against weak baselines. SIGIR Forum, 52(2):40 51, 2018. Lin, T., Fan, C., Ho, N., Cuturi, M., and Jordan, M. I. Projection robust wasserstein distance and riemannian optimization. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS, 2020. Lozupone, C. and Knight, R. Uni Frac: a new phylogenetic method for comparing microbial communities. Appl. Environ. Microbiol., 71(12):8228 8235, 2005. Lu, S., Yu, L., Feng, S., Zhu, Y., and Zhang, W. Co T: Cooperative training for generative modeling of discrete data. In Proceedings of the 36th International Conference on Machine Learning, ICML, pp. 4164 4172, 2019. Ludewig, M., Mauro, N., Latifi, S., and Jannach, D. Performance comparison of neural and non-neural approaches to session-based recommendation. In Proceedings of the 13th ACM Conference on Recommender Systems, Rec Sys, pp. 462 466. ACM, 2019. Michel, P., Ravichander, A., and Rijhwani, S. Does the geometry of word embeddings help document classification? A case study on persistent homology-based representations. In Proceedings of the 2nd Workshop on Representation Learning for NLP, Rep4NLP@ACL, pp. 235 240. Association for Computational Linguistics, 2017. Mollaysa, A., Strasser, P., and Kalousis, A. Regularising non-linear models using feature side-information. In Proceedings of the 34th International Conference on Machine Learning, ICML, pp. 2508 2517, 2017. Munafò, M. R., Nosek, B. A., Bishop, D. V., Button, K. S., Chambers, C. D., Du Sert, N. P., Simonsohn, U., Wagenmakers, E.-J., Ware, J. J., and Ioannidis, J. P. A manifesto for reproducible science. Nature human behaviour, 1(1): 1 9, 2017. Muzellec, B. and Cuturi, M. Generalizing point embeddings using the wasserstein space of elliptical distributions. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS, pp. 10258 10269, 2018. Ni, K., Bresson, X., Chan, T. F., and Esedoglu, S. Local histogram based segmentation using the wasserstein distance. International Journal of Computer Vision, 84(1): 97 111, 2009. Nikolentzos, G., Meladianos, P., Rousseau, F., Stavrakas, Y., and Vazirgiannis, M. Multivariate gaussian document representation from word embeddings for text categorization. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL, pp. 450 455. Association for Computational Linguistics, 2017. Nikolentzos, G., Tixier, A. J., and Vazirgiannis, M. Message passing attention networks for document understanding. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, AAAI, pp. 8544 8551. AAAI Press, 2020. Paty, F. and Cuturi, M. Subspace robust wasserstein distances. In Proceedings of the 36th International Conference on Machine Learning, ICML, pp. 5072 5081, 2019. Pennington, J., Socher, R., and Manning, C. D. Glo Ve: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 1532 1543. ACL, 2014. Petrovich, M., Liang, C., Sato, R., Liu, Y., Tsai, Y. H., Zhu, L., Yang, Y., Salakhutdinov, R., and Yamada, M. Feature robust optimal transport for high-dimensional data. ar Xiv, abs/2005.12123, 2020. URL https://arxiv.org/ abs/2005.12123. Re-evaluating Word Mover s Distance Qu, Y., Ding, Y., Liu, J., Liu, K., Ren, R., Zhao, W. X., Dong, D., Wu, H., and Wang, H. Rocketqa: An optimized training approach to dense passage retrieval for open-domain question answering. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT, pp. 5835 5847. ACL, 2021. Rabin, J., Peyré, G., Delon, J., and Bernot, M. Wasserstein barycenter and its application to texture mixing. In Proceedings of Scale Space and Variational Methods in Computer Vision, pp. 435 446, 2011. Rolet, A., Cuturi, M., and Peyré, G. Fast dictionary learning with a smoothed wasserstein loss. In Proceedings of the 19th International Conference on Artificial Intelligence and Statistics, AISTATS, pp. 630 638, 2016. Salimans, T., Zhang, H., Radford, A., and Metaxas, D. N. Improving gans using optimal transport. In 6th International Conference on Learning Representations, ICLR, 2018. Sanders, N. J. Sanders-twitter sentiment corpus. Sanders Analytics LLC, 242, 2011. Santambrogio, F. Optimal transport for applied mathematicians. Birkhauser, 2015. Sato, R., Cuturi, M., Yamada, M., and Kashima, H. Fast and robust comparison of probability measures in heterogeneous spaces. ar Xiv, abs/2002.01615, 2020a. URL https://arxiv.org/abs/2002.01615. Sato, R., Yamada, M., and Kashima, H. Fast unbalanced optimal transport on a tree. In Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS, 2020b. Schiebinger, G., Shu, J., Tabaka, M., Cleary, B., Subramanian, V., Solomon, A., Gould, J., Liu, S., Lin, S., Berube, P., et al. Optimal-transport analysis of single-cell gene expression identifies developmental trajectories in reprogramming. Cell, 176(4):928 943, 2019. Sculley, D., Snoek, J., Wiltschko, A. B., and Rahimi, A. Winner s curse? on pace, progress, and empirical rigor. In 6th International Conference on Learning Representations, ICLR. Open Review.net, 2018. Sebastiani, F. Machine learning in automated text categorization. ACM Comput. Surv., 34(1):1 47, 2002. Shen, D., Wang, G., Wang, W., Min, M. R., Su, Q., Zhang, Y., Li, C., Henao, R., and Carin, L. Baseline needs more love: On simple word-embedding-based models and associated pooling mechanisms. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL, pp. 440 450. Association for Computational Linguistics, 2018. Singh, S. P., Hug, A., Dieuleveut, A., and Jaggi, M. Context mover s distance & barycenters: Optimal transport of contexts for building representations. In Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, AISTATS, volume 108, pp. 3437 3449. PMLR, 2020. Skianis, K., Malliaros, F. D., Tziortziotis, N., and Vazirgiannis, M. Boosting tricks for word mover s distance. In Artificial Neural Networks and Machine Learning, ICANN, pp. 761 772. Springer, 2020a. Skianis, K., Nikolentzos, G., Limnios, S., and Vazirgiannis, M. Rep the set: Neural networks for learning set representations. In Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, AISTATS, volume 108 of Proceedings of Machine Learning Research, pp. 1410 1420. PMLR, 2020b. Sun, C., Yan, H., Qiu, X., and Huang, X. Gaussian word embedding with a wasserstein distance loss. ar Xiv, abs/1808.07016, 2018. URL http://arxiv.org/ abs/1808.07016. Takezawa, Y., Sato, R., and Yamada, M. Supervised treewasserstein distance. In Proceedings of the 38th International Conference on Machine Learning, ICML, volume 139, pp. 10086 10095, 2021. Trapp, M., Skowron, M., and Schabus, D. Retrieving compositional documents using position-sensitive word mover s distance. In Proceedings of the ACM SIGIR International Conference on Theory of Information Retrieval, ICTIR, pp. 233 236. ACM, 2017. Wang, Q., Wan, J., and Chan, A. B. On diversity in image captioning: Metrics and methods. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020a. Wang, Z., Zhou, D., Yang, M., Zhang, Y., Rao, C., and Wu, H. Robust document distance with wasserstein-fisher-rao metric. In Proceedings of The 12th Asian Conference on Machine Learning, ACML, volume 129 of Proceedings of Machine Learning Research, pp. 721 736. PMLR, 2020b. Weed, J., Bach, F., et al. Sharp asymptotic and finite-sample rates of convergence of empirical measures in wasserstein distance. Bernoulli, 25(4A):2620 2648, 2019. Werner, M. and Laber, E. Speeding up word mover s distance and its variants via properties of distances between Re-evaluating Word Mover s Distance embeddings. In Proceedings of the 24th European Conference on Artificial Intelligence, ECAI, volume 325 of Frontiers in Artificial Intelligence and Applications, pp. 2204 2211. IOS Press, 2020. Wrzalik, M. and Krechel, D. Balanced word clusters for interpretable document representation. In Second International Workshop on Machine Learning, WML@ICDAR, pp. 103 109. IEEE, 2019. Wu, L., Yen, I. E., Xu, K., Xu, F., Balakrishnan, A., Chen, P., Ravikumar, P., and Witbrock, M. J. Word mover s embedding: From word2vec to document embedding. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 4524 4534. Association for Computational Linguistics, 2018. Xu, H., Wang, W., Liu, W., and Carin, L. Distilled wasserstein learning for word embedding and topic modeling. In Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, Neur IPS, pp. 1723 1732, 2018. Yokoi, S., Takahashi, R., Akama, R., Suzuki, J., and Inui, K. Word rotator s distance. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 2944 2960. Association for Computational Linguistics, 2020. Yurochkin, M., Claici, S., Chien, E., Mirzazadeh, F., and Solomon, J. M. Hierarchical optimal transport for document representation. In Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Neur IPS, pp. 1599 1609, 2019. Zhang, M., Liu, Y., Luan, H., Sun, M., Izuha, T., and Hao, J. Building earth mover s distance on bilingual word embeddings for machine translation. In Proceedings of the 30th AAAI Conference on Artificial Intelligence, AAAI, pp. 2870 2876. AAAI Press, 2016. Zhang, M., Liu, Y., Luan, H., and Sun, M. Earth mover s distance minimization for unsupervised bilingual lexicon induction. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 1934 1945. Association for Computational Linguistics, 2017a. Zhang, M., Peng, H., Liu, Y., Luan, H., and Sun, M. Bilingual lexicon induction from non-parallel data with minimal supervision. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, AAAI, pp. 3379 3385. AAAI Press, 2017b. Zhang, R., Chen, C., Gan, Z., Wen, Z., Wang, W., and Carin, L. Nested-wasserstein self-imitation learning for sequence generation. In Proceedings of the 23rd International Conference on Artificial Intelligence and Statistics, AISTATS, volume 108 of Proceedings of Machine Learning Research, pp. 422 433. PMLR, 2020a. Zhang, R., Chen, C., Zhang, X., Bai, K., and Carin, L. Semantic matching via optimal partial transport. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, EMNLP, pp. 212 222. Association for Computational Linguistics, 2020b. Zhang, T., Kishore, V., Wu, F., Weinberger, K. Q., and Artzi, Y. BERTScore: Evaluating text generation with BERT. In 8th International Conference on Learning Representations, ICLR. Open Review.net, 2020c. Zhao, W., Peyrard, M., Liu, F., Gao, Y., Meyer, C. M., and Eger, S. Moverscore: Text generation evaluating with contextualized embeddings and earth mover distance. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP, pp. 563 578. Association for Computational Linguistics, 2019. Zhao, W., Glavas, G., Peyrard, M., Gao, Y., West, R., and Eger, S. On the limitations of cross-lingual encoders as exposed by reference-free machine translation evaluation. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL, pp. 1656 1671. Association for Computational Linguistics, 2020. Re-evaluating Word Mover s Distance A. Duplicate Documents We point out that there are many duplicate samples in the datasets used in the original study. The last two rows in Table 2 report the numbers of duplicate samples. Specifically, duplicate pair reports the number of pairs (s, t) such that s and t are the same, and duplicate samples reports the number of samples s such that s has at least one samples t with the same content. Some of these pairs cross the training and test splits, and some of them have different labels despite having the same contents. This causes problems in the evaluations. If the pairs have different labels, it is impossible to classify both of them correctly. If the pairs have the same label, a k NN classifier places more emphasis on this class for no reason. We found that duplication was caused by different reasons in different datasets. For example, the ohsumed dataset was originally a multi-labeled dataset, and the WMD paper duplicate samples for each label. We found that the original source dataset (Greene & Cunningham, 2006) of bbcsport3 already contained duplicate samples, e.g., athletics/12.txt and athletics/20.txt. We hypothesis this was originated from the data collection process. Although such duplication is not a problem if we intend to measure the performances in noisy environments, the analysis could be misleading if we assume a clean environment. The datasets released by the WMD paper have been used in many studies (Huang et al., 2016; Yurochkin et al., 2019; Le et al., 2019; Werner & Laber, 2020; Wang et al., 2020b; Wu et al., 2018; Skianis et al., 2020b; Mollaysa et al., 2017; Gupta et al., 2020; Skianis et al., 2020a) with the same protocol. We suspect that many readers and authors were unaware of this fact, which might have led to a misleading analysis. We release the indices of duplicate documents and a preprocessing script to remove duplicate samples for the following studies. We believe that sharing this fact within the community is considerably important for aiding in a solid evaluation and analysis. B. Train-test Split As mentioned in the main text, five-fold datasets do not have standard data splits. The WMD paper randomly split whole datasets into train and test data and distributed the splits. We found that the variance of accuracy was high with respect to the randomness of data splits. In some cases, the error rate doubled with bad random seeds. We therefore use the same splits as the original paper to make the results comparable to those in the original paper. However, as we remove the duplicated samples, each split has different number of training data. An alternative experimental setup would be to remove duplicated samples first and then split the dataset into train and test data. This setting is also valid, but one should keep the high variance in mind, avoid copying existing accuracy tables, and distribute the splits when new splits are used. C. Normalized Embeddings We note how we found the embeddings are normalized in the original evaluation. The datasets available at https: //github.com/mkusner/wmd#paper-datasets include the word embeddings, i.e., the X variable. We first found that the norms of these embeddings were all 1. We compared these embeddings with the official 300-dimentional word2vec embeddings trained on the Google News corpus https://code.google.com/archive/p/word2vec/ and found that the directions of these embeddings coincided, i.e., the WMD s embeddings were the normalized version of the official 300-dimentional word2vec embedding. We re-evaluated the classification experiments with both embeddings and found that the accuracy with the normalized embeddings agreed with the original evaluation while the accuracy with the un-normalized embeddings did not. This indicates that the normalized embeddings were used in the experiments in the original paper. D. Nearest Neighbor Classification is Far from Optimal WMD is often used as a weak baseline in document classification. In this section, we point out that the poor performances observed in previous studies are not necessarily due to the problem we found in the main part but rather due to the choices of classifiers. We found that many existing studies, including the original paper on WMD, used k NN classification for distancebased methods (Kusner et al., 2015; Huang et al., 2016; Wang et al., 2020b). This is reasonable owing to its simplicity regardless of suboptimal performance. However, we found that k NN classification was much farther from optimal than we had thought and caused some improper observations. The most prominent issue appears in comparison with non-k NN classifiers. For example, WMD is used with a k NN classifier as a baseline, and other classifiers are adopted for the proposed methods in (Wu et al., 2018; Skianis et al., 2020b; Mollaysa et al., 2017; Li et al., 2019; Gupta et al., 2020; Nikolentzos et al., 2020; 2017). The improvement observed by such evaluations may not be due to the superiority of the proposed methods 3http://mlg.ucd.ie/datasets/bbc.html Re-evaluating Word Mover s Distance Table 6. Weighted k NN classification errors. Lower is better. The last column reports the average relative performance to the normalized BOW with the standard k NN (the first row in Table 2). bbcsport twitter recipe ohsumed classic reuters amazon 20news rel. BOW (L1/L1/wk NN) 3.3 0.6 26.4 1.6 41.1 0.5 41.3 3.6 0.4 4.7 10.3 0.3 23.3 0.888 TF-IDF (L1/L1/wk NN) 2.0 0.7 26.6 1.5 38.3 1.1 36.1 2.8 0.4 5.8 8.2 0.5 20.5 0.787 WMD (wk NN) 4.4 1.2 26.2 2.2 40.5 1.0 41.1 2.7 0.4 3.6 7.1 0.5 21.7 0.823 WMD-TF-IDF (wk NN) 3.2 1.0 25.6 1.5 37.7 0.9 37.0 2.4 0.4 4.0 6.3 0.4 19.6 0.743 against WMD, but due to the superiority of the classifier against the k NN classifier. To mitigate this problem, we found that a weighted k NN classifier (wk NN) with exponential weighting was a good choice for distance-based methods. Specifically, wk NN first gathers k nearest neighbor samples and computes the distances {d1, , dk} to these samples. It then defines the weight for sample i as exp( di/γ) and takes the weighted majority vote, where γ is a hyperparameter. We can think of wk NN as a continuous variant of k NN, which uses a step function as the weight function. As the crux of wk NN, the time and space complexities are the same as in k NN, and it involves no training procedures. Thus, we can easily replace a k NN system with a wk NN system. We conduct experiments to show the suboptimality of k NN. As the drawback of wk NN, it has two hyperparameters k and γ. We find that the k of wk NN plays a similar role as the maximum k of k NN in the hyperparameter tuning. Thus, we fix k of wk NN to 19 and tune only γ in the hyperparameter tuning. Recall that the hyperparameter candidates were k {1, , 19} for k NN in the previous experiments and original paper. We select γ from Γ = {0.005, 0.010, , 0.095, 0.1} (|Γ| = 20) in the hyperparameter tuning4. The remaining settings are the same as in the previous experiments. Table 6 shows the classification errors. We can see that wk NN performs much better than k NN. The benefit of wk NN (the first row in Table 6 versus the first row in Table 2) is more significant than the benefit of WMD (the third versus the first row in Table 2). This indicates that if a proposed method uses non-k NN classifiers, ten percent improvements from the k NN baselines may be due to classifiers, not to the proposed method. This also indicates that there is significant room for improvement in the classifiers before undergoing a design of better similarity measures. It might be acceptable if all baselines and the proposed method use k NN because the relative performances in Tables 2 and 6 do not change significantly. However, it would be better to use more capable classifiers because the results in a paper will be cited as baseline records in following papers, in which other classifiers may be employed. In fact, such comparisons have been carried out in (Wu et al., 2018; Skianis et al., 2020b; Mollaysa et al., 2017; Gupta et al., 2020). Besides, practitioners may underestimate the performance of distance-based methods based on the reported results. We find wk NN is a better choice than k NN because it is as fast as k NN yet much more effective. E. Comparison with other Normalization with Clean Data Table 7 shows the k NN classification errors with clean data. The same tendency as Table 2 can be observed. F. Performance of Low Dimensional Embeddings Table 8 shows the classification errors of WMD with the 5-dimensional embeddings used in Section 5. It shows that WMD with the low dimensional embeddings performs worse than with the original 300-dimensional embeddings, although WMD performs differently from BOW with the low dimensional embeddings (Figure 5). In general, an exact match of words indicates a strong similarity. It is reasonable that counting common words can perform better than relying too much on the underlying geometry. G. Ablation Study on the Cost Function of WMD There are two factors for the cost function of WMD: how to normalize word vectors and how to measure the distance between the normalized vectors. We conducted ablation studies on the cost for WMD using L1 and L2 normalization and L1 and L2 distances. Table 9 shows that the choice of the normalization and distance does not affect the performance much. It 4The magnitude of this range is determined by the empirical distances between samples. This arbitrary choice is a slight drawback of wk NN. We will investigate how to choose it automatically in future work. Re-evaluating Word Mover s Distance Table 7. k NN classification errors with clean data. Lower is better. The same notations as in Table 2. bbcsport twitter recipe ohsumed classic reuters amazon 20news rel. BOW (L1/L1) 3.7 1.0 30.6 1.1 42.9 0.6 39.7 4.2 0.5 5.5 10.6 0.6 29.2 1.000 TF-IDF (L1/L1) 2.3 1.4 30.2 0.7 40.0 1.1 33.4 3.5 0.2 5.9 8.0 0.6 25.9 0.866 WMD 5.5 1.2 30.6 1.2 42.9 0.9 40.6 3.4 0.6 3.8 7.3 0.4 26.9 0.952 WMD-TF-IDF 4.1 1.5 28.8 1.6 40.2 0.9 35.7 2.8 0.3 4.3 6.6 0.3 24.2 0.848 BOW (None/L2) 22.8 1.6 34.2 0.6 59.1 0.9 60.7 36.9 1.1 11.7 28.9 1.1 58.0 3.227 BOW (None/L1) 25.3 2.1 33.9 0.8 64.1 0.8 67.4 55.0 0.5 14.0 32.4 1.3 73.5 4.044 TF-IDF (None/L2) 25.8 1.1 33.5 0.6 65.6 1.0 62.8 41.1 1.1 28.3 43.4 5.4 59.8 4.032 TF-IDF (None/L1) 32.7 1.2 33.6 0.6 70.6 1.8 69.5 55.6 0.2 29.1 42.5 4.8 74.7 4.804 BOW (L1/L2) 11.8 0.7 40.2 2.0 51.4 1.4 55.8 17.5 1.7 12.9 36.8 1.4 46.7 2.336 BOW (L2/L1) 15.0 1.6 34.3 0.9 59.9 0.7 64.3 54.0 0.4 16.2 32.7 2.1 77.1 3.715 BOW (L2/L2) 5.3 0.9 32.6 0.7 45.4 1.1 43.2 6.7 0.7 8.1 13.3 0.5 33.2 1.263 TF-IDF (L1/L2) 20.0 7.4 39.3 1.4 54.1 2.4 55.2 24.8 3.0 24.5 43.2 2.0 45.9 3.168 TF-IDF (L2/L1) 28.7 5.8 33.5 1.0 65.7 2.8 65.7 55.0 0.2 19.9 39.0 5.4 78.6 4.390 TF-IDF (L2/L2) 3.1 1.3 31.5 1.5 43.6 0.8 33.2 5.2 0.4 10.6 11.1 0.6 31.6 1.127 Table 8. k NN classification errors with 5 dimensional embeddings. Lower is better. The first two rows are the same as Table 2 for reference. bbcsport twitter recipe ohsumed classic reuters amazon 20news rel. BOW (L1/L1) 3.9 1.1 30.0 1.1 43.4 0.8 44.1 4.1 0.5 5.7 10.4 0.5 29.1 1.000 WMD 5.1 1.2 29.6 1.5 42.9 0.8 44.5 2.9 0.4 4.0 7.4 0.5 26.8 0.917 WMD (5 dim) 18.5 0.8 32.0 1.0 51.8 1.0 60.2 6.2 0.3 6.9 15.9 0.8 45.4 1.773 Table 9. Ablation study on the cost function. The evaluation protocol is the same as Table 2 in the main text. WMD (x/y) normalizes the word embeddings by x and measures the distance between embeddings by y. WMD (L2/L2) corresponds to the method we used in the main text. It can be seen that the normalization and distance of the cost function do not affect the performance much. bbcsport twitter recipe ohsumed classic reuters amazon 20news WMD (L2/L2) 5.1 1.2 29.6 1.5 42.9 0.8 44.5 2.9 0.4 4.0 7.4 0.5 26.8 WMD (L2/L1) 4.8 0.9 29.6 1.3 43.2 1.0 44.7 2.9 0.3 4.1 7.3 0.3 26.9 WMD (L1/L2) 5.0 1.1 29.5 1.4 42.9 1.0 43.9 2.9 0.4 4.1 7.4 0.5 26.6 WMD (L1/L1) 5.1 1.3 29.7 1.4 43.0 0.8 44.1 2.9 0.4 4.2 7.3 0.4 26.7 should be emphasized that the meaning of (x/y) in Table 9 is different from that for BOW and TF-IDF. The normalization and distance we consider in this section are for the cost function, while the normalization and distance we considered in the main body are for the Bo W representation, which corresponds to the marginal distribution in WMD. Note also that unfortunatelly, we could not conduct an ablation study for WMD (None/L1) and WMD (None/L2) because the embeddings in the released datasets are already normalized as we mention in Appendix C. We expect WMD (None/L1) and WMD (None/L2) degrade performances as Yokoi et al. (2020) reported in STS tasks. H. Proof of Proposition 4.1 We prove the analogy between L1/L1 BOW and OT. Recall that d BOW/L1/L1(x, x ) = n L1(x) n L1(x ) 1, and ( 0 (i = j) 2 (i = j). Proof of Proposition 4.1. Without loss of generality, we assume that the marginals x and x are already L1 normalized, and P i xi = 1 and P i x i = 1 hold. Let P Rm m be the optimal solution of OT(x, x , Cunif). From the marginal constraints, P ii min(xi, x i) holds. If P ii < min(xi, x i), there exists j = i and k = i such that P ij > 0 and P ki > 0 Re-evaluating Word Mover s Distance from the marginal constraints. Then, let Q Rm m be P ii + min(P ij, P ki) (s = i t = i) P ij min(P ij, P ki) (s = i t = j) P ki min(P ij, P ki) (s = k t = i) P kj + min(P ij, P ki) (s = k t = j) P st (otherwise). Then, Q satisfies the constraints of OT and X st Cunif st P st st Cunif st Qst = Cunif ii (P ii (P ii + min(P ij, P ki))) + Cunif ij (P ij (P ij min(P ij, P ki))) + Cunif ki (P ki (P ki min(P ij, P ki))) + Cunif kj (P kj P kj + min(P ij, P ki))) = (4 Cunif kj ) min(P ij, P ki) > 0. The first equation holds from the definition of Q, and the second equation holds because Cunif ii = 0 and Cunif ij = Cunif jk = 2, and the last inequality holds because Cunif kl 2 and min(P ij, P ki) > 0. This contradicts with the optimality of P . Therefore, P ii = min(xi, x i) holds. Therefore, X st Cunif st P st st 2(11 diag(1))st P st st P st 2 X s min(xs, x s) s min(xs, x s) X s min(xs, x s) s (xs min(xs, x s)) + (x s min(xs, x s)) I. Additional Related Work We review additional related work, which we could not place in the main text because of the page limit. First, it has already been known that OT behaves pathologically in high dimensional spaces, mainly owing to the sample complexity. Dudley (1969) showed that the sample complexity of OT grows exponentially with respect to the number of dimensions. Weed et al. (2019) alleviated the bound using other senses of dimensionality, but the dependency is still exponential. Genevay et al. (2019) showed that the sample complexity of the Sinkhorn Divergences (Genevay et al., 2018; Cuturi, 2013) matches that of MMD. The crucial difference between our work and theirs is that we shed light on more practical sides, whereas their interests were on the theoretical side. We believe that our analysis and explanation are insightful, especially for practitioners. Secondly, high-dimensional problems have also been studied from computational aspects. Rabin et al. (2011) and Kolouri et al. (2016) proposed to project embeddings to random one-dimensional spaces to speedup computation because OT in a one-dimensional space can be solved efficiently (Santambrogio, 2015). Paty & Cuturi (2019) and others (Kolouri et al., Re-evaluating Word Mover s Distance 2019; Dhouib et al., 2020; Petrovich et al., 2020; Lin et al., 2020; Huang et al., 2021) proposed robust variants of OT with low dimensional projections. Specifically, their methods adopt the worst distance with respect to candidate projections. Although we tried in early experiments several projection methods, including the principal component analysis as in Section 5 and Appendix F, tree slicing (Le et al., 2019), and rank-based cost matrices instead of actual distance cost matrices (Sato et al., 2020a), we did not find performance improvements. However, exploring these approaches can be one of the promising future directions with the curse of dimensionality in mind.