# all_word_embeddings_from_one_embedding__f822b942.pdf All Word Embeddings from One Embedding Sho Takase Tokyo Institute of Technology sho.takase@nlp.c.titech.ac.jp Sosuke Kobayashi Tohoku University Preferred Networks, Inc. sosk@preferred.jp In neural network-based models for natural language processing (NLP), the largest part of the parameters often consists of word embeddings. Conventional models prepare a large embedding matrix whose size depends on the vocabulary size. Therefore, storing these models in memory and disk storage is costly. In this study, to reduce the total number of parameters, the embeddings for all words are represented by transforming a shared embedding. The proposed method, ALONE (all word embeddings from one), constructs the embedding of a word by modifying the shared embedding with a filter vector, which is word-specific but non-trainable. Then, we input the constructed embedding into a feed-forward neural network to increase its expressiveness. Naively, the filter vectors occupy the same memory size as the conventional embedding matrix, which depends on the vocabulary size. To solve this issue, we also introduce a memory-efficient filter construction approach. We indicate our ALONE can be used as word representation sufficiently through an experiment on the reconstruction of pre-trained word embeddings. In addition, we also conduct experiments on NLP application tasks: machine translation and summarization. We combined ALONE with the current state-of-the-art encoderdecoder model, the Transformer [36], and achieved comparable scores on WMT 2014 English-to-German translation and DUC 2004 very short summarization with less parameters1. 1 Introduction Word embeddings have played a crucial role in the recent progress in the area of natural language processing (NLP) [3, 17, 22, 23]. In particular, word embeddings are necessary to convert discrete input representations into vector representations in neural network-based NLP methods [3]. To convert an input word w into a vector representation in a conventional way, we prepare a one-hot vector vw whose dimension size is equal to the vocabulary size V and an embedding matrix E whose shape is De V (De represents a word embedding size). Then, we multiply E and vw to obtain a word embedding ew. NLP researchers have used word embeddings as input in their models for several applications such as language modeling [38], machine translation [30], and summarization [26]. However, in these methods, the embedding matrix forms the largest part of the total parameters, because V is much larger than the dimension sizes of other weight matrices. For example, the embedding matrix makes up one-fourth of the parameters of the Transformer, a state-of-the-art neural encoder-decoder model, on WMT English-to-German translation [36]. Thus, if we can represent each word with fewer parameters without significant compromises on performance, we can reduce the model size to conserve memory space. If we save the memory space for embeddings, we can also train a bigger model to improve the performance. 1The code is publicly available at https://github.com/takase/alone_seq2seq 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. Conventional way ALONE (proposed method) Source matrices Filter construction Pick an embedding Fixed with random initialization Figure 1: Word embedding constructions by the conventional way and our proposed ALONE. ALONE represents each word with an embedding o and the filter vector mw. We fix source matrices with random initialized values, and thus it is unnecessary to prepare trainable parameters for construction of mw. To increase the expressiveness, we input the embedding into a feed-forward neural network. To reduce the size of neural network models, some studies have proposed the pruning of unimportant connections from a trained network [15, 8, 39]. However, their approaches require twice or much computational cost, because we have to train the network before and after pruning. Another approach [31, 28, 1] limited the number of embeddings by utilizing the composition of shared embeddings. These methods represent an embedding by combining several primitive embeddings, whose number is less than the vocabulary. However, since they require learning the combination of assignments of embeddings to each word, they need additional parameters during the training phase. Thus, prior approaches require multiple training steps and/or additional parameters that are necessary only during the training phase. To address the above issues, we propose a novel method: ALONE (all word embeddings from one), which can be used as a word embedding set without multiple training steps and additional trainable parameters to assign each word to a unique vector. ALONE computes an embedding for each word (as a replacement for ew) by transforming a shared base embedding with a word-specific filter vector and a shared feed-forward network. We represent the filter vectors with combinations of primitive random vectors, whose total is significantly smaller than the vocabulary size. In addition, it is unnecessary to train the assignments, because we assign the random vectors to each word randomly. Therefore, while ALONE retains its expressiveness, its total parameter size is much smaller than the conventional embedding size, which depends on the vocabulary size due to De V . Through experiments, we demonstrate ALONE can be used for NLP with comparable performances to the conventional embeddings but fewer parameters. We first indicate ALONE has enough expressiveness to represent the existing word embedding (Glo Ve [22]) through a reconstruction experiment. In addition, on two NLP applications of machine translation and summarization, we demonstrate ALONE with the state-of-the-art encoder-decoder model trained in an end-to-end manner achieved comparable scores with fewer parameters than models with the conventional word embeddings. 2 Proposed Method: ALONE Figure 1 shows an overview of the conventional and proposed word embedding construction approaches. As shown in this figure, the conventional way requires a large embedding matrix E RDe V , where each column vector is assigned to a word, and we pick a word embedding with a one-hot vector vw R1 V . In other words, each word has a word-specific vector, whose size is De, and the total size summed over the vocabulary is De V . To reduce the size, ALONE make words share some type of vector element with each other. ALONE has a base embedding o R1 Do, which is shared by all words, and represents each word by transforming the base embedding. To concretely obtain a word representation for w, ALONE computes an element-wise product of o and a filter vector mw, and then applies a feed-forward network to increase its expressiveness. The two components are described as follows. Source matrices Pick vectors corresponding to the word w Figure 2: Construction of the filter vector mw. We pick several vectors corresponding to the word w from source matrices and combine them to construct mw. 2.1 Filter Construction To make the result of the element-wise product mw o unique to each word, we have to prepare different filter vectors from each other. In the simplest way, we sample values from a distribution such as Gaussian Do V times, and then we construct the matrix whose size is Do V with the sampled values. However, the above method requires similar memory space to E in the conventional word embeddings. Thus, we introduce a more memory-efficient way herein. Our filter construction approach does not prepare the filter vector for each word explicitly. Instead, we construct a filter vector by combining multiple vectors, as shown in Figure 2. In the first step, we prepare M source matrices such as codebooks, each of which is Do c. Then, we assign one column vector of each matrix to each word randomly. Thus, each word is tied to M (column) vectors. In this step, since the probability of collision between two combinations is much small (1 exp( V 2 2(c M)) based on the birthday problem), each word is probably assigned to the unique set of M vectors. Moreover, the required memory space, i.e., Do c M is smaller than E when we use c M V . To construct the filter vector mw, we pick assigned column vectors from each matrix, and compute the sum of the vectors. Then, we apply a function f( ) to the result. Formally, let m1 w, ..., m M w be column vectors assigned to w, we compute the following equation: In this paper, we use two types of filter vectors: a binary mask and a real number vector based on the following distribution and function. Binary Mask A binary mask is a binary vector whose elements are 0 or 1, such as the dropout mask [29]. Thus, we ignore some elements of o based on the binary mask for each word. For the binary mask construction, we use the Bernoulli distribution. To make the binary mask containing 0 with probability po, we construct the source matrices with sampling from the Bernoulli(1 p Moreover, we use the following Clip( )2 as the function f to trim 1 or more to 1 in the filter vectors. Clip(a) = 1 1 a 0 otherwise (2) In other words, in binary mask construction, mw is computed from the element-wise logical OR operation over all mi w. Real Number Vector In addition to the binary mask, we use the filter vector, which consists of real numbers. We use the Gaussian distribution to construct source matrices, and the identity transformation as the function f. In other words, we use the sum of vectors from source matrices without any transformation as the filter vector. 2Equation (2) represents the operation for a scalar value. If an input is a vector such as Equation (1), we apply Equation (2) to all elements in the input vector. Table 1: Memory space required by each method for word representation. ALONE (naive) represents the case where we prepare filter vectors explicitly. Method Memory usage Conventional way De V ALONE (naive) Do + Dinter (Do + De) + Do V ALONE (proposed) Do + Dinter (Do + De) + M Do c ALONE (proposed (volatile)) Do + Dinter (Do + De) 2.2 Feed-Forward Network We obtain the unique vector to each word by computing the element-wise product of o and mw. However, in this situation, words share the same value in several elements. Thus, we increase the expressiveness by applying a feed-forward network FFN( ) to the result of mw o: FFN(x) = W2(max(0, W1x)), (3) where W1 RDinter Do and W2 RDe Dinter are weight matrices. In short, the feed-forward network in this paper consists of two linear transformations with a Re LU activation function. We use the output of FFN(mw o) as the word embedding for w. 2.3 Discussion on the Number of Parameters and Memory Footprint Table 1 summarizes the memory space required by each method. In this table, ALONE (naive) ignores the filter construction approach introduced in Section 2.1. As described previously, the conventional way requires a large amount of memory space because V is exceptionally large (> 104) in most cases. In contrast, ALONE (proposed) drastically reduces the number of parameters due to Do, Dinter, M c V , and thus, we can reduce the memory footprint when we adopt the introduced filter construction way. Moreover, since ALONE uses random initialized vectors as filter vectors without any training, we can reconstruct them again and again if we store a random seed. Thus, we can ignore the memory space for filter vectors such as ALONE (proposed (volatile)). As an example, consider word embeddings on WMT 2014 English-to-German translation in the setting of the original Transformer [36]. In this setting, the conventional way requires 19M as the memory footprint due to V = 37000 and De = 512. In contrast, ALONE compresses the parameter size to 4M, which is less than a quarter of that footprint, when we set Do = 512 and Dinter = 4096. These values are used in the following experiment on machine translation. For filter vectors, the naive filter construction way requires an additional 19M for the memory footprint, but our introduced approach requires only 262k more when we set c = 64 and M = 8. These values are also used in our experiments. Thus, the proposed ALONE reduces the memory footprint as compared to that of the conventional word embeddings. 3 Experiments In this section, we investigate whether the proposed method, ALONE, can be an alternation of the conventional word embeddings. We first conduct an experiment on the reconstruction of pretrained word embeddings to investigate whether ALONE is capable of mimicking the conventional word embeddings. Then, we conduct experiments on real applications: machine translation and summarization. We train the Transformer, which is the current state-of-the-art neural encoder-decoder model, combined with the ALONE in an end-to-end manner. 3.1 Word Embedding Reconstruction In this experiment, we investigate whether ALONE has a similar expressiveness to the conventional word embeddings. We used the pre-trained 300 dimensional Glo Ve3 [22] as source word embeddings and reconstructed them with ALONE. 3https://nlp.stanford.edu/projects/glove/ (a) Sim Lex-999 (b) Word Sim-353 (c) RG-65 Figure 3: Results on word embedding reconstruction. The dashed line indicates the Spearman s rank correlation coefficient of Glo Ve. Training details: The training objective is to mimic the Glo Ve embeddings with ALONE. Let ew be the conventional word embedding (Glo Ve in this experiment), we minimize the following objective function: w=1 ||ew FFN(mw o)||2. (4) We optimized the above objective function with Adam [13] whose hyper-parameters are default settings in Py Torch [21]. We set mini-batch size 256 and the number of epochs 1000. We constructed each mini-batch with uniformly sampling from vocabulary and regard training on the whole vocabulary as one epoch. For c, M, and po in the binary mask, we set 64, 8, and 0.5 respectively4. We used the same dimension size as Glo Ve (300) for Do and conducted experiments with varying Dinter in {600, 1200, 1800, 2400}. In each setting, the total number of parameters is 0.4M, 0.7M, 1.1M, and 1.4M. We selected the top 5k words based on the frequency in English Wikipedia as target words5. We used five random seeds to initialize ALONE, and report the average of five scores. Test data: We used the word similarity task, which is widely used to evaluate the quality of word embeddings [22, 4, 37]. The task investigates whether similarity based on trained word embeddings corresponds to the human-annotated similarity. In this paper, we used three test sets: Sim Lex-999 [9], Word Sim-353 [5], and RG-65 [25]. We computed Spearman s rank correlation (ρ) between the cosine similarities of word embeddings and the human annotations as in previous studies. Results: Figure 3 shows the Spearman s ρ of ALONE in both filter vectors. This figure also shows the Spearman s ρ without training o (Fix o). The dashed line indicates the Spearman s ρ of Glo Ve, i.e., the upper bound of word embedding reconstruction. Figure 3 indicates that ALONE achieved comparable scores to Glo Ve on all datasets in Dinter = 2400. These results indicate that ALONE has the expressive power to represent the conventional word embeddings. In the comparison between filter vectors, we cannot find the significant difference in Dinter = 2400, but the real number vectors slightly outperformed binary masks in small Dinter. Thus, we should use real number vectors as the filter vectors if the number of parameters is strongly restricted. Figure 3 shows that the setting without training o is defeated against training o in most cases. Therefore, it is better to train o to obtain superior representations. 3.2 Machine Translation Section 3.1 indicates ALONE can mimic pre-trained word embeddings, but there remain two questions: 1. Can we train ALONE in an end-to-end manner? 2. In the realistic situation, can ALONE reduce the number of parameters related to word embeddings while maintaining performance? 4We tried several values, but we cannot observe any significant differences among the results. 5We can use the whole vocabulary of pre-trained Glo Ve embeddings but we restricted vocabulary size to shorten the training time. Table 2: Results of WMT En-De translation. Method Dinter Embed BLEU Conv S2S [6] 66.0M 25.2 Transformer [36] 16.8M 27.3 Transformer+De FINE [16] - 27.01 Transformer (conventional word embeddings) 16.8M 27.12 Transformer (factorized embed) 4.3M 26.43 Transformer (factorized embed) 8.5M 26.56 Transformer+ALONE (Binary) 4096 4.2M 26.97 Transformer+ALONE (Real number) 4096 4.2M 26.93 Transformer+ALONE (Binary) 8192 8.4M 27.55 Transformer+ALONE (Real number) 8192 8.4M 27.61 ALONE without training o Transformer+ALONE (Binary) 4096 4.2M 26.75 Transformer+ALONE (Real number) 4096 4.2M 26.85 Transformer+ALONE (Binary) 8192 8.4M 26.90 Transformer+ALONE (Real number) 8192 8.4M 26.95 To answer these questions, we conduct experiments on machine translation and summarization. Training details: In machine translation, we used ALONE as word embeddings instead of the conventional embedding matrix E in the Transformer [36]. We adopted the base model setting and thus shared embeddings with the pre-softmax linear transformation matrix. We used the fairseq implementation [19] and followed the training procedure described in its documentation 6. We set Do the same number as the dimension of each layer in the Transformer (dmodel, i.e., 512) and varied Dinter. For other hyper-parameters, we set as follows: c = 64, M = 8, and po = 0.5. Moreover, we applied the dropout after the Re LU activation function in Equation (3). Dataset: We used WMT En-De dataset since it is widely used to evaluate the performance of machine translation [6, 36, 18]. Following previous studies [36, 18], we used WMT 2016 training data, which contains 4.5M sentence pairs, newstest2013, newstest2014 for training, validation, and test respectively. We constructed a vocabulary with the byte pair encoding [27] whose merge operation is 32K in sharing source and target. We measured case-sensitive BLEU with Sacre BLEU [24]. Result: Table 2 shows BLEU scores of the Transformer with ALONE in the case Dinter = 4096, 8192. This table also shows BLEU scores of previous studies [6, 36, 16] and the Transformer trained in our environment with the same hyper-parameters as the original one [36]7. De FINE [16] uses a factorization approach, which constructs embeddings from small matrices instead of one large embedding matrix, to reduce the number of parameters8. In addition, we trained the Transformer with a simple factorization approach as a baseline. In the simple factorization approach, we construct word embeddings from one small embedding matrix and weight matrix to expand a small embedding to one with the larger dimensionality De. We modified the dimension size to make the number of parameters related to embeddings almost equal to those in ALONE. Table 2 shows the Transformer with ALONE in Dinter = 8192 outperformed other methods even though the embedding parameter size is half that of the Transformer with the conventional embeddings. This result indicates that our ALONE can reduce the number of parameters related to embeddings without negatively affecting the performance on machine translation. 6https://github.com/pytorch/fairseq/tree/master/examples/scaling_nmt 7The number of parameters for embeddings in the Transformer is different from that in the original one [36] owing to the difference of vocabulary size. 8Table 2 shows the reported score. We cannot demonstrate the embedding parameter size for Transformer+De FINE because its vocabulary size is unreported but the parameter size of Transformer+De FINE is 68M, which is larger than that of the original Transformer (60.9M). We consider that the original Transformer (and our experiments) save the total parameter size by sharing embeddings with the pre-softmax linear transformation matrix. Table 3: Results on DUC 2004 task 1. The scores in bold are superior to the previous top score. Method Dinter Embed R-1 R-2 R-L ABS [26] 42.1M 28.18 8.49 23.81 LSTM Enc Dec+WFE [32] 37.7M 32.28 10.54 27.80 Transformer+LRPE+PE [34] 8.3M 32.29 11.49 28.03 +ALONE (Binary) 512 0.5M 31.60 11.12 27.25 +ALONE (Real number) 512 0.5M 31.96 11.50 27.74 +ALONE (Binary) 1024 1.0M 32.51 11.48 28.08 +ALONE (Real number) 1024 1.0M 32.57 11.63 28.24 In comparison with the factorized embedding approaches, Transformer with ALONE (Dinter = 8192) achieved superior BLEU scores as compared to Transformer+De FINE [16], while the total parameter size of Transformer with ALONE, 52.5M, is smaller than it (68M). In other words, Transformer+ALONE achieved better performance even though it had a disadvantage in the parameter size. Moreover, Transformer+ALONE outperformed Transformer (factorized embed) despite the number of parameters related to embeddings in both being almost equal. These results imply that ALONE is superior to a simple approach using a small embedding matrix and expansion with a weight matrix. Table 2 also shows BLEU scores of the Transformer with ALONE in the case without training the base embedding o. The setting without training o didn t achieve BLEU scores as high as training o. Therefore, we also have to train o in the training ALONE with an end-to-end manner. 3.3 Summarization Training details: We also conduct an experiment on the headline generation task, which is one of the abstractive summarization tasks [26, 34]. The purpose of this task is to generate a headline of a given document with the desired length. Thus, we introduced ALONE into the Transformer with the length control method [34]. For other model details, we used the same as the experiment on machine translation. We used the publicly available implementation9 and followed their training procedure. As the length control method, we used the combination of LRPE and PE [34]. Moreover, we re-ranked decoded candidates based on the content words following the previous study [34]. Dataset: As in previous studies [26, 34], we used pairs of the first sentence and headline extracted from the annotated English Gigaword with the same pre-processing script provided by Rush et al. [26]10 as the training data. The training set contains about 3.8M pairs. For the source-side, we used the byte pair encoding [27] to construct a vocabulary. We set the hyper-parameter to fit the vocabulary size 16K. For the target-side, we used a set of characters as a vocabulary to control the number of output characters. We shared the source-side and target-side vocabulary. We used the DUC 2004 task 1 [20] as the test set. Following the evaluation protocol [20], we truncated the characters of the output headline to 75 bytes and computed the recall-based ROUGE score. Result: Table 3 shows the ROUGE scores of the previous best method (Transformer+LRPE+PE [34]) with ALONE in the case Dinter = 512, 102411. This table indicates that ALONE (Real number) achieved the comparable ROUGE-2 score to the previous best in Dinter = 512. In addition, Transformer+LRPE+PE+ALONE in Dinter = 1024 outperformed the previous best score except for ROUGE-2 in ALONE (Binary) despite the embedding parameter size being one-eighth of that of the original Transformer+LRPE+PE. Figure 4 shows loss (negative log likelihood) values on validation sets of machine translation and summarization (newstest2013 and a set sampled from English Gigaword, respectively) for each word embedding method. This figure indicates that the convergence of the Transformer with ALONE is 9https://github.com/takase/control-length 10https://github.com/facebookarchive/NAMAS 11In the abstractive summarization task, the vocabulary size is much smaller than that in the machine translation experiment because we used characters for the target-side vocabulary, and the source-side vocabulary size is also small. Thus, we set a smaller Dinter. 0 50 100 150 200 250 300 Epoch Valid loss (NLL) Conventional embeddings Binary (4096) Real number (4096) Binary (8192) Real number (8192) (a) Machine Translation 0 20 40 60 80 100 Epoch Valid loss (NLL) Conventional embeddings Binary (512) Real number (512) Binary (1024) Real number (1024) (b) Summarization Figure 4: Validation loss of each method. slower than the conventional embeddings. In particular, the convergence of Binary (512) is much slower than other methods in summarization. We consider this slow convergence harmed ROUGE scores of ALONE (Binary) in DUC 2004. Meanwhile, the Transformer+ALONE in Dinter = 8192, which outperformed the original Transformer in machine translation, achieved superior validation loss values as compared to the conventional embeddings. In conclusion, based on the machine translation and summarization results, the answers for the previously mentioned two questions are as follows; 1. Yes, we can train ALONE with a neural method from scratch. 2. Yes, our ALONE can reduce the parameter size for embeddings without sacrificing performance. 4 Related Work Researchers have proposed several strategies to compress neural networks such as pruning [15, 8, 39], knowledge distillation [10, 11], and quantization [7, 2, 28, 1, 31, 33]. Han et al. [8] proposed iterative pruning, which consists of the following three steps: (1) train a neural network to find important connections, (2) prune the unimportant connections, and (3) re-train the neural network to tune the remaining connections. Zhang et al. [39] iteratively performed steps (2) and (3) to compress a neural machine translation model. Knowledge distillation approaches train a small network to mimic a pre-trained network by minimizing the difference between the outputs of the small network and pre-trained original network [10, 11]. Kim and Rush [11] applied knowledge distillation to a neural machine translation model and obtained a smaller model that achieves comparable scores to the original network. These approaches require additional computational costs to acquire a compressed model because they need to train a base network before compression. In contrast, our proposed ALONE does not require a pre-trained network because we can train it with an end-to-end manner in the same way as the conventional word embeddings. In addition, we can also apply the above approaches to compress ALONE because the approaches are orthogonal to it. Chen et al. [2] proposed Hashed Net, which constructs the weight matrix from a few parameters. Hashed Net decides the assignment of trainable parameters to an element of the weight matrix based on a hash function. Some studies applied such a parameter assignment approach to word embeddings [31, 28, 1]. For example, Suzuki and Nagata [31] constructs word embeddings with the concatenation of several sub-vectors (trainable parameters). Their method optimizes the parameters and assignments of sub-vectors through training. While those methods can represent each word with small parameters after training, they require additional parameters during the training phase. In fact, Suzuki and Nagata [31] uses conventional word embeddings during training. In contrast, such additional parameters are unnecessary in ALONE. Svenstrup et al. [33] proposed the method to assign each word to several vectors using a hash function. Their method also has no additional parameters during training, but it requires weight vectors for each word. Thus, its parameter size depends on the vocabulary size V . In contrast, since the parameters of ALONE are independent from V , ALONE can represent each word with fewer parameters. To reduce the number of parameters related to word embeddings, some studies have reduced the vocabulary size V [12, 37, 35]. In these days, it is common to construct the vocabulary with subwords [27, 14] for generation tasks such as machine translation. Furthermore, some studies use characters (or character n-grams) as their vocabulary and construct word embeddings from character embeddings [12, 37, 35]. This study does not conflict with these studies because ALONE is used to represent the word, sub-word, and character embeddings in our experiments. 5 Conclusion This paper proposes ALONE, a novel method to reduce the number of parameters related to word embeddings. ALONE constructs embeddings for each word from one embedding in contrast to the conventional way that prepares a large embedding matrix whose size depends on the vocabulary size V . Through experiments, we indicated that ALONE can represent each word while maintaining the similarity of pre-trained word embeddings. In addition, we can train ALONE in an end-to-end manner on real NLP applications. We combined ALONE with the strong neural encoder-decoder method, Transformer [36], and achieved comparable scores on WMT 2014 English-to-German translation and DUC 2004 very short summarization with fewer parameters. Broader Impact This study addresses the reduction of trainable parameters for word embeddings. Word embeddings are fundamental component of various neural network-based NLP methods because we need them to convert a symbolic input into vector representations. Thus, the proposed method, ALONE, has potential to reduce the parameter size of existing neural network-based NLP methods. In this paper, we combined ALONE with a neural encoder-decoder model but we expect that it also has a positive effect on other methods such as large-scale neural language models. Acknowledgments and Disclosure of Funding We thank Hiroshi Noji, Hitomi Yanaka, Koki Washio, and Saku Sugawara for constructive discussions. We thank anonymous reviewers for their useful suggestions. This work was supported by JSPS KAKENHI Grant Number JP18K18119. The first author is supported by Microsoft Research Asia (MSRA) Collaborative Research Program. [1] Ting Chen, Martin Renqiang Min, and Yizhou Sun. Learning k-way d-dimensional discrete codes for compact embedding representations. In Proceedings of the 35th International Conference on Machine Learning (ICML), pages 854 863, 2018. [2] Wenlin Chen, James Wilson, Stephen Tyree, Kilian Weinberger, and Yixin Chen. Compressing neural networks with the hashing trick. In Proceedings of the 32nd International Conference on Machine Learning (ICML), 2015. [3] Ronan Collobert, Jason Weston, Léon Bottou, Michael Karlen, Koray Kavukcuoglu, and Pavel Kuksa. Natural language processing (almost) from scratch. Journal of Machine Learning Research, 12:2493 2537, 2011. [4] Manaal Faruqui, Jesse Dodge, Sujay Kumar Jauhar, Chris Dyer, Eduard Hovy, and Noah A. Smith. Retrofitting word vectors to semantic lexicons. In Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pages 1606 1615, 2015. [5] Lev Finkelstein, Evgeniy Gabrilovich, Yossi Matias, Ehud Rivlin, Zach Solan, Gadi Wolfman, and Eytan Ruppin. Placing search in context: The concept revisited. In Proceedings of the 10th international conference on World Wide Web (WWW), pages 406 414, 2001. [6] Jonas Gehring, Michael Auli, David Grangier, Denis Yarats, and Yann N. Dauphin. Convolutional sequence to sequence learning. In Proceedings of the 34th International Conference on Machine Learning (ICML), volume 70, pages 1243 1252, 2017. [7] Yunchao Gong, Liu Liu, Ming Yang, and Lubomir D. Bourdev. Compressing deep convolutional networks using vector quantization. Co RR, abs/1412.6115, 2014. [8] Song Han, Jeff Pool, John Tran, and William Dally. Learning both weights and connections for efficient neural network. In Advances in Neural Information Processing Systems 28 (NIPS), pages 1135 1143. 2015. [9] Felix Hill, Roi Reichart, and Anna Korhonen. Sim Lex-999: Evaluating semantic models with (genuine) similarity estimation. Computational Linguistics, 41(4):665 695, 2015. [10] Geoffrey Hinton, Oriol Vinyals, and Jeffrey Dean. Distilling the knowledge in a neural network. In NIPS Deep Learning and Representation Learning Workshop, 2015. [11] Yoon Kim and Alexander M. Rush. Sequence-level knowledge distillation. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1317 1327, 2016. [12] Yoon Kim, Yacine Jernite, David Sontag, and Alexander M. Rush. Character-aware neural language models. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI), pages 2741 2749, 2016. [13] Diederik P. Kingma and Jimmy Ba. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations (ICLR), 2015. [14] Taku Kudo. Subword regularization: Improving neural network translation models with multiple subword candidates. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (ACL), pages 66 75, 2018. [15] Yann Le Cun, John S. Denker, and Sara A. Solla. Optimal brain damage. In D. S. Touretzky, editor, Advances in Neural Information Processing Systems 2 (NIPS), pages 598 605. 1990. [16] Sachin Mehta, Rik Koncel-Kedziorski, Mohammad Rastegari, and Hannaneh Hajishirzi. De FINE: Deep factorized input token embeddings for neural sequence modeling. In Proceedings of the 8th International Conference on Learning Representations (ICLR), 2020. [17] Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. Distributed representations of words and phrases and their compositionality. In Advances in Neural Information Processing Systems 26 (NIPS), pages 3111 3119. 2013. [18] Myle Ott, Sergey Edunov, David Grangier, and Michael Auli. Scaling neural machine translation. In Proceedings of the Third Conference on Machine Translation (WMT), pages 1 9, 2018. [19] Myle Ott, Sergey Edunov, Alexei Baevski, Angela Fan, Sam Gross, Nathan Ng, David Grangier, and Michael Auli. fairseq: A fast, extensible toolkit for sequence modeling. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pages 48 53, 2019. [20] Paul Over, Hoa Dang, and Donna Harman. Duc in context. Information Processing & Management, 43(6):1506 1520, 2007. [21] Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32 (NIPS), pages 8024 8035. 2019. [22] Jeffrey Pennington, Richard Socher, and Christopher D. Manning. Glove: Global vectors for word representation. In Empirical Methods in Natural Language Processing (EMNLP), pages 1532 1543, 2014. [23] Matthew Peters, Mark Neumann, Mohit Iyyer, Matt Gardner, Christopher Clark, Kenton Lee, and Luke Zettlemoyer. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pages 2227 2237, 2018. [24] Matt Post. A call for clarity in reporting BLEU scores. In Proceedings of the Third Conference on Machine Translation (WMT), pages 186 191, 2018. [25] Herbert Rubenstein and John B. Goodenough. Contextual correlates of synonymy. Communications of the ACM, 8(10):627 633, 1965. [26] Alexander M. Rush, Sumit Chopra, and Jason Weston. A Neural Attention Model for Abstractive Sentence Summarization. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 379 389, 2015. [27] Rico Sennrich, Barry Haddow, and Alexandra Birch. Neural machine translation of rare words with subword units. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (ACL), pages 1715 1725, 2016. [28] Raphael Shu and Hideki Nakayama. Compressing word embeddings via deep compositional code learning. In Proceedings of the 6th International Conference on Learning Representations (ICLR), 2018. [29] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. Dropout: A simple way to prevent neural networks from overfitting. Journal of Machine Learning Research, 15:1929 1958, 2014. [30] Ilya Sutskever, Oriol Vinyals, and Quoc V. Le. Sequence to Sequence Learning with Neural Networks. In Advances in Neural Information Processing Systems 27 (NIPS), pages 3104 3112, 2014. [31] Jun Suzuki and Masaaki Nagata. Learning compact neural word embeddings by parameter space sharing. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI), pages 2046 2052, 2016. [32] Jun Suzuki and Masaaki Nagata. Cutting-off redundant repeating generations for neural abstractive summarization. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics (EACL), pages 291 297, 2017. [33] Dan Tito Svenstrup, Jonas Hansen, and Ole Winther. Hash embeddings for efficient word representations. In Advances in Neural Information Processing Systems 30 (NIPS), pages 4928 4936. 2017. [34] Sho Takase and Naoaki Okazaki. Positional encoding to control output sequence length. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT), pages 3999 4004, 2019. [35] Sho Takase, Jun Suzuki, and Masaaki Nagata. Character n-gram embeddings to improve rnn language models. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence (AAAI), pages 5074 5082, 2019. [36] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in Neural Information Processing Systems 30 (NIPS), pages 5998 6008. 2017. [37] John Wieting, Mohit Bansal, Kevin Gimpel, and Karen Livescu. Charagram: Embedding words and sentences via character n-grams. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1504 1515, 2016. [38] Wojciech Zaremba, Ilya Sutskever, and Oriol Vinyals. Recurrent neural network regularization. Co RR, abs/1409.2329, 2014. [39] Xiaowei Zhang, Wei Chen, Feng Wang, Shuang Xu, and Bo Xu. Towards compact and fast neural machine translation using a combined method. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1475 1481, 2017.