# fpets_fully_parallel_endtoend_texttospeech_system__bc9676e5.pdf The Thirty-Fourth AAAI Conference on Artificial Intelligence (AAAI-20) FPETS : Fully Parallel End-to-End Text-to-Speech System Dabiao Ma, 1 Zhiba Su, 1 Wenxuan Wang, 2 Yuhao Lu 1 1Turing Robot Co.,Ltd. Beijing, China {madabiao, suzhiba, luyuhao}@uzoo.cn 2The Chinese University of Hong Kong, Shenzhen. Guangdong, China wenxuanwang1@link.cuhk.edu.cn End-to-end Text-to-speech (TTS) system can greatly improve the quality of synthesised speech. But it usually suffers form high time latency due to its auto-regressive structure. And the synthesised speech may also suffer from some error modes, e.g. repeated words, mispronunciations, and skipped words. In this paper, we propose a novel non-autoregressive, fully parallel end-to-end TTS system (FPETS). It utilizes a new alignment model and the recently proposed U-shape convolutional structure, UFANS. Different from RNN, UFANS can capture long term information in a fully parallel manner. Trainable position encoding and two-step training strategy are used for learning better alignments. Experimental results show FPETS utilizes the power of parallel computation and reaches a significant speed up of inference compared with state-of-the-art end-to-end TTS systems. More specifically, FPETS is 600X faster than Tacotron2, 50X faster than DCTTS and 10X faster than Deep Voice3. And FPETS can generates audios with equal or better quality and fewer errors comparing with other system. As far as we know, FPETS is the first end-to-end TTS system which is fully parallel. Introduction TTS systems aim to generate human-like speeches from texts. End-to-end TTS system is a type of system that can be trained on (text,audio) pairs without phoneme duration annotation(Wang et al. 2017). It usually contains 2 components, an acoustic model and a vocoder. Acoustic model predicts acoustic intermediate features from texts. And vocoder, e.g. Griffin-Lim (Griffin et al. 1984), WORLD (MORISE, YOKOMORI, and OZAWA 2016), Wave Net (van den Oord et al. 2016b), synthesizes speeches with generated acoustic features. The advantages of end-to-end TTS system are threefold: 1) reducing manual annotation cost and being able to utilize raw data, 2) preventing the error propagation between Copyright c 2020, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. 1Dabiao Ma, Zhiba Su and Wenxuan Wang have equal contributions. Yuhao Lu is the corresponding author. 2Codes and demos will be released at https://github.com/ suzhiba/Full-parallel 100x real time End2End TTS different components, 3) reducing the need of feature engineering. However, without the annotation of duration information, end-to-end TTS systems have to learn the alignment between text and audio frame. Most competitive end-to-end TTS systems have an encoder-decoder structure with attention mechanism, which is significantly helpful for alignment learning. Tacotron (Wang et al. 2017) uses an autoregressive attention (Bahdanau, Cho, and Bengio 2014) structure to predict alignment, and uses CNNs and GRU (Cho et al. 2014) as encoder and decoder, respectively. Tacotron2(Shen et al. 2018), which is a combination of the modified Tacotron system and Wave Net, also use an autoregressive attention. However, the autoregressive structure greatly limits the inference speed in the context of parallel computation. Deep voice 3 (Ping et al. 2018) replaces RNNs with CNNs to speed up training and inference. DCTTS (Tachibana, Uenoyama, and Aihara 2017) greatly speeds up the training of attention module by introducing guided attention. But Deep Voice 3 and DCTTS still have autoregressive structure. And those models also suffer from serious error modes e.g. repeated words, mispronunciations, or skipped words (Ping et al. 2018). Low time latency is required in real world application. Autoregressive structures, however, greatly limit the inference speed in the context of parallel computation. (Ping et al. 2018) claims that it is hard to learn alignment without a autoregressive structure. So the question is how to design a non-autoregressive structure that can perfectly determine alignment? In this paper, we propose a novel fully parallel endto-end TTS system (FPETS). Given input phonemes, our model can predict all acoustic frames simultaneously rather than autoregressively. Specifically, we follow the commonly used encoder-decoder structure with attention mechanism for alignment. But we replace autoregressive structures with a recent proposed U-shaped convolutional structure (UFANS)(Ma et al. 2018), which can be fully parallel and has stronger representation ability. Our fully parallel alignment structure inference alignment relationship between all phonemes and audio frames at once. Our novel trainable position encoding method can utilize position information better and two-step training strategy improves the alignment Figure 1: Model architecture. The light blue blocks are input/output flow. quality. Experimental results show FPETS utilizes the power of parallel computation and reaches a significant speed up of inference compared with state-of-the-art end-to-end TTS systems. More specifically, FPETS is 600X faster than Tacotron2, 50X faster than DCTTS and 10X faster than Deep Voice3. And FPETS can generates audios with equal or better quality and fewer errors comparing with other system. As far as we know, FPETS is the first end-to-end TTS system which is fully parallel. Model Architecture Most competitive end-to-end TTS systems have an encoderdecoder structure with attention mechanism(Wang et al. 2017) (Ping et al. 2018). Following this overall architecture, our model consists of three parts, shown in Fig.1. The encoder converts phonemes into hidden states that are sent to decoder; The alignment module determines the alignment width of each phoneme, from which the number of frames that attend on that phoneme can be induced; The decoder receives alignment information and converts the encoder hidden states into acoustic features. Encoder The encoder encodes phonemes into hidden states. It consists of 1 embedding layer , 1 dense layer, 3 convolutional layers, and a final dense layer. Some of TTS systems(Li et al. 2018) use self attention network as encoder. But we find that it dose not make significant difference, both in loss value and MOS. Alignment module determines the mapping from phonemes to acoustic features. We discard autoregressive structure, which is widely used in other alignment modules(Ping et al. 2018)(Wang et al. 2017)(Shen et al. 2018), for time latency issue. Our novel alignment module consists of 1 embedding layer, 1 UFANS (Ma et al. 2018) structure, trainable position encoding and several matrix multiplications, as depicted in Fig.1. Fully parallel UFANS structure UFANS is a modified version of U-Net for TTS task aiming to speed up inference. The structure is shown in Fig.2.In alignment structure, UFANS is used to predict alignment width, which is similar to phoneme duration.Those pooling and up-sampling operation along the spatial dimension make the receptive field increases exponentially and high way connection enables the combination of different scales features. For each phoneme i, we define the Alignment Width ri which represents its relationship with frame numbers. Suppose the number of phonemes in an utterance is N, and UFANS outputs a sequence of N scalars : [r0, r1, ..., r N 1]; Then we relate the alignment width ri to the acoustic frame index j. The intuition is that the acoustic frame with index j= i 1 k=0 rk + 1 2ri should be the one that attends most on i-th phoneme. And we need a structure that satisfies the intuition. Position Encoding Function Position encoding (Gehring et al. 2017) is a method to embed a sequence of absolute positions into a sequence of vectors. Sine and cosine position encoding has two very important properties that make it suitable for position encoding. In brief, function g(x) = f cos( x s f ) has a heavy tail that enables one acoustic frame to receive phoneme information very far away; The gradient function | g(s)| = | f )| is insensitive to the term x s. We give a more detailed illustration in Appendix. Trainable Position Encoding Some end-to-end TTS system, like deep voice 3 and Tacotron2, use sine and cosine functions of different frequencies and add those position encoding vectors to input embedding. But they both take position encoding as a supplement to help the training of attention module and the position encoding vectors remain constant. We propose a trainable position encoding, which is better than absolute position encoding in getting position information. We define the absolute alignment position si of i-th phoneme as : 2ri, i = 0, ..., Tp 1, r 1 = 0 (1) Figure 2: UFANS Model Architecture Now choose L float numbers log uniformly from range [1.0, 10000.0] and get a sequence of frequencies [f0, ..., f L 1]. For i-th phoneme, the position encoding vector vpi of this phoneme is defined as : vpi = [vpi,sin, vpi,cos], [vpi,sin]k = sin( si [vpi,cos]k = cos( si fk ), k = 0, ..., L 1 Concatenating vpi, i = 0, ..., Tp 1 together, we get a matrix P that represents position information of all the phonemes, denoted as Key , see Fig.1 : P = [vp T 0 , ..., vp T Tp 1] (3) And similarly, for the j-th frame of the acoustic feature, the position encoding vector vaj is defined as : vaj = [vaj,sin, vaj,cos], [vaj,sin]k = sin( j [vai,cos]k = cos( j fk ), k = 0, ..., L 1 Concatenating all the vectors, we get the matrix F that represents position information of all the acoustic frames, denoted as Query , see Fig.1: F = [va T 0 , ..., va T Ta 1] (5) And now define the attention matrix A as : A = FP T , Aji = vpiva T j , i = 0, ..., Tp 1, j = 0, ..., Ta 1 (6) That is, the attention of j-th frame on i-th phoneme is proportional to the inner product of their encoding vectors. This inner product can be rewritten as : vpiva T j = f ) + sin(si It is clear when j = si, the j-th frame is the one that attends most on i-th phoneme. The normalized attention matrix ˆA is : ˆA, ˆAji = Aji Now ˆAji represents how much j-th frame attends on i-th phoneme. Then we use argmax to build new attention matrix A: 1 if i = argmax k [0,...,Tp 1] Ajk 0 otherwise (9) Now define the number of frames that attend more on i-th phoneme than any other phoneme to be its attention width wi. From the definition of attention width, A is actually a matrix representing attention width wi. The alignment width ri and wi are different but related. For two adjacent absolute alignment positions si and si+1, consider the two functions: f ), g2(x) = f cos(x si+1 The values of the two functions only depend on the relative position of x to si and si+1. It is known function g1 decreases when x moves away from si (locally, but it is sufficient here). So we have: Figure 3: UFANS Decoder g1(x) > g2(x) when x [si, 1 2(si + si+1)) g1(x) < g2(x) when x ( 1 2(si + si+1), si+1] Thus x = 1 2(si + si+1) is the right attention boundary of phoneme i, similarly the left attention boundary is x = 1 2(si 1 + si). It can be deduced that : 2(si + si+1) 1 2(si 1 + si) (10) 4(ri 1 + ri+1 + 2ri) (11) i = 0, ..., Tp 1, r 1 = r0, r Tp = r TP 1 (12) which means attention width and alignment width can be linearly transformed to each other. And it is further deduced that : Tp 1 k=0 wk = Ta (13) UFANS Decoder The decoder receives alignment information and converts the encoded phonemes information to acoustic features, see Figure 3. Relative position is the distance between the phoneme and previous phoneme. Our model use it to enhance position relationship. Following (Ma et al. 2018), we use UFANS as our decoder. The huge receptive field enables to capture long-time information dependency and the highway skip connection structure enables the combination of different level of features. It generates good quality acoustic features in a fully parallel manner. Training Strategy We use Acoustic Loss, denoted as LOSSacou, to evaluate the quality of generated acoustic features, which is L2 norm between predicted acoustic features and ground truth features. In order to train a better alignment model, we propose a two-stage training strategy. Our model focus more on alignment learning in stage 1. In stage 2 we fix the alignment module and train the whole system. Stage 1 :Alignment Learning In order to enhance the quality of alignment learning, we use convolutional decoder and design an alignment loss. Convolutional Decoder: UFANS has stronger representation ability than vanilla CNN. But the learning of alignment will be greatly disturbed if using UFANS as decoder. The experimental evidences and analysis are shown in the next section. So we replace UFANS decoder with a convolutional decoder. The convolutional decoder consists of several convolution layers with gated activation (van den Oord et al. 2016a), several Dropout (Srivastava et al. 2014) operations and one dense layer. Alignment Loss: We define an Alignment Loss, denoted as LOSSalign, based on the fact that the summation of alignment width should be equal or close to the frame length of acoustic features. We relax this restriction by using a threshold γ : LOSSalign = γ, if | Tp 1 k=0 rk Ta| < γ | Tp 1 k=0 rk Ta|, otherwise (14) The final loss LOSS is a weighted sum of LOSSacou and LOSSalign : LOSS = LOSSacou + σLOSSalign (15) We choose 0.02 as alignment loss weight based on grid search from 0.005 to 0.3. Stage 2 : Overall Training In stage 2, we fix the welltrained alignment module and use UFANS as decoder to train the overall end-to-end system. Only Acoustic Loss is used as objective function in this stage. Experiments and Results Dataset LJ speech(Ito 2017) is a public speech dataset consisting of 13100 pairs of text and 22050 HZ audio clips. The clips vary from 1 to 10 seconds and the total length is about 24 hours. Phoneme-based textual features are given. Two kinds of acoustic features are extracted. One is based on WORLD vocoder that uses mel-frequency cepstral coefficients(MFCCs). The other is linear-scale log magnitude spectrograms and mel-band spectrograms that can be feed into Griffin-Lim algorithm or a trained Wave Net vocoder. The WORLD vocoder uses 60 dimensional melfrequency cepstral coefficients, 2 dimensional band aperiodicity, 1 dimensional logarithmic fundamental frequency, Table 1: Hyper-Parameter Structure value Encoder/DNN Layers 1 Encoder/CNN Layers 3 Encoder/CNN Kernel 3 Encoder/CNN Filter Size 1024 Encoder/Final DNN Layers 1 Alignment/UFANS layers 4 Alignment/UFANS hidden 512 Alignment/UFANS Kernel 3 Alignment/UFANS Filter Size 1024 CNN Decoder/CNN Layers 3 CNN Decoder/CNN Kernel 3 CNN Decoder/CNN Filter Size 1024 UFANS Decoder/UFANS layers 6 UFANS Decoder/UFANS hidden 512 UFANS Decoder/UFANS Kernel 3 UFANS Decoder/UFANS Filter Size 1024 Droupout 0.15 their delta, delta-delta dynamic features and 1 dimensional voice/unvoiced feature. It is 190 dimensions in total. The WORLD vocoder based feature uses FFT window size 2048 and has a frame time 5 ms. The spectrograms are obtained with FFT size 2048 and hop size 275. The dimensions of linear-scale log magnitude spectrograms and mel-band spectrograms are 1025 and 80. Implementation Details Hyperparameters of our model are showed in Table 1. Tacotron2, DCTTS and Deep Voice3 are used as baseline . The model configurations are shown in Appendix. Adam are used as optimizer with β1 = 0.9, β2 = 0.98, ϵ = 1e 4. Each model is trained 300k steps. All the experiments are done on 4 GTX 1080Ti GPUs, with batch size of 32 sentences on each GPU. Main Results We aim to design a TTS system that can synthesis speech quickly, high quality and with fewer errors. So we compare our FPUTS with baseline on inference speed, MOS and error modes. Inference Speed The inference speed evaluates time latency of synthesizing a one-second speech, which includes data transfer from main memory to GPU global memory, GPU calculations and data transfer back to main memory. As is shown in Table 2, our FPETS model is able to greatly take advantage of parallel computations and is significantly faster than other systems. MOS Harvard Sentences List 1 and List 2 are used to evaluate the mean opinion score (MOS) of a system. The synthesized audios are evaluated on Amazon Mechanical Turk using crowd MOS method (Protasio Ribeiro et al. 2011). The score ranges from 1 (Bad) to 5 (Excellent). As is shown in Table 2: Inference speed comparison Method Autoregressive Inference speed (ms) Tacotron2 Yes 6157.3 DCTTS Yes 494.3 Deep Voice 3 Yes 105.4 FPETS No 9.9 Table 3: MOS results comparison Method Vocoder MOS Tacotron2 Griffin Lim 3.51 0.070 DCTTS Griffin Lim 3.55 0.107 Deep Voice 3 Griffin Lim 2.79 0.096 FPETS Griffin Lim 3.65 0.082 Tacotron2 Wave Net 3.04 0.103 DCTTS Wave Net 3.43 0.109 FPETS Wave Net 3.27 0.108 FPETS WORLD 3.81 0.122 Table 3, Our FPETS is no worse than other end-to-end system. The MOS of Wave Net-based audios are lower than expected since background noise exists in these audios. Robustness Analysis Attention-based neural TTS systems may run into several error modes that can reduce synthesis quality. For example, repetition means repeated pronunciation of one or more phonemes, mispronunciation means wrong pronunciation of one or more phonemes and skip word means one or more phonemes are skipped. In order to track the occurrence of attention errors, 100 sentences are randomly selected from Los Angeles Times, Washington Post and some fairy tales. As is shown in Table 4, Our FPETS system is more robust than other systems. Alignment Learning Analysis Alignment learning is essential for end-to-end TTS system which greatly affects the quality of generated audios. So we further discuss the factors that can affect the alignment quality. 100 audios are randomly selected from training data, denoted as origin audios. Their utterances are fed to our system to generate audios, denoted as re-synthesized audios. The method to evaluate the alignment quality is objectively computing the difference of the phoneme duration between origin audios and their corresponding re-synthesized audios. The phoneme durations are obtained by hand. Figure 4 is the labeled phonemes of audio LJ048-0033 . Here only results with mel-band spectrograms using Griffin-Lim algorithm are shown. For MFCCs, results are similar. We compare alignment quality between different alignment model configurations. Table 6 shows the overall results on 100 audios. Table 5 is a case study which shows how phoneme-level duration is affected by different model. Position Encoding Function and Alignment Quality We replace the Sine and Cosine position encoding function with Gaussian function. As Table 6 shows, the experimental results show that the model can not learn correct alignment Table 4: Robustness Comparison Repeats Mispronunciation Skip Tacotron2 2 5 4 DCTTS 2 10 1 Deep Voice 3 1 5 3 FPETS 1 2 1 Figure 4: The upper is real audio of LJ048-0033 , the lower is the re-synthesized audio from alignment learning model. text : prior to November twenty two nineteen sixty three phoneme : P R AY ER T UW N OW V EH M B ER T W EH N T IY T UW N AY N T IY N S IH K S T IY TH R IY with Gaussian function. We give a theoretical analysis in Appendix. Trainable Position Encoding and Alignment Quality We replace the trainable position encoding with a fixed position encoding. The experimental results show that the model can learn better alignment with trainable position encoding. Decoder and Alignment Quality In order to identify the relationship between decoder and alignment quality in stage 1, we replace simple convolutional decoder by UFANS with 6 down-sampling layers. Experiments show the computed attention width is much worse than that with the simple convolutional decoder. And the synthesized audios also suffer from error modes like repeated words and skipped words. The results show the simple decoder may be better in alignment learning stage. More details are shown in Table 6. With UFANS decoder, our model can get comparable loss no matter that the alignment is accurate or not. Therefore, alignment isn t well trained with UFANS decoder. Human is sensitive to phoneme speed, so speech will be terrible if duration in inaccurate. To solve the problem, we train the alignment with simple CNNs, then fix the alignment structure. With the fixed alignment and UFANS decoder, our model can generate high quality audio in a parallel way. Related Works Fast Speech (Ren et al. 2019),which is proposed in same period, can also generate acoustic features in a parallel way. Specifically, it extract attention alignments from an auto-regressive encoder-decoder based teacher model for phoneme duration prediction, which is used by a length regulator to expand the source phoneme sequence to match the length of the target mel-spectrogram sequence for parallel mel-spectrogram generation. Using the phoneme duration extracted from an teacher model is a creative work to solve Figure 5: Attention plot of text : This is the destination for all things related to development at stack overflow. Phoneme : DH IH S IH Z DH AH D EH S T AH N EY SH AH N F AO R AO L TH IH NG Z R IH L EY T IH D T UW D IH V EH L AH P M AH N T AE T S T AE K OW V ER F L OW . the problem that model can t inference in a parallel way. However, it s speed is still not fast enough to satisfy industrial application, especially it can t speed up when batch size is increased. Our FPETS has lower time latency and faster than Fast Speech. On average FPETS generates 10ms per sentence under GTX 1080ti GPU and Fasts Speech is 170ms per sentence under Tesla V100 GPU, which is known faster than GTX 1080ti. And FPETS can also automatically specify the phoneme duration by trainable position encoding. In this paper, a new non-autoregressive, fully parallel endto-end TTS system, FPETS, is proposed. Given input phonemes, FPETS can predict all acoustic frames simultaneously rather than autoregressively. Specifically FPETS utilize a recent proposed U-shaped convolutional structure, which can be fully parallel and has stronger representation ability. The fully parallel alignment structure inference alignment relationship between all phonemes and audio frames at once. The novel trainable position encoding method can utilize position information better and two-step training strategy improves the alignment quality. FPETS can utilize the power of parallel computation and reach a significant speed up of inference compared with state-of-the-art end-to-end TTS systems. More specifically, FPETS is 600X faster than Tacotron2, 50X faster than DCTTS and 10X faster than Deep Voice3. And FPETS can generates audios with equal or better quality and fewer errors comparing with other system. As far as we know, FPETS is the first end-to-end TTS system which is fully parallel. Table 5: A case study about phoneme-level comparison of alignment quality. Real duration and the predicted duration by our alignment method, using Gaussian as position encoding function, using fixed position encoding, using UFANS as decoder are shown. P R AY ER T UW N OW V EH M B ER T W EH real 5.35 7.28 15.48 13.43 4.96 3.44 3.36 5.44 4.72 7.20 4.56 1.92 7.12 5.36 3.36 3.84 resynth 3.55 7.97 13.28 11.37 4.88 4.00 6.19 5.27 5.46 6.39 3.56 2.08 6.13 5.69 4.34 3.03 resynth-Gauss 6.31 6.03 5.78 6.11 6.59 6.73 6.74 6.76 6.75 6.75 6.77 6.80 6.84 6.82 6.79 6.78 resynth-fixenc 7.41 7.35 11.40 10.46 4.04 4.60 2.95 6.41 4.30 7.86 5.26 2.45 9.21 6.77 3.90 2.85 resynth-UFANS 4.08 8.09 9.41 8.45 6.90 5.70 5.21 5.71 5.98 5.29 4.87 5.20 5.43 5.14 5.10 5.37 IY T UW N AY N T IY N S IH K S T IY TH real 10.80 9.76 9.76 6.80 6.08 6.16 7.28 5.28 5.36 6.56 6.16 4.08 3.52 6.32 9.36 9.76 resynth 10.89 11.26 9.69 7.72 5.33 6.55 7.30 5.90 5.81 5.43 5.11 4.33 3.57 6.81 10.57 11.54 resynth-Gauss 6.78 6.79 6.77 6.75 6.76 6.74 6.72 6.74 6.76 6.80 6.84 6.82 6.79 6.78 6.77 6.81 resynth-fixenc 12.05 9.21 8.26 5.76 6.90 7.63 6.47 3.20 4.74 4.11 5.85 2.97 4.01 5.29 11.26 10.60 resynth-UFANS 7.14 7.38 8.96 8.20 5.39 5.54 7.31 6.34 5.56 6.42 5.84 4.76 5.62 7.61 8.26 8.17 Table 6: Comparison of alignment quality between different configurations. Sine-Cosine or Gaussian encoding function, trainable or fixed position encoding and CNN or UFANS decoder are evaluated based on their average difference between their duration prediction and real duration length on 100 audios. Encoding func Trainable? Decoder Average-diff Gaussian Trainable CNN 2.58 Sin-Cos Fixed CNN 1.96 Sin-Cos Trainable UFANS 1.80 Sin-Cos Trainable CNN 0.85 References Bahdanau, D.; Cho, K.; and Bengio, Y. 2014. Neural machine translation by jointly learning to align and translate. ar Xiv e-prints abs/1409.0473. Cho, K.; van Merri enboer, B.; G ulc ehre, C .; Bahdanau, D.; Bougares, F.; Schwenk, H.; and Bengio, Y. 2014. Learning phrase representations using rnn encoder decoder for statistical machine translation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), 1724 1734. Doha, Qatar: Association for Computational Linguistics. Gehring, J.; Auli, M.; Grangier, D.; Yarats, D.; and Dauphin, Y. 2017. Convolutional sequence to sequence learning. In ICML. Griffin, D. W.; Jae; Lim, S.; and Member, S. 1984. Signal estimation from modified short-time fourier transform. IEEE Trans. Acoustics, Speech and Sig. Proc 236 243. Ito, K. 2017. The lj speech dataset. https://keithito.com/ LJ-Speech-Dataset/. Li, N.; Liu, S.; Liu, Y.; Zhao, S.; Liu, M.; and Zhou, M. 2018. Close to human quality tts with transformer. Ar Xiv abs/1809.08895. Ma, D.; Su, Z.; Lu, Y.; Wang, W.; and Li, Z. 2018. Ufans: Ushaped fully-parallel acoustic neural structure for statistical parametric speech synthesis with 20x faster. ar Xiv preprint ar Xiv:1811.12208. MORISE, M.; YOKOMORI, F.; and OZAWA, K. 2016. World: A vocoder-based high-quality speech synthesis sys- tem for real-time applications. IEICE Transactions on Information and Systems E99.D(7):1877 1884. Ping, W.; Peng, K.; Gibiansky, A.; Arik, S. O.; Kannan, A.; Narang, S.; Raiman, J.; and Miller, J. 2018. Deep voice 3: 2000-speaker neural text-to-speech. In International Conference on Learning Representations. Protasio Ribeiro, F.; Florencio, D.; Zhang, C.; and Seltzer, M. 2011. Crowdmos: An approach for crowdsourcing mean opinion score studies. In ICASSP. IEEE. Ren, Y.; Ruan, Y.; Tan, X.; Qin, T.; Zhao, S.; Zhao, Z.; and Liu, T. 2019. Fastspeech: Fast, robust and controllable text to speech. Co RR abs/1905.09263. Shen, J.; Pang, R.; Weiss, R. J.; Schuster, M.; Jaitly, N.; Yang, Z.; Chen, Z.; Zhang, Y.; Wang, Y.; Skerrv-Ryan, R.; et al. 2018. Natural tts synthesis by conditioning wavenet on mel spectrogram predictions. In 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 4779 4783. IEEE. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; and Salakhutdinov, R. 2014. Dropout: A simple way to prevent neural networks from overfitting. Journal of Machine Learning Research 15:1929 1958. Tachibana, H.; Uenoyama, K.; and Aihara, S. 2017. Efficiently trainable text-to-speech system based on deep convolutional networks with guided attention. Co RR abs/1710.08969. van den Oord, A.; Kalchbrenner, N.; Espeholt, L.; kavukcuoglu, k.; Vinyals, O.; and Graves, A. 2016a. Conditional image generation with pixelcnn decoders. In Lee, D. D.; Sugiyama, M.; Luxburg, U. V.; Guyon, I.; and Garnett, R., eds., Advances in Neural Information Processing Systems 29. Curran Associates, Inc. 4790 4798. van den Oord, A.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A. W.; and Kavukcuoglu, K. 2016b. Wavenet: A generative model for raw audio. Co RR abs/1609.03499. Wang, Y.; Skerry-Ryan, R. J.; Stanton, D.; Wu, Y.; Weiss, R. J.; Jaitly, N.; Yang, Z.; Xiao, Y.; Chen, Z.; Bengio, S.; Le, Q. V.; Agiomyrgiannakis, Y.; Clark, R.; and Saurous, R. A. 2017. Tacotron: A fully end-to-end text-to-speech synthesis model. Co RR abs/1703.10135.