# retrospective_reader_for_machine_reading_comprehension__728fe553.pdf Retrospective Reader for Machine Reading Comprehension Zhuosheng Zhang1,2,3, Junjie Yang2,3,4, Hai Zhao1,2,3, 1Department of Computer Science and Engineering, Shanghai Jiao Tong University 2Key Laboratory of Shanghai Education Commission for Intelligent Interaction and Cognitive Engineering, Shanghai Jiao Tong University, Shanghai, China 3Mo E Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, Shanghai, China 4SJTU-Paris Tech Elite Institute of Technology, Shanghai Jiao Tong University, Shanghai, China zhangzs@sjtu.edu.cn, jj-yang@sjtu.edu.cn, zhaohai@cs.sjtu.edu.cn Machine reading comprehension (MRC) is an AI challenge that requires machines to determine the correct answers to questions based on a given passage. MRC systems must not only answer questions when necessary but also tactfully abstain from answering when no answer is available according to the given passage. When unanswerable questions are involved in the MRC task, an essential verification module called verifier is especially required in addition to the encoder, though the latest practice on MRC modeling still mostly benefits from adopting well pre-trained language models as the encoder block by only focusing on the reading . This paper devotes itself to exploring better verifier design for the MRC task with unanswerable questions. Inspired by how humans solve reading comprehension questions, we proposed a retrospective reader (Retro-Reader) that integrates two stages of reading and verification strategies: 1) sketchy reading that briefly investigates the overall interactions of passage and question, and yields an initial judgment; 2) intensive reading that verifies the answer and gives the final prediction. The proposed reader is evaluated on two benchmark MRC challenge datasets SQu AD2.0 and News QA, achieving new state-of-the-art results. Significance tests show that our model is significantly better than strong baselines. 1 Introduction Be certain of what you know and be aware what you don t. That is wisdom. Confucius (551 BC - 479 BC) Machine reading comprehension (MRC) aims to teach machines to answer questions after comprehending given passages (Hermann et al. 2015; Joshi et al. 2017; Rajpurkar, Jia, and Liang 2018), which is a fundamental and longstanding goal of natural language understanding (NLU) (Zhang, Zhao, and Wang 2020). It has significant application scenarios such as question answering and dialog sys- Corresponding author. This paper was partially supported by National Key Research and Development Program of China (No. 2017YFB0304100), Key Projects of National Natural Science Foundation of China (U1836222 and 61733011), Huawei-SJTU long term AI project, Cutting-edge Machine reading comprehension and language model. Copyright c 2021, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Passage: Computational complexity theory is a branch of the theory of computation in theoretical computer science that focuses on classifying computational problems according to their inherent difficulty, and relating those classes to each other. A computational problem is understood to be a task that is in principle amenable to being solved by a computer, which is equivalent to stating that the problem may be solved by mechanical application of mathematical steps, such as an algorithm. Question: What cannot be solved by mechanical application of mathematical steps? Gold Answer: no answer Plausible answer: algorithm Table 1: An unanswerable MRC example. tems (Zhang et al. 2018; Choi et al. 2018; Reddy, Chen, and Manning 2019; Zhang, Huang, and Zhao 2018; Xu, Zhao, and Zhang 2021; Zhu et al. 2018). The early MRC systems (Kadlec et al. 2016; Chen, Bolton, and Manning 2016; Dhingra et al. 2017; Wang et al. 2017; Seo et al. 2016) were designed on a latent hypothesis that all questions can be answered according to the given passage (Figure 1-[a]), which is not always true for real-world cases. The recent progress on the MRC task has required that the model must be capable of distinguishing those unanswerable questions to avoid giving plausible answers (Rajpurkar, Jia, and Liang 2018). MRC task with unanswerable questions may be usually decomposed into two subtasks: 1) answerability verification and 2) reading comprehension. To determine unanswerable questions requires a deep understanding of the given text and requires more robust MRC models, making MRC much closer to real-world applications. Table 1 shows an unanswerable example from SQu AD2.0 MRC task (Rajpurkar, Jia, and Liang 2018). So far, a common reading system (reader) which solves MRC problem generally consists of two modules or building steps as shown in Figure 1-[a]: 1) building a robust language model (LM) as encoder; 2) designing ingenious mechanisms as decoder according to MRC task characteristics. The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21) Answer Prediction scorehas scorenull scoreext Sketchy Reading Module (External Front Verification, E-FV) Intensive Reading Module (Internal Front Vefification, I-FV) Input Decoder (Rear Verification) Encoding Interaction Encoding Interaction Part I: Model Desgins Part II: Retrospective Reader Architecture Sketchy Reading Intensive Reading [e] (Encoder+FV)+FV-(Decoder+RV) Decoder (R-V) Verifier Decoder [c] Encoder-(Decoder+I-FV) [b] (Encoder+E-FV)-Decoder [a] Encoder+Decoder Encoder Decoder [d] Sketchy and Internsive Reading Figure 1: Reader overview. For the left part, models [a-c] summarize the instances in previous work, and model [d] is ours, with the implemented version [e]. In the names of models [a-e], ( ) represents a module, + means the parallel module and - is the pipeline. The right part is the detailed architecture of our proposed Retro-Reader. Pre-trained language models (Pr LMs) such as BERT (Devlin et al. 2019) and XLNet (Yang et al. 2019) have achieved success on various natural language processing tasks, which broadly play the role of a powerful encoder (Zhang et al. 2019; Li et al. 2020; Zhou, Zhang, and Zhao 2019). However, it is quite time-consuming and resource-demanding to impart massive amounts of general knowledge from external corpora into a deep language model via pre-training. Recently, most MRC readers keep the primary focus on the encoder side, i.e., the deep Pr LMs (Devlin et al. 2019; Yang et al. 2019; Lan et al. 2020), as readers may simply and straightforwardly benefit from a strong enough encoder. Meanwhile, little attention is paid to the decoder side1 of MRC models (Hu et al. 2019; Back et al. 2020; Reddy et al. 2020), though it has been shown that better decoder or better manner of using encoder still has a significant impact on MRC performance, no matter how strong the encoder it is (Zhang et al. 2020a; Liu et al. 2021; Li et al. 2019, 2018; Zhu, Zhao, and Li 2020). For the concerned MRC challenge with unanswerable questions, a reader has to handle two aspects carefully: 1) give the accurate answers for answerable questions; 2) effectively distinguish the unanswerable questions, and then refuse to answer. Such requirements lead to the recent reader s design by introducing an extra verifier module or answer-verification mechanism. Most readers simply stack the verifier along with encoder and decoder parts in 1We define decoder here as the task-specific part in an MRC system, such as passage and question interaction and answer verification. a pipeline or concatenation way (Figure 1-[b-c]), which is shown suboptimal for installing a verifier. As a natural practice of how humans solve complex reading comprehension (Zheng et al. 2019; Guthrie and Mosenthal 1987), the first step is to read through the full passage along with the question and grasp the general idea; then, people re-read the full text and verify the answer if not so sure. Inspired by such a reading and comprehension pattern, we proposed a retrospective reader (Retro-Reader, Figure 1- [d]) that integrates two stages of reading and verification strategies: 1) sketchy reading that briefly touches the relationship of passage and question, and yields an initial judgment; 2) intensive reading that verifies the answer and gives the final prediction. Our major contributions are three folds:2 1. We propose a new retrospective reader design which is capable of effectively performing answer verification instead of simply stacking verifier in existing readers. 2. Experiments show that our reader can yield substantial improvements over strong baselines and achieve new state-of-the-art results on benchmark MRC tasks. 3. For the first time, we apply the significance test for the concerned MRC task and show that our models are significantly better than the baselines. 2 Related Work The research of machine reading comprehension have attracted great interest with the release of a variety of bench- 2Our source code is available at https://github.com/cooelf/ Awesome MRC. mark datasets (Hill et al. 2015; Hermann et al. 2015; Rajpurkar et al. 2016; Joshi et al. 2017; Rajpurkar, Jia, and Liang 2018; Lai et al. 2017). The early trend is a variety of attention-based interactions between passage and question, including Attention Sum (Kadlec et al. 2016), Gated attention (Dhingra et al. 2017), Self-matching (Wang et al. 2017), Attention over Attention (Cui et al. 2017) and Biattention (Seo et al. 2016). Recently, Pr LMs dominate the encoder design for MRC and achieve great success. These Pr LMs include ELMo (Peters et al. 2018), GPT (Radford et al. 2018), BERT (Devlin et al. 2019), XLNet (Yang et al. 2019), Ro BERTa (Liu et al. 2019), ALBERT (Lan et al. 2020), and ELECTRA (Clark et al. 2020). They show strong capacity for capturing the contextualized sentence-level language representations and greatly boost the benchmark performance of current MRC. Following this line, we take Pr LMs as our backbone encoder. In the meantime, the study of the decoder mechanisms has come to a bottleneck due to the already powerful Pr LM encoder. Thus this work focuses on the non-encoder part, such as passage and question attention interactions, and especially the answer verification. To solve the MRC task with unanswerable questions is though important, only a few studies paid attention to this topic with straightforward solutions. Mostly, a treatment is to adopt an extra answer verification layer, the answer span prediction and answer verification are trained jointly with multi-task learning (Figure 1-[c]). Such an implemented verification mechanism can also be as simple as an answerable threshold setting broadly used by powerful enough Pr LMs for quickly building readers (Devlin et al. 2019; Zhang et al. 2020b). Liu et al. (2018) appended an empty word token to the context and added a simple classification layer to the reader. Hu et al. (2019) used two types of auxiliary loss, independent span loss to predict plausible answers and independent no-answer loss the to decide answerability of the question. Further, an extra verifier is adopted to decide whether the predicted answer is entailed by the input snippets (Figure 1-[b]). Back et al. (2020) developed an attention-based satisfaction score to compare question embeddings with the candidate answer embeddings (Figure 1- [c]). Zhang et al. (2020c) proposed a verifier layer, which is a linear layer applied to context embedding weighted by start and end distribution over the context words representations concatenated to [CLS] token representation for BERT (Figure 1-[c]). Different from these existing studies which stack the verifier module in a simple way or just jointly learn answer location and non-answer losses, our Retro-Reader adopts a two-stage humanoid design (Zheng et al. 2019; Guthrie and Mosenthal 1987) based on a comprehensive survey over existing answer verification solutions. 3 Our Proposed Model We focus on the span-based MRC task, which can be described as a triplet P, Q, A , where P is a passage, and Q is a query over P, in which a span is a right answer A. Our system is supposed to not only predict the start and end po- sitions in the passage P and extract span as answer A but also return a null string when the question is unanswerable. Our retrospective reader is composed of two parallel modules: a sketchy reading module and an intensive reading module to conduct a two-stage reading process. The intuition behind the design is that the sketchy reading makes a coarse judgment (external front verification) about the answerability, whether the question is answerable; and then the intensive reading jointly predicts the candidate answers and combines its answerability confidence (internal front verification) with the sketchy judgment score to yield the final answer (rear verification). 3 3.1 Sketchy Reading Module Embedding We concatenate question and passage texts as the input, which is firstly represented as embedding vectors to feed an encoder (i.e., a Pr LM). In detail, the input texts are first tokenized to word pieces (subword tokens). Let T = {t1, . . . , tn} denote a sequence of subword tokens of length n. For each token, the input embedding is the sum of its token embedding, position embedding, and token-type embedding. Let X = {x1, . . . , xn} be the outputs of the encoder, which are embedding features of encoding sentence tokens of length n. The input embeddings are then fed to the interaction layer to obtain the contextual representations. Interaction Following Devlin et al. (2019), the encoded sequence X is processed to a multi-layer Transformer (Vaswani et al. 2017) for learning contextual representations. For the following part, we use H = {h1, . . . , hn} to denote the last-layer hidden states of the input sequence. External Front Verification After reading, the sketchy reader will make a preliminary judgment, whether the question is answerable given the context. We implement this reader as an external front verifier (E-FV) to identify unanswerable questions. The pooled first token (the special symbol, [CLS]) representation h1 H, as the overall representation of the sequence, is passed to a fully connection layer to get classification logits ˆyi composed of answerable (logitans) and unanswerable (logitna) elements. We use cross entropy as training objective: i=1 [yi log ˆyi + (1 yi) log(1 ˆyi)] (1) where ˆyi Soft Max(FFN(h1)) denotes the prediction and yi is the target indicating whethter the question is answerbale or not. N is the number of examples. We calculate the difference as the external verification score: scoreext = logitna logitans, which is used in the later rear verification as effective indication factor. 3Intuitively, our model is supposed to be designed as shown in Figure 1-[d]. In the implementation, we find that modeling the entire reading process into two parallel modules is both simple and practicable with basically the same performance, which results in a parallel reading module design at last as shown in Figure 1-[e]. 3.2 Intensive Reading Module The objective of the intensive reader is to verify the answerability, produce candidate answer spans, and then give the final answer prediction. It employs the same encoding and interaction procedure as the sketchy reader, to obtain the representation H. In previous studies (Devlin et al. 2019; Yang et al. 2019; Lan et al. 2020), H is directly fed to a linear layer to yield the prediction. Question-aware Matching Inspired by previous success of explicit attention matching between passage and question (Kadlec et al. 2016; Dhingra et al. 2017; Wang et al. 2017; Seo et al. 2016), we are interested in whether the advance still holds based on the strong Pr LMs. Here we investigate two alternative question-aware matching mechanisms as an extra layer. Note that this part is only used for ablation in Table 7 as a reference for interested readers. We do not use any extra matching part in our submission on test evaluations (e.g., in Tables 2-3) for the sake of simplicity as our major focus is the verification. To obtain the representation of each passage and question, we split the last-layer hidden state H into HQ and HP as the representations of the question and passage, according to its position information. Both of the sequences are padded to the maximum length in a minibatch. Then, we investigate two potential question-aware matching mechanisms, 1) Transformer-style multi-head cross attention (CA) and 2) traditional matching attention (MA). Cross Attention We feed the HQ and H to a revised one-layer multi-head attention layer inspired by Lu et al. (2019). Since the setting is Q = K = V in multi-head self attention, which are all derived from the input sequence, we replace the input to Q with H, and both of K and V with HQ to obtain the question-aware context representation H . Matching Attention Another alternative is to feed HQ and H to a traditional matching attention layer (Wang et al. 2017), by taking the question presentation HQ as the attention to the representation H: M = Soft Max(H(WHQ + b e)T), H = MHQ, (2) where W and b are learnable parameters. e is a all-ones vector and used to repeat the bias vector into the matrix. M denotes the weights assigned to the different hidden states in the concerned two sequences. H is the weighted sum of all the hidden states and it represents how the vectors in H can be aligned to each hidden state in HQ. Finally, the representation H is used for the later predictions. If we do not use the above matching like the original use in BERT models, then H = H for the following part. Span Prediction The aim of span-based MRC is to find a span in the passage as answer, thus we employ a linear layer with Soft Max operation and feed H as the input to obtain the start and end probabilities, s and e: s, e Soft Max(FFN(H )). (3) The training objective of answer span prediction is defined as cross entropy loss for the start and end predictions, i [log(ps ys i ) + log(pe ye i )] (4) where ys i and ye i are respectively ground-truth start and end positions of example i. N is the number of examples. Internal Front Verification We adopted an internal front verifier (I-FV) such that the intensive reader can identify unanswerable questions as well. In general, a verifier s function can be implemented as a cross-entropy loss (I-FVCE), binary cross-entropy loss (I-FV-BE), or regressionstyle mean square error loss (I-FV-MSE). The pooled representation h 1 H , is passed to a fully connected layer to get the classification logits or regression score. Let ˆyi Linear(h1) denote the prediction and yi is the answerability target, the three alternative loss functions are as defined as follows: (1) We use cross entropy as loss function for the classification verification: yi,k = Soft Max(FFN(h 1)), k=1 [yi,k log yi,k] , (5) where K means the number of classes (K = 2 in this work). N is the number of examples. (2) For binary cross entropy as loss function for the classification verification: yi = Sigmoid(FFN(h1)), i=1 [yi log yi + (1 yi) log(1 yi)] . (6) (3) For the regression verification, mean square error is adopted as its loss function: yi = FFN(h 1), (7) i=1 (yi yi)2. (8) During training, the joint loss function for FV is the weighted sum of the span loss and verification loss: L = α1Lspan + α2Lans, (9) where α1 and α2 are weights. Threshold-based Answerable Verification Following previous studies (Devlin et al. 2019; Yang et al. 2019; Liu et al. 2019; Lan et al. 2020), we adopt threshold based answerable verification (TAV), which is a heuristic strategy to decide whether a question is answerable according to the predicted answer start and end logits finally. Given the output start and end probabilities s and e, and the verification probability v, we calculate the has-answer score scorehas and the no-answer score scorenull: scorehas = max(sk + el), 1 < k l n, scorenull = s1 + e1, (10) We obtain a difference score between scorenull and the scorehas as the final no-answer score: scorediff = scorenull scorehas. An answerable threshold δ is set and determined according to the development set. The model predicts the answer span that gives the has-answer score if the final score is above the threshold δ, and null string otherwise. TAV is used in all our models as the last step for the answerability decision. We denote it in our baselines with (+TAV) as default in Table 2-3, and omit the notation for simplicity in analysis part to avoid misunderstanding to keep on the specific ablations. 3.3 Rear Verification Rear verification (RV) is the combination of predicted probabilities of E-FV and I-FV, which is an aggregated verification for final answer. v = β1scorediff + β2scoreext, (11) where β1 and β2 are weights. Our model predicts the answer span if v > δ, and null string otherwise. 4 Experiments We use the available Pr LMs as encoder to build baseline MRC models: BERT (Devlin et al. 2019), ALBERT (Lan et al. 2020), and ELECTRA (Clark et al. 2020). Our implementations of BERT and ALBERT are based on the public Pytorch implementation from Transformers.4 ELECTRA is based on the Tensorflow release.5 We use the pre-trained LM weights in the encoder module of our reader, using all the official hyperparameters.6 For the fine-tuning in our tasks, we set the initial learning rate in {2e-5, 3e-5} with a warmup rate of 0.1, and L2 weight decay of 0.01. The batch size is selected in {32, 48}. The maximum number of epochs is set in 2 for all the experiments. Texts are tokenized using wordpieces (Wu et al. 2016), with a maximum length of 512. Hyper-parameters were selected using the dev set. The manual weights are α1 = α2 = β1 = β2 = 0.5 in this work. For answer verification, we follow the same setting according to the corresponding literatures (Devlin et al. 2019; Lan et al. 2020; Clark et al. 2020), which simply adopts the answerable threshold method described in 3.2. In the following part, +TAV (for all the baseline modes) denotes the baseline verification for easy reference, which is equivalent to the baseline implementations in public literatures (Devlin et al. 2019; Lan et al. 2020; Clark et al. 2020). 4https://github.com/huggingface/transformers. 5https://github.com/google-research/electra. 6BERTlarge; ALBERTxxlarge; ELECTRAlarge. Model Dev Test EM F1 EM F1 Human - - 86.8 89.5 BERT (Devlin et al. 2019) - - 82.1 84.8 Neur Qu RI (Back et al. 2020) 80.0 83.1 81.3 84.3 XLNet (Yang et al. 2019) 86.1 88.8 86.4 89.1 Ro BERTa (Liu et al. 2019) 86.5 89.4 86.8 89.8 SG-Net (Zhang et al. 2020c) - - 87.2 90.1 ALBERT (Lan et al. 2020) 87.4 90.2 88.1 90.9 ELECTRA (Clark et al. 2020) 88.0 90.6 88.7 91.4 Our implementation ALBERT (+TAV) 87.0 90.2 - - Retro-Reader on ALBERT 87.8 90.9 88.1 91.4 ELECTRA (+TAV) 88.0 90.6 - - Retro-Reader on ELECTRA 88.8 91.3 89.6 92.1 Table 2: The results (%) for SQu AD2.0 dataset. The results are from the official leaderboard. TAV: threshold-based answerable verification ( 3.2). 4.2 Benchmark Datasets Our proposed reader is evaluated in two benchmark MRC challenges. SQu AD2.0 As a widely used MRC benchmark dataset, SQu AD2.0 (Rajpurkar, Jia, and Liang 2018) combines the 100,000 questions in SQu AD1.1 (Rajpurkar et al. 2016) with over 50,000 new, unanswerable questions that are written adversarially by crowdworkers to look similar to answerable ones. The training dataset contains 87k answerable and 43k unanswerable questions. News QA News QA (Trischler et al. 2017) is a questionanswering dataset with 100,000 human-generated questionanswer pairs. The questions and answers are based on a set of over 10,000 news articles from CNN supplied by crowdworkers. The paragraphs are about 600 words on average, which tend to be longer than SQu AD2.0. The training dataset has 20k unanswerable questions among 97k questions. 4.3 Evaluation Metrics Two official metrics are used to evaluate the model performance: Exact Match (EM) and a softer metric F1 score, which measures the average overlap between the prediction and ground truth answer at the token level. Significance Test With the rapid development of deep MRC models, the dominant models have achieved very high results (e.g., over 90% F1 scores on SQu AD2.0), and further advance has been very marginal. Thus a significance test would be beneficial for measuring the difference in model performance. For selecting evaluation metrics for the significance test, since answers vary in length, using the F1 score would have Model Dev Test EM F1 EM F1 Existing Systems BARB (Trischler et al. 2017) 36.1 49.6 34.1 48.2 m LSTM (Wang and Jiang 2017) 34.4 49.6 34.9 50.0 Bi DAF (Seo et al. 2016) - - 37.1 52.3 R2-Bi LSTM (Weissenborn 2017) - - 43.7 56.7 AMANDA (Kundu and Ng 2018) 48.4 63.3 48.4 63.7 DECAPROP (Tay et al. 2018) 52.5 65.7 53.1 66.3 BERT (Devlin et al. 2019) - - 46.5 56.7 Neur Qu RI (Back et al. 2020) - - 48.2 59.5 Our implementation ALBERT (+TAV) 57.1 67.5 55.3 65.9 Retro-Reader on ALBERT 58.5 68.6 55.9 66.8 ELECTRA (+TAV) 56.3 66.5 54.0 64.5 Retro-Reader on ELECTRA 56.9 67.0 54.7 65.7 Table 3: Results (%) for News QA dataset. The results except ours are from Tay et al. (2018) and Back et al. (2020). TAV: threshold based answerable verification ( 3.2). a bias when comparing models, i.e., if one model fails on one severe example though works well on the others. Therefore, we use the tougher metric EM as the goodness measure. If the EM is equal to 1, the prediction is regarded as right and vice versa. Then the test is modeled as a binary classification problem to estimate the answer of the model is exactly right (EM=1) or wrong (EM=0) for each question. According to our task setting, we used Mc Nemar s test (Mc Nemar 1947) to test the statistical significance of our results. This test is designed for paired nominal observations, and it is appropriate for binary classification tasks (Ziser and Reichart 2017). The p-value is defined as the probability, under the null hypothesis, of obtaining a result equal to or more extreme than what was observed. The smaller the p-value, the higher the significance. A commonly used level of reliability of the result is 95%, written as p = 0.05. 4.4 Results Tables 2-3 compare the leading single models on SQu AD2.0 and News QA. Retro-Reader on ALBERT and Retro-Reader on ELECTRA denote our final models (i.e., our submissions to SQu AD2.0 online evaluation), which are respectively the ALBERT and ELECTRA based retrospective reader composed of both sketchy and intensive reading modules without question-aware matching for simplicity. According to the results, we make the following observations: 1) Our implemented ALBERT and ELECTRA baselines show the similar EM and F1 scores with the original numbers reported in the corresponding papers (Lan et al. 2020; Clark et al. 2020), ensuring that the proposed method can be fairly evaluated over the public strong baseline systems. 2) In terms of powerful enough Pr LMs like ALBERT and ELECTRA, our Retro-Reader not only significantly outper- Model All Has Ans No Ans EM F1 EM F1 EM F1 BERT 78.8 81.7 74.6 80.3 83.0 83.0 + E-FV 79.1 82.1 73.4 79.4 84.8 84.8 + I-FV-CE 78.6 82.0 73.3 79.5 84.5 84.5 + I-FV-BE 78.8 81.8 72.6 78.7 85.0 85.0 + I-FV-MSE 78.5 81.7 73.0 78.6 84.8 84.8 + RV 79.6 82.5 73.7 79.6 85.2 85.2 ALBERT 87.0 90.2 82.6 89.0 91.4 91.4 + E-FV 87.4 90.6 82.4 88.7 92.4 92.4 + I-FV-CE 87.2 90.3 81.7 87.9 92.7 92.7 + I-FV-BE 87.2 90.2 82.2 88.4 92.1 92.1 + I-FV-MSE 87.3 90.4 82.4 88.5 92.3 92.3 + RV 87.8 90.9 83.1 89.4 92.4 92.4 Table 4: Results (%) with different answer verification methods on the SQu AD2.0 dev set. CE, BE, and MSE are short for the two classification and one regression loss functions defined in 3.2. Method Prec. Rec. F1 Acc. ALBERT 91.70 93.42 92.55 86.14 Retro-Reader on ALBERT 94.30 92.38 93.33 87.49 ELECTRA 92.71 92.58 92.64 86.30 Retro-Reader on ELECTRA 93.27 93.51 93.39 87.60 Table 5: Performance on the unanswerable questions from SQu AD2.0 dev set. forms the baselines with p-value < 0.01,7 but also achieves new state-of-the-art on the SQu AD2.0 challenge.8 3) The results on News QA further verifies the general effectiveness of our proposed Retro-Reader. Our method shows consistent improvements over the baselines and achieves new state-of-the-art results. 5 Ablations 5.1 Evaluation on Answer Verification Table 4 presents the results with different answer verification methods. We observe that either of the front verifiers boosts the baselines, and integrating both as rear verification works the best. Note that we show the Has Ans and No Ans only for completeness. Since the final predictions are based on the threshold search of answerability scores ( 3.2), there exists a tradeoff between the Has Ans and No Ans accuracies. We see that the final RV that combines E-FV and I-FV shows 7Besides the Mc Nemar s test, we also used paired t-test for significance test, with consistent findings. 8When our models were submitted (Jan 10th 2020 and Apr 05, 2020 for ALBERTand ELECTRA-based models, respectively), our Retro-Reader achieved the first place on the SQu AD2.0 Leaderboard (https://rajpurkar.github.io/SQu AD-explorer/ ) for both single and ensemble models. Method EM F1 ALBERT 87.0 90.2 Two-model Ensemble 87.6 90.6 Retro-Reader 87.8 90.9 Table 6: Comparisons with Equivalent Parameters on the dev set of SQu AD2.0. Method SQu AD2.0 News QA EM F1 EM F1 BERT 78.8 81.7 51.8 62.5 + CA 78.8 81.7 52.1 62.7 + MA 78.3 81.4 52.4 62.6 ALBERT 87.0 90.2 57.1 67.5 + CA 87.3 90.3 56.0 66.3 + MA 86.8 90.0 55.8 66.1 Table 7: Results (%) with matching interaction methods on the dev sets of SQu AD2.0 and News QA. the best performance, which we select as our final implementation for testing. We further conduct the experiments on our model performance of the 5,945 unanswerable questions from the SQu AD 2.0 dev set. Results in Table 5 show that our method improves the performance on unanswerable questions by a large margin, especially in the primary F1 and accuracy metrics. 5.2 Comparisons with Equivalent Parameters When using sketchy reading module for external verification, we have two parallel modules that have independent parameters. For comparisons with equivalent parameters, we add an ensemble of two baseline models, to see if the advance is purely from the increase of parameters. Table 6 shows the results. We see that our model can still outperform two ensembled models. Although the two modules share the same design of the Transformer encoder, the training objectives (e.g., loss functions) are quite different, one for answer span prediction, the other for answerable decision. The results indicate that our two-stage reading modules would be more effective for learning diverse aspects (verification and span prediction) for solving MRC tasks with different training objectives. From the two modules, we can easily find the effectiveness of either the span prediction or answer verification, to improve the modules correspondingly. We believe this design would be quite useful for real-world applications. 5.3 Evaluation on Matching Interactions Table 7 shows the results with different interaction methods described in 3.2. We see that merely adding extra layers could not bring noticeable improvement, which indicates that simply adding more layers and parameters would not substantially benefit the model performance. The results verified the Pr LMs strong ability to capture the relationships Passage: Southern California consists of a heavily developed urban environment, home to some of the largest urban areas in the state, along with vast areas that have been left undeveloped. It is the third most populated megalopolis in the United States, after the Great Lakes Megalopolis and the Northeastern megalopolis. Much of southern California is famous for its large, spread-out, suburban communities and use of automobiles and highways. The dominant areas are Los Angeles, Orange County, San Diego, and Riverside-San Bernardino, each of which are the centers of their respective metropolitan areas... Question: What are the second and third most populated megalopolis after Southern California? Answer: Gold: no answer ALBERT (+TAV): Great Lakes Megalopolis and the Northeastern megalopolis. Retro-Reader over ALBERT: no answer scorehas = 0.03, scorena = 1.73, δ = 0.98 Table 8: Answer prediction examples from the ALBERT baseline and Retro-Reader. between passage and question after processing the paired input by deep self-attention layers. In contrast, answer verification could still give consistent and substantial advance. 5.4 Comparison of Predictions To have an intuitive observation of the predictions of Retro Reader, we give a prediction example on SQu AD2.0 from baseline and Retro-Reader in Table 8, which shows that our method works better at judging whether the question is answerable on a given passage and gets rid of the plausible answer. 6 Conclusion As machine reading comprehension tasks with unanswerable questions stress the importance of answer verification in MRC modeling, this paper devotes itself to better verifieroriented MRC task-specific design and implementation for the first time. Inspired by human reading comprehension experience, we proposed a retrospective reader that integrates both sketchy and intensive reading. With the latest Pr LM as encoder backbone and baseline, the proposed reader is evaluated on two benchmark MRC challenge datasets SQu AD2.0 and News QA, achieving new state-of-the-art results and outperforming strong baseline models in terms of newly introduced statistical significance, which shows the choice of verification mechanisms has a significant impact for MRC performance and verifier is an indispensable reader component even for powerful enough Pr LMs used as the encoder. In the future, we will investigate more decoder-side problem-solving techniques to cooperate with the strong encoders for more advanced MRC. References Back, S.; Chinthakindi, S. C.; Kedia, A.; Lee, H.; and Choo, J. 2020. Neur Qu RI: Neural Question Requirement Inspector for Answerability Prediction in Machine Reading Comprehension. In International Conference on Learning Representations. Chen, D.; Bolton, J.; and Manning, C. D. 2016. A Thorough Examination of the CNN/Daily Mail Reading Comprehension Task. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2358 2367. Choi, E.; He, H.; Iyyer, M.; Yatskar, M.; Yih, W.-t.; Choi, Y.; Liang, P.; and Zettlemoyer, L. 2018. Qu AC: Question Answering in Context. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2174 2184. Clark, K.; Luong, M.-T.; Le, Q. V.; and Manning, C. D. 2020. ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators. In International Conference on Learning Representations. Cui, Y.; Chen, Z.; Wei, S.; Wang, S.; Liu, T.; and Hu, G. 2017. Attention-over-Attention Neural Networks for Reading Comprehension. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 593 602. Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2019. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), 4171 4186. Dhingra, B.; Liu, H.; Yang, Z.; Cohen, W. W.; and Salakhutdinov, R. 2017. Gated-Attention Readers for Text Comprehension. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 1832 1846. Guthrie, J. T.; and Mosenthal, P. 1987. Literacy as Multidimensional: Locating Information and Reading Comprehension. Educational Psychologist 22(3-4): 279 297. Hermann, K. M.; Kocisky, T.; Grefenstette, E.; Espeholt, L.; Kay, W.; Suleyman, M.; and Blunsom, P. 2015. Teaching machines to read and comprehend. Advances in neural information processing systems 28: 1693 1701. Hill, F.; Bordes, A.; Chopra, S.; and Weston, J. 2015. The Goldilocks Principle: Reading Children s Books with Explicit Memory Representations. ar Xiv preprint ar Xiv:1511.02301 . Hu, M.; Wei, F.; Peng, Y.; Huang, Z.; Yang, N.; and Li, D. 2019. Read+ verify: Machine reading comprehension with unanswerable questions. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, 6529 6537. Joshi, M.; Choi, E.; Weld, D.; and Zettlemoyer, L. 2017. Trivia QA: A Large Scale Distantly Supervised Challenge Dataset for Reading Comprehension. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 1601 1611. Kadlec, R.; Schmid, M.; Bajgar, O.; and Kleindienst, J. 2016. Text Understanding with the Attention Sum Reader Network. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 908 918. Kundu, S.; and Ng, H. T. 2018. A Question-Focused Multi Factor Attention Network for Question Answering. In Mc Ilraith, S. A.; and Weinberger, K. Q., eds., AAAI, 5828 5835. Lai, G.; Xie, Q.; Liu, H.; Yang, Y.; and Hovy, E. 2017. RACE: Large-scale Re Ading Comprehension Dataset From Examinations. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, 785 794. Lan, Z.; Chen, M.; Goodman, S.; Gimpel, K.; Sharma, P.; and Soricut, R. 2020. ALBERT: A Lite BERT for Selfsupervised Learning of Language Representations. In International Conference on Learning Representations. Li, Z.; He, S.; Cai, J.; Zhang, Z.; Zhao, H.; Liu, G.; Li, L.; and Si, L. 2018. A unified syntax-aware framework for semantic role labeling. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2401 2411. Li, Z.; He, S.; Zhao, H.; Zhang, Y.; Zhang, Z.; Zhou, X.; and Zhou, X. 2019. Dependency or span, end-to-end uniform semantic role labeling. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, 6730 6737. Li, Z.; Wang, R.; Chen, K.; Utiyama, M.; Sumita, E.; Zhang, Z.; and Zhao, H. 2020. Explicit Sentence Compression for Neural Machine Translation. Proceedings of the AAAI Conference on Artificial Intelligence 34(05): 8311 8318. Liu, L.; Zhang, Z.; ; Zhao, H.; Zhou, X.; and Zhou, X. 2021. Filling the Gap of Utterance-aware and Speaker-aware Representation for Multi-turn Dialogue. In The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21). Liu, X.; Li, W.; Fang, Y.; Kim, A.; Duh, K.; and Gao, J. 2018. Stochastic Answer Networks for SQu AD 2.0. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 1694 1704. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; and Stoyanov, V. 2019. Ro BERTa: A robustly optimized BERT pretraining approach. ar Xiv preprint ar Xiv:1907.11692 . Lu, J.; Batra, D.; Parikh, D.; and Lee, S. 2019. Vi LBERT: Pretraining Task-Agnostic Visiolinguistic Representations for Vision-and-Language Tasks. In Advances in Neural Information Processing Systems, 13 23. Mc Nemar, Q. 1947. Note on the sampling error of the difference between correlated proportions or percentages. Psychometrika 12(2): 153 157. Peters, M. E.; Neumann, M.; Iyyer, M.; Gardner, M.; Clark, C.; Lee, K.; and Zettlemoyer, L. 2018. Deep contextualized word representations. In NAACL-HLT, 2227 2237. Radford, A.; Narasimhan, K.; Salimans, T.; and Sutskever, I. 2018. Improving language understanding by generative pre-training. Technical report . Rajpurkar, P.; Jia, R.; and Liang, P. 2018. Know What You Don t Know: Unanswerable Questions for SQu AD. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), 784 789. Rajpurkar, P.; Zhang, J.; Lopyrev, K.; and Liang, P. 2016. SQu AD: 100,000+ Questions for Machine Comprehension of Text. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, 2383 2392. Reddy, R. G.; Sultan, M. A.; Kayi, E. S.; Zhang, R.; Castelli, V.; and Sil, A. 2020. Answer Span Correction in Machine Reading Comprehension. In Findings of the Association for Computational Linguistics: EMNLP 2020. Reddy, S.; Chen, D.; and Manning, C. D. 2019. Co QA: A Conversational Question Answering Challenge. Transactions of the Association for Computational Linguistics 7: 249 266. Seo, M.; Kembhavi, A.; Farhadi, A.; and Hajishirzi, H. 2016. Bidirectional Attention Flow for Machine Comprehension. In International Conference on Learning Representations. Tay, Y.; Luu, A. T.; Hui, S. C.; and Su, J. 2018. Densely connected attention propagation for reading comprehension. In Advances in neural information processing systems, 4906 4917. Trischler, A.; Wang, T.; Yuan, X.; Harris, J.; Sordoni, A.; Bachman, P.; and Suleman, K. 2017. News QA: A Machine Comprehension Dataset. In Proceedings of the 2nd Workshop on Representation Learning for NLP, 191 200. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A. N.; Kaiser, L.; and Polosukhin, I. 2017. Attention is All you Need. Advances in neural information processing systems 30: 5998 6008. Wang, S.; and Jiang, J. 2017. Machine Comprehension Using Match-LSTM and Answer Pointer. In International Conference on Learning Representations. Wang, W.; Yang, N.; Wei, F.; Chang, B.; and Zhou, M. 2017. Gated Self-Matching Networks for Reading Comprehension and Question Answering. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 189 198. Weissenborn, D. 2017. Reading Twice for Natural Language Understanding. Co RR abs/1706.02596. Wu, Y.; Schuster, M.; Chen, Z.; Le, Q. V.; Norouzi, M.; Macherey, W.; Krikun, M.; Cao, Y.; Gao, Q.; Macherey, K.; et al. 2016. Google s neural machine translation system: Bridging the gap between human and machine translation. ar Xiv preprint ar Xiv:1609.08144 . Xu, Y.; Zhao, H.; and Zhang, Z. 2021. Topic-aware multiturn dialogue modeling. In The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21). Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R. R.; and Le, Q. V. 2019. XLNET: Generalized autoregressive pretraining for language understanding. In Advances in neural information processing systems, 5753 5763. Zhang, S.; Zhao, H.; Wu, Y.; Zhang, Z.; Zhou, X.; and Zhou, X. 2020a. DCMN+: Dual Co-Matching Network for Multichoice Reading Comprehension. Proceedings of the AAAI Conference on Artificial Intelligence 34(05): 9563 9570. Zhang, Z.; Huang, Y.; and Zhao, H. 2018. Subwordaugmented Embedding for Cloze Reading Comprehension. In Proceedings of the 27th International Conference on Computational Linguistics (COLING 2018), 1802 1814. Zhang, Z.; Li, J.; Zhu, P.; Zhao, H.; and Liu, G. 2018. Modeling Multi-turn Conversation with Deep Utterance Aggregation. In Proceedings of the 27th International Conference on Computational Linguistics, 3740 3752. Zhang, Z.; Wu, Y.; Li, Z.; and Zhao, H. 2019. Explicit Contextual Semantics for Text Comprehension. In Proceedings of the 33rd Pacific Asia Conference on Language, Information and Computation (PACLIC 33). Zhang, Z.; Wu, Y.; Zhao, H.; Li, Z.; Zhang, S.; Zhou, X.; and Zhou, X. 2020b. Semantics-aware BERT for language understanding. Proceedings of the AAAI Conference on Artificial Intelligence 34(05): 9628 9635. Zhang, Z.; Wu, Y.; Zhou, J.; Duan, S.; Zhao, H.; and Wang, R. 2020c. SG-Net: Syntax-Guided Machine Reading Comprehension. Proceedings of the AAAI Conference on Artificial Intelligence 34(05): 9628 9635. Zhang, Z.; Zhao, H.; and Wang, R. 2020. Machine Reading Comprehension: The Role of Contextualized Language Models and Beyond. ar Xiv preprint ar Xiv:2005.06249 . Zheng, Y.; Mao, J.; Liu, Y.; Ye, Z.; Zhang, M.; and Ma, S. 2019. Human Behavior Inspired Machine Reading Comprehension. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, 425 434. Zhou, J.; Zhang, Z.; and Zhao, H. 2019. LIMIT-BERT: Linguistic Informed Multi-Task BERT. In Findings of the Association for Computational Linguistics: EMNLP 2020, 4450 4461. Zhu, P.; Zhang, Z.; Li, J.; Huang, Y.; and Zhao, H. 2018. Lingke: A Fine-grained Multi-turn Chatbot for Customer Service. In Proceedings of the 27th International Conference on Computational Linguistics (COLING 2018), System Demonstrations, 108 112. Zhu, P.; Zhao, H.; and Li, X. 2020. Dual multi-head coattention for multi-choice reading comprehension. ar Xiv preprint ar Xiv:2001.09415 . Ziser, Y.; and Reichart, R. 2017. Neural Structural Correspondence Learning for Domain Adaptation. In Proceedings of the 21st Conference on Computational Natural Language Learning (Co NLL 2017), 400 410.