# logigan_learning_logical_reasoning_via_adversarial_pretraining__16eb4a31.pdf Logi GAN: Learning Logical Reasoning via Adversarial Pre-training Xinyu Pi1 , Wanjun Zhong2 , Yan Gao3, Nan Duan3, Jian-Guang Lou3 1University of Illinois Urbana-Champaign, Urbana, USA 2Sun Yat-Sen University 3Microsoft Research Asia xinyupi2@illinois.edu, zhongwj25@mail2.sysu.edu.cn {yan.gao, jlou, nanduan}@microsoft.com We present Logi GAN, an unsupervised adversarial pre-training framework for improving logical reasoning abilities of language models. Upon automatic identification of logical reasoning phenomena in massive text corpus via detection heuristics, we train language models to predict the masked-out logical statements. Inspired by the facilitation effect of reflective thinking in human learning, we analogically simulate the learning-thinking process with an adversarial Generator-Verifier architecture to assist logic learning. Logi GAN implements a novel sequential GAN approach that (a) circumvents the non-differentiable challenge of the sequential GAN by leveraging the Generator as a sentence-level generative likelihood scorer with a learning objective of reaching scoring consensus with the Verifier; (b) is computationally feasible for large-scale pre-training with longer target length. Both base and large size language models pre-trained with Logi GAN demonstrate obvious performance improvement on 12 datasets requiring general reasoning abilities, revealing the fundamental role of logic in broad reasoning, as well as the effectiveness of Logi GAN. Ablation studies on Logi GAN components reveal the relative orthogonality between linguistic and logic abilities and suggest that reflective thinking s facilitation effect might also generalize to machine learning 2. 1 Introduction Learning without thinking is labor lost; thinking without learning is perilous. Confucius Pre-trained Language Models (PLMs) (Devlin et al., 2018; Brown et al., 2020; Raffel et al., 2020) are approaching human-level performance in numerous tasks requiring basic linguistic abilities (Rajpurkar et al., 2016; Wang et al., 2018), setting off a huge wave of interest in Natural Language Processing (NLP). Despite the emerging fervor, researchers soon realized that PLMs are relatively incompetent in their reasoning abilities, which seems to be an insurmountable bottleneck for PLMs with even better linguistic abilities (Kassner & Schütze, 2019; Helwe et al., 2021). Following this, researchers delve into reasoning from multitudinous aspects, striving to improve PLMs reasoning abilities. From our perspective, reasoning (in natural language) is essentially an inferential process where an unstated statement is drawn based on several presented statements, and Logic is the systemic set of principles that provides reasoning with correctness and consistency assurance (Hurley, 1982). Regardless of the variability of contents, logical reasoning generally incorporates two invariant forms: drawing conclusions based on some premises (aka. deduction & induction, (Reichertz, 2013)), or indicates equal contribution. Work done during internship at Microsoft Research Asia. 2The code is released in https://github.com/microsoft/Contextual SP/tree/master/logigan 36th Conference on Neural Information Processing Systems (Neur IPS 2022). hypothesizing premises to explain some conclusions (aka. abduction (Douven, 2021)). Most existing tasks requiring general reasoning ability, such as natural language inference (Nie et al., 2019) and complex machine reading comprehension (Lai et al., 2017), can be readily interpreted by this criterion. Other tasks requiring specialized reasoning skills can be considered either as (i) providing sufficient premises but requiring specific ways of premise extraction to draw conclusions, such as multi-hop (Yang et al., 2018b) or hybrid (Chen et al., 2020) reasoning; or (ii) requires external knowledge, such as commonsense (Sap et al., 2020) or numerical (Dua et al., 2019) knowledge, as premises to draw conclusions, hence could also be interpreted by the two forms of logical reasoning. Following this analysis on the relation between logic and reasoning, Logic ability will be an essential foundation for a broad scope of reasoning, and should be prioritized in improving PLMs reasoning abilities 3. Conventional pre-training via randomized Masked Language Modeling (MLM) and auxiliary tasks are generally developed upon Firth (1957) s distributional hypothesis of semantics "a word is characterized by the company it keeps." Under this paradigm, models efficiently learn to capture grammatical structures and contextualized semantics. However, since logical consistency is beyond correctness on a linguistic level, it is less obvious how MLM could help with logical reasoning abilities. Do models harvest logic ability for free from MLM? Or is that something that needs further learning beyond language acquisition? Motivated by these questions, we propose an unsupervised pre-training method aiming at enhancing the logical reasoning ability of PLMs: we automatically identify occurrences of logical reasoning phenomena in large corpus via detection heuristics, and then require PLMs to predict the masked-out logical statements made in the original context (Section 3). For example, in the case Bob recently made up his mind to lose weight. Therefore, [MASK] , the prediction goal is the masked original statement he decides to go on a diet . However, statements different from the original statement could also be logically consistent with respect to a given context. For example, he decides to exercise from today on is also a reasonable inference in the case above. Since Generators trained merely from recovering original statements are not encouraged to explore the possibilities of other reasonable statements, their overall learning effectiveness of logic could potentially be degraded. Therefore, it makes sense to provide additional feedback based on the degree of logical consistency between statements predicted beforehand and the original context we realize this much resembles humans reflective thinking process. Inspired by research from cognitive psychology (Moon, 2013; Boud et al., 2013; Di Stefano et al., 2016) advocating for the vital role of reflective thinking in improving the experiential efficiency of human learning, we hypothesize that machines might also benefit from reflective thinking in their learning processes. Following this hypothesis, we analogically simulate humans learning-thinking process with a Generator-Verifier architecture, and propose Logi GAN, a novel adversarial training approach tailored for sequential GAN training to further facilitate the learning of logical reasoning. In Logi GAN s design, the Generator learns not only to recover the original masked statements, but also to score candidate statements (based on their generative likelihood) in a manner that could reach scoring consensus with the Verifier, who learns to make judgments on the logical consistency between premises and conclusions. The more logically consistent the Verifier thinks of a statement w.r.t. the input context, the higher generative likelihood score is expected to be assigned by the Generator. To encourage the exploration of broader possibilities of reasonable statements other than the original one, we also apply a diversified sampling strategy for candidate statement generation. Both Generator and Verifier scoring processes are continuous throughout the adversarial training, thereby circumventing the non-differentiable barrier in sequential GAN posed by the discrete beam-search. Moreover, Logi GAN does not involve token-wise Monte Carlo Search for policy gradient estimation, and scoring processes of Generator and Verifier are decoupled, so that parallel score computation is possible. This makes large-scale pre-training with longer target length computationally feasible. To test the effectiveness of Logi GAN, we extensively experiment on 12 datasets requiring general reasoning ability. The apparent performance improvements of both base and large size PLMs across all tasks reveal models harvest of logic ability, shoring up the fundamental role of logic in general reasoning. We also carry out ablation studies to understand the functionality of Logi GAN components, the results of which shed light on the relative orthogonality between linguistic and logic ability and suggest that the facilitation effect of reflective thinking is also generalizable to machine learning. 3We expand this analysis in-depth in App. A, and refer intrigued readers there. 2 Logic Pre-training In this work, we primarily focus on improving PLMs ability of informal logic (Groarke, 2021). We include the three most classical types of logical reasoning: deductive, inductive, and abductive reasoning conducted in the form of natural language (Reichertz, 2004; Kennedy & Thornberg, 2018; Reichertz, 2007; Douven, 2021) in our consideration. Note that our coverage is broader than the informal logic strictly defined in the philosophy community (Munson, 1976) that primarily focuses on analyzing the soundness and cogency of the application of the aforementioned reasoning in real-life arguments. The other half of logic investigation the normative study of formal logic (typically conducted in a symbolic form), such as truth-function logic (Buvac & Mason, 1993), modal logic (Priest, 2008), and fuzzy logic (Dote, 1995), is beyond the scope of this paper. Logic Indicators as Detection Heuristics. To set up an unsupervised pre-training aiming at improving models logic ability, the very first step will be to identify logical reasoning phenomena from a vast-scale unstructured text corpus. While invocations of logic are not explicitly stated in most cases, writers usage of logic indicators usually marks their logical reasoning processes with high precision (Hurley, 1982), thereby serving as an ideal heuristic device for our detection purpose. We consider two standard types of logic indicators: (i) conclusion indicator such as Therefore , We may infer that , which denotes drawing conclusion deductively or inductively from given premises; And (ii) premise indicator such as Due to , The reason that , which denotes abductively hypothesizing premises that explain or provide evidence to some stated conclusions. Corpus Construction. For a text corpus, we detect every occurrence of pre-defined logic indicators (listed in App. C), and mask out (i.e., replace with [MASK]) the entire statement subsequent to the indicator (each training example will have exactly one masked-out statement). Then models task will be to perform language modeling and predict the masked statement. We emphasize that statements are declarative sentences or declarative clauses, owning complete subject and predicate structures, and are capable of being factually true or false. To supply sufficient context information for these predictions, we keep x complete sentences previous to the [MASK], as well as y sentences after the [MASK], where x and y can be sampled from a geometric distribution with pre-defined hyper-parameters. Fig. 1 illustrates the input and output format, and we discuss details in Sec. 4. Masked Logical Statement Prediction. In the simplest setting, the Generator learns to infill the masked statement via a single-task pre-training, which fulfills the training process of a typical masked language modeling task. The only difference is that models no longer predict randomly masked tokens or spans but instead logic-targeted masked complete statements. Models are trained to perform Max Likelihood Estimation (MLE) for masked statements under a typical teacher forcing loss. Practically, generative pre-trained language models such as T5 (Raffel et al., 2020) could take up the position of Generator G. Given a single input context / output statement pair (c, s), the teacher forcing loss can be mathematically expressed as 4: Ltf(c, s) = 1 t=1 log p Gθ (wt(s) | w1:t 1(s); c) (1) 3 The Adversarial Training Framework Since Generators trained merely from recovering masked original statements miss out opportunities of exploring other reasonable statements, Logi GAN implements an adversarial mechanism for providing Generators with extra signals based on logical consistency between pseudo-statements (sampled from Eq. 3) and context to encourage explorations. The adversarial framework has two major components: (i) a Verifier V that learns to judge logical consistency between statements and context; (ii) a Generator G that learns both to recover masked original statements, and scores pseudo-statements (based on their generative likelihood) in a manner that could reach scoring consensus with the Verifier The more logically consistent the Verifier thinks of a statement w.r.t. the input context, the more likely the Generator is expected to generate the statement under the input context (i.e., assign higher generative likelihood score). The overall objective of Logi GAN can be expressed as the minimax objective: JG .V = min θ max ϕ Es+ ptrue(.|c)[log Vϕ(c, s+)] + Es pneg(.|Gθ,c,s+)[log(1 Vϕ(c, s ))]. (2) 4wt(.) denotes the tth token of a input string. Context with Logic Indicator All men are mortal, and Socrates is a man. Therefore, [MASK] . All men are mortal, and Socrates is a man. Therefore, Socrates is mortal. 𝓖score (Generative Likelihood Score) 𝓥score (Logical Consistency Score) External Source Original Text Ground Truth Self-Sampling Socrates will eventually die in a certain sense, Socrates is dead a mortal can never be a god Socrates was among the mortal things Socrates is mortal Pseudo-Statements Ground-truth Statement Figure 1: Logi GAN Overview. Generator targets to predict the masked-out logical statement and scores candidate statements, while Verifier justifies the logical correctness of statements. The blue path indicates the process where the Generator helps Verifier learning, while the yellow path denotes the process of giving Verifier feedback for Generator training. where Gθ and Vϕ denote the Generator and the Verifier with model parameters θ and ϕ, respectively. s+/s represents ground-truth statements from original text / sampled pseudo-statement 5. We discuss sampling details of pseudo-statements later in this section in Eq. 3. Classical GAN settings (Goodfellow et al., 2014; Zhu et al., 2017) fall short in sequential generation because the gradient propagation from the Verifier to the Generator is blocked by a non-differentiable beam-search during text generation. Previous approaches such as (Yu et al., 2017) address this challenge by token-wise policy gradient estimation via Monte Carlo Search. However, since the sampling run-time grows exponentially with the length of the target sequence, their original implementations are not applicable to million-scale pre-training with relatively longer target length as in our scenario. Different from them, Logi GAN omits the token-wise Monte Carlo Search for policy gradient estimation, and realizes a similar goal via measuring the similarity of scoring distributions between Verifier and Generator. The main procedures of Logi GAN can be summarized in four steps: (a) several candidate pseudo-statements are sampled on a sentence level; (b) the Verifier assigns the logical consistency scores Vscore based on how logical consistent these candidates are w.r.t the original context; (c) the Generator assigns the sentence-level generative likelihood score Gscore to each candidate to reflect how likely it will produce the pseudo-statement under the given context. (d) The similarity between Generator and Verifier score distributions is computed as a new signal to encourage the Generator to reach scoring consensus with the Verifier i.e., the more logically consistent the Verifier thinks of the statement, the higher likelihood score the Generator is expected to assign. Since both scoring processes are continuous, the non-differentiable barrier is successfully bypassed. Meanwhile, this design does not involve sequential token-level sampling and decouples the Generator and Verifier scoring processes, thereby enabling parallel score computations. This makes large-scale pre-training with relatively longer target sequence length computationally feasible. The overall framework overview is illustrated in Fig. 1, and the detailed training procedure is summarized in Algorithm 1. To diversify the candidate pseudo-statements, we sample pseudostatements from two sources: (i) self-sampling via diversified beam-search from the Generator; or (ii) retrieving similar statements from the corpus, and the sampling process can be summarized as: pneg(. | Gθ, c, s+) = {sα Gθ(. | c) sβ R(s+)}, (3) where Gθ(. | c) denotes self-sampled statement sα given context c from Generator Gθ, and R(s+) denotes a retriever6 that retrieves textually similar statements sβ with ground-truth statement s+ from the corpus. Note that this process is conducted separately for the corpus of Verifier and Generator. 5Note: in real practice, there is a gap between sampled pseudo-statements s and logically inconsistent statements. We keep current symbolic denotations for simplicity only and discuss this issue in App. B. 6Any retriever is feasible and we adopt BM25 as the retrieving method here. Algorithm 1: Adversarial Training Process Dependencies :(1) A Pre-trained Generative Language Model as Generator G0 (2) A Pre-trained Discriminative Language Model as Verifier V0 (3) Generator Source Training Coprus CG with size M (4) Verifier Source Training Corpus CV with size N (5) Pre-defined Warmup epochs E, max iterations of GAN training Q (6) Pre-defined training sample size m, n for V, G per iteration 1 Random partition CG into CGα, CGβ with size Mα, Mβ; 2 G0 Warmup Gα on CG0 for E epochs with Ltf; 3 for i in 1:Q do 5 CV i, CGi Sample m examples from CV, and n examples from CGβ, w/o replacement; 6 g CV i, g CGi Sample pseudo-statements for CV i, CGi with Gi and BM25, as in Eq. 3; 7 Vi Train Vi 1 on g CV i for 1 epoch with Lver, as in Eq. 4; (Verifier Training) 8 for ec in batch ( g CGi ) do 9 Vscore, Gscore Vi, Gi do scoring on ec, as in Eq. 5 and 6; 10 Lgen λ1 Ltf(s+ from ec) + λ2 DKL(Vscore || Gscore) , as in Eq. 7; 11 Gi Update Gi for 1 step with Lgen; (Generator Training) 3.1 Training of Verifier The Verifier serves as a critic to judge whether a statement is logically consistent w.r.t. the context. Therefore, the training task of Verifier can be viewed as a binary classification problem. Pre-trained language models that could perform discriminative classification tasks such as BERT (Devlin et al., 2018), ALBERT (Lan et al., 2019), and Ro BERTa (Liu et al., 2019), will be suitable for the role of Verifier. With y = 1 for both ground-truth and logically consistent pseudo-statements, and y = 0 for other pseudo-statements, the binary classification loss for a single pair of context/statement/label (c, s, y) of Verifier can be mathematically expressed as (omitting average): Lver(c, s, y) = y log Vϕ(y | [c; s]) (1 y) log(1 Vϕ(y | [c; s])), (4) 3.2 Training of Generator The Generator targets both to recover the original masked statements, and to score pseudo-statements based on their generative likelihood in a manner that could reach sentence-level scoring consensus with the Verifier. This corresponds to the two sources of learning signals received by the Generator: (i) the original generative objective with teacher forcing loss defined in Eq.1 as a signal; and (ii) the distribution similarity between sentence-level generative likelihood score assigned by Generator and logic consistency score assigned by Verifier. To achieve the goal of (ii), we first sample pseudostatements {s 1 , ..., s n } from pneg(. | Gθ, c, s+). Then the Verifier assigns logical consistency score Vscore based on how logically consistent the pseudo-statements are w.r.t. the context, expressed as: Vscore(c; s 1 , ..., s n ) = [Vϕ(s 1 , c); Vϕ(s 2 , c); ...; Vϕ(s n , c)], (5) After this, the Generator assigns a sentence-level generative likelihood score Gscore for each pseudostatement to reflect how likely the pseudo-statement will be produced under the given context: Gscore(c; s 1 , ..., s n ) = [ℓθ(s 1 | c); ℓθ(s 2 | c); ...; ℓθ(s n | c)], (6) where ℓθ(s | c) is the accumulated log-likelihood of the statement s conditioned on the context c. Afterward, each statement with a high Verifier score Vϕ(s, c) is also expected to receive a high generative score ℓθ(s | c) to facilitate the Generator s capturing of the Verifier s judgment criterion based on logic consistency. KL-divergence (Kullback & Leibler, 1951) DKL is therefore a appropriate measure for the similarity between the score distribution of Vscore and Gscore. For the purpose of smoothing the gradient to stabilize the GAN training process, we gather both the ground-truth 12% considering 2% because of 3% in order 3% thanks to as a result other premise indicators 2% other conclusion indicators Distribution of Logic Indicators since considering because of because due to now that in order thanks to therefore hence thus so that so as a result other premise indicators other conclusion indicators Frequency Frequency Figure 2: Corpus statistics. Histograms on the left side display length of masked statements (bottom) and prev-and-post statement context (top). The right-side pie chart displays indicators distribution. (learned with teacher-forcing loss) and pseudo statements (learned with KL loss) inside the same batch w.r.t. a single input context c. In our case, there is exactly one ground-truth statement and n pseudo-statements for each input context c. For a batch of (c; s+, s 1 , ..., s n ), the overall objective of the Generator is defined as (in App. F we show how Eq. 7 commits to the optimization of Eq. 2): Lgen(c; s+, s 1 , ..., s n ) = λ1 Ltf(c, s+) + λ2 DKL(Vscore(s 1 , ..., s n )) || Gscore(s 1 , ..., s n ). (7) 4 Experiment Setup 4.1 Datasets To test the effectiveness of Logi GAN, we extensively experiment on 12 datasets requiring reasoning via natural language. Specifically, Re Clor (Yu et al., 2020), Logi QA (Liu et al., 2021a), Adversarial NLI - ANLI, (Nie et al., 2019), focuses especially on logical reasoning, Tell Me Why (Lal et al., 2021) on abuductive reasoning, Hotpot QA (Yang et al., 2018a) on multi-hop reasoning, Quo Ref (Dasigi et al., 2019) on reasoing with co-reference resolution, Mu Tual (Cui et al., 2020), DREAM (Sun et al., 2019)), SAMSum (Gliwa et al., 2019) on reasoning in conversational scenarios, and Narrative QA (s Koˇ ciský et al., 2018), RACE (Lai et al., 2017), XSum (Narayan et al., 2018) on general verbal reasoning. These datasets make most, if not all, necessary premises for drawing logically consistent conclusions available in their provided context, and require few external premises like commonsense or numerical knowledge. Hence, they fit nicely for testing our hypothesis that Logi GAN brings PLMs logic ability beyond their intrinsic linguistic ability, which could benefit general reasoning processes. 4.2 Pre-training Corpus We apply the corpus construction methodology ( 2) on the widely used Book Corpus (Kobayashi, 2018), which consists of e-books and movies with topics crawled from general domains. Although some corpus featuring debates and arguments (Walker et al., 2012; Abbott et al., 2016; Swanson et al., 2015) appears to be more suitable for our emphasis on logic, we do not elect them due to their high domain specificity in fields such as politics, law, and economics. We discard overly short statements and instances where indicators do not indicate logical reasoning (e.g., since 2010 indicating a time point rather than premises, so happy indicating degree of the subsequent adjective rather than conclusions). This results in 3.14 million (1.43 and 1.71 million from conclusion and premise indicators, respectively) instances. Corpus statistics are visualized in Fig. 2. Baseline Choice. Since our primary goal of the experiment is to test the effectiveness of Logi GAN and test our hypothesis that logic ability can be further enhanced beyond PLMs intrinsic linguistic ability, we only compare models pre-trained with Logi GAN against their vanilla versions. After Logi GAN pre-training, we discard the auxiliary Verifier (discussed in Sec. 6) and employ the Generator only to solve all downstream tasks in a purely end-to-end manner. For our main experiments, we initialize Generators from both base and large size pre-trained T5 (Raffel et al., 2020), and Verifier from pre-trained ALBERT-large (Lan et al., 2019). We leave discussions of the rest implementation details and hyper-parameter settings of pre-training and downstream fine-tuning in Appendix D. Elastic Search vs. Self Sampling. As stated earlier in section 3.2, candidate pseudo-statements have two possible sources they could either be sampled via beam search from the Generator s selfdistribution, or could be retrieved from some external resources. We carry out two variant versions of Logi GAN whose Generator is trained purely from self-sampled sentences as pseudo-statements (Logi GAN base(ss)), and from extra pseudo-statements retrieved from corpus by Elastic Search Gormley & Tong (2015) (Logi GAN base(ss+es)). For the large model, we use Logi GAN large (es+ss) as default. Our database consists of 3.14 million sentences discovered by the corpus construction process, and we keep the top-5 similar retrieved sentences along with self-samples from Generator. 5 Experiments 5.1 Experimental Results Table 1: Main results of Logi GAN on 12 downstream tasks (development sets). Multiple Choice & Classification Datasets Models / Dataset Re Clor Logi QA RACE DREAM ANLI Mu Tual Avg. Metrics Acc Acc Acc Acc Acc Acc Vanilla T5 base 35.20 27.19 63.89 59.36 44.10 67.38 49.52 Logi GAN base (ss) 40.20 34.72 67.13 63.38 49.50 69.41 54.06 Logi GAN base (ss+es) 40.00 37.02 67.27 63.73 49.70 69.98 54.62 Vanilla T5 large 50.40 38.56 78.99 78.98 58.00 76.41 63.56 Logi GAN large 54.80 40.55 80.67 81.42 63.50 77.88 66.47 Generation Datasets Models / Dataset Quo Ref Hotpot QA Narrative QA Tell Me Why SAMSum XSum Avg. Metrics EM / F1 EM / F1 Rouge L Rouge L Rouge L Rouge L Vanilla T5 base 70.76 / 74.58 61.11 / 74.86 48.11 30.03 39.32 29.14 36.65 Logi GAN base (ss) 75.02 / 78.68 62.68 / 76.14 49.44 31.18 39.92 30.26 37.70 Logi GAN base (ss+es) 74.94 / 78.40 62.80 / 76.18 49.46 31.15 40.21 30.27 37.77 Vanilla T5 large 80.06 / 83.25 66.11 / 79.80 51.09 31.42 41.40 31.58 38.87 Logi GAN large 81.92 / 85.25 67.04 / 80.36 51.79 32.72 43.13 33.49 40.28 As presented in Table 1, both base and large size PLMs further pre-trained with Logi GAN surpass their vanilla baselines across both discriminative and generative task formats, through a wide scope of downstream tasks requiring general reasoning abilities. We can make the following observations: Among all observed improvements, those on tasks with particular emphasis on logic (Re Clor, Logi QA, and ANLI) are most noticeable. These positive results manifest the effectiveness of Logi GAN in injecting logic ability into PLMs, while testifying to our primary hypothesis that logic ability is fundamental to general reasoning as well. This conclusion answers the two questions in the intro section 7, suggesting that randomized MLM pre-training might fall short in endowing language models with logic ability, and a logic-targeted pre-training approach like Logi GAN may further assist logic learning beyond language acquisition. Furthermore, extra retrieved pseudo-statements (ss+es) bring some additional performance improvement compared with the pure self-sampling (ss) Logi GAN variant, revealing the important role of pseudo-statements diversity in adversarial training. 5.2 Ablation Study and Analysis Observing the apparent performance enhancement, we now aim at pinpointing the truly functional components of Logi GAN through ablation studies and deriving the origins of observed improvements. 7Is logic ability obtained for free from MLM? Could it be further learned beyond language acquisition? For fair comparison purposes, we hold all pre-training and downstream settings (including hyperparameters, implementation designs, and evaluations) unchanged from full Logi GAN. All variations are initialized from T5base, and we report performance variance on 7 datasets. Table 2: Ablation Results on 7 datasets. The last column shows average performance variance, along with relative percentage improvement against vanilla T5base as the baseline. Models / Dataset Re Clor Logi QA RACE DREAM ANLI Quo Ref Narrative QA Metrics Acc Acc Acc Acc Acc EM / F1 Rouge L Average Vanilla T5 base 35.20 27.19 63.89 59.36 44.10 70.76 / 74.58 48.11 49.80(+0.0%) Logi GAN base (ss+es) 40.00 37.02 67.27 63.73 49.70 74.94 / 78.40 49.46 54.59(+9.6%) I. Random Sentence 36.00 30.56 61.26 58.15 45.40 70.96 / 74.50 48.38 50.10(+0.6%) II. MLE Logic Pre-train 38.80 35.02 64.55 61.71 46.00 73.61 / 76.96 49.30 52.71(+5.9%) III. Iterative Multi-task 37.20 34.25 64.01 62.06 46.20 71.67 / 75.14 49.15 52.08(+4.6%) I. Random Masked Sentence Prediction Pre-training. To explain the observed improvements, our first hypothesis is: Models harvest extra linguistic ability from masked statement prediction compared with masked token (or span) prediction. Quite intuitively, filling entire sentences with complete subject-predicate structures might put additional demands on models to capture more abundant syntactic information beyond the coverage of masked token (or span) prediction. Since Logi GAN involves recovering masked sentences, it is then necessary to determine to what degree, if any, that the observed performance gain is attributable to models plausible linguistic ability improvement. We therefore carry out a variant pre-training where the prediction objects are randomly masked sentence. Results (shown in Table 2) displays that masked sentence prediction training barely brings improvement against the vanilla baseline. This suggests it is unlikely that masked sentence prediction empowers PLM trained from masked token prediction significantly better linguistic ability, nor likely that the extra pre-training corpus per se significantly raises the performance. Therefore, we reject the first hypothesis and conclude that observed improvements should derive from somewhere else. II. MLE-only Logic Pre-training. Our second hypothesis is that logic-guided masked statement prediction enhances models intrinsic ability of logical reasoning, thereby lifting the downstream performance. Having addressed the potential impact of learning randomized complete sentence generation, we next aim to check how learning logic-targeted statement generation affects models behavior. We ablate the entire adversarial training process, and train models to perform maximum likelihood estimation (MLE) with teacher-forcing loss only on masked-out logical statements. Results 2 of MLE-only logic pre-training reveals quite a notable improvement across almost all datasets against both vanilla baseline and I., suggesting that learning to generate logical statements indeed injects extra abilities into the model. Since results of I. eliminate the possibility that models harvest stronger linguistic abilities from complete sentence prediction, it is safe to partially ascribe the better downstream performance to models enhanced ability in modeling logical reasoning. This reveals the relative orthogonality between logic ability and models inherent linguistic ability, suggesting that logic ability could be enhanced through further logic-targeted pre-training. III. Iterative Multi-task Pre-training. Since II. only partially explains the observed improvements, here is our last hypothesis: the adversarial training procedure of Logi GAN explains the unexplained rest part beyond the coverage of II. Here a multi-task pre-training with both generation and verification tasks will be the most natural intermediate setting between the single-model generation-only setting of II. and Logi GAN s dual-model adversarial setting. However, since the verification task relies on Generator s self-sampled statements, we adopt an iterative self-critic pre-training manner following Nijkamp et al. (2021). Unlike typical multi-tasking training that simultaneously carries different tasks and then sums the losses, our generation and verification tasks happen alternately 8. Surprisingly, the iterative multi-task pre-training barely brings any positive effects to models downstream performance compared with II. One possible explanation for this might be that the drastically different mechanisms between the verification and generations task intervene with each other, making the single-model & multi-task setting non-beneficial. Now that we have confirmed that an extra verification task fails to explain the rest improvement, we can accept our final hypothesis and conclude that it is indeed the adversarial mechanism between the Generator and Verifier that truly facilitate learning of logical reasoning, thereby further improving the downstream performance beyond II. 8Verification is formulated as a generation task model outputs natural language token good and bad . 6 Discussion A Psycholinguistic Interpretation of Logic-oriented MLM From the first glance, the idea of Logic-oriented MLM seems to be naive and simply. However, we argue that to fully appreciate the value and potential of logic-oriented MLM, going beyond the superficial appearance of masked text and touching down to their underlying psycholinguistic essence is necessary. It is neither the linguistic patterns of the masked-out text, nor the masking technique to corrupt them that makes logic-oriented MLM (and its potential follow-ups) unique. What truly matters is the distinctive cognitive processes proceeding in the minds of writers when they put down different pieces of text language is a window into human minds (Pinker, 2007). Consider the following examples when a human is filling in the [MASK] s: (1) 19 + 69 = [MASK]." (Numerical cognition). (2) Windows is founded by [MASK]. (Declarative memory retrieval). (3) Socrates is a mortal, so he will eventually [MASK] (Logical reasoning). (4) If I feed my dog more than 2 treats per day, it will get [MASK]. (Causal inference). (5) "A crow immediately stands out in swans because it s [MASK]." (Common sense reasoning). (6) Mike deeply bows to his teacher to show his [MASK]. (Social perception). Though answers to these [MASK] s are similar in string length, they nevertheless involve different information pathways and substantially distinctive cognitive processes in writers minds. Logic-oriented MLM shines in that it consistently captures exactly one type of such cognitive process, and trains LMs to model humans logic reasoning mechanism, which is well beyond modeling language per se. On the contrary, Randomized MLM does not capture consistent underlying cognitive processes, which could significantly lower LMs efficiency of learning advanced intelligence mechanisms beyond language itself. The empirical effectiveness of Logic-oriented MLM provides positive evidence for the practicability of this paradigm, suggesting that LMs might be able to learn various advanced human cognitive processes other than logical reasoning via a similar approach. Combined with Logi GAN s natural analogy of humans learning-thinking mechanism during logic development, Logi GAN made an encouraging attempt to unify cognitive modeling and language model pre-training. Adversarial Training Might Assist Downstream Generation Tasks. Although in our experiments, we discard the Verifier and solve downstream tasks with the Generator only, some previous works (Shen et al., 2021; Cobbe et al., 2021) reveal that the Verifier can be used for ranking multiple generation results, thereby effectively enhancing overall downstream accuracy. However, in their paradigm, the information propagates unidirectionally from the Generator to the Verifier, and the Generator cannot directly benefit from the Verifier s discriminative feedback. In contrast, our Logi GAN adversarial training paradigm surmounts the non-differentiable obstacle and could potentially enlighten a new paradigm of both pre-training and downstream fine-tuning. Improving Logical Pre-training. Our paper demonstrates that PLMs logic ability can be further enhanced beyond their inherent linguistic ability, and adversarial training may bring extra benefits beyond the learning of logic-targeted masked statement prediction. However, our heuristic-based approach to identifying logical phenomena in a text corpus and the single mask prediction setting can be further improved. Logic recognition methods with higher recall and better unsupervised task designs (e.g., logical indicator prediction, or logic-guided sentence shuffling) are worthwhile to explore in the further work. Besides, since we are adopting a general domain pre-training corpus (i.e., Book Corpus) with bare emphasis on logic, understanding the impacts of extending pre-training to the domain-specific corpus (e.g., law corpus) or others emphasizing logical reasoning is also substantial. 7 Related Works Generative Adversarial Training in NLP. Unlike conventional GAN (Goodfellow et al., 2014; Mirza & Osindero, 2014; Zhu et al., 2017) that generates continuous output such as images, sequential GAN generates discrete sequences via non-differential searches. This makes feedback from the discriminator not propagatable to the generator. To tackle this challenge, Seq GAN (Yu et al., 2017) borrows an idea from reinforcement learning, treating each output token as a single action, and estimates token-wise policy gradient via Monte Carlo search. Rank GAN (Lin et al., 2017) adopts a similar approach but breaks the binary-classification assumption of discriminator task design, and a ranker provides feedback to the generator. Their generator attempts to generate verisimilar sentences to deceive the ranker into ranking synthetic sentences higher over multiple human-written ones. In our scenario, however, the gold ranking is hard to determine because measuring which statements are more logically consistent w.r.t. context than others is non-trivial, and multi-gold cases are possible. While successfully enabling communication between generator and discriminator, the original designs of Seq GAN, Rank GAN, as well as other works such as (Guo et al., 2017; Fedus et al., 2018; Caccia et al., 2018; Rekabdar et al., 2019), generally formulate text generation as a sequential action decision problem, thereby involving heavy sampling for policy gradient estimation, and are sensitive to the length of the target sequence. Since large-scale pre-training (with arbitrary target length) puts a high demand on scalability and computational efficiency, the above approaches are not readily applicable in our scenario. Furthermore, previous work leverages adversarial training to improve qualities of generated examples, whereas our focus is on enhancing models intrinsic logic ability. A recent work, AR2 (Zhang et al., 2021), leverages adversarial training to improve dense document retrieval. With a retriever-ranker architecture, the learning objective of retriever is to maximize the agreeableness between its own score assignment and that of the ranker for input documents. This is conceptually similar to Logi GAN, as our Generator also aims at reaching consensus with Verifier. However, AR2 does not fall into the sequential GAN paradigm, since it does not involve any sequential text generation, and there is no non-differentiable barrier between the retriever and ranker. Pre-training for Reasoning Ability Improvement. Previous works have extensively investigated the possibility of injecting specific type of reasoning via pre-training, such as numerical (Geva et al., 2020; Yoran et al., 2021; Pi et al., 2022), commonsense (Zhong et al., 2019; Tamborrino et al., 2020; Staliunaite et al., 2021), formal logic (Wang et al., 2021; Pi et al., 2022), multi-hop (Deng et al., 2021; Zhong et al., 2022), and tabular (Liu et al., 2021b) reasoning. Different from them, Logi GAN focuses on logic reasoning, which plays a fundamental role in general reasoning via natural language. 8 Conclusion In this work, we hypothesize that (i) logic ability plays a key role in a wide scope of tasks requiring general reasoning; and (ii) PLMs logic ability can be further improved beyond their original linguistic ability. We correspondingly propose Logi GAN, an unsupervised adversarial pre-training framework for logical reasoning enhancement. Logi GAN circumvents the non-differentiable challenge of sequential GAN via a novel Generator-Verifier scoring consensus mechanism, and enables largescale pre-training with longer target length. Extensive experiments and ablation studies reveal the effectiveness and functional components of Logi GAN, providing evidence to our major hypothesis. Rob Abbott, Brian Ecker, Pranav Anand, and Marilyn A. Walker. Internet argument corpus 2.0: An sql schema for dialogic social media and the corpora to go with it. In LREC, 2016. David Boud, Rosemary Keogh, and David Walker. Reflection: Turning experience into learning. Routledge, 2013. Tom B. Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel M. Ziegler, Jeffrey Wu, Clemens Winter, Christopher Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam Mc Candlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language models are few-shot learners. Co RR, abs/2005.14165, 2020. URL https://arxiv.org/abs/2005.14165. Sasa Buvac and Ian A Mason. Propositional logic of context. In AAAI, pp. 412 419, 1993. Massimo Caccia, Lucas Caccia, William Fedus, Hugo Larochelle, Joelle Pineau, and Laurent Charlin. Language gans falling short. Co RR, abs/1811.02549, 2018. URL http://arxiv.org/abs/ 1811.02549. Wenhu Chen, Hanwen Zha, Zhiyu Chen, Wenhan Xiong, Hong Wang, and William Yang Wang. Hybrid QA: A dataset of multi-hop question answering over tabular and textual data. In Findings of the Association for Computational Linguistics: EMNLP 2020, pp. 1026 1036, Online, November 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.findings-emnlp.91. URL https://aclanthology.org/2020.findings-emnlp.91. Karl Cobbe, Vineet Kosaraju, Mohammad Bavarian, Mark Chen, Heewoo Jun, Lukasz Kaiser, Matthias Plappert, Jerry Tworek, Jacob Hilton, Reiichiro Nakano, Christopher Hesse, and John Schulman. Training verifiers to solve math word problems. ar Xiv preprint ar Xiv:2110.14168, 2021. Leyang Cui, Yu Wu, Shujie Liu, Yue Zhang, and Ming Zhou. Mutual: A dataset for multi-turn dialogue reasoning. Co RR, abs/2004.04494, 2020. URL https://arxiv.org/abs/2004.04494. Pradeep Dasigi, Nelson F. Liu, Ana Marasovi c, Noah A. Smith, and Matt Gardner. Quoref: A reading comprehension dataset with questions requiring coreferential reasoning. In EMNLP, 2019. Xiang Deng, Yu Su, Alyssa Lees, You Wu, Cong Yu, and Huan Sun. Reason BERT: Pre-trained to reason with distant supervision. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 6112 6127, Online and Punta Cana, Dominican Republic, November 2021. Association for Computational Linguistics. URL https://aclanthology. org/2021.emnlp-main.494. Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: pre-training of deep bidirectional transformers for language understanding. Co RR, abs/1810.04805, 2018. URL http://arxiv.org/abs/1810.04805. Giada Di Stefano, Francesca Gino, Gary P Pisano, and Bradley R Staats. Making experience count: The role of reflection in individual learning. Harvard Business School NOM Unit Working Paper, (14-093):14 093, 2016. Y. Dote. Introduction to fuzzy logic. In Proceedings of IECON 95 - 21st Annual Conference on IEEE Industrial Electronics, volume 1, pp. 50 56 vol.1, 1995. doi: 10.1109/IECON.1995.483332. Igor Douven. Abduction. In Edward N. Zalta (ed.), The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University, Summer 2021 edition, 2021. Dheeru Dua, Yizhong Wang, Pradeep Dasigi, Gabriel Stanovsky, Sameer Singh, and Matt Gardner. Drop: A reading comprehension benchmark requiring discrete reasoning over paragraphs. In NAACL, 2019. William Fedus, Ian Goodfellow, and Andrew M. Dai. Maskgan: Better text generation via filling in the, 2018. URL https://arxiv.org/abs/1801.07736. J. Firth. A synopsis of linguistic theory 1930-1955. In Studies in Linguistic Analysis. Philological Society, Oxford, 1957. reprinted in Palmer, F. (ed. 1968) Selected Papers of J. R. Firth, Longman, Harlow. Mor Geva, Ankit Gupta, and Jonathan Berant. Injecting numerical reasoning skills into language models. Co RR, abs/2004.04487, 2020. URL https://arxiv.org/abs/2004.04487. Bogdan Gliwa, Iwona Mochol, Maciej Biesek, and Aleksander Wawer. SAMSum corpus: A humanannotated dialogue dataset for abstractive summarization. In Proceedings of the 2nd Workshop on New Frontiers in Summarization, pp. 70 79, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: 10.18653/v1/D19-5409. URL https://aclanthology.org/ D19-5409. Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial networks, 2014. URL https: //arxiv.org/abs/1406.2661. Clinton Gormley and Zachary J. Tong. Elasticsearch: The definitive guide. 2015. Leo Groarke. Informal Logic. In Edward N. Zalta (ed.), The Stanford Encyclopedia of Philosophy. Metaphysics Research Lab, Stanford University, Fall 2021 edition, 2021. Jiaxian Guo, Sidi Lu, Han Cai, Weinan Zhang, Yong Yu, and Jun Wang. Long text generation via adversarial training with leaked information. Co RR, abs/1709.08624, 2017. URL http: //arxiv.org/abs/1709.08624. Chadi Helwe, Chloé Clavel, and Fabian M Suchanek. Reasoning with transformer-based models: Deep learning, but shallow reasoning. In 3rd Conference on Automated Knowledge Base Construction, 2021. Patrick Hurley. A Concise Introduction to Logic. Belmont, CA, USA: Wadsworth, 1982. Nora Kassner and Hinrich Schütze. Negated LAMA: birds cannot fly. Co RR, abs/1911.03343, 2019. URL http://arxiv.org/abs/1911.03343. Brianna L Kennedy and Robert Thornberg. Deduction, induction, and abduction. The SAGE handbook of qualitative data collection, pp. 49 64, 2018. Sosuke Kobayashi. Homemade bookcorpus. https://github.com/BIGBALLON/cifar-10-cnn, 2018. Solomon Kullback and Richard A Leibler. On information and sufficiency. The annals of mathematical statistics, 22(1):79 86, 1951. Guokun Lai, Qizhe Xie, Hanxiao Liu, Yiming Yang, and Eduard Hovy. Race: Large-scale reading comprehension dataset from examinations. ar Xiv preprint ar Xiv:1704.04683, 2017. Yash Kumar Lal, Nathanael Chambers, Raymond Mooney, and Niranjan Balasubramanian. Tellmewhy: A dataset for answering why-questions in narratives. Co RR, abs/2106.06132, 2021. URL https://arxiv.org/abs/2106.06132. Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, and Radu Soricut. ALBERT: A lite BERT for self-supervised learning of language representations. Co RR, abs/1909.11942, 2019. URL http://arxiv.org/abs/1909.11942. Kevin Lin, Dianqi Li, Xiaodong He, Ming-Ting Sun, and Zhengyou Zhang. Adversarial ranking for language generation. In NIPS, 2017. Jian Liu, Leyang Cui, Hanmeng Liu, Dandan Huang, Yile Wang, and Yue Zhang. Logiqa: A challenge dataset for machine reading comprehension with logical reasoning. In Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI 20, 2021a. ISBN 9780999241165. Qian Liu, Bei Chen, Jiaqi Guo, Zeqi Lin, and Jian-Guang Lou. TAPEX: table pre-training via learning a neural SQL executor. Co RR, abs/2107.07653, 2021b. URL https://arxiv.org/abs/2107. 07653. Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized BERT pretraining approach. Co RR, abs/1907.11692, 2019. URL http://arxiv.org/abs/1907.11692. Mehdi Mirza and Simon Osindero. Conditional generative adversarial nets. Co RR, abs/1411.1784, 2014. URL http://arxiv.org/abs/1411.1784. Jennifer A Moon. Reflection in learning and professional development: Theory and practice. Routledge, 2013. Ronald Munson. The Way of Words an Informal Logic. Boston, MA, USA: Houghton Mifflin School, 1976. Shashi Narayan, Shay B. Cohen, and Mirella Lapata. Don t give me the details, just the summary! Topic-aware convolutional neural networks for extreme summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 2018. Yixin Nie, Adina Williams, Emily Dinan, Mohit Bansal, Jason Weston, and Douwe Kiela. Adversarial NLI: A new benchmark for natural language understanding. Co RR, abs/1910.14599, 2019. URL http://arxiv.org/abs/1910.14599. Erik Nijkamp, Bo Pang, Ying Nian Wu, and Caiming Xiong. SCRIPT: Self-critic Pre Training of transformers. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 5196 5202, Online, June 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.naacl-main. 409. URL https://aclanthology.org/2021.naacl-main.409. Xinyu Pi, Qian Liu, Bei Chen, Morteza Ziyadi, Zeqi Lin, Yan Gao, Qiang Fu, Jian-Guang Lou, and Weizhu Chen. Reasoning like program executors. Co RR, abs/2201.11473, 2022. URL https://arxiv.org/abs/2201.11473. S. Pinker. The Stuff of Thought: Language as a Window Into Human Nature. Viking, 2007. ISBN 9780670063277. URL https://books.google.com/books?id=jyl SITT9ZNUC. G. Priest. An Introduction to Non-Classical Logic: From If to Is. Cambridge Introductions to Philosophy. Cambridge University Press, 2008. ISBN 9781139469678. URL https://books. google.com/books?id=r MXVbm Aw3Yw C. Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter J. Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. Journal of Machine Learning Research, 21(140):1 67, 2020. URL http://jmlr.org/papers/v21/20-074.html. Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. Squad: 100, 000+ questions for machine comprehension of text. Co RR, abs/1606.05250, 2016. URL http://arxiv.org/abs/ 1606.05250. Jo Reichertz. 4.3 abduction, deduction and induction in qualitative research. A Companion to, pp. 159, 2004. Jo Reichertz. Abduction: The logic of discovery of grounded theory. Sage London, 2007. Jo Reichertz. Induction, deduction. The SAGE handbook of qualitative data analysis, pp. 123 135, 2013. Banafsheh Rekabdar, Christos Mousas, and Bidyut Gupta. Generative adversarial network with policy gradient for text summarization. 2019 IEEE 13th International Conference on Semantic Computing (ICSC), pp. 204 207, 2019. Tomáš Koˇ ciský, Jonathan Schwarz, Phil Blunsom, Chris Dyer, Karl Moritz Hermann, Gábor Melis, and Edward Grefenstette. The Narrative QA reading comprehension challenge. Transactions of the Association for Computational Linguistics, TBD:TBD, 2018. URL https://TBD. Maarten Sap, Vered Shwartz, Antoine Bosselut, Yejin Choi, and Dan Roth. Commonsense reasoning for natural language processing. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: Tutorial Abstracts, pp. 27 33, Online, July 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.acl-tutorials.7. URL https://aclanthology.org/2020.acl-tutorials.7. Jianhao Shen, Yichun Yin, Lin Li, Lifeng Shang, Xin Jiang, Ming Zhang, and Qun Liu. Generate & rank: A multi-task framework for math word problems. Co RR, abs/2109.03034, 2021. URL https://arxiv.org/abs/2109.03034. Ieva Staliunaite, Philip John Gorinski, and Ignacio Iacobacci. Improving commonsense causal reasoning by adversarial training and data augmentation. Co RR, abs/2101.04966, 2021. URL https://arxiv.org/abs/2101.04966. Kai Sun, Dian Yu, Jianshu Chen, Dong Yu, Yejin Choi, and Claire Cardie. DREAM: A challenge dataset and models for dialogue-based reading comprehension. Transactions of the Association for Computational Linguistics, 2019. URL https://arxiv.org/abs/1902.00164v1. Reid Swanson, Brian Ecker, and Marilyn Walker. Argument mining: Extracting arguments from online dialogue. In Proceedings of the 16th Annual Meeting of the Special Interest Group on Discourse and Dialogue, pp. 217 226, Prague, Czech Republic, September 2015. Association for Computational Linguistics. doi: 10.18653/v1/W15-4631. URL https://aclanthology.org/ W15-4631. Alexandre Tamborrino, Nicola Pellicano, Baptiste Pannier, Pascal Voitot, and Louise Naudin. Pre-training is (almost) all you need: An application to commonsense reasoning. Co RR, abs/2004.14074, 2020. URL https://arxiv.org/abs/2004.14074. Marilyn Walker, Jean Fox Tree, Pranav Anand, Rob Abbott, and Joseph King. A corpus for research on deliberation and debate. In Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC 12), pp. 812 817, Istanbul, Turkey, May 2012. European Language Resources Association (ELRA). URL http://www.lrec-conf.org/proceedings/ lrec2012/pdf/1078_Paper.pdf. Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R. Bowman. GLUE: A multi-task benchmark and analysis platform for natural language understanding. Co RR, abs/1804.07461, 2018. URL http://arxiv.org/abs/1804.07461. Siyuan Wang, Wanjun Zhong, Duyu Tang, Zhongyu Wei, Zhihao Fan, Daxin Jiang, Ming Zhou, and Nan Duan. Logic-driven context extension and data augmentation for logical reasoning of text. ar Xiv preprint ar Xiv:2105.03659, 2021. Zhilin Yang, Peng Qi, Saizheng Zhang, Yoshua Bengio, William W. Cohen, Ruslan Salakhutdinov, and Christopher D. Manning. Hotpotqa: A dataset for diverse, explainable multi-hop question answering. Co RR, abs/1809.09600, 2018a. URL http://arxiv.org/abs/1809.09600. Zhilin Yang, Peng Qi, Saizheng Zhang, Yoshua Bengio, William W. Cohen, Ruslan Salakhutdinov, and Christopher D. Manning. Hotpot QA: A dataset for diverse, explainable multi-hop question answering. In Conference on Empirical Methods in Natural Language Processing (EMNLP), 2018b. Ori Yoran, Alon Talmor, and Jonathan Berant. Turning tables: Generating examples from semistructured tables for endowing language models with reasoning skills. Co RR, abs/2107.07261, 2021. URL https://arxiv.org/abs/2107.07261. Lantao Yu, Weinan Zhang, Jun Wang, and Yong Yu. Seqgan: Sequence generative adversarial nets with policy gradient. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, AAAI 17, pp. 2852 2858. AAAI Press, 2017. Weihao Yu, Zihang Jiang, Yanfei Dong, and Jiashi Feng. Reclor: A reading comprehension dataset requiring logical reasoning. In International Conference on Learning Representations (ICLR), April 2020. Hang Zhang, Yeyun Gong, Yelong Shen, Jiancheng Lv, Nan Duan, and Weizhu Chen. Adversarial retriever-ranker for dense text retrieval. Co RR, abs/2110.03611, 2021. URL https://arxiv. org/abs/2110.03611. Wanjun Zhong, Duyu Tang, Nan Duan, Ming Zhou, Jiahai Wang, and Jian Yin. Improving question answering by commonsense-based pre-training. In CCF International Conference on Natural Language Processing and Chinese Computing, pp. 16 28. Springer, 2019. Wanjun Zhong, Junjie Huang, Qian Liu, Ming Zhou, Jiahai Wang, Jian Yin, and Nan Duan. Reasoning over hybrid chain for table-and-text open domain qa. ar Xiv preprint ar Xiv:2201.05880, 2022. Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A. Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. Co RR, abs/1703.10593, 2017. URL http://arxiv. org/abs/1703.10593. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] (c) Did you discuss any potential negative societal impacts of your work? [N/A] (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [N/A] (b) Did you include complete proofs of all theoretical results? [N/A] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] We will release the code and data upon acceptance. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [No] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [No] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [Yes] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]