# improving_openended_text_generation_via_adaptive_decoding__32ef7915.pdf Improving Open-Ended Text Generation via Adaptive Decoding Wenhong Zhu 1 Hongkun Hao 1 Zhiwei He 1 Yiming Ai 1 Rui Wang 1 Current language models decode text token by token according to probabilistic distribution, and determining the appropriate candidates for the next token is crucial to ensure generation quality. This study introduces adaptive decoding, a mechanism that dynamically empowers language models to ascertain a sensible candidate set during generation. Specifically, we introduce an entropy-based metric called confidence and conceptualize determining the optimal candidate set as a confidenceincreasing process. The rationality of including a token in the candidate set is assessed by leveraging the increment of confidence. Experimental results reveal that our method balances diversity and coherence well. The human evaluation shows that our method can generate human-preferred text. Additionally, our method can potentially improve the reasoning ability of language models. 1. Introduction When it comes to open-ended text generation tasks, the aim is to produce coherent and fluent output based on limited input information (Holtzman et al., 2020). These tasks span various applications, including casual conversation (Thoppilan et al., 2022), storytelling (Mostafazadeh et al., 2016), and similar domains. Language models (LMs) presently estimate the likelihood of the next token in text generation, and they have gained extensive application in various tasks (Brown et al., 2020; Touvron et al., 2023). One factor that determines the quality of generation is the decoding algorithm. For LMs, training solely on maximum likelihood estimation (MLE) without additional fine-tuning, such as human preference alignment, often raises distinct degradation issues (Holtzman et al., 2020). These challenges are closely 1MT Lab, Department of Computer Science and Engineering , Shanghai Jiao Tong University, Shanghai, China. Correspondence to: Rui Wang . Proceedings of the 41 st International Conference on Machine Learning, Vienna, Austria. PMLR 235, 2024. Copyright 2024 by the author(s). intertwined with decoding algorithms. The decoding algorithm can be seen as a mechanism to truncate the entire vocabulary space into the next candidate token set. When this set is limited, as seen in greedy decoding and beam search, it leads to heightened repetition as the model favors previously decoded tokens (Li et al., 2020; Zhu et al., 2023). Conversely, when this set is expansive, as in top-p sampling, it can result in incoherence and hallucination (Xu et al., 2023). This occurs because certain improbable tokens still contribute to the cumulative probability, causing issues when sampling these tokens. Numerous efforts have been devoted to addressing these challenges. Some endeavors focus on enhancing diversity within the specified sampling space (Su et al., 2022; Xu et al., 2023), while others strive to identify a rational sampling space (Meister et al., 2023b; Hewitt et al., 2022). We provide one human-written text example for the GPT2XL model and use teacher-forcing decoding to demonstrate the distinction between the human-preferred and LMpreferred tokens. In doing so, we save the probability of human-preferred and LM-prefered tokens. As shown in Figure 1(a), the choices made by humans and the LM predominantly align at high probability points. The distinction lies in that the portion of the token selected by humans tends to fall within the low probability region predicted by the LM. Simultaneously, the model tends to choose tokens with relatively higher probability. This suggests that the selection of the next token could have many possibilities. On the other hand, the distribution predicted by the LM exhibits a long-tailed pattern (Holtzman et al., 2020), indicating that choosing the next token following the previous context is constrained. The main difference in the long tail distribution is the head shape. For example, as shown in Figure 1(b), the distribution is flattened, indicating that the LM has multiple choices of the next token. The diversity of generated text may be compromised if our candidate set is too limited. While in Figure 1(c), the distribution is on a sharp shape, implying the restricted choices. In this scenario, an excessively large candidate set could lead to a loss of coherence in the generated text. These observations raise the question: how do we determine the rational size of the candidate set to ensure both diversity and coherence in the generated text? Improving Open-Ended Text Generation via Adaptive Decoding (a) The probability of human-preferred token and LM-preferred token. (b) Head distribution given the prefix: Growing up, Obama faced the challenges (c) Head distribution given the prefix: attended Columbia University in New Figure 1. Human written text: Barack Obama was born in 1961. He was raised in Hawaii by his mother and grandparents. Growing up, Obama faced the challenges of being biracial, with a Kenyan father and an American mother. Despite these challenges, he excelled academically and eventually attended Columbia University in New York City. We provide this human-written text for GPT2-XL and use teacher-forcing decoding. In this study, we present an adaptive decoding algorithm designed to dynamically determine the size of a suitable candidate set during generation. Entropy is a metric to quantify the disorder or randomness within a system. It would be useful in gauging the confidence of an LM in predicting the probability distribution of the next token. We conceptualize identifying the optimal candidate set as an confidence-increasing procedure. Specifically, we establish a system embracing two sets: a candidate set, initially empty, and an ordered set, initially encompassing the entire vocabulary. One metric named confidence was formulated to gauge system confidence based on entropy. Continuously selecting tokens with the highest probability from the ordered set and appending them to the candidate set significantly enhances the confidence of this system. The rationality of a token to join the candidate set is assessed based on the increment of confidence, allowing the model to dynamically determine a suitable set of candidates. In our experiments, we performed two open-ended text generation tasks: document continuation and story generation. The results suggest that our approach significantly enhances diversity while preserving coherence in GPT2-XL(1.5B) and Llama2-7B models. On Llama2-7B-chat, coherence has been enhanced alongside increased diversity. Additionally, two mathematical reasoning tasks demonstrate that our method potentially improves the reasoning ability of language models. Human evaluation results show that our approach produces text that is more human-preferred. In summary, the main contributions of this paper are as follows: We integrate the concept of entropy reduction into the text decoding process and design a confidence increment as the metric to assess the rationality of a token as a candidate. We introduce a novel adaptive decoding mechanism that enables LMs to determine an appropriately sized set of candidates autonomously during the generation. 1 Adaptive decoding effectively generates text in a style resembling human expression, ensuring a balance between coherence and diversity. 2. Problem Formulation 2.1. Language Models We follow the definitions given in Hewitt et al. (2022). Let the random variable X = (X1, X2, ..., XT ) represents a sequence of tokens, where each Xt belongs to a finite vocabulary V. We use x ϵ)) top-k sampling xt sample from V(k) G G {xt} end for Return: G 5. Experiments 5.1. Setups Models We assess the performance of our method on the GPT2-XL model (Radford et al., 2019), the Llama2-7B Improving Open-Ended Text Generation via Adaptive Decoding Table 3. Main experiments are conducted by repeating the sampling process five times for each prefix, and the average score is reported. Metrics where the algorithm outperforms others are highlighted in bold. * denotes a significant difference compared to other methods (p < 0.05). LM Decoding Writing Prompts Wiki Text-103 rep-2 rep-3 rep-4 Diversity MAUVE Coherence rep-2 rep-3 rep-4 Diversity MAUVE Coherence Human 3.63 0.87 0.39 95.27 27.12 5.42 1.20 0.41 93.18 62.94 greedy 82.00 80.62 79.66 1.82 0.61 56.27 71.86 68.33 65.95 6.98 2.03 57.81 top-k 12.78 6.68 4.70 80.50 22.28 47.69 9.96 3.57 1.96 86.08 77.15 61.19 mirostat 58.48 51.81 47.32 18.17 5.61 50.10 45.38 36.21 30.74 31.11 36.06 60.51 top-p 12.72 6.86 4.96 80.24 21.20 48.01 8.11 2.74 1.56 88.82 78.22 60.00 typical 13.28 7.62 5.83 79.51 20.36 46.98 6.73 2.13 1.19 90.71 78.49 59.66 η-sampling 8.92 4.43 3.33 86.26 22.12 45.15 7.80 2.55 1.45 89.44 77.94 59.90 adaptive (ours) 6.86 3.66 3.07 89.30 25.18 37.71 5.20 1.43 0.80 93.23 79.57 55.98 Human 7.44 1.59 0.64 90.67 32.84 4.73 1.00 0.33 94.10 60.34 greedy 77.32 74.67 72.84 5.58 1.51 46.40 43.71 36.70 32.09 34.54 22.28 55.85 top-k 9.13 4.45 3.27 87.18 23.57 23.02 7.18 2.44 1.20 90.01 74.80 58.49 top-p 8.33 4.10 3.29 88.38 23.60 23.05 7.07 2.49 1.33 90.05 77.19 58.19 mirostat 13.93 8.49 6.65 79.71 21.39 56.23 12.45 6.14 3.92 80.63 69.22 65.79 typical 7.16 3.55 2.91 90.33 25.78 22.56 7.06 2.48 1.27 90.08 76.04 57.96 η-sampling 6.60 2.96 2.35 90.99 24.43 23.19 5.93 1.84 0.90 91.96 78.26 57.46 adaptive (ours) 7.02 2.63 1.58 89.98 35.26 35.10 6.81 2.31 1.24 90.55 76.74 58.22 model, and its associated chat version (Touvron et al., 2023). Datasets We explore two open-ended text generation applications: document continuation using the Wiki Text-103 dataset (Merity et al., 2017), which contains a large collection of Wikipedia articles. Another is story generation on the Writing Prompts dataset (Fan et al., 2018), a notably challenging endeavor. This task involves crafting imaginative continuations based on abstract, high-level story prompts contributed by online users. Baselines Our proposed algorithm is compared against various decoding algorithms. Top-k method (Fan et al., 2018) maximizes P x Ax η}, where η = min (ϵ, α exp H (Xt | X p2 > ... > p|V| with P|V| i pi = 1. If, however, pk < 1 Pk 1 i pi |V| k+1 , it implies the existence of j within the range [k + 1, |V|] such that pj > pk. This assertion results in a contradiction. Corollary C.2. Given that the sum of probabilities for all tokens in the vocabulary is expressed as P|V| i pi = 1, we can deduce that 1 Pk 1 i pi pk = P|V| k+1 pi 0. Observing this inequality, it is evident that 1 Pk 1 i pi pk. Corollary C.3. The function Confk(X) is monotonically increasing, as it is always guaranteed that Conf 0. Improving Open-Ended Text Generation via Adaptive Decoding C.2. Proof of Section 3.3 Proof. It is intuitively correct that H(X) = PV i pi log pi converges to Pk i pi log pi under the condition Pθ(X > k + m|X > k) 1. In accordance with Corollary C.1, the following inequality holds: log |V| Conf pk log pk(|V| k) i pi) log 1 i pi) log (1 Pk i pi) |V| k . Moreover, based on Corollary C.2, we can further derive: log |V| Conf pk log pk(|V| k) + pk log (1 Pk i pi) |V| k + pk log 1 pk pk log 1 = 0. Detailed mathematical proof in Corollary 3.4 is as follows. i=1 Pθ(xi) log Pθ(xi) k+1 Pθ(xi) log Pθ(xi). As Pθ(X > k + m|X > k) 1, k+1 Pθ(xi) log Pθ(xi) 0. (7) We define the binary random variable B(u) = 1 if X u 0 if X > u. Then, we can express H(X) in terms of B(u) as: H(X) = H(X | B(u)) + I(X; B(u)) = H(X | B(u) = 1) Pr{B(u) = 1} + H(X | B(u) = 0) Pr{B(u) = 0} + I(X; B(u)) = H(X(u)) Pr{B(u) = 1} + H(X | B(u) = 0) Pr{B(u) = 0} + I(X; B(u)). As u k, we have Pr(B(u) = 1) 1, leading to H(B(u)) 0. Consequently, I(X; B(u)) 0 due to I(X; B(u)) H(B(u)). Additionally, H(X|B(u) = 0) Pr(B(u) = 0) can be simplified as: Improving Open-Ended Text Generation via Adaptive Decoding H(X | B(u) = 0) Pr{B(u) = 0} i=k+1 Pθ(xi) log Pθ(xi) Pr{B(u) = 0} i=k+1 Pθ(xi) log Pθ(xi) log Pr{B(u) = 0}) i=k+1 Pθ(xi) log Pθ(xi) + i=k+1 Pθ(xi) log Pr{B(u) = 0} i=k+1 Pθ(xi) log Pθ(xi) + Pr{B(u) = 0} log Pr{B(u) = 0}. As u k, we have Pr(B(u) = 0) 0 and Equation (7), then both terms lead to zero. In conclusion, we obtain H(X) = H(X(k)) + ϵ, where ϵ is a smaller number. D. Hyperparameter Choosing hyperparameters seems a challenging task, and the process of our parameter selection is illustrated in Table 10 and Table 11. Here are some interesting things we found: The top-k and top-p algorithms enhance diversity by expanding the truncation space, leading to higher MAUVE values. This occurrence can be attributed to the ability of LMs to sample certain low-probability tokens, which are proximate to the probabilities preferred by humans. This proximity results in a closer match between the two curves illustrated in Figure 1(a). However, it is crucial to recognize that the concurrent increase in diversity corresponds to a decrease in coherence. Mirostat excels in effectively managing text coherence by controlling model perplexity. The ease of controlling parameters further enhances its utility, even though some of the observed increase in coherence may be attributed to the rise in indices resulting from repetition. Additionally, our investigation reveals that Mirostat, when applied to the Llama2-7B model, maintains high coherence and ensures a certain level of diversity. E. Instrcution In Section 5.5, we use the following instruction, where the sentence is derived from the prompt in Writing Pompts. [INST] <> You re a writer. You will write a story given my prefix. <> Prefix: {sentence} [/INST] In Section 5.6, we use the following instruction, where the problem is derived from the dataset GSM8K and Multi Arith. [INST] <>You are a help assistant and a math expert. Please solve the following question and directly return me the answer.<> Problem: {sentence} Let s think step by step\n[/INST] F. Human Evaluation The human evaluation adheres to the principles of blind review, ensuring that the evaluator remains unaware of the decoding algorithms employed in the two generations or the human writing in advance. The evaluator determines the superior outcome based on three guiding principles. Our application for human evaluation is illustrated in Appendix F. Improving Open-Ended Text Generation via Adaptive Decoding Table 10. Optimal hyperparameter selection for various decoding algorithms in the GPT2-XL model. The line where the bold is located designates the optimal parameter value. We repeat the sampling process five times for each prefix and report the average score. Decoding Writing Prompts Wiki Text-103 rep-2 rep-3 rep-4 Diversity MAUVE Coherence rep-2 rep-3 rep-4 Diversity MAUVE Coherence Human 7.70 1.82 0.79 90.06 29.57 5.14 1.02 0.33 93.66 63.49 k = 5 32.86 24.03 19.55 49.37 23.94 53.16 19.18 10.37 6.90 70.40 82.29 62.57 k = 10 19.60 11.83 8.71 69.82 33.65 50.04 11.50 4.44 2.44 83.41 82.49 62.25 k = 15 14.91 8.35 6.16 76.86 44.92 49.51 9.90 3.64 2.01 86.24 88.75 61.41 k = 20 12.23 6.26 4.35 81.22 48.25 47.87 8.52 2.78 1.48 88.51 88.67 61.61 p = 0.80 25.39 18.79 15.90 61.61 30.29 49.13 12.98 5.94 3.84 80.89 77.74 61.11 p = 0.85 19.78 13.14 10.80 70.02 35.10 48.20 11.23 4.87 3.02 83.73 79.61 61.46 p = 0.90 15.16 9.14 7.07 76.96 38.52 47.15 9.57 3.67 2.17 86.62 84.46 61.13 p = 0.95 12.21 7.04 5.53 81.43 51.69 47.35 7.70 2.70 1.49 89.24 88.50 60.87 mirostat 2.0 59.05 52.33 47.75 17.26 13.75 50.21 45.09 35.99 30.57 31.43 56.59 60.34 mirostat 3.0 58.73 51.91 47.27 17.82 16.52 49.63 45.13 36.09 30.61 31.27 55.29 60.73 mirostat 4.0 58.62 51.84 47.17 17.69 16.30 49.28 45.14 36.07 30.60 31.11 56.56 61.01 mirostat 5.0 58.55 51.70 47.04 18.17 16.91 49.71 45.46 36.46 31.03 31.11 53.55 61.09 τ = 0.90 15.24 9.24 7.25 76.74 38.10 46.64 9.38 3.60 2.14 86.94 85.35 61.27 τ = 0.92 14.02 8.09 6.27 78.65 43.92 46.63 9.08 3.35 2.06 87.47 84.99 61.00 τ = 0.95 12.05 6.96 5.32 81.47 38.01 46.99 8.02 2.73 1.52 89.10 85.39 60.89 τ = 0.99 9.22 4.63 3.64 85.42 40.57 46.59 6.74 1.93 0.95 91.13 90.07 59.93 η = 0.004 12.33 7.08 5.66 80.95 41.86 46.69 7.83 2.63 1.48 89.22 90.40 61.10 η = 0.002 9.61 4.98 3.63 85.42 47.24 45.60 7.58 2.75 1.80 89.81 82.18 59.55 η = 0.0009 9.10 4.58 3.41 86.58 59.34 45.67 6.45 1.75 0.98 91.38 87.17 60.33 η = 0.0006 8.99 4.38 3.67 85.98 46.23 44.38 6.44 1.72 0.82 91.64 86.45 60.37 ϵ = 0.0005 5.46 3.16 2.75 91.38 47.19 36.38 4.82 1.55 1.00 93.81 83.72 56.58 ϵ = 0.001 7.07 3.85 3.11 89.32 55.19 37.72 5.22 1.51 0.93 93.00 84.54 56.39 ϵ = 0.005 40.69 33.57 29.95 39.80 11.02 47.51 21.10 12.50 9.10 67.63 69.27 61.30 ϵ = 0.01 71.14 67.75 65.28 9.81 2.89 50.71 45.17 37.82 33.96 34.02 23.08 61.04 Consistency: The extent to which the produced text maintains semantic coherence with the given context. Smoothness: The degree to which the generated text flows seamlessly and is readily comprehensible. Richness: The presence of diverse and captivating content in the generated text G. Case Study Given the prefix sampled from Wiki Text-103, which provides relevant enough context, we give the truncated human continuation and the generation of our proposed adaptive decoding. Cases can be found in Table 12 and Table 13. Improving Open-Ended Text Generation via Adaptive Decoding Table 11. Optimal hyperparameter selection for various decoding algorithms in the Llama2-7B model. The line where the bold is located designates the optimal parameter value. We repeat the sampling process five times for each prefix and report the average score. Decoding Writing Prompts Wiki Text-103 rep-2 rep-3 rep-4 Diversity MAUVE Coherence rep-2 rep-3 rep-4 Diversity MAUVE Coherence Human 7.70 1.82 0.79 90.06 29.57 5.14 1.02 0.33 93.66 63.49 k = 5 22.69 14.76 11.10 66.96 39.66 27.15 12.14 5.56 3.20 81.69 63.82 59.43 k = 10 14.21 7.26 5.05 79.94 45.34 27.44 9.35 3.70 2.07 86.51 74.15 59.51 k = 15 11.00 5.42 3.65 84.46 40.40 25.98 7.88 2.76 1.40 89.11 82.88 59.38 k = 20 9.62 4.19 2.94 86.29 47.36 26.25 7.44 2.55 1.40 89.54 87.07 59.27 p = 0.80 17.85 10.92 8.23 74.17 46.40 27.23 11.22 5.05 2.94 83.40 75.67 59.33 p = 0.85 13.51 7.66 5.58 80.26 44.91 27.55 9.64 4.01 2.29 85.84 79.15 58.94 p = 0.90 11.02 5.41 3.91 83.58 55.53 31.11 8.20 3.02 1.58 88.37 80.49 59.54 p = 0.95 8.39 4.32 3.26 87.96 58.75 26.99 7.13 2.52 1.32 90.12 82.22 58.89 mirostat 2.0 13.58 7.99 6.04 79.94 44.55 55.91 12.25 6.10 3.94 80.99 81.55 66.51 mirostat 3.0 13.46 7.85 5.88 80.04 50.51 54.83 12.36 6.14 3.99 80.72 82.06 65.87 mirostat 4.0 14.95 8.59 6.85 79.11 42.97 55.10 12.70 6.50 4.32 80.06 81.09 66.31 mirostat 5.0 15.08 9.82 7.88 77.82 45.22 54.84 12.49 6.23 4.06 80.53 78.36 66.46 τ = 0.90 10.83 5.36 3.84 84.65 38.07 25.23 8.24 3.24 1.77 88.10 77.49 59.23 τ = 0.92 9.37 4.37 3.38 86.86 33.88 27.80 7.68 2.80 1.46 89.17 82.78 58.79 τ = 0.95 9.54 5.03 3.78 87.09 40.18 26.80 6.81 2.20 1.06 90.61 87.75 58.83 τ = 0.99 8.67 4.53 4.13 88.93 53.26 25.76 6.00 1.98 1.00 91.75 80.68 58.50 η = 0.004 7.86 3.42 2.50 88.69 48.93 26.30 7.15 2.47 1.30 90.04 77.47 58.76 η = 0.002 7.17 3.58 2.62 89.76 48.34 26.88 6.41 2.21 1.17 91.15 81.49 58.76 η = 0.0009 7.10 3.04 2.14 90.42 61.30 26.78 5.68 1.72 0.85 92.41 86.14 58.76 η = 0.0006 6.69 2.93 2.42 90.43 63.80 26.35 5.87 1.73 0.76 92.04 88.28 58.75 ϵ = 0.0005 7.21 2.84 1.74 89.54 69.00 35.40 5.90 1.78 0.79 92.00 87.18 58.78 ϵ = 0.001 8.38 3.54 2.14 87.76 59.34 35.92 6.74 2.16 1.03 90.65 87.81 59.26 ϵ = 0.005 19.22 11.44 8.43 70.75 58.23 42.35 10.42 4.61 2.69 84.76 76.91 60.03 ϵ = 0.01 35.60 27.20 22.96 47.12 43.69 44.02 14.11 7.34 4.65 78.55 68.76 58.69 Figure 6. Our application for human evaluation. Improving Open-Ended Text Generation via Adaptive Decoding Table 12. Case study: Text continuation with the GPT2-XL model. Entities are in blue fonts, logical relationships are in red, and coherent contents are highlighted in green. Prefix: Following a period of rest in England, Little turned down a desk assignment and volunteered to return to action on the Western Front, joining Lieutenant Colonel Collishaw s Human Continuation: No. 3 Squadron RNAS in March 1918. The unit evolved into No. 203 Squadron of the new Royal Air Force on 1 April, formed after the merger of the RNAS and the Royal Flying Corps. Now ranked captain, and again flying Sopwith Camels, Little gained a further nine successes, beginning with a Fokker Triplane on 1 April, and concluding with two kills in one day on 22 May, an Albatros and a DFW. During this stretch of victories, on 21 April 1918, he was brought down unharmed by Friedrich Ehmann. On 27 May, Little received reports of German Gotha bombers in the vicinity, and took off on a moonlit evening to intercept the raiders. As he closed with one of the bombers, his plane was caught in a searchlight beam and he was struck by a bullet that passed through both his thighs. He crash-landed in a field near Nœux, and bled to death before he was discovered the following morning by a passing gendarme. Little s skull and ankle had also been fractured in the impact; his body was identified by his friend and fellow ace, Charles Dawson Booker. Adaptive decoding: 49th Rifles. On his first day back, he led a cavalry charge which was repulsed in the village of Culmore Wood in June 1916, while advancing across an open plain on the company s way to Vimy Ridge. In late 1916, He had been promoted from sergeant and given a smaller objective position in the near-by village of Meadowhall. His company was soon attacked by a cavalry squad of such ferocity that the archers and machine gunners could have opened up at once. Little was bitten, painfully, by a cavalry lance. He received the nicknames Noire and The Snake , because of his long knife, thin face, and reptilian nature. He later hired a pet donkey to pull his belongings back to camp at the end of each day s march. In late May 1917, three weeks after the Heed affair, Little was deployed for the first time in front of the Hindenburg Line when, as part of the Anglo-French alliance, his squad was to capture and destroy German flak guns in Belgium. As if on cue, two machine gun crews came through a broken field and attacked Little s men, who retreated through a canal. Prefix: The total population size of the African river martin is unknown. In the late 1980s, it appeared to be common, if local, and large numbers were Human Continuation: seen on migration in Gabon. However, it is particularly poorly studied in the Democratic Republic of the Congo (DRC), and it is not known if there is any relationship between the birds breeding in the DRC and those breeding in coastal areas of Gabon and Congo. Several hundred birds were seen at Conkouati-Douli National Park in Congo in 1996 and a flock of 15,000 birds was seen in Gabon in 1997. Investigations in Gabon in September 2003 extended the known range of this martin. More than 300 birds were found with hundreds of rosy bee-eaters in the Ombou e area and the newly established Igu ela National Park, and a mixed flock with rosy bee-eaters at Loango National Park in Gabon was estimated at 100,000 birds in total. In January 2010, 250 martins were seen at a new location in Bakoumba, and single birds were observed in hirundine flocks near Moanda, at Mounana, and at Lekoni. Adaptive decoding: seen during spring breeding in the Cape province of South Africa. The species is also common in South America, but these differences between species and range regions is unclear. The African river martin is a robust and broad-bodied species that has a dark body and light face, which enhances its camouflage. Its chest is often painted or spotted with black. The coloration and markings can differ between individuals. They have very large ears and a long tail, and in female martins have a pattern that resembles a dorsal stripe, which appears more yellow than brown when united with dark bars, thus giving them the common name of yellow-backed or yellow-fronted. The tail is large, and in males has a faint red pink spot at the base that is very difficult to see unless the animal is under a light. The neck varies from strong to weak, and males often have longer teeth. Prefix: The water temperature of Briar Creek near its source ranges from 35.96 to 64.94 F (2.20 to 18.30 C), Human Continuation: with a median of 51.44 F (10.80 C). This is the highest minimum temperature of any studied location in the watershed. Near West Branch Briar Creek s confluence with Briar Creek, the latter creek has a temperature range of 32 to 67.46 F (0.00 to 19.70 C) and a median of 49.28 F (9.60 C). This is the lowest minimum temperature of any studied location in the drainage basin. Near its mouth, the temperature ranges from 32.18 to 80.96 F (0.10 to 27.20 C). The median temperature is 57.56 F (14.20 C). On West Branch Briar Creek, the temperature is a minimum of 35.6 F (2.0 C), a median of 57.38 F (14.10 C), and a maximum of 71.6 F (22.0 C). Two miles downstream of East Branch Briar Creek s headwaters, the median temperature is 49.73 F (9.85 C) and the range is 32.18 to 71.42 F (0.10 to 21.90 C). Adaptive decoding: with a seasonal mean range of 40.52 F (3.68 C). Seasonal variation in temperature, as well as production of sulfides, can significantly affect food source quality, quality of water, and the quality and quantity of sediment that may be trapped in the porous pipes and culverts that lead into groundwater. Relatively constant temperatures between 40 F (4 C) and 55 F (13 C) are common in the watershed. Temperatures above these extremes may occur occasionally when the watershed is dry, but the average annual maximum temperature in Briar Creek is 52.92 F (12.55 C). Average summer temperatures on the creek s western slopes are generally less than 60 F (15 C) throughout most of the year. Sustaining temperature above 60 F (15 C) is usually limited to the summer months and are relatively infrequent. Briar Creek drains nine watersheds with areas in the northern half of the watersheds containing more than 70% of the plants (wheat, barley, rice, rye and bulgur), but the predominant tree species (huckleberry, aspen, birch, and pine) is oak. Improving Open-Ended Text Generation via Adaptive Decoding Table 13. Case study: Text continuation with the Llama2-7B model. Entities are in blue fonts, logical relationships are in red, and coherent contents are highlighted in green. Prefix: Liszt found his method through two compositional practices, which he used in his symphonic poems. The first practice was cyclic form, a Human Continuation: procedure established by Beethoven in which certain movements are not only linked but actually reflect one another s content. Liszt took Beethoven s practice one step further, combining separate movements into a single-movement cyclic structure. Many of Liszt s mature works follow this pattern, of which Les pr eludes is one of the best-known examples. The second practice was thematic transformation, a type of variation in which one theme is changed, not into a related or subsidiary theme but into something new, separate and independent. Adaptive decoding: form that repeats the thematic material several times in different musical settings, as with the work of Haydn and Beethoven. It also was used in works such as Wagner s opera cycle, Der Ring des Nibelungen and Brahms Alto Rhapsody. Liszt s second practice was program music, which uses musical form in order to support programmatic concepts. Liszt applied both compositional practices to his symphonic poems, using cyclic form to construct musical episodes that reflect programmatic ideas. The music reflects the program throughout the piece. Prefix: Rebellions can occur if the loyalty of a particular province falls too low, with a rebel army appearing in the province to attempt to assume control Human Continuation: from the owners. Civil wars may also take place if several generals commanding large armies have sufficiently low loyalty. In the event of a civil war, the player is given the choice to back either the current rulers or the rebels. It had been planned to allow other factions who had established a prior claim to the throne by marriage to princesses to join in a civil war to claim the throne for themselves; however, this was never implemented. Naval warfare is carried out upon the campaign map, where ships can be built and organised into fleets. These fleets can be used to control the game s sea regions and form sea lanes, allowing trade and troop movement between provinces that have constructed a port. Fleets can engage in sea battles with foreign fleets, although unlike land battles these are resolved by the computer. Religion plays an important aspect in Medieval: Total War, with religious differences between the Catholic, Orthodox and Muslim factions affecting diplomacy and population loyalty. Adaptive decoding: . If the loyalty falls too far, the province will secede, and become a separate, unallied nation (this may sometimes be undesired, so it is important to know how to increase loyalty and to always pay attention to the needs of all provinces). As a province is loyal, its loyalty will be decreased by any war that the province takes part in, and will be increased by having any kind of benefit in the province. As a result of the fall of the West Roman Empire, most of the Western Mediterranean provinces were either conquered or claimed by the Christian Roman States, or by Germanic tribes such as the Visigoths, Ostrogoths or Vandals. The Eastern Mediterranean however is largely in the hands of Arabic Caliphates, or the Empire of Japan. The Western Mediterranean is the primary area where player wars begin, although it is possible to war in the eastern mediterranean, the arctic, or most recently, in the Indian Ocean. The Western Mediterranean however is also the richest area in the game, with a large amount of resources and a large number of wealthy cities which can be attacked. Prefix: In 1947 Bessie was elected to the Labour Party s National Executive Committee (NEC). She was generally identified with the left Human Continuation: wing of the party, and was for a time associated with a grouping known as the Socialist Fellowship , which espoused a programme of colonial freedom, workers control and reduced arms expenditure. She resigned from the Fellowship in 1950, along with fellow-MPs Fenner Brockway and Ellis Smith, when it condemned the United Nations intervention in the Korean War. She continued her wholehearted campaigning on behalf of the poorest in the country, pleading with parliament to remember the queues outside the Poor Relief offices , and castigating the New Look fashion of 1948 as wasteful, the ridiculous whim of idle people . Bessie s fiery reputation did not harm her electorally; in the February 1950 general election, with the Exchange constituency greatly increased by boundary changes, her majority rose to 5,344. Nationally Labour lost 76 seats, and its parliamentary majority was reduced to five. Attlee s second government was short-lived; Adaptive decoding: wing of the Labour Party. She was also active in organisations supporting Jewish refugees from the Nazis. With the help of another Jewish member of the Labour Party, Maurice Jacobson, Bessie had her own radio programme (Bessie Braddock s Book) and also became a popular speaker at meetings. In 1947 she was approached by the National Union of General and Municipal Workers (NUGM) and asked if she would run for parliament in an election. In the general election of 1950 Bessie Braddock was elected as Member of Parliament for Liverpool Exchange. She stood for the National Union of General and Municipal Workers, Labour Co-Operative and the Socialist Medical Association (of which she was a founder member). Her election campaign became more and more left wing. In Parliament she was very outspoken and unpredictable, using colourful language in her speeches. A political opponent once described her as the only man in the House of Commons . In 1954 she was appointed as parliamentary private secretary to Hugh Gaitskell.