# keywordguided_neural_conversational_model__cbdc6b1b.pdf Keyword-Guided Neural Conversational Model Peixiang Zhong,1,2 Yong Liu, 1,2 Hao Wang, 3* Chunyan Miao 1,2* 1 Alibaba-NTU Singapore Joint Research Institute, Nanyang Technological University (NTU), Singapore 2 Joint NTU-UBC Research Centre of Excellence in Active Living for the Elderly, NTU, Singapore 3 Alibaba Group, China {peixiang001, stephenliu, ascymiao}@ntu.edu.sg, cashenry@126.com We study the problem of imposing conversational goals/keywords on open-domain conversational agents, where the agent is required to lead the conversation to a target keyword smoothly and fast. Solving this problem enables the application of conversational agents in many real-world scenarios, e.g., recommendation and psychotherapy. The dominant paradigm for tackling this problem is to 1) train a next-turn keyword classifier, and 2) train a keyword-augmented response retrieval model. However, existing approaches in this paradigm have two limitations: 1) the training and evaluation datasets for next-turn keyword classification are directly extracted from conversations without human annotations, thus, they are noisy and have low correlation with human judgements, and 2) during keyword transition, the agents solely rely on the similarities between word embeddings to move closer to the target keyword, which may not reflect how humans converse. In this paper, we assume that human conversations are grounded on commonsense and propose a keyword-guided neural conversational model that can leverage external commonsense knowledge graphs (CKG) for both keyword transition and response retrieval. Automatic evaluations suggest that commonsense improves the performance of both next-turn keyword prediction and keyword-augmented response retrieval. In addition, both self-play and human evaluations show that our model produces responses with smoother keyword transition and reaches the target keyword faster than competitive baselines. Introduction Building a human-like open-domain conversational agent (CA) has been one of the milestones in artificial intelligence (AI). Early conversational agents are primarily based on rules (Weizenbaum 1966; Colby, Weber, and Hilf 1971), e.g., Eliza (Weizenbaum 1966), the first CA developed in 60 s, simulates a Rogerian psychotherapist based on hand-crafted pattern matching rules. In recent years, with the advancement of data-driven neural networks, neural open-domain conversational models are becoming dominant (Vinyals and Le 2015; Lowe et al. 2015; Gao, Galley, and Li 2018). *Corresponding author Copyright 2021, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Hi, I like drinking dr. pepper and comics drawings and you? Wow nice combination, I am eating pasta, italian is my favorite kind of food. I do love pasta but not egg noodles, since I eat only vegan. That is a good one are you a vegan? I prefer meat hamburger is amazing. I really love jelly sandwiches, but hold the pb because i'm allergic. Me too, I enjoy ham and cheese sandwiches with orange juice. Figure 1: Illustration of keyword-guided conversations from self-play simulations. Keywords are highlighted in bold. Given a random starting keyword comics , the agent (red) leads the conversation to the target keyword juice smoothly and fast. Recent efforts in open-domain neural conversational models are primarily aiming to improve the response diversity (Li, Monroe, and Jurafsky 2016; Zhang et al. 2018b) and endowing responses with knowledge (Zhou et al. 2018b; Dinan et al. 2019b), personality (Li et al. 2016a; Zhang et al. 2018a), emotion (Zhou et al. 2018a; Zhong, Wang, and Miao 2019) and empathy (Rashkin et al. 2019; Zhong et al. 2020). All the efforts mentioned above are focusing on models that passively respond to user messages. However, in many real-world scenarios, e.g., conversational recommendation, psychotherapy and education, conversational agents are required to actively lead the conversation by smoothly changing the conversation topic to a designated one. For example, during a casual conversation, the agent may actively lead the user to a specific product or service that the agent wants to introduce and recommend. In this paper, we follow the line of research in (Tang et al. 2019; Qin et al. 2020) and study the problem of imposing conversational goals/keywords on open-domain conversational agents, where the agent is required to lead the conversation to a target keyword smoothly and fast. As illustrated in Figure 1, given a target keyword juice and a random starting keyword comics , the agent is required to converse with the user in multiple exchanges and lead the conversation to juice . The challenge of this problem lies in how to balance the tradeoff between maximizing keyword transition smoothness and minimizing the number of turns taken to reach the target. On the one hand, passively responding to the user solely based on the conversation context would The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21) Hi, how was ur weekend? Hey, it was good! How was yours? Party weekend, it was amazing What kind of party? Kind of a get together with friends Nice. I like to ride my bike if I've time on the weekend Traffic is major hassle here. I get mad You should ride a bike instead of drive! Haha My work place is a bit far Where do you work? I sell insurance I work in a bank Figure 2: Illustration of keyword (in bold) transitions in a sample conversation from Conv AI2 (Zhang et al. 2018a). Transitions indicated by arrows are considered relevant. The rest keyword transitions, e.g., friends ride, are irrelevant (but used in the training and evaluation datasets of existing studies). achieve high smoothness but may take many turns to reach the target, but on the other hand, directly jumping to the target word by ignoring the conversation context would minimize the number of turns but produce non-smooth keyword transitions. Tang et al. (2019) proposed to break down the problem into two sub-problems: next-turn keyword selection and keyword-augmented response retrieval. Tang et al. (2019) proposed a next-turn keyword predictor and a rule-based keyword selection strategy to solve the first sub-problem, allowing the agent to know what is the next keyword to talk about given the conversation history and the target keyword. In addition, Tang et al. (2019) proposed a keywordaugmented response retrieval model to solve the second subproblem, allowing the agent to produce a response that is relevant to the selected keyword. However, there are two major limitations in existing studies (Tang et al. 2019; Qin et al. 2020). First, the training and evaluation datasets for next-turn keyword prediction are directly extracted from conversations without human annotations, thus, the majority of the ground-truth keyword transitions are noisy and have low correlations with human judgements. As illustrated in Figure 2, only a few keyword transitions in a conversation are considered relevant. In fact, in our human annotation studies of over 600 keyword transitions, we found that around 70% of keyword transitions in the next-turn keyword prediction datasets are rated as not relevant, which renders the trained next-turn keyword predictor in existing studies less reliable. Second, the rule-based keyword selection strategy primarily leverages the cosine similarity between word embeddings to select keywords that are closer to the target keyword. Word embeddings are trained based on the distributional hypothesis that words that have similar contexts have similar meanings, which may not reflect how humans relate words in conversational turn-taking. In this paper, we assume that human conversations are grounded on commonsense and propose a keyword-guided neural conversational model that can leverage external com- monsense knowledge graphs (CKG) for both next-turn keyword selection and keyword-augmented response retrieval. Humans rely on commonsense to reason, and commonsense reasoning plays an important role in the cognitive process of conversational turn-taking (Schegloff 1991; Stocky, Faaborg, and Lieberman 2004; Lieberman et al. 2004). Relying on a CKG for keyword transition would allow the agent to select a more target-related keyword for the nextturn. Moreover, we leverage commonsense triplets from the CKG using Graph Neural Networks (GNN) for both nextturn keyword prediction and keyword-augmented response retrieval to achieve more accurate predictions. In summary, our contributions are as follows: We identify two limitations of existing studies in nextturn keyword selection: 1) noisy training and evaluation datasets, and 2) unreliable keyword transition based on the similarity between word embeddings. For the first time in this task, we propose to use CKG for keyword transition and propose two GNN-based models to incorporate commonsense knowledge for next-turn keyword prediction and keyword-augmented response retrieval, respectively. We propose a large-scale open-domain conversation dataset for this task, obtained from Reddit. The linguistic patterns in Reddit are far more diverse than the Conv AI2 (Zhang et al. 2018a) used in existing studies, which are collected from only hundreds of crowd-workers. We conduct extensive experiments and the results show that grounding keyword transitions on CKG improves overall conversation smoothness and allows the agent to reach the target faster. In addition, leveraging commonsense triplets substantially improves the performance of both next-turn keyword prediction and keywordaugmented response retrieval. Finally, self-play and human evaluations show that our model produces smoother responses and reaches the target keyword faster than competitive baselines. Related Work In recent years, several studies proposed to build conversational agents that can actively lead a conversation to a designated target keyword/goal (Tang et al. 2019; Wu et al. 2019). Our work follows the task definition in (Tang et al. 2019), which has been discussed in Introduction. Very recently, Qin et al. (2020) improved (Tang et al. 2019) in 1) next-turn keyword prediction by only considering keyword transitions that are present in the training dataset and 2) keyword-augmented response retrieval by constraining that the selected response must contain the predicted keyword or a keyword closer to the target keyword. As a result, Qin et al. (2020) obtained the state-of-the-art performance on this task in terms of task success rate and transition smoothness. Another line of research (Wu et al. 2019) focused on the specific movie domain and proposed to use factoid knowledge graph to proactively lead the conversation from a random entity to a given entity. Our work differs from (Wu et al. 2019) in that 1) we focus on open-domain conversations whereas they focus on movie domain; 2) we leverage commonsense knowledge graph for keyword transitions whereas they leverage factoid knowledge graph for entity transitions1; and 3) we allow the target to be any arbitrary keyword whereas they constrain the target to be at most twohop away from the starting entity. Following the line of research in (Wu et al. 2019), Xu et al. (2020a) proposed to use hierarchical reinforcement learning (HRL) to incorporate factoid knowledge graph for high-level topic selection and low-level in-depth topic-related conversation. Xu et al. (2020b) proposed a framework to represent prior information as a conversation graph (CG) and leverage policy learning to incorporate the CG into conversation generation. Commonsense has been studied extensively in recent neural conversational models (Young et al. 2018; Zhou et al. 2018b; Zhang et al. 2020; Zhong et al. 2021). Zhou et al. (2018b) proposed graph attentions to statically incorporate one-hop knowledge triplets into conversation understanding and dynamically generate knowledge-aware responses. Recently, Zhang et al. (2020) extended (Zhou et al. 2018b) to multi-hop knowledge triplets by proposing an attention mechanism to incorporate outer triplets and a GNN model to aggregate central triplets. Different from existing studies that leverage commonsense to improve the diversity and informativeness of responses, we incorporate commonsense into our approach for more reasonable keyword transition and more accurate response retrieval. Our Approach In this section, we first introduce our task definition, and then describe the CKG used in our paper, and finally propose the Commonsense-aware Keyword-guided neural Conversational model (CKC). Task Definition Given a conversation history of n utterances: x1:n = x1, ..., xn, we denote the sequence of keywords for xi as ki, and the response to x1:n as y. Briefly, given a target keyword t and a random initial utterance x1 with its keywords k1, the task of the agent is to chat with the user and lead the conversation to the target keyword smoothly and fast. The target is only presented to the agent and unknown to the user. We consider the target is achieved when an utterance (either by the user or by the agent) mentions the target keyword2. We break down the task into two sub-problems: nextturn keyword selection and keyword-augmented response retrieval. We propose a CKG-aware next-turn keyword predictor and a CKG-guided keyword transition strategy to solve the first sub-problem. We then propose a CKG-aware keyword-augmented response retrieval to solve the second sub-problem. 1In our work, a keyword can be a named entity, e.g., AAAI2021, or a generic content word, e.g., conference. 2This is different from (Tang et al. 2019) where mentioning a synonym of the target can be considered as success because we found that synonyms are unreliable to measure the task success. keywords & concepts Hello, how are you today? Hi, i'm justin. I'm on tour and just came in from a performance. That's awesome, how do you like performing? I'm currently looking for a new job. It is my life. I was born into it. Both of my parents are musicians. Sounds like it was meant to be. My daughter loves to preform she's a prodigy really. living Part Of play_a_guitar Capable Of ... Gated Graph Neural Network Hierarchical GRU Hierarchical Pooling Concat, Linear & Softmax Word Embedding Word Embedding Figure 3: Illustration of our proposed CKG-aware next-turn keyword prediction. We only use the most recent two utterances and their concepts and keywords as input. Words in bold denote keywords. Concepts are words or multi-word expressions extracted from utterances based on the CKG vocabulary. Commonsense Knowledge Graph (CKG) In this paper, we use Concept Net (Speer, Chin, and Havasi 2017) as our CKG. Concept Net is a large-scale multilingual semantic graph that describes general human knowledge in natural language. Each node/concept on Concept Net can be a single word, e.g., food or a multi-word expression, e.g., having lunch . The edges on Concept Net represent the semantic relations between nodes and have weights suggesting the confidence score, e.g., having lunch, Has Prerequisite, food with a weight of 2.83. The majority of edge weights are in [0, 10]. We only include triplets that satisfy the following requirements into our CKG: 1) the edge weight is at least 1, 2) at least one node is in our keyword vocabulary3, and 3) the other node is in our word vocabulary4. CKG-Aware Next-Turn Keyword Prediction Given a history of n utterances x1:n and n sequences of keywords k1:n, we propose a model that can predict the nextturn keywords kn+1. Note that kn+1 can include multiple keywords, hence this is a multi-label classification problem. One major limitation of existing studies is that the training and evaluation datasets for next-turn keyword prediction are noisy, as discussed in Introduction. In this paper, we assume that human conversations are grounded on commonsense and leverage commonsense to 1) clean the training and evaluation datasets; and 2) propose a CKG-aware model for more accurate next-turn keyword prediction. Specifically, for each example in both training and evaluation datasets, we remove next-turn keywords that are not in the immediate neighborhood of historical keywords. During 3The keyword vocabulary is a subset of our word vocabulary containing frequent content words. 4For a multi-word expression, we require that each single word to be in our word vocabulary. keywords top keywords Great. Very good to start young. I've been playing guitar since I was three. Hello, how are you today? Hi, i'm justin. I'm on tour and just came in from a performance. That's awesome, how do you like performing? I'm currently looking for a new job. It is my life. I was born into it. Both of my parents are musicians. Sounds like it was meant to be. My daughter loves to preform she's a prodigy really. living Part Of Related To play_a_guitar Capable Of ... Gated Graph Neural Network Word Embedding Word Embedding Word Embedding Trained Next-Turn Keyword Predictor Context Utterance Representation concat concepts Candidate Utterance Representation concat concepts Utterance Matching Keyword Matching Figure 4: Illustration of our proposed CKG-aware response retrieval model. model prediction in both training and evaluation, we also only output keywords that are in the immediate neighborhood of input keywords. In other words, our model only outputs CKG-grounded keyword predictions. We then propose a CKG-aware model that takes as input xn 1, xn, kn 1, kn and the CKG, and output kn+1. Note that existing studies only use kn 1, kn and GRU (Cho et al. 2014) to predict kn+1 (Tang et al. 2019; Qin et al. 2020). Using longer context information does not improve performance in our experiments. An illustration of our model is presented in Figure 3. Utterance Representation We obtain the utterance representation x Rd1 from the contextual utterances xn 1 and xn using a hierarchical GRU (HGRU) encoder, where d denotes the final hidden state size of HGRU. CKG Graph Representation We obtain a CKG graph representation G RN d2 using a Gated Graph Neural Network (GGNN) (Li et al. 2016b), where N denotes the number of nodes in the CKG and d2 denotes the hidden size of GGNN. For each node on the CKG, the convolution operation in GGNN first computes a parameterized weighted average of neighboring node representations and then updates its own representation using a GRU. The nodes in CKG are represented via word embeddings. Multi-word nodes are represented via averaged word embeddings. The CKG representation is learned jointly with the next-turn keyword prediction and the gradients on the CKG are directly back-propagated to the word embeddings. Both utterances and CKG share the same word embedding layer, which can effectively reduce the number of model parameters and enable knowledge transfer on word embeddings. Keyword and Concept Representation We extract the keyword and concept representations K RNk d2 and C RNc d2 from G, respectively, where Nk = |kn 1| + |kn| and Nc denote the number of concepts in xn 1 and xn. Concepts are extracted from utterances via string matching with the CKG. We then apply hierarchical pooling where we first use mean pooling to aggregate K and C and obtain k Rd2 and c Rd2, respectively, and then apply max pooling to combine k and c and obtain the final representation kc Rd2. Essentially, kc represents the CKG-aware representation learned from the utterances xn 1 and xn. Classification Finally, we concatenate the utterance representation x Rd1 and the CKG-aware keyword and concept representation kc Rd2, and then feed it into a linear transformation layer, followed by a softmax layer. The entire model is optimized by minimizing the negative loglikelihoods of all ground-truth next-turn keywords. CKG-Guided Keyword Selection Strategy After obtaining a keyword distribution of the next utterance using the proposed next-turn keyword predictor, we propose a CKG-guided keyword selection strategy to select the most appropriate keyword for subsequent keyword-augmented response retrieval. Specifically, we select the keyword that is closer to the target than current keywords and has the highest probability. The distance between keywords is measured as the weighted path length between keywords on the CKG, computed by the Floyd-Warshall algorithm (Floyd 1962). Note that the edge weights on Concept Net correlate positively with concept relatedness. Hence, we apply a reciprocal operation to the weights before computing path lengths. Essentially, our proposed strategy allows the agent to chat smoothly (by selecting the most likely next-turn keyword) while leading the conversation closer to the target keyword (by traversing to the target keyword via the most reasonable path on the CKG). Keyword-Augmented Response Retrieval The last module in our approach is a keyword-augmented response retrieval model, as illustrated in Figure 4. At a highlevel, it is a response retrieval model that selects the best candidate response given the context utterances and the predicted keywords. Utterance Representations The context utterance representation X RNx d is obtained by the concatenation of two representations: 1) the flattened GRU encoded contextual representation and 2) the CKG-aware contextual concept representation, where Nx denotes the total number of tokens and concepts in the context and d denotes the hidden size of GRU and GGNN. Similarly, the candidate utterance representation Y RNy d is obtained by: 1) the GRU encoded candidate representation and 2) the CKG-aware candidate concept representation, where Ny denotes the total number of tokens and concepts in the candidate. Keyword Representations Besides utterance-based matching, we learn keyword-based matching to allow keyword-augmented response retrieval. To this end, we aim to select the candidate that best matches the predicted next-turn keywords given contextual utterances. Specifically, we first obtain the top predicted next-turn keywords using a trained next-turn keyword predictor. We then obtain the CKG-aware predicted keyword representation Kx RNkx d and candidate keyword representation Ky RNky d from GGNN, where Nkx and Nky denotes the number of predicted keywords and candidate keywords, respectively. In practice, following (Tang et al. 2019), we set Nkx = 3, allowing top-3 keywords to be matched with candidate keywords. Matching We compute the matching score su R between context utterance representation X RNx d and candidate utterance representation Y RNy d as follows: su = dot(max(X), max(Y)) (1) where max denotes max pooling along the sequence dimension, and dot denotes dot product. Similarly, the matching score sk R between predicted keyword representation Kx RNkx d and candidate keyword representation Ky RNky d is computed as follows: sk = dot(max(Kx), max(Ky)) (2) The final matching score s R is obtained as follows: s = su + λksk (3) where λk denotes a hyper-parameter controlling the weight for keyword scores. We optimize the entire response retrieval model by minimizing the negative log-likelihood of the ground-truth response among all candidates. Experimental Settings In this section, we introduce the datasets, evaluation metrics, baselines and model settings. Dataset Split #Conv. #Utter. #Key. Avg. #Key. Conv AI2 Train 8950 132601 2678 1.78 Valid 485 7244 2069 1.79 Test 500 7194 1571 1.50 Reddit Train 112693 461810 2931 2.27 Valid 6192 25899 2851 2.25 Test 5999 24108 2846 2.30 Table 1: Dataset statistics. #Key. denotes the number of unique keywords and Avg. #Key. denotes the average number of keywords per utterance. Dataset We use the Conv AI2 dataset proposed in (Zhang et al. 2018a; Dinan et al. 2019a) and preprocessed in (Tang et al. 2019) in our experiments. Conversations in Conv AI2 are open-domain and cover a broad range of topics. In addition, we collect a large-scale open-domain conversation dataset from the social media Reddit5. The proposed Reddit dataset is collected from casual chats on the Casual Conversation6 and Casual UK7 subreddits, where users chat freely with each other in any topic. Reddit is significantly larger and more diverse than Conv AI2. Following (Tang et al. 2019), we use TF-IDF and partof-speech (POS) features to extract keywords from both datasets. We use a maximum of 8 contextual utterances and each utterance is truncated to 30 tokens. The number of keywords for each utterance is capped at 10. We limit the vocabulary of both datasets to the most frequent 20K tokens. In the task of next-turn keyword prediction, we remove keyword transitions not covered by our CKG, as discussed in Our Approach. In addition, we remove self-loops, i.e., a keyword transit to itself, in both training and evaluation examples to prevent the model from predicting keywords that exist in the context, because predicting self-loops would not lead the conversation to the target. After preprocessing, the average number of keyword candidates for Conv AI2 and Reddit are 158 and 201, respectively. The number of nodes/edges on CKG are 87K/221K and 97K/273K for Conv AI2 and Reddit, respectively. The statistics of the two datasets are presented in Table 1. Evaluation Metrics Turn-Level Evaluation Following (Tang et al. 2019; Qin et al. 2020), we use R@k, the recall at position k (=1, 3, 5) over all neighboring keywords, and P@1, the precision at the first position, for next-turn keyword prediction. Note that we have a smaller set of candidate keywords than that in (Tang et al. 2019) because we only keep neighboring keywords as candidates. We use R@k, the recall at position k (=1, 3, 5) over all 20 candidate responses (a ground-truth response and 19 negative candidates), and MRR, the mean reciprocal rank, for keyword-augmented response retrieval. 5https://www.reddit.com/. We use the Pushshift dataset on Google Big Query. 6https://www.reddit.com/r/Casual Conversation 7https://www.reddit.com/r/Casual UK/ Conv AI2 Reddit Model R@1 R@3 R@5 P@1 R@1 R@3 R@5 P@1 Random 1.03 0.09 2.99 0.12 4.83 0.04 1.18 0.12 0.60 0.06 1.88 0.24 3.35 0.34 0.69 0.04 PMI 16.96 34.15 46.39 19.11 6.90 16.06 22.98 7.79 Neural 17.81 0.35 34.59 0.42 44.88 0.66 19.91 0.57 7.22 0.26 16.81 0.20 23.89 0.21 8.12 0.35 Kernel 16.23 0.50 32.07 0.84 42.62 0.76 17.57 0.87 7.38 0.17 17.10 0.28 24.81 0.70 8.24 0.22 DKRN 18.03 0.15 34.60 0.56 45.06 0.95 20.09 0.38 7.11 0.21 16.47 0.72 23.42 0.98 8.08 0.29 Ours (CKC) 19.31 0.44 36.26 0.45 46.32 0.57 21.98 0.66 8.23 0.31 17.83 0.25 24.89 0.12 9.17 0.28 Table 2: Test results (in %) for next-turn keyword prediction. Results are averaged over 3 random seeds. Conv AI2 Reddit Model R@1 R@3 R@5 MRR R@1 R@3 R@5 MRR PMI 48.67 0.25 75.88 0.49 86.38 0.15 64.74 0.26 45.31 0.70 68.93 0.37 79.75 0.46 60.42 0.50 Neural 47.93 0.47 75.53 0.62 86.36 0.20 64.25 0.38 44.96 0.21 68.75 0.27 79.59 0.23 60.18 0.22 Kernel 48.55 0.51 75.57 0.32 86.04 0.04 64.47 0.37 44.55 0.33 68.47 0.24 79.66 0.38 59.92 0.30 DKRN 48.44 0.34 75.78 0.20 86.83 0.16 64.64 0.17 44.92 0.45 68.84 0.45 79.59 0.65 60.19 0.44 Ours (CKC) 59.90 0.41 83.03 0.31 92.15 0.17 73.50 0.26 50.02 0.41 72.94 0.33 82.87 0.22 64.33 0.35 Table 3: Test results (in %) for keyword-augmented response retrieval. Results are averaged over 3 random seeds. Dialogue-Level Evaluation Following (Tang et al. 2019), we measure the target success rate (Succ.) and number of turns (#Turns) to reach the target for keyword-guided conversation evaluation using self-play simulations. We run self-play simulations for 1K conversations between each model and a base response retrieval model8. In addition, we measure target success rate (Succ.) and conversation smoothness (Smo.) using human evaluations with three annotators on 100 conversations for each model. The smoothness is rated in the [1, 5] scale, higher is better. Baselines and Model Settings We compare our model with the following baselines: PMI (Tang et al. 2019), Neural (Tang et al. 2019), Kernel (Tang et al. 2019) and DKRN (Qin et al. 2020). We follow their released implementations9. All baselines are trained and evaluated using the same filtered datasets as our model. We initialize the embedding layer of all models using Glo Ve embedding of size 200 (Pennington, Socher, and Manning 2014). All hidden sizes in GRU and GGNN are set to 200. We use one layer in GGNN and set λk = 0.01. We optimize our model using Adam (Kingma and Ba 2014) with batch size of 32, an initial learning rate of 0.001 and a decay rate of 0.9 for every epoch. Result Analysis In this section, we present the experimental results, model analysis, case study and limitations. Next-Turn Keyword Prediction The results for next-turn keyword prediction are presented in Table 2. Among all baselines except Random, the nonparameterized PMI performs worst, and Neural, Kernel and 8This model respond passively to messages, which is the same as the base model used in (Tang et al. 2019). 9We fixed a bug in DKRN where the keyword transition mask is obtained using train+valid+test datasets. Conv AI2 Reddit Model Succ. (%) #Turns Succ. (%) #Turns PMI 14.6 5.83 5.1 4.88 Neural 18.9 6.07 11.1 5.99 Kernel 20.7 5.89 10.6 5.83 DKRN 25.6 4.54 18.4 4.42 Ours (CKC) 28.9 4.23 22.7 4.19 Table 4: Self-play simulation results. Conv AI2 Reddit Model Succ. (%) Smo. Succ. (%) Smo. PMI 16.0 3.05 6.3 2.68 Neural 17.3 2.77 11.0 2.85 Kernel 22.3 2.88 12.3 2.57 DKRN 25.0 3.01 17.7 2.81 Ours (CKC) 29.3 3.27 22.3 3.08 Table 5: Human evaluation results. Smo. denotes conversation smoothness. DKRN performs comparably on both datasets. Our proposed model achieves consistent better performance than all baselines across all metrics and datasets, suggesting that incorporating CKG triplets into keyword prediction helps. Keyword-Augmented Response Retrieval The results for keyword-augmented response retrieval are presented in Table 3. The baselines differ in which nextturn keyword prediction model is used. It is surprising that all baselines perform comparably regardless of the nextturn keyword prediction model. This may suggest that the baselines are unable to effectively leverage the predicted keyword information into response retrieval. Our model achieves substantially better performance than all baselines on both datasets. The performance improvement can be primarily attributed to 1) we additionally incorporate utterancerelated CKG triplets into utterance representation learning; Next-Turn Keyword Prediction Model R@1 Ours (CKC) 19.31 0.44 - concepts 18.56 0.31 Keyword-Augmented Response Retrieval Model R@1 Ours (CKC) 59.90 0.41 - concepts 53.11 0.43 - keywords 52.30 0.54 Self-Play Simulation Model Succ. (%) #Turns Ours (CKC) 28.9 4.23 - CKG-based strategy 22.3 4.42 Table 6: Ablation study (in %) on Conv AI2. Target: music A: Hey, how are you doing? H: I m well, thanks. Working on a party I m planning. A: I am sitting here listening to pearl jam, my favorite band. H: Super cool! Do you sing? I was just singing in my shower. A: No, but I was in jazz band in hs. H: Congrats! I love music and playing my guitar and violin. A: That s awesome! However, my favorite is country music. Table 7: Case study from self-play simulations on Conv AI2. A denotes our model and H denotes the base model. and 2) we propose an additional keyword matching module to match the predicted keywords with candidate keywords, whereas baselines directly match predicted keywords with candidate utterances. Keyword-Guided Conversation The self-play simulation results for keyword-guided conversation are presented in Table 4. DKRN performs best among all baselines, which can be primarily attributed to its strategy of selecting keyword-related responses. This strategy requires a pool of confident candidates to select from. A larger pool will lead to higher success rate but lower smoothness because potentially less likely candidates can be selected. In all experiments, we set the pool size to 100. Our model also leverages this strategy but instead use weighted path lengths to measure keyword relatedness. Our model outperforms all baselines in both metrics on both datasets. Note that the success rates on Conv AI2 are consistently larger than that on Reddit across all models, which can be partially due to the higher next-turn keyword prediction accuracy on Conv AI2. The human evaluation results are presented in Table 5. The results for success rate are similar to that in self-play simulations. Among all baselines, DKRN has slightly more robust performance in smoothness on both datasets. Our model obtains consistently better performance in both success rate and smoothness on both datasets, suggesting that our model can select confident candidates that are also related to the target keyword. Model Analysis Table 6 presents the ablation study of our model across multiple tasks on the Conv AI2 test set. In both next-turn key- word prediction and keyword-augmented response retrieval, removing concepts representation from our model leads to degraded performance in R@1, suggesting that CKG triplets are helpful in learning the semantic representation of utterances. In keyword-augmented response retrieval, unlike other baselines that do not leverage keywords effectively, our model performs noticeably worse when keywords are removed, showing that our design of matching keywords separately indeed contribute to the overall matching. Finally, we examine the impact of our CKG-guided keyword selection strategy on self-play simulations. The results in Table 6 show that replacing our CKG-based strategy by the embedding-based strategy (Tang et al. 2019; Qin et al. 2020) leads to worse performance in both success rate and number of turns. We present a case study from our self-play simulations in Table 7. Our model can lead the conversation from a starting keyword party to the target keyword music smoothly and fast. Limitations One major limitation of existing approaches including ours is the mediocre accuracy of retrieving keyword-related responses (this is different from keyword-augmented response retrieval where the ground-truth responses do not necessarily correlate with the input keywords), which bottlenecks the overall target success rate. In fact, for both DKRN and our model, the target keyword can be successfully selected most of the time during self-play simulations, however, both models can not retrieve the keyword-related responses given the selected target keyword accurately. A potential solution to this problem is to train the keyword-augmented response retrieval model on datasets where input keywords and groundtruth responses are correlated, which is left to future work. We study the problem of imposing conversational goals/keywords on open-domain conversational agents. The keyword transition module in existing approaches suffer from noisy datasets and unreliable transition strategy. In this paper, we propose to ground keyword transitions on commonsense and propose two GNN-based models for the tasks of next-turn keyword transition and keywordaugmented response retrieval, respectively. Extensive experiments show that our proposed model obtains substantially better performance on these two tasks than competitive baselines. In addition, the model analysis suggests that CKG triplets and our proposed CKG-guided keyword selection strategy are helpful in learning utterance representation and keyword transition, respectively. Finally, both self-play simulations and human evaluations show that our model can achieve better success rate, reach the target keyword faster, and produce smoother conversations than baselines. Acknowledgments This research is supported, in part, by Alibaba Group through Alibaba Innovative Research (AIR) Program and Alibaba-NTU Singapore Joint Research Institute (JRI) (Alibaba-NTU-AIR2019B1), Nanyang Technological University, Singapore. This research is also supported, in part, by the National Research Foundation, Prime Minister s Office, Singapore under its AI Singapore Programme (AISG Award No: AISG-GC-2019-003) and under its NRF Investigatorship Programme (NRFI Award No. NRF-NRFI052019-0002). Any opinions, findings and conclusions or recommendations expressed in this material are those of the authors and do not reflect the views of National Research Foundation, Singapore. This research is also supported, in part, by the Singapore Ministry of Health under its National Innovation Challenge on Active and Confident Ageing (NIC Project No. MOH/NIC/COG04/2017 and MOH/NIC/HAIG03/2017). References Cho, K.; van Merrienboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; and Bengio, Y. 2014. Learning Phrase Representations using RNN Encoder Decoder for Statistical Machine Translation. In EMNLP, 1724 1734. Colby, K. M.; Weber, S.; and Hilf, F. D. 1971. Artificial paranoia. Artificial Intelligence 2(1): 1 25. Dinan, E.; Logacheva, V.; Malykh, V.; Miller, A.; Shuster, K.; Urbanek, J.; Kiela, D.; Szlam, A.; Serban, I.; Lowe, R.; et al. 2019a. The second conversational intelligence challenge (convai2). ar Xiv preprint ar Xiv:1902.00098 . Dinan, E.; Roller, S.; Shuster, K.; Fan, A.; Auli, M.; and Weston, J. 2019b. Wizard of Wikipedia: Knowledge-Powered Conversational Agents. In ICLR. Floyd, R. W. 1962. Algorithm 97: shortest path. Communications of the ACM 5(6): 345. Gao, J.; Galley, M.; and Li, L. 2018. Neural approaches to conversational AI. In SIGIR, 1371 1374. Kingma, D. P.; and Ba, J. 2014. Adam: A method for stochastic optimization. ar Xiv preprint ar Xiv:1412.6980 . Li, J.; Galley, M.; Brockett, C.; Spithourakis, G.; Gao, J.; and Dolan, B. 2016a. A Persona-Based Neural Conversation Model. In ACL, 994 1003. Li, J.; Monroe, W.; and Jurafsky, D. 2016. A Simple, Fast Diverse Decoding Algorithm for Neural Generation. ar Xiv preprint ar Xiv:1611.08562 . Li, Y.; Tarlow, D.; Brockschmidt, M.; and Zemel, R. S. 2016b. Gated Graph Sequence Neural Networks. In Bengio, Y.; and Le Cun, Y., eds., ICLR. Lieberman, H.; Liu, H.; Singh, P.; and Barry, B. 2004. Beating common sense into interactive applications. AI Magazine 25(4): 63 76. Lowe, R.; Pow, N.; Serban, I. V.; and Pineau, J. 2015. The Ubuntu Dialogue Corpus: A Large Dataset for Research in Unstructured Multi-Turn Dialogue Systems. In SIGDIAL, 285 294. Pennington, J.; Socher, R.; and Manning, C. D. 2014. Glove: Global vectors for word representation. In EMNLP, 1532 1543. Qin, J.; Ye, Z.; Tang, J.; and Liang, X. 2020. Dynamic Knowledge Routing Network for Target-Guided Open Domain Conversation. In AAAI, 8657 8664. Rashkin, H.; Smith, E. M.; Li, M.; and Boureau, Y.-L. 2019. Towards Empathetic Open-domain Conversation Models: A New Benchmark and Dataset. In ACL, 5370 5381. Schegloff, E. A. 1991. Conversation analysis and socially shared cognition. American Psychological Association . Speer, R.; Chin, J.; and Havasi, C. 2017. Concept Net 5.5: an open multilingual graph of general knowledge. In AAAI, 4444 4451. Stocky, T.; Faaborg, A.; and Lieberman, H. 2004. A commonsense approach to predictive text entry. In CHI, 1163 1166. Tang, J.; Zhao, T.; Xiong, C.; Liang, X.; Xing, E.; and Hu, Z. 2019. Target-Guided Open-Domain Conversation. In ACL, 5624 5634. Vinyals, O.; and Le, Q. 2015. A neural conversational model. ar Xiv preprint ar Xiv:1506.05869 . Weizenbaum, J. 1966. ELIZA a computer program for the study of natural language communication between man and machine. Communications of the ACM 9(1): 36 45. Wu, W.; Guo, Z.; Zhou, X.; Wu, H.; Zhang, X.; Lian, R.; and Wang, H. 2019. Proactive Human-Machine Conversation with Explicit Conversation Goal. In ACL, 3794 3804. Xu, J.; Wang, H.; Niu, Z.; Wu, H.; and Che, W. 2020a. Knowledge graph grounded goal planning for open-domain conversation generation. In AAAI, 9338 9345. Xu, J.; Wang, H.; Niu, Z.-Y.; Wu, H.; Che, W.; and Liu, T. 2020b. Conversational Graph Grounded Policy Learning for Open-Domain Conversation Generation. In ACL, 1835 1845. Young, T.; Cambria, E.; Chaturvedi, I.; Zhou, H.; Biswas, S.; and Huang, M. 2018. Augmenting end-to-end dialogue systems with commonsense knowledge. In AAAI, 4970 4977. Zhang, H.; Liu, Z.; Xiong, C.; and Liu, Z. 2020. Grounded Conversation Generation as Guided Traverses in Commonsense Knowledge Graphs. In ACL. Zhang, S.; Dinan, E.; Urbanek, J.; Szlam, A.; Kiela, D.; and Weston, J. 2018a. Personalizing Dialogue Agents: I have a dog, do you have pets too? In ACL, 2204 2213. Zhang, Y.; Galley, M.; Gao, J.; Gan, Z.; Li, X.; Brockett, C.; and Dolan, B. 2018b. Generating informative and diverse conversational responses via adversarial information maximization. In NIPS, 1810 1820. Zhong, P.; Wang, D.; Li, P.; Zhang, C.; Wang, H.; and Miao, C. 2021. CARE: Commonsense-Aware Emotional Response Generation with Latent Concepts. In AAAI. Zhong, P.; Wang, D.; and Miao, C. 2019. An Affect-Rich Neural Conversational Model with Biased Attention and Weighted Cross-Entropy Loss. In AAAI, 7492 7500. Zhong, P.; Zhang, C.; Wang, H.; Liu, Y.; and Miao, C. 2020. Towards Persona-Based Empathetic Conversational Models. In EMNLP, 6556 6566. Zhou, H.; Huang, M.; Zhang, T.; Zhu, X.; and Liu, B. 2018a. Emotional Chatting Machine: Emotional Conversation Generation with Internal and External Memory. In AAAI, 730 739. Zhou, H.; Young, T.; Huang, M.; Zhao, H.; Xu, J.; and Zhu, X. 2018b. Commonsense Knowledge Aware Conversation Generation with Graph Attention. In IJCAI, 4623 4629.