# collaborative_alignment_of_nlp_models__6dfa7207.pdf Collaborative Alignment of NLP models Fereshte Khani Microsoft fkhani@microsoft.com Marco Tulio Ribeiro Google Deep Mind marcotcr@gmail.com Despite substantial advancements, Natural Language Processing (NLP) models often require post-training adjustments to enforce business rules, rectify undesired behavior, and align with user values. These adjustments involve operationalizing "concepts" dictating desired model responses to certain inputs. However, it s difficult for a single entity to enumerate and define all possible concepts, indicating a need for a multi-user, collaborative model alignment framework. Moreover, the exhaustive delineation of a concept is challenging, and an improper approach can create shortcuts or interfere with original data or other concepts. To address these challenges, we introduce Co Align, a framework that enables multiuser interaction with the model, thereby mitigating individual limitations. Co Align aids users in operationalizing their concepts using Large Language Models, and relying on the principle that NLP models exhibit simpler behaviors in local regions. Our main insight is learning a local model for each concept, and a global model to integrate the original data with all concepts. We then steer a large language model to generate instances within concept boundaries where local and global disagree. Our experiments show Co Align is effective at helping multiple users operationalize concepts and avoid interference for a variety of scenarios, tasks, and models. 1 Introduction NLP models have showcased remarkable capabilities, yet they are not exempt from imperfections. Unacceptable values embedded in their training data, persistent errors, or violations of business rules highlight the need to teach certain concepts to these models. A concept relates a set of inputs to desired behaviors, e.g. in sentiment analysis, a concept may dictate that religion does not connote sentiment (e.g., I m Muslim is Neutral). Similarly, in natural language inference (NLI), the broader concept of downward monotonicity specifies entailment relations when certain parts of expressions are made more specific (e.g., All cats like tuna entails All small cats like tuna ). The standard way of teaching concepts to models is adding new training data that exemplifies the concept, e.g. adding neutral sentences containing religious words [1], or adding entailment pairs that illustrate downward monotonicity [2]. However, it is hard to ensure that the data provided does not lead to shortcuts, i.e. spurious correlations or heuristics that allow models to make predictions without capturing the true underlying concept, such as all sentences with religious terms are neutral , or going from general to specific leads to entailment . Further, the model may overfit fail to generalize from the instances provided to the actual concept, e.g. only recognizing pairs in the form ( all X... , all ADJECTIVE X... ), and not pairs like ( all animals... , all cats... ). Finally, both shortcuts and overfitting can lead to interference with the original data or other concepts, e.g. leading to failures on I love Islam or pairs like ( Some cats like tuna , Some small cats like tuna ). In sum, operationalizing concepts is challenging, since users typically cannot think of all concept boundaries and interactions in advance. Work done while at Microsoft 37th Conference on Neural Information Processing Systems (Neur IPS 2023). Figure 1: Co Align loop for operationalizing a single concept. (a) The user starts by providing some seed data from the concept and their labels, (b) they are used to learn a local concept model. (c) GPT-3 is then prompted to generate new examples, prioritizing examples where the local model disagrees with the global model. (d) Actual disagreements are shown to the user for labeling, and (e) each label improves either the local or the global model. The loop c-d-e is repeated until convergence, i.e. until the user has operationalized the concept and the global model has learned it. One possible solution is to ask domain experts to create data that covers the concept as comprehensively and precisely as possible, e.g. the GLUE diagnostics dataset [3] or the Fra Ca S test suite [4]. However, these datasets are often expensive to create, small (and thus not suitable for training), and not exhaustive, as even experts still fail to think about all aspects and nuances of a concept [5]. Another solution is to have users provide data incrementally while they receive feedback from the model, e.g. adversarial training [6] or adaptive testing [5]. These do not require users to think about everything in advance, and can expose and correct model weaknesses. However, adversarial training does not include the notion of concepts explicitly, and adaptive testing does not address the interaction between different concepts or between a concept and the original data. As a result, users may not explore concept boundaries efficiently, and thus do not know when they have sufficiently covered a concept or whether they have introduced interference with other concepts. In this paper, we introduce Collaborative Alignment of NLP Models (Co Align). Co Align leverages the collective knowledge of multiple users to cover many concepts instead of relying on a single user. Co Align allows users to collaborate with AI systems and other users in specifying concepts and teaching them to models. Our key insight is making concepts explicit by learning a local model for each concept, in addition to a global model that integrates the original data and all additional concepts. When operationalizing a new concept, we rely on a large language model (LLM) to generate new instances where the local and global model disagree, and ask users to label such examples (Figure ??). Intuitively, these examples are either cases where the local model is not yet fully specified, or where the global model still makes errors on the concept (due to overfitting or shortcut reliance). As users label these examples, both models are updated until convergence, i.e. until the concept has been learned in a way that does not conflict with prior data or prior concepts. In a sense, each local model is ever-improving cheap expert in its respective concept. The speed of prediction of local models and diversity of examples generated by the LLM enable users to explore the boundaries between concepts and existing data, an exploration that could be challenging for users to do without aid. Our experimental results demonstrate the effectiveness of Co Align in operationalizing concepts and handling interference. We first show Co Align outperforms Ada Test [5], a SOTA tool for debugging NLP models that also uses GPT-3 by revealing more bugs and fixing them without interference. We then demonstrate that Co Align operatinalize concepts even when user starts with biased data, outperforming a model that relies solely on data collection. We compare the data selection mechanism of Co Align to random selection and uncertainty sampling by running simplified version of Co Align where instead of using GPT-3 we iteratively select examples from an unlabeled pool. We show Co Align outperforms both baselines in teaching an NLI model about downwardand upwardmonotone concepts [3], as well as teaching a sentiment analysis model about Amazon products reviews. Finally, in a pilot study, we demonstrated that Co Align helped users clarify their concepts. Let x be a text string, and y a categorical label, e.g. sentiment (positive, negative, neutral). We assume there is a true function f(x) = y that we aim to approximate with a model ˆf(x). Assume we have access to a base dataset D0 = {(x1, y1), ..., (xn, yn)}, e.g. of movie reviews, from base distribution P0. We refer to the model trained on D0 as the base model ˆf0. A concept Ci is associated with a distribution Pi over the input space, e.g. a concept might place probability mass exclusively on sentences containing religious words. We say x Ci if Pi(x) > 0. Since it s hard for users to be exhaustive, we do not assume users can generate from Pi, but that they can label any x with f(x), and as to whether it is in Ci or not. We further assume users can provide a very small number samples in the support of Pi. Without loss of generality, we assume that we have k users developing a model collaboratively, each with their own concept. Our goal is to train a model ˆf that does well on both the base distribution and all concepts, i.e., minimizing 1 k+1 Pk i=0 Ex Pi[ ˆf(x) = f(x)]. 3 Collaborative Alignment of NLP models In this section we describe how a single user operationalizes their concept by producing a dataset Di in the context of an existing global model ˆf trained on the base dataset D0 and previous concepts D1:i 1. If ˆf0 is already aligned within the concept, i.e. ˆf0(x) = f(x) for all x Ci, we would be done and there would be no need for Di. Thus, what we really want is for Di to specify the boundary around failures , cases not currently handled by ˆf0. We abstract away the choice model and learning procedure, assuming that the model can learn the concept given Di (in practice, we use neural networks that are expressive enough). 3.1 Sampling from the concept Since by assumption we cannot sample from the concept distribution Pi, it is a challenge to find regions of Pi where ˆf0 fails. To address this, we use GPT-3 [7] as a generator G to simulate a random walk within the concept. To do so, we construct a prompt with m in-concept examples, and use this prompt as input to G to generate more samples. Then, we ask the user in the loop to accept or reject each generated sample x (x is accepted if x Ci), and also to label the accepted x with f(x ). The value of m controls the tradeoff between precision and recall, with high m generating more in-concept examples and low m exploring the space more broadly. Under some conditions it can be shown that G simulates a Markov chain with stationary distribution Pi (Appendix A), but the weaker condition of connectivity suffices for finding the concept failures, i.e. there must be a path between any x , x Ci with nonzero transition probabilities according to G and the prompt. That is, if the concept is connected, with enough time we should be able to find regions of Pi that are not already learned by ˆf0. While sampling from G eventually leads to the yet-unlearned concept regions, it is an inefficient use of human effort, as it does not use the user labels to guide generation (i.e. the user has to label many examples that the current model already handles correctly). A better approach would be ask the user to label in a way that maximizes the expected information gain for the concept, to which we now turn. 3.2 Local Concept Models Complex functions can be approximated by simpler functions in a local neighborhood, as evidenced by theoretical results (e.g. Taylor expansion) and empirical applications [8, 9]. Since a concept is a natural local neighborhood, we use this insight and learn a local model ˆfi to approximate f(x) in Ci. We present a toy example for intuition in Figure 2, where we show toy distributions P0 (Figure 2a) and P0 with an additional concept P1 (Figure 2b). In 2b, ˆf0 learned on samples from P0 (dashed line) predicts most of P1 correctly, except for a small region in the bottom left quadrant. However, we would need many random samples from P1 in order to learn that region, and reach the best model ˆf (solid line in Figure 2b). In contrast, we can learn a good local model ˆf1 for P1 with a trivial number of samples (Figure 2c). This local model can be used to produce a disagreement region between ˆf1 and ˆf0 (Figure 2d), and sampling from that region would lead to failure discovery much faster. (a) The initial data (P0) is elliptical Gaussian (b) The user concept (Pi) is spherical Gaussian (c) local function for the user data (d) disagreement regions between local and global models Figure 2: (a) A model is trained on two elliptical Gaussian (b) The spherical Gaussian are showing a user concept, who wants to find bugs in the model and teach the model about the new concept. However, since the model (dashed line) have high accuracy on the user concept it is hard to find bugs and teaching the model requires data from the low probability region. (c) We fit a classifier to the Spherical Gaussian which can be done with only a few data points (i.e., local model), (d) we then focus on disagreements between these two models to find bugs in the user concept. More generally, we define a score function as the disagreement between the local and global function. This score function is used to steer generation such as to maximize the score of generated samples x , by adding instances to the prompt for G with probability proportional to their score (similar to Ribeiro and Lundberg [5], who use a different score function). We note that models may present false agreement on some samples, i.e. ˆfi(x ) = ˆf0(x ) = f(x ). To account for this, we also sample from the agreement region sometimes, with a probability that decays over time as we gain more confidence in the local model. 3.3 Operationalizing a concept: from disagreement to convergence The local and global models disagree on regions where the concept has not yet been learned (global is wrong) or the local model does not fit the user s concept correctly (local is wrong). Thus, every label from the disagreement region results in an improvement in whichever model is incorrect. As labeling progresses, we update both the concept model ˆfi and the global model ˆf until we reach convergence (i.e. ˆf has learned the concept Ci). Note that while ˆfi is trained on data from Pi, ˆf is trained to do well on the base distribution P0, as well as on all concepts P1:k. (b) local fit (c) generation (d) selection Figure 3: This figure illustrates the main steps in Co Align: (a) The user starts by providing a small number of data points and their labels within their domain (represented by the black circle), (b) we fit a simple model (shown as a linear model) to represent the user s concept, (c) we use the generator (G) to generate data points towards the region where the local model disagrees with the global model, (d) a diverse set of data points are selected for the user to label, (e) based on the user s feedback, the local and global models are updated until convergence. We present pseudo-code for operationalizing a concept in Algorithm 1, and illustrate it in Figure 3. The local and global models are initialized in lines 1-2 (Figure 3a and b). In line 5 (Figure 3c), G is prompted such as to generate in-concept samples that are likely to be in the disagreement region. Then, in line 6 (Figure 3d), we select a batch of b examples for labeling according to the disagreement score of the generated instances. This generation-labeling loop is repeated for L iterations, after which both the local and global models get updated in lines 7-8 (Figure 3e). For the sake of interactivity, in practice we do not train from scratch when updating the models, and instead finetune them from a previous checkpoint. After a few iterations, ˆf and ˆfi converge (i.e. the user has operationalized the concept and the global model has learned it). If the number of generated samples b L is large enough, we can assume that no disagreement on generated sample between ˆf and ˆfi means they have converged, and thus we stop and output the generated data Di (line 9). Algorithm 1: Operationalizing a new concept i input :Base dataset D0, Concept datasets D1:i 1, A small set of samples from concept Di 1 Init local and global: Train ˆfi on Di, and train ˆf on D0:i 1; 3 for L iterations do 4 Generation: Prompt G with subset from Di chosen with probability | ˆf(x) ˆfi(x)| ; 5 Labeling: Select b samples with prob. | ˆf(x ) ˆfi(x )|. Users reject x if out of concept, or add to Di with label f(x ); 6 Update local and global: Train ˆfi on Di, and train ˆf on D0:i; 7 while Di was updated this round; output :A dataset Di 3.4 Handling interference between concepts When aligning a model to multiple users (or training a model on different distributions), two types of conflicts can happen: (1) Literal disagreements: where two users disagree in labeling the same input. (2) Interference: where over-generalization in one concept interfere with another concept (e.g., if a user adds a lot of neutral examples for the religion does not connote sentiment concept, then training the model on these new examples can cause the model to predict neutral for sentences like "I love Islam" which interfere with a concept that says love X is positive). As stated in Section 2, we assume there is a single true function f (thus no literal disagreement) and we only focus on handling interference. 2 Handling interference is crucial since any local change in ML models can interfere with other parts of the model [10 12]. Having local functions (cheap experts) enable us to check interference efficiently. Every time that a user operationalizes a concept according to Algorithm 1, we check the resulting global model ˆf against the local models for all previous concepts. To do so, we re-run Algorithm 1, only asking the user for labels if there is a newfound disagreement region. In practice, this means that a user adding a new concept needs to make sure it does not break any concepts from other users, a process similar to regression testing in software engineering. While we want to handle interference, having multiplicity of concepts can be beneficial in refining a concept. In particular, we want the global model to not overfit to a concept (i.e., only memorize the training data) and generalize well; however generalizing well is dependent on other concepts and the previous data. Take, for instance, a world where only bananas are yellow. A user might guide an ML model to recognize bananas solely based on their yellow color. Now if a new user introduces another yellow object, like corn, the model must discern other distinguishing features. Merely combining training data for two such concepts doesn t suffice (as shown in [11]); the boundaries must be distinguished. Furthermore, interference can be beneficial by exposing false agreement regions between the global model and any individual concept model. In other words, while both ˆf and ˆfi may be relying on the same shortcut to solve some regions of Pi, it is unlikely that this shortcut does not cause interference between ˆf and all local models ˆfj for j = i. In this case, the interference caused by adding Cj is actually beneficial, as it further refines concept Ci. In practice the original dataset (D0) is often very large and we cannot fine-tune the model on D0:k. Instead of choosing the whole D0, every time we sample data points with highest disagreement between ˆf0 and ˆf from D0. In other word, we treat ˆf0 as a user with concept distribution P0, this enable us to deal with interference with original model as well (as an example in Figure 2 we might only choose the data points from elliptical Gaussian in disagreement region). 2We like to note that while we haven t introduced a specific mechanism to resolve literal disagreements, our method can surface such disagreements. This can pave the way for resolution through discussions, voting, or even tweaking the model to reflect multiple perspectives, especially in cases where there isn t a consensus among users. 4 Linear regression analysis We now study Co Align in a noiseless overparametrized linear regression scenario. This particular setup is used in recent literature to gain some insights into the behaviors of deep networks [10, 11, 13, 14]. For more details and proofs see Appendix B. Setup. Let x Rd where only a subset the data points are valid. We assume θ Rd such that y = θ x. Let Si denote the smallest subspace containing all valid data points of Ci. Given k examples in Ci, let Sobv i denote the subspace observed by the training data, Suno i denote the unobserved subspace (thus Si = Sobv i + Suno i ), and Sinv i is the subspace that concept i does not have any variation in it. We consider overparametrized noiseless linear regression, where the number of features (d) exceeds the number of training data, enabling us to always interpolate all training data. We assume local and global models infer the min L2 norm interpolant. As a running example, let x R3 and consider a concept where data points belonging to that concept satisfies x1 = x2. Let s assume we observed x = [1, 1, 0] with label y = 2. In this case we have: Sobv 1 = {[1, 1, 0]}, Suno 1 = {[0, 0, 1]} and Sinv 1 = {[1, 1, 0]}. The min-norm solution interpolating the concept is ˆθi = [1, 1, 0]. Operationalizing a concept. The min-norm solution can also be constructed by introducing constraints that projection of ˆθi on Suno i and Sinv i has to be zero. In the above example, the minnorm solution is the unique answer of the following linear equations ([0, 0, 1]θ = 0, [1, 1, 0]θ = 0, [1, 1, 0]θ = 2). When we teach the global model about a concept, the naive combination of data points could violate these constraints, leading to disagreements between local and global models. For our running example, suppose the original data point is x = [0, 1, 1], y = 2. Combining the concept data point with the original data point for inferring global model results in ˆθglobal = [ 2 3], which leads to a disagreement between local and global model for data points varying in the [0, 0, 1] direction. In case of disagreement, two scenarios could arise: (1) the direction is non-zero and the local model needs more specification (happens a lot at the beginning), or (2) the direction is zero, but an explicit constraint is needed to prevent the global model from assuming other values. In our running example, the generator can help us to find a data point that the two model disagree 3. Let s assume the data point is x = [0, 0, 1] where local model predicts 0 but global model predicts 2 3. We show this data point to the user and either the user confirms [0, 0, 1] as non-zero or makes the zero explicit, resulting in a global prediction of ˆθglobal = [0, 2, 0]. Once we learn the local concept (i.e., all unobserved directions are indeed zero), how many of Suno i directions need to be added as explicit constraints? Intuitively, we do not need to query user for any direction in Suno i that has already been observed in Sobv 0 or are orthogonal Sobv 0 since there is no interference. The following proposition indicates the maximum number of disagreements for teaching global model about the local concept. Proposition 1. If proj Suno i (θ ) = 0, then the maximum number of disagreement between local and global models is dim(proj Sobv 0 (Suno i (Suno i Sobv 0 ) )). Handling interference between concepts. During the operationalization of concept i, we maintain Sobv 0 unchanged. However, in handling interference (Section 3.4), we add data to other concepts and the original data, potentially leading to new conflicts with concept i. Therefore, in addition to considering the projection of Suno i on observed subspaces we need to consider the unobserved subspaces as well. Similar to above, if an unobserved direction in one concept has been observed in another concept we do not need to query user. With notation of Sobv 0:k denoting sum of all the Sobv i , and S i denotes sum of all subspaces except i, the following proposition bounds number of times users need to add data to their concepts due to interference. Proposition 2. If for all i, proj Suno i (θ ) = 0 then the maximum number of times that we need to handle interference is Pk i=1 dim proj S i Suno i (Suno i Sobv 0:k ) . 3Recall that only some of the data points are valid thus we need the generator to find a data points with variation in [0, 0, 1] direction. Concept Examples Examples of bugs found by Co Align X person = not antonym (X) person How can I become a positive person? How can I become a person who is not negative? predicts duplicate shortcut bugs ( How can I become a mysterious person? How can I become someone with no mystery? predicts non-duplicate overfit bugs ( How can I become a blind person? How can I become someone who has lost his (physical) vision? Modifiers changes question intent Is Mark Wright a photographer? Is Mark Wright an accredited photographer? predicts not-duplicate shortcut bugs ( Is he an artist? Is he an artist among other people? predicts duplicate overfit bugs ( Is Joe Bennett a famous court case? Is Joe Bennett a famous American court case? Table 1: Examples of bugs found by Co Align in the concepts introduced by Check List, which were subsequently debugged using Ada Test, demonstrating that Ada Test had not yet fully operationalized these concepts. Corig: X = not antonym (X) , Cnew: Modifiers changes question intent Corig: X = synonym (X) , Cnew: less X = more antonym (X) Co Align Ada Test Co Align Ada Test broken by new concept 7/50 24/50 9/50 18/50 fixed by new concept 5/50 2/50 20/50 18/50 Table 2: Comparison of Co Align and Ada Test in terms of handling interference. For both methods, we labeled 50 sentences in the disagreement region of a model that only learned the original concept and a model that learned both original and the new concepts. Co Align outperforms Ada Test when the new concept conflicts (top) or is similar (bottom) to the original concept. 5 Experiments We first show Co Align outperforms Ada Test [5], a leading NLP debugging tool that also uses GPT3, by revealing more bugs and resolving them without interference. We then show Co Align can operationalize a concept even with biased seed data. We then show the selection mechanism of Co Align outperforms baselines such as random sampling and uncertainty sampling.We conclude with a small pilot study of using Co Align.4 Co Align outperforms Ada Test[5]. We consider a scenario where a user finds multiple bugs and wants to fix them. Following Ribeiro and Lundberg [5], we use Quora Question Pairs (QQP) dataset where the goal is to predict if two questions are duplicate or not. We finetune Ro BERTa-Large model on the QQP dataset, despite high accuracy of the model (92.2%), Ribeiro et al. [15] identified multiple concepts where the model has low performance. We use the 6 concepts with highest failure (see Table 1 for two of the concepts example). For each of the 6 concepts, Ada Test iteratively adds data with GPT-3 and adaptively find failures until finding failures becomes qualitatively difficult . For each concept, we initialize Co Align with the Ada Test generated data as Di. Even though the model has been debugged with Ada Test, Co Align quickly reveals 5 semantically meaningful sub-categories of bugs for each concept (with many failures within each sub-category). We show a few examples from different sub-categories in Table 1, which illustrate that Ada Test had not yet operationalized various regions of concepts where the model was still failing. We now compare Co Align with Ada Test in terms of handling interference. To do so, we pick pairs of concepts Corig, Cnew that might cause interference with one another, but that were debugged and reported as fixed by Ribeiro and Lundberg [5]. We then run Co Align on these pairs, noting that the output of both Ada Test and Co Align are small concept datasets Dorig and Dnew. We then train two models for each method, one finetuned on Dorig and one finetuned on the union of Dorig and Dnew. Finally, we generate new data from Porig and manually label 50 instances where there is disagreement between the models, to check if adding Dnew caused interference on Porig. We present the proportion of sentences that were broken (right to wrong) or fixed (wrong to right) when the new concept is added in Table 2, disregarding instances that are invalid, out of domain, or for which the label is unclear. The top pair seems more liable to interference between concepts, but we note that Ada Test data results in much more interference than Co Align data. In the bottom pair, 4Code and data will be released in https://github.com/fereshte-khani/Co Align. adding a concept with Co Align actually improves the original concept more often than interferes with it, while Ada Test data has a neutral effect. biased-SB SB Base 86.7 2.5 82.6 1.7 Data collection 98.6 0.9 80.7 1.6 Co Align 94.9 1.7 94.5 1.1 Table 3: When we have access to a biased dataset (biased SB) with consistently negative skin reviews and positive battery reviews. Co Align outperforms naive data collection across the entire skin and battery reviews (SB) dataset by efficiently conceptualizing concepts and avoiding shortcuts. Co Align works even with biased seed data. To start working with Co Align, a user should first provide some seed data (the first data points to train the local model). One natural question is that how much the model is dependent on the this seed data? This question is especially important when data is very biased and naive data collection might lead to very skewed data. Our goal is to understand if Co Align can cover a concept even when the initial data points are very biased. For evaluating this question, we simulate a concept consists of reviews containing the phrases battery life or my skin (from now on, we refer to this concept as SB). We then simulate a very extreme scenario where instances with battery life are always positive, and those with my skin are always negative, from now on we refer to this distribution biased-SB. We train a weak base model by finetuning BERT-base-uncased on reviews that contain the word install . Our goal is to see if Co Align can lead to good accuracy in SB while only having data points from biased-SB. We initialize Co Align with 10 instances from biased-SB (5 positive and 5 negative sentence). We then run Co Align for 5 rounds adding 10 data points in each round. To avoid the need of user labels, we train a high accuracy model as an oracle to simulate user labels. As an oracle, we train Ro BERTa-Large on the whole Amazon Review dataset where we only keep reviews with rating 5 (positive sentiment) and rating 1 (negative sentiment). The accuracy of oracle on validation dataset is 98.6%. As shown in Table 3, despite starting with biased samples, Co Align achieves high accuracy across all of SB. Conversely, the same number of random instances from biased-SB (simulating naive data collection) only increases accuracy for biased-SB while decreasing accuracy on the whole concept (SB). Qualitatively, GPT-3 starts generating in-concept instances without the bias, as that is where the local concept model disagrees with the base model. This controlled experiment illustrates how a generator focused on disagreements can be helpful in exploring the boundaries of concepts, even when we start from a biased sample. Figure 4: Co Align outperforms other data selection baselines when learning downward-monotone concept in MNLI task. MNLI Amazon Overall average Worst case Overall average Worst case Random 89.6 87.7 87.6 85.3 Uncertainty 91.6 90.6 87.9 85.9 Co Align 91.9 90.7 88.2 86.0 Table 4: Co Align outperforms data selection baselines when simultaneously learning two concepts for MNLI (upward and downward monotone) and 4 concepts (product categories) for sentiment analysis. Results shown are an average of 10 runs. Co Align outperforms random and uncertainty sampling. We adapt Co Align so that it selects instances from a pool of unlabeled data based on the disagreement score rather than using GPT-3 as a generator (lines 5-6 in Algorithm 1). We compare Co Align with two data selection baselines: random selection and uncertainty sampling [16, 17]. Each method selects a batch of 50 examples per iteration (Co Align selects examples randomly in the first batch), after which models are retrained. We use Ro BERTa-Large [18] finetuned on MNLI (binary) [19] as a base model, and use the downwardmonotone concept from an NLI monotonicity dataset [2] as a pool of unlabeled data. The base model starts with high accuracy on the base validation dataset (93.5%), and low accuracy on the concept (23.5%). We present accuracy on the concept over iterations in Figure 4. While accuracy on the base data remains constant throghout iterations, Co Align s disagreement-based sampling is more efficient than uncertainty sampling or random selection. In order to evaluate interference between concepts, we try to learn various concepts simultaneously. We use the same NLI model and concept as the previous paragraph, and add upward monotone [2] as an additional concept. We also use Ro BERTa-LARGE finetuned on the digital ebook category of Amazon Reviews as a base model, and use 4 additional product categories as concepts to be learned (pet-product, home, home-improvement, and wireless). We run 10 iterations of each method, with random sampling and uncertainty sampling selecting an equal number of instances from each concept per round. Table 4 shows the per-concept average and worst case accuracy of each method, where we note that Co Align produces better models on average. Task & Topic Seed data Initial rounds Final rounds Sentiment & Islam Lewis is a terrorist negative Alice practices Islam neutral Marco goes to the mosque neutral Alice practices radical Islam Alice is a fundamentalist Muslim Alice wants to convert everyone to Islam Alice prays for ISIS Lewis is Muslim extremist Lara, president of ISIS, said . .. Toxicity & Gender I m a man non-toxic men are superior to women toxic I m feminine women like to use emotion to manipulate Men have caused most of the wars Toxicity & Islam ISIS is violent non-toxic Muslims are violent toxic Dealing with Muslim countries can be tricky Taliban is violent Quron is a political manifesto Table 5: Some examples of a pilot study that qualitatively examines the effectiveness of the Co Align in assisting users in conceptualizing their ideas and debugging the model. Users typically begin with simple sentences in their concept as seed data along with their labels. As users continue to use Co Align, the Co Align suggestions (disagreements between locally fitted model and the global model) become increasingly complex showing the diverse coverage of GPT-3. Pilot Study (qualitative). We conducted a very small pilot study to show humans need assistance to operationalize their concept (i.e., we show that users might not even know the exact boundaries of their concept beforehand). Four users (one computer scientist and three with no prior knowledge of AI) used Co Align to align a model within their chosen concept in either sentiment analysis or toxicity task. Each participant interacted with Co Align for 5-7 rounds, and reported an improved alignment in their concept and an increase in sentence complexity over time. For example, for the Sentiment & Islam, the user-provided seed data matched the global model perfectly, indicating no initial bugs. However, in the first round of using Co Align, some disagreement were found between the fitted local model and the global model. The user identified some of these as bugs in the global model (e.g. Alice practices Islam daily" where the global model predicted negative) and some as bugs in the local model (e.g. Alice is a radical Islamist" where the local model predicted neutral). As the user made repeated adjustments to both models, the disagreements gradually centered around the concept boundaries, reaching a point where users could no longer determine the correct behavior. This pattern of initial misalignment correction followed by uncertainty in label determination was consistent across all users, suggesting successful concept operationalization in clear label regions. Table 5 provides examples of user interactions with Co Align5. 6 Related Work Our work relates to three areas of research: debugging, alignment, and interference management. We briefly review the relevant literature and highlight the differences and contributions of our work. Debugging. Numerous efforts aim to find and fix models failures. Checklist [15] uses templates to test and improve models in different areas; however, these templates have low coverage for finding bugs. Dynabench [6] iteratively discovers bugs in a model by using human-generated adversarial examples; however, this approach requires human creativity.The closest work to us is Ada Test [5], which uses an LLM (and a few prompts) as a proxy for the user s behavior. Thus, unlike Co Align Ada Test proxies do not adapt to the user s feedback and is susceptible to the biases and limitations of the LLM, resulting in lower performance than Co Align. Moreover, Ada Test do not consider the interference among different concepts. 5We acknowledge that in sentiment and toxicity tasks a single instance can have multiple acceptable labels according to different users [20, 21], here we show that even a single user needs assistance to understand the boundary of their concept. Alignment. The objective is to align a model with human intentions, a complex task since humans often cannot articulate their intentions. A prevalent method to tackle this is Reinforcement Learning with Human in the Loop (RLHF) [22 25], where the model learns a reward function based on human feedback on various outputs. The key distinction between our work and RLHF lies in our use of a local function for each concept, rather than a universal function for all concepts, and our generation of inputs in the model s disagreement region to assist users in better operationalizing their concept. Constitutional AI [26] works with multiple "expert" models in different domains, but they aggregate all the data generated by these models without addressing potential interference between them. Interference management. Interference is a common problem in ML since it is hard to change a model s behavior locally without affecting other areas. The trade-off between accuracy and robustaccuracy where robust-training led to decrease in accuracy shown to be mitigated with self-training with the original model and using unlabeled data [11, 27]. Unlike their work, we do not have access to a reliable model for self-training and we need to improve the models while handling interference. Another type of interference is catastrophic forgetting [28, 29], in which learning on a new task may degrade the performance on the previously learned tasks. Some possible mitigation is multi-task learning [30, 31], or weight averaging between the original and the fine-tuned model [32]. Unlike these works, we are interested in exploiting the interference between models, as they can help the user operationalize their concept in the context of the model better. Lastly, this work primarily addresses interference arising from shortcut learning, there may be literal disagreements among users[20, 21]. Although a crucial issue, it falls outside the scope of this paper. However, we like to note that our method can surface such disagreements. This could potentially facilitate resolution through discussions, voting, or modifying the model to reflect a range of perspectives, especially in situations where there isn t a consensus among users. 7 Conclusion Specifying model behavior on specific concepts is crucial in the development of NLP models, as it allows users to encode business rules, fix undesirable behavior, and force alignment with user values. Operationalizing concepts in a way that avoids shortcuts, overfitting, and interference with prior data is challenging. In this paper we presented Co Align, a framework that leverages local concept models and large language models to help users operationalize concepts effectively while avoiding interference. We showed that Co Align is more effective than prior work at exploring problematic concept regions, even prior work that uses the same language model and relies on interactive user feedback. We envision a future where NLP models are developed in a collaborative fashion, similar to open source software or Wikipedia, and speculate that harnessing the perspectives and expertise of a large and diverse set of users would lead to better models, both in terms of overall quality and in various fairness dimensions. For this scenario to materialize, we need ways to help users express their knowledge, and verify the impact of their proposed changes to models (the equivalent of diffs or regression tests ). We believe Co Align is a step in this direction. 8 Broader Impact and Limitations Co Align aids in operationalizing concepts without filtering the values a user wishes the model to align with, which might inadvertently allow a malicious user to encode harmful behavior into the NLP model, a risk for which we currently have no safeguards. Next, we only handled interference that arises from machine learning shortcomings and can be addressed by adding more data. However, there might be literal disagreements between users (i.e., two users prefer different labels for the same sentence). Although our method can surface such disagreements, we lack a definitive solution to resolve disagreements between users. Furthermore, our efforts were primarily on classification tasks, leaving out generative tasks (e.g., next word prediction). A possible workaround is to use a local classifier where a user only indicates whether an output aligns with their concept or not, then train the global model accordingly (similar to Ouyang et al. [23] but with multiple reward models instead of just one). Lastly, our theoretical framework is limited but our goal was to gain some initial insights into why interference occurs and estimates the number of instances required to address it. Handling malicious users, resolving literal disagreements, studying Co Align for generative tasks and more general theoretical analysis of alignment are compelling further research directions. 9 Acknowledgment We gratefully acknowledge the contribution of Scott Lundberg, whose insightful discussions and assistance in utilizing the Ada Test repository greatly enhanced our research. Further, we thank Brent Hecht and Zexue He for providing invaluable early feedback on this work. [1] Lucas Dixon, John Li, Jeffrey Sorensen, Nithum Thain, and Lucy Vasserman. Measuring and mitigating unintended bias in text classification. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, pages 67 73, 2018. [2] Hitomi Yanaka, Koji Mineshima, Daisuke Bekki, and Kentaro Inui. Do neural models learn systematicity of monotonicity inference in natural language? ar Xiv preprint ar Xiv:2004.14839, 2020. [3] Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel Bowman. GLUE: A multi-task benchmark and analysis platform for natural language understanding. In Proceedings of the 2018 EMNLP Workshop Blackbox NLP: Analyzing and Interpreting Neural Networks for NLP, pages 353 355, Brussels, Belgium, November 2018. Association for Computational Linguistics. doi: 10.18653/v1/W18-5446. URL https://aclanthology. org/W18-5446. [4] Robin Cooper, Dick Crouch, Jan Van Eijck, Chris Fox, Johan Van Genabith, Jan Jaspars, Hans Kamp, David Milward, Manfred Pinkal, Massimo Poesio, et al. Using the framework. Technical report, Technical Report LRE 62-051 D-16, The Fra Ca S Consortium, 1996. [5] Marco Tulio Ribeiro and Scott Lundberg. Adaptive testing and debugging of nlp models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3253 3267, 2022. [6] Douwe Kiela, Max Bartolo, Yixin Nie, Divyansh Kaushik, Atticus Geiger, Zhengxuan Wu, Bertie Vidgen, Grusha Prasad, Amanpreet Singh, Pratik Ringshia, Zhiyi Ma, Tristan Thrush, Sebastian Riedel, Zeerak Waseem, Pontus Stenetorp, Robin Jia, Mohit Bansal, Christopher Potts, and Adina Williams. Dynabench: Rethinking benchmarking in NLP. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 4110 4124, Online, June 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.naacl-main.324. URL https:// aclanthology.org/2021.naacl-main.324. [7] Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. Language models are few-shot learners. Advances in neural information processing systems, 33:1877 1901, 2020. [8] Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. "why should i trust you?": Explaining the predictions of any classifier. In Knowledge Discovery and Data Mining (KDD), 2016. [9] Scott M Lundberg and Su-In Lee. A unified approach to interpreting model predictions. Advances in neural information processing systems, 30, 2017. [10] Fereshte Khani and Percy Liang. Removing spurious features can hurt accuracy and affect groups disproportionately. In ACM Conference on Fairness, Accountability, and Transparency (FAcc T), 2021. [11] Aditi Raghunathan, Sang Michael Xie, Fanny Yang, John C. Duchi, and Percy Liang. Adversarial training can hurt generalization. ar Xiv preprint ar Xiv:1906.06032, 2019. [12] Megha Srivastava, Besmira Nushi, Ece Kamar, Shital Shah, and Eric Horvitz. An empirical analysis of backward compatibility in machine learning systems. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 3272 3280, 2020. [13] Preetum Nakkiran. More data can hurt for linear regression: Sample-wise double descent. ar Xiv preprint ar Xiv:1912.07242, 2019. [14] Aditi Raghunathan, Sang Michael Xie, Fanny Yang, John Duchi, and Percy Liang. Understanding and mitigating the tradeoff between robustness and accuracy. ar Xiv preprint ar Xiv:2002.10716, 2020. [15] Marco Tulio Ribeiro, Tongshuang Wu, Carlos Guestrin, and Sameer Singh. Beyond accuracy: Behavioral testing of NLP models with Check List. In Association for Computational Linguistics (ACL), pages 4902 4912, 2020. [16] David D Lewis. A sequential algorithm for training text classifiers: Corrigendum and additional data. In Acm Sigir Forum, volume 29, pages 13 19. ACM New York, NY, USA, 1995. [17] Vu-Linh Nguyen, Mohammad Hossein Shaker, and Eyke Hüllermeier. How to measure uncertainty in uncertainty sampling for active learning. Machine Learning, 111(1):89 122, 2022. [18] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized bert pretraining approach. ar Xiv preprint ar Xiv:1907.11692, 2019. [19] Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R Bowman. Glue: A multi-task benchmark and analysis platform for natural language understanding. ar Xiv preprint ar Xiv:1804.07461, 2018. [20] Ruyuan Wan, Jaehyung Kim, and Dongyeop Kang. Everyone s voice matters: Quantifying annotation disagreement using demographic information. ar Xiv preprint ar Xiv:2301.05036, 2023. [21] Aida Mostafazadeh Davani, Mark Díaz, and Vinodkumar Prabhakaran. Dealing with disagreements: Looking beyond the majority vote in subjective annotations. Transactions of the Association for Computational Linguistics, 10:92 110, 2022. [22] Paul Christiano, Jan Leike, Tom B. Brown, Miljan Martic, Shane Legg, and Dario Amodei. Deep reinforcement learning from human preferences. In Advances in Neural Information Processing Systems (Neur IPS), 2017. [23] Long Ouyang, Jeff Wu, Xu Jiang, Diogo Almeida, Carroll L Wainwright, Pamela Mishkin, Chong Zhang, Sandhini Agarwal, Katarina Slama, Alex Ray, et al. Training language models to follow instructions with human feedback. ar Xiv preprint ar Xiv:2203.02155, 2022. [24] Daniel M Ziegler, Nisan Stiennon, Jeffrey Wu, Tom B Brown, Alec Radford, Dario Amodei, Paul Christiano, and Geoffrey Irving. Fine-tuning language models from human preferences. ar Xiv preprint ar Xiv:1909.08593, 2019. [25] Wangchunshu Zhou and Ke Xu. Learning to compare for better training and evaluation of open domain natural language generation models. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 9717 9724, 2020. [26] Yuntao Bai, Saurav Kadavath, Sandipan Kundu, Amanda Askell, Jackson Kernion, Andy Jones, Anna Chen, Anna Goldie, Azalia Mirhoseini, Cameron Mc Kinnon, et al. Constitutional ai: Harmlessness from ai feedback. ar Xiv preprint ar Xiv:2212.08073, 2022. [27] Yair Carmon, Aditi Raghunathan, Ludwig Schmidt, Percy Liang, and John C. Duchi. Unlabeled data improves adversarial robustness. In Advances in Neural Information Processing Systems (Neur IPS), 2019. [28] Michael Mc Closkey and Neal J Cohen. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of learning and motivation, volume 24, pages 109 165. Elsevier, 1989. [29] James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, et al. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences, 114(13):3521 3526, 2017. [30] Yu Zhang and Qiang Yang. An overview of multi-task learning. National Science Review, 5(1): 30 43, 2018. [31] Emilio Parisotto, Jimmy Lei Ba, and Ruslan Salakhutdinov. Actor-mimic: Deep multitask and transfer reinforcement learning. ar Xiv preprint ar Xiv:1511.06342, 2015. [32] Gabriel Ilharco, Mitchell Wortsman, Samir Yitzhak Gadre, Shuran Song, Hannaneh Hajishirzi, Simon Kornblith, Ali Farhadi, and Ludwig Schmidt. Patching open-vocabulary models by interpolating weights. ar Xiv preprint ar Xiv:2208.05592, 2022. [33] Suriya Gunasekar, Blake E Woodworth, Srinadh Bhojanapalli, Behnam Neyshabur, and Nati Srebro. Implicit regularization in matrix factorization. Advances in Neural Information Processing Systems, 30, 2017. A Stationary distribution of the generator Let PG denote the transition probability of the generator, where PG(x | r = x ) denote probability of generating x condition on the prompt being x . We want to create a Markov chain to simulate a random walk within the user s distribution (Pi). Specifically, we begin with a user-provided seed data point x and use it as a prompt for G to generate a new data point x . If Pi(x ) > 0, we accept x , otherwise we remain at x and repeat the process. We are interested in conditions under which the stationary distribution of this markov chain is Pi Let s assume that support of Pi is finite and denote it by S, let P G(x |x) be the transition function over S S where P G(x | x) = PG(x | x) + P PG(x | x)I[Pi(x ) = 0] and for x, x S where x = x we have P G(x | x ) = PG(x | x ). If the transition graph generated by P is irreducible (any state can be reached from any other state) and all its states are positive recurrent (the expected time to return to a state is finite), then the unique stationary distribution using G is Pi if the following equality holds: Ex Pi [P (x | x)] = Pi(x ) (1) The above statement states that the probability of a data point (Pi(x)) should be proportional to the probability of reaching to that point with the transition function of P . If instead of only one data point we use m data points for prompt, we can create a graph where each node is m data points, and then analyse the stationary distribution of Markov chain on such graph. In this case, when we start from a node with m examples and prompt the language model with them in this case the probability of going to (x , xm, . . . , x2) from (xm, . . . , x1) is equal to P G(x | r = (xm, . . . , x1)). B Linear regression analysis In this section, we examine the performance of Co Align in a simple linear regression scenario. Specifically, we aim to investigate following aspects: (1) the number of data points required to teach a local concept to a global model, and (2) the reasons behind interference among concepts and the number of steps necessary to resolve it. We consider each input x Rd where only some of the data points are valid. There exist a true function θ Rd such that y = θ x. The support of each concept (Pi) lays on a subspace (Si) and all valid data points on that subspace belongs to Ci. Given k examples in Ci, let Sobv i denote the subspace observed by the training data, Suno i denote the unobserved subspace, thus Si = Sobv i +Suno i is the smallest subspace containing all data points in Ci. Finally Sinv i denote the subspace that concept i does not have any variation in it. As a running example, let x R3 and consider a concept where data points belonging to that concept satisfies x1 = x2. Recall that only some of the data points in this subspace are valid e.g., a point is valid if x1 is odd thus [1, 1, 0] is valid while [2, 2, 1] is not. Let s assume we observed x = [1, 1, 0] in that subspace with label y = 2. In this case we have: Sobv 1 = [1, 1, 0], Suno 1 = [0, 0, 1] and Sinv 1 = [1, 1, 0]. We consider the overparametrized noiseless linear regression, where number of features (d) is larger than number of acquired training examples (n) ( therefore, we can always interpolate all the training data) and there is no noise in observed targets. Following work of [33] which showed gradient descent on linear regression lead to min L2-norm, we assume local and global models infer the min L2 norm interpolant. As an example, for our running example the min-norm solution interpolating the concept is ˆθ = [1, 1, 0]. B.2 Operationalizing a concept: from disagreement to convergence An alternate interpretation of the min-norm involves inferring the parameters by taking into account explicit constraints that require ˆθ s projection on Suno i and Sinv i to be zero. For instance, in our current example, we can deduce the min-norm solution by solving these linear equations: ([0, 0, 1]θ = 0, [1, 1, 0]θ = 0, [1, 1, 0]θ = 2). These constraints are generally valid as the unseen directions often do not affect the output. However, these constraints may be violated when we combine local concept data with global data, as the projection of Suno i and Sobv 0 may not be zero. This implies that the output could change with variations in the unseen directions, leading to local models typically outperforming global models within a local concept. To ensure both local and global models perform equally well in the local concept, we need to enforce the invariance constraints explicitly. This involves adding new data that exhibit variations in the unseen directions and demonstrating that these variations do not affect the output. Furthermore, we presume that Suno i is significantly large, making methods that attempt to examine all possible directions inefficient. Therefore, it s more advantageous to only verify directions that are affected by the merge. Consider the previous example where we observed x = [1, 1, 0], y = 2 for the local concept. Now, imagine the we observed x = [0, 1, 1], y = 2 in global dataset. When we combine this data point with the concept data point, we get ˆθglobal = [ 2 3]. This causes a disagreement in data points that vary in the [0, 0, 1] direction within the local concept, the local model predicts 0 while the global model predicts 2 3. Note that both the global and local predictions align for variations in the [1, 1, 0] direction. In the event of such a disagreement, we have two options: (1) The variation in this direction is indeed non-zero, suggesting the local model requires further refinement - a frequent occurrence in early stages, or (2) The variation is zero, but it needs to be specified as such; otherwise, the global model assumes other values due to its implicit bias towards generating the simplest model. Note that there is no disagreements in the common directions between Suno 0 and Suno i or their orthogonal subspaces. Referring to the above example, the generator identifies a data point where the two models disagree. Let s assume this data point is x = [0, 0, 1], where the local model predicts 0, but the global model predicts 2 3. In such a case, we present this data point to the user. Let s assume user specify that the label for this data point is 0. In this case by adding this new data point the global prediction adjusts to ˆθglobal = [0, 2, 0]. After we learn the local concept (i.e., all the unobserved directions are indeed zero), how many of them do we need to add as explicit constraints? the following proposition shows maximum number of disagreements after learning a local concept. Proposition 1. If proj Suno i (θ ) = 0, then the maximum number of disagreement between local and global models is dim(proj Sobv 0 (Suno i (Suno i Sobv 0 ) )). Proof. The global and local models agree on all observed directions (i.e., Sobv i and Sobv 0 ). However, there is a disagreement for any vector u in Suno i such that ˆθglobal = proj Sobv 0 (θ ) u = 0 since ˆθ i u = 0. Let s assume we add k examples such that local and global disagree. We now prove that k dim(proj Sobv 0 (Suno i (Suno i Sobv 0 ) )). For the k added examples, only consider their components in (Suno i (Suno i Sobv 0 ) ) (we can remove the Sobv i components by subtracting their projection on Sobv i similarly remove any component in (Suno i Sobv 0 ) by subtracting their projection in Sobv 0 ). In order to have a disagreement these data points should have non-zero projection on Sobv i otherwise there will be no disagreements. As a result the maximum number of data points is dim(proj Sobv 0 (Suno i (Suno i Sobv 0 ) )). B.3 Handling interference between concepts In previous section, we explained why disagreement can happen between local and global model and how we can resolve the disagreements by querying user of the local concept. We bound number of disagreement with dimension of projection of Suno i on Sobv 0 . In previous section we did not need to change Sobv 0 but when concept j has conflicts with concept i we also add data to concept j (thus changing Sobv j ) which can lead to new conflicts with concept i. Figure 5: Co Align loop for operationalizing a concept. (a) The user starts by providing some seed data from the concept and their labels, they are then used to learn a local concept model. (b) GPT-3 is then prompted to generate new examples, prioritizing examples where the local model disagrees with the global model. (c) Actual disagreements are shown to the user for labeling, and (d) each label improves either the local or the global model. The loop b-c-d is repeated until convergence, i.e. until the user has operationalized the concept and the global model has learned it. Each local model acts as a proxy for the user and step b always check if global and local model are aligned with each other. Therefore, if user on the left changes the global model such that it causes interference with user on the right step b finds such interference and asks the right user to intervene and add more data. The following proposition state that in addition to the dimension of projection of Suno i on observed subspace we also need to calculate projection on the unobserved space of different concepts as they might get added in the future. With notation of Sobv 0:k denoting sum of all the Sobv i , and S i denotes sum of all subspaces except i, the following proposition bounds number of times users need to add data to their concepts due to interference. Proposition 2. If for all i, proj Suno i (θ ) = 0 then the maximum number of times that we need to handle interference is Pk i=1 dim proj S i Suno i (Suno i Sobv 0:k ) . Proof. The proof is similar to Proposition 1. Here we need to deal with conflicts with all other topics and since it is possible that we add their unobserved subspace as well we need to compute the dimension of Suno i on the whole Sj subspace not only Sobv j . Let assume we added t example from concept i to handle interference, we now prove that t dim(proj S i(Suno i (Suno i Sobv 0:k ) )). For every data point that we add we first remove Sobv 0:k components by removing its projection on Sobv 0:k . Now in order to have a conflict this data point should have non-zero projection on S i. As a result the maximum number of data points we can add is less or equal than dim(proj S i(Suno i (Suno i Sobv 0:k ) )), summing over all the concept result in maximum number of interference that needs to be handled. C Extra figures and tables Figure 5 is showing the Co Align loop for multiple users. Table 6 shows the 6 concepts used in our experiment along with examples for each concept. Concept Check List Examples Roberta-Large accuracy Synonyms in simple templates How can I become more vocal? How can I become more outspoken? More X = Less antonym(X) How can I become more optimistic? How can I become less pessimistic? X person = not antonym(X) person How can I become a positive person? How can I become a person who is not negative? Orders is irrelevant in symmetric relations Are tigers heavier than insects? What is heavier, insects or tigers? Active / Passive swap Does Anna love Benjamin? Is Benjamin loved by Anna? Modifiers changes question intent Is Mark Wright a photographer? Is Mark Wright an accredited photographer? Table 6: The 6 concepts used for debugging Roberta-Large model fine-tuned on QQP dataset.