# counterfactual_generation_with_identifiability_guarantees__8a569b79.pdf Counterfactual Generation with Identifiability Guarantees Hanqi Yan1,4 , Lingjing Kong2 , Lin Gui3, Yuejie Chi2, Eric Xing2,4, Yulan He1,3, Kun Zhang2,4 1University of Warwick,2 Carnegie Mellon University, 3King s College London, 4Mohamed Bin Zayed University of Artificial Intelligence Counterfactual generation lies at the core of various machine learning tasks, including image translation and controllable text generation. This generation process usually requires the identification of the disentangled latent representations, such as content and style, that underlie the observed data. However, it becomes more challenging when faced with a scarcity of paired data and labeling information. Existing disentangled methods crucially rely on oversimplified assumptions, such as assuming independent content and style variables, to identify the latent variables, even though such assumptions may not hold for complex data distributions. For instance, food reviews tend to involve words like tasty , whereas movie reviews commonly contain words such as thrilling for the same positive sentiment. This problem is exacerbated when data are sampled from multiple domains since the dependence between content and style may vary significantly over domains. In this work, we tackle the domain-varying dependence between the content and the style variables inherent in the counterfactual generation task. We provide identification guarantees for such latent-variable models by leveraging the relative sparsity of the influences from different latent variables. Our theoretical insights enable the development of a do Main Adap Tive coun Terfactual g Eneration model, called (MATTE). Our theoretically grounded framework achieves state-of-the-art performance in unsupervised style transfer tasks, where neither paired data nor style labels are utilized, across four large-scale datasets. 1 Introduction Counterfactual generation serves as a crucial component in various machine learning applications, such as controllable text generation and image translation. These applications aim at producing new data with desirable style attributes (e.g., sentiment, tense, or hair color) while preserving the other core information (e.g., topic or identity) [Li et al., 2019, 2022, Xie et al., 2023, Isola et al., 2017, Zhu et al., 2017]. Consequently, the central challenge in counterfactual generation is to learn the underlying disentangled representations. To achieve this goal, prior work leverages either paired data that only differ in style components [Rao and Tetreault, 2018, Shang et al., 2019, Xu et al., 2019b, Wang et al., 2019b], or utilizes style labeling information [John et al., 2019, He et al., 2020, Dathathri et al., 2020, Yang and Klein, 2021, Liu et al., 2022]. However, collecting paired data or labels can be labour-intensive and even infeasible in many real-world applications [Chou et al., 2022, Calderon et al., 2022, Xie et al., 2023]. This has prompted recent work [Kong et al., 2022, Xie et al., 2023] to delve into unsupervised identification of latent variables by tapping into multiple domains. To attain identifiability guarantees, a prevalent assumption made in these works [Kong et al., 2022, Xie et al., 2023] is that the content and the style Equal Contribution. Work was done when Hanqi Yan was a visiting student at MBZUAI. Correspondence to: Yulan He (yulan.he@kcl.ac.uk) and Kun Zhang (kunz1@cmu.edu). 37th Conference on Neural Information Processing Systems (Neur IPS 2023). latent variables are independent of each other. However, this assumption is often violated in practical applications. First, the dependence between content and style variables can be highly pronounced. For example, to express a positive sentiment, words such as tasty and flavor are typically used in conjunction with food-related content. In contrast, words like thrilling are more commonly used with movie-related content [Li et al., 2019, 2022]. Moreover, the dependence between content and style often varies across different distributions. For example, a particular cuisine may be highly favored locally but not well received internationally. This varying dependence between content and style variables poses a significant challenge in obtaining the identifiability guarantee. To the best of our knowledge, this issue has not been addressed in previous studies. In this paper, we address the identification problem of the latent-variable model that takes into account the varying dependence between content and style (see Fig 1). To this end, we adopt a natural notion of influence sparsity inherent to a range of unstructured data, including natural languages, for which the influences from the content and the style differ in their scopes. Specifically, the influence of the style variable on the text is typically sparser compared to that of the content variable, as it is often localized to a smaller fraction of the words [Li et al., 2018] and plays a secondary role in word selection. For instance, the tense of a sentence is typically reflected in only its verbs which are affected by the sentence content information. Our contributions can be summarised as: 1) We show identification guarantees for both the content and the style components, even when their interdependence varies. This approach removes the necessity for a large number of domains with specific variance properties [Kong et al., 2022, Xie et al., 2023]. 2) Guided by our theoretical findings, we design a do Main Adap Tive coun Terfactual g Eneration model (MATTE). It does not require paired data or style annotations but allows style intervention, even across substantially different domains. 3) We validate our theoretical discoveries by demonstrating state-of-the-art performance on the unsupervised style transfer task, which demands representation disentanglement, an integral aspect of counterfactual generation. 2 Related work Label-free Style Transfer on variation autoencoders (VAEs). To perform style transfer, existing methods that use parallel or non-parallel labelled data often rely on style annotations to refine the attribute-bearing representations, although some argue that disentanglement is not necessary for style edit [Sudhakar et al., 2019, Wang et al., 2019a, Dai et al., 2019]. In practice, disentangled methods typically employ adversarial objectives to ensure the content encoder remains independent of style information or leverage style discriminators to refine the derived style variables [Hu et al., 2017, Shen et al., 2017, Li et al., 2018, Keskar et al., 2019, John et al., 2019, Sudhakar et al., 2019, Dathathri et al., 2020, Hu et al., 2017, Dathathri et al., 2020, Yang and Klein, 2021, Liu et al., 2022]. Several studies have tackled this task without style labels. Riley et al. [2020] emphasized the implicit style connection between adjacent sentences and used T5 [Raffel et al., 2020] to extract the style vector for conditional generation. CPVAE Xu et al. [2020] split the latent variable into content and style variables and mapped them to a k-dimensional simplex for k-way sentiment modeling. Our work aligns more closely with CPVAE and follows VAE-based label-free disentangled learning from data-generation perspective Higgins et al. [2017], Kumar et al. [2018], Mathieu et al. [2019], Xu et al. [2020], Mita et al. [2021], Fan et al. [2023]. Latent-variable identification. Representation learning serves as the cornerstone for generative models, where the goal is to create representations that effectively capture underlying factors in the disentangled data-generation process. In the realm of linear generative functions, Independent component analysis (ICA) [Comon, 1994, Bell and Sejnowski, 1995] is a classical approach known for its identifiability. However, when dealing with nonlinear functions, ICA is proven unidentifiable without the inclusion of additional assumptions [Hyvärinen and Pajunen, 1999]. To tackle this problem, recent work incorporated supplementary information [Hyvarinen et al., 2019, Sorrenson et al., 2020, Hälvä and Hyvarinen, 2020, Khemakhem et al., 2020, Kong et al., 2022], e.g., class/domain labels. However, these approaches require a number of domains/classes that are twice the number of latent components, which can be unfeasible when dealing with high-dimensional representations. Another line of work [Zimmermann et al., 2022, von Kügelgen et al., 2021, Locatello et al., 2020, Gresele et al., 2019, Kong et al., 2023a,b] leverages paired data (e.g., two rotated versions of the same image) to identify the shared latent factor within the pair. The third line of work [Lachapelle et al., 2022, Yang et al., 2022, Zheng et al., 2022] makes sparsity assumptions on the nonlinear generating function. Although their sparsity assumption alleviates some of the explicit requirements in the previous two types of work, it may not hold for complex data distributions. For instance, in the case of generating text, each topic-related latent factor may influence a large number of components. Instead, we adopt a relative sparsity assumption, where we only require the influence of one subspace to be sparser than the other. Unlike prior work [Zheng et al., 2022], each latent variable is allowed to influence a non-sparse set of components, and the influence can overlap arbitrarily within each subspace. Importantly, we necessitate neither many domains/classes nor paired data as prior work mentioned above. 3 Disentangled Representation for Counterfactual Generation Figure 1: The data generation process: The grey shading indicates the variable is observed. The observed variable (i.e., text) x is generated from content c and style s. Both content c and style s are influenced by the domain variable u and the content also influences the style. s is the exogenous variable of s, representing the independent information of s. In this section, we discuss the connection between counterfactual generation and the identification of the data-generating process shown in Fig 1. Disentangled latent representation. The data-generating process in Figure 1 can be expressed in Equation 1: c p(c|u), s p(s|c, u), x = g(c, s), (1) where the data (e.g., text) x X are generated by latent variables z := [c, s] Z Rdz through a smooth and invertible generating function g( ) : Z X. The latent space comprises two subspaces: the content variable c C Rdc and the style variable s S Rds. We define c as the description of the main topic, e.g., We ordered the steak recommended by the waitress, , and s comprises supplementary details connected to the primary topic, e.g., the sentiment towards the dish, as exemplified in it was delicious! . Consequently, the counterfactual generation task here is to preserve the content information c while altering the stylistic aspects represented by s. Content-style dependence. In many real-world problems, content c can significantly impact style s. For instance, when it comes to the positive descriptions of food (content), words like delicious are more prevalent than terms like effective . Intuitively, the content c acts to constrain the range of vocabulary choices for style s. As a result, the counterfactually generated data should preserve the inherent relationship between c and s. We directly model this dependence as: s := gs( s; c, u), (2) where gs characterizes the influence from c to s and the exogenous variable s accounts for the inherent randomness of s. In the running example, s can be interpreted as the randomness involved in choosing a word from the vocabulary defined by content c, encompassing words similar to delicious such as tasty , yummy . In contrast, prior work [Kong et al., 2022, Xie et al., 2023] assumes the independence between c and s thus neglecting this dependence. Challenges from multiple domains. As we outlined in Section 1, the ability to handle domain shift is crucial for unsupervised counterfactual generation. Domain embedding u represents a specific domain and the domain distribution shift influences both the marginal distribution of content p(c|u) and the dependence of content on style p(s|c, u). The change in content distribution p(c|u) across different domains u reflects the variability in the subjects of sentences across these domains. For example, it can manifest as a change from discussing food in restaurant reviews to actors in movie reviews. The change in content-style dependence p(s|c, u) signifies that identical sentence subjects (i.e., content) could be associated with disparate stylistic descriptions in different domains. For instance, the same political question could provoke significantly different sentiments among various demographic groups. Such considerations are absent in prior work [Kong et al., 2022]. Here, we learn a shared model (ˆp(c| ), gs( s, c, )) and domain-specific embeddings u. This approach enables effective knowledge transfer across domains and manages distribution shifts efficiently. For a target domain τ, which may have limited available data, we can learn uτ using a small amount of unlabeled data xτ while preserving the multi-domain information in the shared model. In light of the above discussion, the essence of counterfactual generation now revolves around the task of discerning the disentangled representation (c, s) within the data-generating process (Fig 1) across various domains with unlabeled data (x, u): if we can successfully identify (c, s), we could perform counterfactual reasoning by manipulating s while preserving both the content information and the content-style dependence. 4 Identifiability of the Latent Variables In this section, we introduce the identification theory for the content c and the style s sequentially and then discuss their implications for methodological development. We introduce notations and definitions that we use throughout this work. When working with matrices, we adopt the following indexing notations: for a matrix M, the i-th row is denoted as Mi,:, the j-th column is denoted as M:,j, and the (i, j) entry is denoted as [M]i,j. We can also use this notation to refer to specific sets of indices within a matrix. For an index set I {1, . . . , m} {1, . . . , n}, we use Ii,: to denote the set of column indices whose row coordinate is i, i.e., Ii,: := {j : (i, j) I}, and analogously I:,j to denote the set of row indices whose column coordinate is j. In addition, we define a subspace of Rn using an index set S: Rn S = {z Rn| i / S, zi = 0}, i.e., it consists of all vectors in Rn whose entries are zero for all indices not in S. Finally, we can define the support of a matrix-valued function M(x) : X Rm n as the set of indices whose corresponding entries are nonzero for some input value x, i.e., Supp(M) := {(i, j) : x X, s.t., [M(x)]i,j = 0}. 4.1 Influence Sparsity for Content Identification 𝑐! 𝑐" 𝑐# 𝑠! 𝑠" sparser denser partially intersecting Figure 2: Sparsity in Jg. We show that the subspace c can be identified. That is, we can estimate a generative model (pˆc, pˆs|ˆc, ˆg) 2 following the data-generating process in Equation 1 and the estimated variable ˆc can capture all the information of c without interference from s. In the following, we denote the Jacobian matrices Jg(z) s and Jˆg(ˆz) s supports as G and ˆG respectively. Further, we denote as T a set of matrices with the same support as that of the matrix-valued function J 1 g (c)Jˆg(ˆc). Assumption 1 (Content identification). i. g is smooth and invertible and its inverse g 1 is also smooth. ii. For all i {1, . . . , dx}, there exist {z(ℓ)}|Gi,:| ℓ=1 and T T , such that span({Jg(z(ℓ))i,:}|Gi,:| ℓ=1 ) = Rdz Gi,: and [Jg(z(ℓ))T]i,: R dz ˆGi,:. iii. For every pair (cjc, sjs) with jc [dc] and js {dc + 1, . . . , dz}, the influence of sjs is sparser than that of cjc, i.e., G:,jc 0 > G:,js 0. Theorem 1. We assume the data-generating process in Equation 1 with Assumption 1. If for given dimensions (dc, ds), a generative model (pˆc, pˆs|ˆc, ˆg) follows the same generating process and achieves the following objective: arg min pˆc,pˆs,ˆg jˆs {dc+1,...,dz} ˆG:,jˆs 0 subject to: pˆx(x) = px(x), x X, (3) then the estimated variable ˆc is an one-to-one mapping of the true variable c. That is, there exists an invertible function hc( ) such that ˆc = hc(c). A proof can be found in Appendix A.5. Interpretation. Theorem 1 states that by matching the marginal distribution px(x) under a sparsity constraint of ˆs subspace, we can successfully eliminate the influence of s from the estimated ˆc. This warrants that the content information can be fully retained without being entangled with the style information for a successful counterfactual generation. We can further identify s from that of s when the dependence gs function is invertible in its argument s [Kong et al., 2022]. Discussion on assumptions. Assumption 1-i. ensures that the information of all latent variables [c, s] is preserved in the observed variables x, which is a necessary condition for latent-variable identification [Hyvarinen et al., 2019, Kong et al., 2022]. Assumption 1-ii. ensures that the influence from the latent variable z varies sufficiently over its domain. This excludes degenerate cases where the Jacobian matrix is partially constant, and thus, its support fails to faithfully capture the influence between latent variables and the observed variables. Assumption 1-iii. encodes the observation that the s subspace exerts a relatively sparser influence on the observed data than the subspace c (Fig 2.). 2As our theory is not contingent on the availability of multiple domains, we drop the domain index u in our notations in this section for ease of exposition. This is reasonable for language, where the main event largely predominant the sentence and the stylistic variable play complementary and local information for particular attributes, e.g., tense, sentiment, and formality [Xu et al., 2019a, Wang et al., 2021, Ross et al., 2021]. For the language data, x corresponds to a piece of text (e.g., a sentence) with its dimension dx equal to the number of words multiplied by the word embedding dimension, i.e., multiple dimensions of x correspond to a single word. Therefore, even if a word is simultaneously influenced by both c and s, the influence from the content c tends to be denser on this word s embedding dimension, as content usually takes precedence in word selection over style. Contrast with prior work. Zheng et al. [2022] impose sparse influence constraints on the generating function g in an absolute sense each latent component should have a very sparse influence on the observed data. In contrast, Theorem 1 only calls for relative sparsity between two subspaces where each latent component s influence may not be sparse and unique as in Zheng et al. [2022]. We believe this is reasonable for many real-world applications like languages. Kong et al. [2022] assume the independence between the two subspaces and identify the content subspace by resorting to its invariance and sufficient variability of the style subspace over multiple domains. However, as discussed in Section 3, the invariance of the content subspace is often violated, and so is the independence assumption. In contrast, we permit the content subspace to vary over domains and allow for the dependence between the two subspaces. 4.2 Partially Intersecting Influences for Style Identification In this section, we show the identifiability for the style subspace s, under one additional condition: the influences from the two subspaces c and s do not fully overlap. Assumption 2 (Partially intersecting influence supports). For every pair (cjc, sjs), the supports of their influences on x do not fully intersect, i.e., G:,jc G:,js 0 < min{ G:,jc 0 , G:,js 0}. Theorem 2. We follow the data-generating process Equation 1 and Assumption 1 and Assumption 2. We optimize the objective function in Equation 3 together with (jˆc,jˆs) {1,...,dc} {dc+1,...,dz} ˆG:,jˆc ˆG:,jˆs 0 . (4) The estimated style variable ˆs is a one-to-one mapping to the true variable s. That is, there exists an invertible mapping hs( ) between s and ˆs, i.e., ˆs = hs(s). The proof can be found in Appendix A.6. Interpretation. Theorem 2 states that we can recover the style subspace s if the influences from the two subspaces do not interfere with each other (Fig 2). This condition endows the subspaces distinguishing footprints and thus forbids the content information in c from contaminating the estimated style variable ˆs. The identification of s is crucial to counterfactual generation tasks: if the estimated style variable ˆs does capture all the true style variable s, intervening on ˆs cannot fully alter the original style that is intended to be changed. Discussion on assumptions. Assumption 2 prescribes that each content component cjc and each style component sjs do not fully contain each other s influence support. Together with Theorem 1, this assumption is essential to the identification of s, without which ˆsjs may absorb the influence from cjc. Assumption 2 does not demand the supports of the entire subspaces c and s to be partially intersecting or even disjoint, and the latter directly implies Assumption 2. This assumption is plausible for many real-world data distributions, especially for unstructured data like languages and images certain dimensions in the pixels and word embeddings may reflect the information of either the content or the style. Contrast with prior work. Kong et al. [2022] obtains the identifiability of the style subspace s by exploiting the access to multiple domains over which the marginal distribution of s (i.e., p(s|u)) varies substantially over domains u. This hinges on the independence between the two subspaces and is not applicable when the marginal distribution of s only varies over the content c, i.e., p(s|c, u). 5 A Framework for Unsupervised Counterfactual Generation In this section, we translate the theoretical insights outlined in Section 4 into an unsupervised counterfactual generation framework. Guided by the theory, we can approximate the underlying data-generating process depicted in Fig 1 and recover the disentangled latent components. 𝑔!"# 𝑓"$# 𝑥 𝑥!"# 𝑠 𝑠~𝑁(0,1) 𝑟$ 𝑐~𝑁(0,1) 𝑐 𝑟% 𝑧 𝑥&!'()*"! Figure 3: Our VAE-based framework MATTE. During training, the input x is fed to the encoder fenc to derive the latent variable z = [c, s], which is then passed to the decoder gdec for reconstruction. Flow modules, denoted as rc and rs, are implemented to model the causal influences on c and s respectively, which yields the creation of exogenous variables c and s. To generate transferred data xtransfer, we intervene on the style exogenous variable s while keeping the original content variable c unchanged (indicated by the green arrows). In the following, we will describe each module in our VAE-based estimation framework (Fig 3), the learning objective, and the procedure for counterfactual generation. 5.1 VAE-based Estimation Framework Given input sentences x from various domains, we use the VAE encoder fenc to parameterize the posterior distribution qfenc(z|x) and sample z qfenc(z|x). 3 The posterior sample z is then fed into the VAE decoder gdec for reconstruction xrec = gdec(z), as in conventional VAE training. We split z into two components: c and s. As shown in Fig 1, both c and s encompass information of a particular domain u, and s is also influenced by c. We parameterize such influences using flow-based models [Dolatabadi et al., 2020, Durkan et al., 2019] rc and rs, respectively: c = rc(c; u), s = rs(s; u, c), (5) where c and s are exogenous variables that are independent of each other, and u and (u, c) act as contextual information for the flow models rc( ; u) and rs( ; u, c). This design promotes parameter sharing across domains, as we only need to learn a domain embedding u (c.f., a separate flow model per domain). As part of the evidence-lower-bound (ELBO) objective in VAE, we regularize the distributions of z = [ c, s] to align with the prior p( z) using Kullback Leibler (KL) divergence. Consequently, the VAE learning objective can be expressed as: LVAE := log pfenc,gdec(xrec) + KL(qfenc,rc,rs( z|x)|p( z)), (6) where the prior p( z) is set to a standard Gaussian distribution, N(0, I), consistent with typical VAE implementations. 5.2 Sparsity Regularization for Identification Guarantees Guided by the insights from Theorem 1 and Theorem 2, the sparsity constraint on the influence of s (i.e., Equation 3) and the partially intersecting influence constraint (i.e., Equation 4) are crucial to faithfully recover and disentangle the latent representations c and s. Sparsity of the style influence. To implement Equation 3, we compute the Jacobian matrix Jgdec(z) for the decoder function on-the-fly and apply ℓ1 regularization to the columns corresponding to the style variable [Jgdec(z)]:,dc+1:dz to control its sparsity. That is, Lsparsity = [Jgdec(z)]:,dc+1:dz 1. Partially intersecting influences. To encourage sparsity in the intersection of influence between c and s (as defined in Equation 4), we select K output dimensions Is of Jgdec(z) that capture the most substantial influence from s and another set of K output dimensions Ic that receive the least influence from c. Subsequently, we apply ℓ1 regularization to the influence from c on the output dimensions at the intersection Is Ic, i.e., Lpartial = [Jgdec(z)]Is Ic,1:dc 1. Content variable masking. In practice, the content dimensionality dc is a design choice. When dc is set excessively large, the sparsity regularization term Lsparsity may cause the style variable s to lose its influence, squeezing the information of s into the content variable c. To handle this issue, we apply a trainable soft mask that operates on c to dynamically control its dimensionality. In sum, the overall training objective is as follows: L := LVAE + λsparsity Lsparsity + λpartial Lpartial + λc-mask Lc-mask, (7) where λ s are weight parameters to balance various loss terms. 3For the sake of simplicity, in this section, we discuss estimated variables without theˆ notation, as in 4. 5.3 Style Intervention As discussed in Section 3, the content-style dependence should be preserved when generating counterfactual text to ensure linguistic consistency. This can be achieved by intervening on the exogenous style variable s of the original sample. Specifically, we feed the original sample x to the encoder fenc to obtain variables [c, s]. Subsequently, we pass the style variable s through the flow models rs to obtain its exogenous counterpart s, i.e., s = rs(s; c, u). To carry out style transfer, we set the original variable s to the desired style value stransfer, which is the average of the exogenous style values of randomly selected samples with the desired style. As the flow model rs is invertible, we can obtain the transferred style variable stransfer = r 1 s ( stransfer; c, u), which, together with the original content variable c, generates the new sample xtransfer = gdec([c, stransfer]). This process is illustrated in Fig 3 using green arrows. We demonstrate the importance of preserving the content-style dependence and provide evidence that our approach can indeed fulfill this purpose (Fig 4). 6 Experimental Results We validate our theoretical findings by conducting experiments on multiple-domain sentiment transfer tasks, which require effective disentanglement of factors, a concept at the core of our identifiability theory. Table 1: Dataset on four domains. Domains Train Dev Test IMDB 344,175 27,530 27,530 Yelp 444,102 63,484 1000 Amazon 554,998 2,000 1,000 Yahoo 4,000 4,000 4,000 Datasets and Evaluation Schema. The proposed method is trained on four-domain datasets (Tab 1), i.e, movie (Imdb) [Diao et al., 2014], restaurant (Yelp) [Li et al., 2018], ecommerce (Amazon) [Li et al., 2018] and news (Yahoo) [Zhang et al., 2015, Li et al., 2019]. 4 From common practice [Yang et al., 2018, Lample et al., 2019], we evaluate the generated sentences in terms of the four automatic metrics: (1) Accuracy. We train a CNN classifier on the original style-labelled dataset, which has over 95.0% accuracy when evaluated on the four separate validation datasets. Subsequently, we employ it to evaluate the transformed sentences, gauging how effectively they convey the intended attributes. (2) BLEU [Papineni et al., 2002]. It compares the n-grams in the generated text with those in the original text, measuring how well the original content is retained 5. (3) G-score. It represents the geometric mean of the predicted probability for the ground-truth style category and the BLEU score. Due to its comprehensive nature, it is our primary metric. (4) Fluency. It is the perplexity score of GPT-2 [Radford et al., 2019] lower perplexity values indicate a higher levels of fluency. For human evaluation, we invited three evaluators proficient in English to rate the sentiment reverse, semantic preservation, fluency and overall transfer quality using a 5-point Likert scale, where higher scores signify better performance. Furthermore, they were asked to rank the generated sentences produced from different models, with the option to include tied items in their ranking. 6.1 Sentiment transfer Baselines. We compare our model with the state-of-the-art text transfer models that do not rely on style labels, along with a supervised model, B-GST [Sudhakar et al., 2019], which is based on GPT2 [Radford et al., 2019] and accomplishes style transfer through a combination of deletion, retrieval, and generation. The other VAE-based baselines can be divided into two groups based on their architecture: those with LSTM backbones and those utilizing pretrained language models (PLM). Within the LSTM group, β-VAE [Higgins et al., 2017] encourages disentanglement by progressively increasing the latent code capacity. Joint Train [Li et al., 2022] uses the Glo Ve embedding to initialize s and learns c through LSTM. CPVAE [Xu et al., 2020] is the state-of-the-art unsupervised style transfer model, which maps the style variable to a k-dimensional probability simplex to model different style categories. In the PLM group, we use GPT2 [Radford et al., 2019] as the backbone and introduce an additional variational layer after fine-tuning its embedding layer to generate the latent variable z, referred to as GPT2-FT. Also, we consider Optimus [Li et al., 2020], which is one of the most widely-used pretrained VAE models, utilizing BERT [Devlin et al., 2019] as the encoder and GPT2 as the decoder. 4Dataset and detailed experiment configurations can be found in Appendix, A.1. 5We also adopt CTC score [Deng et al., 2021], to mitigate potential issues brought by the word-overlap measurements in BLEU, as it considers the matching embeddings. The evaluation results are shown in Table 9. IMDB Yelp Model Acc( ) BLEU( ) G-score( ) PPL( ) Acc( ) BLEU( ) G-score( ) PPL( ) B-GST [Sudhakar et al., 2019] 36.20 0.80 50.45 2.62 32.09 1.81 48.58 2.08 82.00 0.20 32.06 1.34 35.43 0.92 50.45 2.38 β-VAE [Higgins et al., 2017] 38.27 1.03 11.37 3.03 9.05 1.19 43.59 3.07 40.30 0.92 7.58 2.73 6.86 1.08 59.34 3.81 Joint Train [Li et al., 2022] 24.13 0.52 23.26 2.85 12.28 1.91 70.11 2.76 14.20 0.62 31.72 1.91 12.74 0.95 84.07 2.37 CPVAE [Xu et al., 2020] 20.15 0.40 49.82 1.25 20.01 0.96 70.18 2.78 14.50 0.30 51.47 1.81 16.84 0.92 72.81 2.26 GPT2-FT [Radford et al., 2019] 15.20 0.25 28.93 3.16 12.19 2.84 71.08 2.19 12.00 0.42 39.62 1.92 14.49 1.35 78.37 2.14 Optimus [Li et al., 2020] 14.07 0.20 59.04 1.68 17.47 1.21 61.90 2.61 13.60 0.30 69.82 1.92 21.24 1.83 52.56 2.01 MATTE 32.43 0.28 45.10 2.91 25.92 1.62 50.08 2.02 34.30 0.26 50.14 2.51 26.34 1.37 51.51 2.09 Amazon Yahoo Model Acc( ) BLEU( ) G-score( ) PPL( ) Acc( ) BLEU( ) G-score( ) PPL( ) B-GST [Sudhakar et al., 2019] 60.45 0.65 56.02 2.36 47.67 1.68 49.01 3.18 84.30 0.40 40.39 2.81 38.65 1.64 58.20 2.19 β-VAE Higgins et al. [2017] 50.08 0.68 8.04 2.62 9.39 1.24 33.09 2.53 55.47 0.40 3.77 1.32 5.85 1.71 52.17 3.06 Joint Train Li et al. [2022] 32.90 0.42 23.21 2.16 18.33 1.07 84.63 2.76 35.33 0.28 14.04 1.72 11.62 0.92 67.34 2.84 CPVAE Xu et al. [2020] 32.60 0.20 41.08 1.28 30.08 1.15 77.61 3.12 43.92 0.30 25.44 1.37 20.28 0.95 76.28 2.67 GPT2-FT [Radford et al., 2019] 30.46 0.30 40.34 2.82 26.72 1.93 79.36 2.63 17.90 0.40 44.19 1.86 15.99 1.11 70.99 2.37 Optimus [Li et al., 2020] 24.80 0.20 62.50 1.55 28.53 1.22 74.66 3.10 27.10 0.15 32.73 1.82 19.17 1.69 73.18 2.76 MATTE 34.50 0.24 52.25 1.48 35.73 1.14 63.37 2.22 38.45 0.20 42.40 1.35 29.01 2.30 56.12 2.57 Table 2: Comparison with unsupervised methods across four domain datasets with supervised B-GST as an upper bound reference. The top and the second-best results are in bold. Quantitative performance. Among LSTM baselines in Table 2, β-VAE shows high sentiment transfer accuracy and fluency but poor content preservation. We observed that many generated sentences follow simple but repetitive patterns, e.g., 2.2% transferred sentences in Yelp containing the phrase I highly recommend while only 0.6% original sentences do. They are fluent and correctly sentiment flipped but their semantics are significantly different from the original sentences, indicating a generation degradation problem 6. CPVAE achieves an overall better G-Score than all the other baselines across three domains (except for Yelp). Compared with the other LTSM-based methods, its superiority in content preservation is pronounced. PLMs models achieve overall better BLEU scores compared with the LSTM group. Optimus outperforms GPT2-FT, which can be partly explained by the fact that the variational layer in Optimus has been pretrained on 1,990K Wikipedia sentences. Our model is built on top of CPVAE with the proposed causal influence modules and sparsity regularisations. It gains consistent improvements in G-score and fluency across all datasets over all the other unsupervised methods. Compared with the supervised method, despite a relatively large gap in accuracy due to the lack of supervision, our approach achieves comparable BLEU scores. The human evaluation results in Table 3 show that human annotators favour Optimus in terms of content preservation and fluency, but MATTE ranks the best-performing method with 58.5% support set, compared to 41.00% for Optimus. Style Content Fluency Best rank(%) CPVAE 1.30 2.78 3.20 21.50 Optimus 1.41 3.79 4.12 41.00 MATTE 1.99 2.91 3.53 58.50 Table 3: Human evaluation results from three annotators. The Cohen s Kappa coefficient among every two annotators over the Best-rank is 0.46. Human annotators favour Optimus in terms of content preservation and fluency, but MATTE ranks the best-performing method with more than 58% support set after considering the style reverse success rate. Src1: This guy is an awful actor. . % CPVAE The guy is very flavorful. ! Optimus The guy is an amazing actor. ! MATTE This guy is an amazing actor! ! Src2: I had it a long time now and I still love it. ! CPVAE I had it a long time before and I ve never eaten it. % Optimus It is a long time now and I always get this food. ! MATTE I had it a long time now and I never played it. % Src3 These come in handy with those tender special moments. ! CPVAE These come in handy with those sexy care employees. ! Optimus These are fateless in their only safe place. % MATTE These come in handy with those poorly executed characteristics. % Table 4: Generated style-transferred Sentences. !, %represent sentiment polarity. Qualitative results. We randomly selected three sentences from the test sets and analyzed the results generated by the top-performing baselines in Table 4. For Src 1, although all the methods successfully transfer the original sentiment of the sentence, CPVAE generates the word flavourful for the content guy, resulting in an unnatural sentence. This issue arises because CPVAE fails to identify the domain-specific content-style dependency, i.e., the Src domain is in IMDB, while the transferred sentence incorrectly uses the style word flavourful which is commonly used in Yelp. While Optimus can generate relatively fluent sentences partly due to its powerful decoder, it hardly 6The diversity-n [Li et al., 2016] also indicates repetitious pattern and the evaluation results are in Table 10. Model Acc( ) BLEU( ) G-score( ) PPL( ) Acc( ) BLEU( ) G-score( ) PPL( ) Backbone 20.15 49.82 20.01 70.18 14.50 51.47 16.84 72.81 Indep [Kong et al., 2022] 45.00 30.88 19.89 61.85 61.90 25.67 21.24 73.78 Causal Dep 28.71 39.63 21.85 53.25 22.10 48.98 25.98 55.14 :w. / Lsparsity 21.55 56.59 20.90 65.26 13.20 56.26 14.59 54.10 :w. / Lpartial 30.18 51.95 25.57 54.66 33.70 49.09 25.81 52.87 :w. / Lc-mask(Full) 32.43 45.10 25.92 50.08 34.30 50.14 26.34 51.51 Table 5: Ablation results on sentiment transfer on two domains. Causal Dep incorporates style flow rs to model dependency of c on s, while Indep assumes the independence between the two variables. marks the improvements over Backbone, while over the Causal Dep. maintains the original semantics (Src 2, 3), indicating a lack of effective disentanglement between c and s. These failure modes demonstrate the importance of a proper disentanglement of content and style and modelling the causal influence between the two across domains. Benefiting from theoretical insights, our approach manages to reflect the content influence across different domains in Src 1 and retain the content information in Src2, 3. 6.2 Ablation Study Figure 4: Histograms of negative log-likelihood (NLL) of 1000 Amazon test samples evaluated on the original latent variable and intervened ones. left: flips s, right flips s. The table shows the corresponding sentences. Ablation studies in Table 5 are used to verify our theoretical results in 4 7. On top of the backbone, CPVAE, we incrementally add each component of our method: (1)Indep considers the domain influence on c (i.e., the rc module in Fig 3), while neglecting the independence between c and s. It experiences a large accuracy boost in conjunction with a significant degradation in BLEU, suggesting poor retention of the content information. (2) Causal Dep takes into account the dependency between content and style by incorporating the module rs in Fig 3. This ameliorates the content retention problem and strikes an overall better balance, as reflected by the raised BLEU score and G-score, although causal dependence in Causal Dep is not sufficient for identification without proper regularization. (3) After introducing the style sparsity regularization Lsparsity as specified in Theorem 1, we observe a significant increase of BLEU over Causal Dep, verifying Theorem 1 that the style influence sparsity facilitates content identification 4.1. (4) We further introduce Lpartial inspired by Theorem 2, which controls the intersection of content and style influence supports. This improvement in style identification, i.e., the recovery of accuracy over Lsparsity corroborates Theorem 2. (5) The incorporation of Lc-mask arrive at our full model, which further improves the style identification, consistent with our motivation in 5.3. It also exhibits the best G-score across all the datasets, with the most predominant improvement over Causal Dep on the Yahoo dataset, where the G-score increases from 21.39% to 29.01 %. The importance of content-style dependence. We demonstrate the importance of content-style dependence by visualizing the changes in negative log-likelihood (NLL) induced by different ways of style intervention, namely flipping s as in our method and flipping s which breach the content-style dependence. If the NLL increases after the style transfer, it indicates that the new variables are located in a lower density region [Zheng et al., 2022, Xu et al., 2020]. Fig 4 shows the histograms of NLLs for all the Amazon test samples, both before and after a style transfer. We can see that the NLL distribution changes negligibly when we flip s, in contrast with the significant change caused by flipping s. This implies that flipping s enables better preservation of the joint distribution of the 7The results on Amazon and Yahoo are in Appendix, Table 8. original sentence. The generated sentences resulting from flipping s exhibit a higher level of semantic fidelity to the original sentence, with a clear inverse sentiment. 6.3 Comparison with large language model As widely recognized, large language models (LLMs) have demonstrated an impressive capability in text generation. However, we consider the principles of counterfactual generation to be complementary to the development of LLMs. We aim to leverage our theoretical insights to further enhance the capabilities of LLMs. We provide examples in Table 6 where LLMs struggle with sentiment transfer, primarily due to their tendency to overlook the broader and implicit sentiments while accurately altering invidivual sentiment words. Consequently, it is reasonable to anticipate that LLMs could benefit from the principles of representation learning, as developed in our work. Src: The buttons to extend the arms worked exactly one time before breaking. Chat GPT-p1: The buttons to extend the arms failed to work from the beginning, never functioning even once. Chat GPT-p2: The buttons to extend the arms never worked, even once, and remained functional until they broke. Our: The buttons to extend the arms worked exactly as described. Src: I love that it uses natural ingredients but it was ineffective on my skin. Chat GPT-p1: I dislike that it uses natural ingredients, but it was highly effective on my skin. Chat GPT-p2: I dislike that it uses natural ingredients, but it was highly effective on my skin. Our: I like that it uses natural ingredients, and it was also good. Src: This case is cute however this is the only good thing about it. Chat GPT-p1: This case is not cute; however, it is the only good thing about it. Chat GPT-p2: This case is not cute at all; however, it is the only bad thing about it. Ours:This case is cute and overall a valuable product. Table 6: A Sentiment transfer example, on which Chat GPT fails to completely reverse the overall sentiment of the sentence, although it successfully negates individual words within text. In contrast, our method achieves the sentiment reversal with minimal changes. Chat GPT-p1 and Chat GPT-p2 represent results obtained from two different prompts, i.e., p1: Flip the sentiment of the following sentences, but keep the content unchanged as much as possible. ; p2: Please invert the sentiment while preserving content as much as possible in the following sentence that originates from the original domain. . 6.4 Visualization of style variable Figure 5: The style variables of sentences in past-tense (blue) and presenttense (red) following a UMAP projection. Left: CPVAE; Right: MATTE. We further validate our theoretical insights within additional content-style disentanglement scenarios. As tense has a relatively sparse influence on sentences compared to their content, we choose tense (past and present) as another style for illustration. Specifically, we collect 1000 sentences in either past or present tense from the Yelp Dev set and derive their style representations, denoted as s, by feeding these sentences into our well-trained model. The projection results of CPVAE and MATTE are shown in 5. The distinct separation between the red and blue data points indicates a more discriminative and better disentangled style variabl. However, in the case of CPVAE, some blue data points are mixed within the lower portion of the red region. 7 Conclusion and limitations Prior work [Kong et al., 2022, Xie et al., 2023] have employed multiple domains to achieve unsupervised representation disentanglement. However, the assumed independence between the content and style variables often does not hold in real-world data distributions, particularly in natural languages. To tackle this challenge, we address the identification problem in latent-variable models by leveraging the sparsity structure in the data-generating process. This approach provides identifiability guarantees for both the content and the style variables. We have implemented a controllable text generation method based on these theoretical guarantees. Our method outperforms existing methods on various large-scale style transfer benchmark datasets, thus validating our theory. It is important to note that while our method shows promising empirical results for natural languages, the sparsity assumption (Assumption 1-iii.) may not hold for certain data distributions like images, where the style component could exert dense influences on pixel values. In such cases, we may explore other forms of inherent sparsity in the given distribution, e.g., sparse dependencies between content and style or sparse changes over multiple domains, to achieve identifiability guarantees and develop empirical approaches accordingly. Acknowledgements We thank anonymous reviewers for their constructive feedback. This work was funded by the UK Engineering and Physical Sciences Research Council (grant no. EP/T017112/1, EP/T017112/2, EP/V048597/1, EP/X019063/1). YH is supported by a Turing AI Fellowship funded by the UK Research and Innovation (grant no. EP/V020579/1, EP/V020579/2). The work of LK and YC is supported in part by NSF under the grants CCF-1901199 and DMS-2134080. This project is also partially supported by NSF Grant 2229881, the National Institutes of Health (NIH) under Contract R01HL159805, a grant from Apple Inc., a grant from KDDI Research Inc., and generous gifts from Salesforce Inc., Microsoft Research, and Amazon Research. A. J. Bell and T. J. Sejnowski. An information-maximization approach to blind separation and blind deconvolution. Neural computation, 7(6):1129 1159, 1995. N. Calderon, E. Ben-David, A. Feder, and R. Reichart. Do Co Gen: Domain counterfactual generation for low resource domain adaptation. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 7727 7746, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.acl-long.533. URL https://aclanthology.org/2022.acl-long.533. Y.-L. Chou, C. Moreira, P. Bruza, C. Ouyang, and J. Jorge. Counterfactuals and causability in explainable artificial intelligence: Theory, algorithms, and applications. Information Fusion, 81: 59 83, 2022. P. Comon. Independent component analysis, a new concept? Signal processing, 36(3):287 314, 1994. N. Dai, J. Liang, X. Qiu, and X. Huang. Style transformer: Unpaired text style transfer without disentangled latent representation. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5997 6007, Florence, Italy, July 2019. Association for Computational Linguistics. doi: 10.18653/v1/P19-1601. URL https://aclanthology.org/ P19-1601. S. Dathathri, A. Madotto, J. Lan, J. Hung, E. Frank, P. Molino, J. Yosinski, and R. Liu. Plug and play language models: A simple approach to controlled text generation. In International Conference on Learning Representations, 2020. URL https://openreview.net/forum?id=H1ed Ey BKDS. M. Deng, B. Tan, Z. Liu, E. Xing, and Z. Hu. Compression, transduction, and creation: A unified framework for evaluating natural language generation. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 7580 7605, Online and Punta Cana, Dominican Republic, Nov. 2021. Association for Computational Linguistics. doi: 10.18653/v1/ 2021.emnlp-main.599. URL https://aclanthology.org/2021.emnlp-main.599. J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171 4186, Minneapolis, Minnesota, June 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1423. URL https://aclanthology.org/N19-1423. Q. Diao, M. Qiu, C.-Y. Wu, A. J. Smola, J. Jiang, and C. Wang. Jointly modeling aspects, ratings and sentiments for movie recommendation (jmars). In Proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining, pages 193 202, 2014. H. M. Dolatabadi, S. M. Erfani, and C. Leckie. Invertible generative modeling using linear rational splines. In S. Chiappa and R. Calandra, editors, The 23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy], volume 108 of Proceedings of Machine Learning Research, pages 4236 4246. PMLR, 2020. URL http://proceedings.mlr.press/v108/dolatabadi20a.html. C. Durkan, A. Bekasov, I. Murray, and G. Papamakarios. Neural spline flows. Advances in neural information processing systems, 32, 2019. D. Fan, Y. Hou, and C. Gao. Cf-vae: Causal disentangled representation learning with vae and causal flows. ar Xiv preprint ar Xiv:2304.09010, 2023. L. Gresele, P. K. Rubenstein, A. Mehrjou, F. Locatello, and B. Schölkopf. The incomplete rosetta stone problem: Identifiability results for multi-view nonlinear ica, 2019. H. Hälvä and A. Hyvarinen. Hidden markov nonlinear ica: Unsupervised learning from nonstationary time series. In Conference on Uncertainty in Artificial Intelligence, pages 939 948. PMLR, 2020. J. He, X. Wang, G. Neubig, and T. Berg-Kirkpatrick. A probabilistic formulation of unsupervised text style transfer. Co RR, abs/2002.03912, 2020. URL https://arxiv.org/abs/2002.03912. I. Higgins, L. Matthey, A. Pal, C. Burgess, X. Glorot, M. Botvinick, S. Mohamed, and A. Lerchner. beta-VAE: Learning basic visual concepts with a constrained variational framework. In International Conference on Learning Representations, 2017. URL https://openreview.net/ forum?id=Sy2fz U9gl. Z. Hu, Z. Yang, X. Liang, R. Salakhutdinov, and E. P. Xing. Toward controlled generation of text. In International conference on machine learning, pages 1587 1596. PMLR, 2017. C.-W. Huang, D. Krueger, A. Lacoste, and A. Courville. Neural autoregressive flows. In International Conference on Machine Learning, pages 2078 2087. PMLR, 2018. A. Hyvärinen and P. Pajunen. Nonlinear independent component analysis: Existence and uniqueness results. Neural networks, 12(3):429 439, 1999. A. Hyvarinen, H. Sasaki, and R. Turner. Nonlinear ica using auxiliary variables and generalized contrastive learning. In The 22nd International Conference on Artificial Intelligence and Statistics, pages 859 868. PMLR, 2019. P. Isola, J.-Y. Zhu, T. Zhou, and A. A. Efros. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017. V. John, L. Mou, H. Bahuleyan, and O. Vechtomova. Disentangled representation learning for non-parallel text style transfer. In A. Korhonen, D. R. Traum, and L. Màrquez, editors, Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, July 28August 2, 2019, Volume 1: Long Papers, pages 424 434. Association for Computational Linguistics, 2019. doi: 10.18653/v1/p19-1041. URL https://doi.org/10.18653/v1/p19-1041. N. S. Keskar, B. Mc Cann, L. R. Varshney, C. Xiong, and R. Socher. CTRL: A conditional transformer language model for controllable generation. Co RR, abs/1909.05858, 2019. URL http://arxiv. org/abs/1909.05858. I. Khemakhem, D. Kingma, R. Monti, and A. Hyvarinen. Variational autoencoders and nonlinear ica: A unifying framework. In International Conference on Artificial Intelligence and Statistics, pages 2207 2217. PMLR, 2020. L. Kong, S. Xie, W. Yao, Y. Zheng, G. Chen, P. Stojanov, V. Akinwande, and K. Zhang. Partial disentanglement for domain adaptation. In International Conference on Machine Learning, pages 11455 11472. PMLR, 2022. L. Kong, B. Huang, F. Xie, E. Xing, Y. Chi, and K. Zhang. Identification of nonlinear latent hierarchical models, 2023a. L. Kong, M. Q. Ma, G. Chen, E. P. Xing, Y. Chi, L.-P. Morency, and K. Zhang. Understanding masked autoencoders via hierarchical latent variable models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 7918 7928, June 2023b. A. Kumar, P. Sattigeri, and A. Balakrishnan. Variational inference of disentangled latent concepts from unlabeled observations. In 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, April 30 - May 3, 2018, Conference Track Proceedings. Open Review.net, 2018. URL https://openreview.net/forum?id=H1k G7GZAW. S. Lachapelle, P. Rodriguez, Y. Sharma, K. E. Everett, R. Le Priol, A. Lacoste, and S. Lacoste-Julien. Disentanglement via mechanism sparsity regularization: A new principle for nonlinear ica. In Conference on Causal Learning and Reasoning, pages 428 484. PMLR, 2022. G. Lample, S. Subramanian, E. Smith, L. Denoyer, M. Ranzato, and Y.-L. Boureau. Multiple-attribute text rewriting. In International Conference on Learning Representations, 2019. C. Li, X. Gao, Y. Li, B. Peng, X. Li, Y. Zhang, and J. Gao. Optimus: Organizing sentences via pre-trained modeling of a latent space. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 4678 4699, Online, Nov. 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.emnlp-main.378. URL https://aclanthology.org/2020.emnlp-main.378. D. Li, Y. Zhang, Z. Gan, Y. Cheng, C. Brockett, B. Dolan, and M. Sun. Domain adaptive text style transfer. In K. Inui, J. Jiang, V. Ng, and X. Wan, editors, Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, pages 3302 3311. Association for Computational Linguistics, 2019. doi: 10.18653/v1/D19-1325. URL https://doi.org/10.18653/v1/D19-1325. J. Li, M. Galley, C. Brockett, J. Gao, and B. Dolan. A diversity-promoting objective function for neural conversation models. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 110 119, San Diego, California, June 2016. Association for Computational Linguistics. doi: 10.18653/v1/N16-1014. URL https://aclanthology.org/N16-1014. J. Li, R. Jia, H. He, and P. Liang. Delete, retrieve, generate: a simple approach to sentiment and style transfer. In M. A. Walker, H. Ji, and A. Stent, editors, Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, New Orleans, Louisiana, USA, June 1-6, 2018, Volume 1 (Long Papers), pages 1865 1874. Association for Computational Linguistics, 2018. doi: 10.18653/ v1/n18-1169. URL https://doi.org/10.18653/v1/n18-1169. X. Li, X. Long, Y. Xia, and S. Li. Low resource style transfer via domain adaptive meta learning. In M. Carpuat, M. de Marneffe, and I. V. M. Ruíz, editors, Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022, Seattle, WA, United States, July 10-15, 2022, pages 3014 3026. Association for Computational Linguistics, 2022. doi: 10.18653/v1/2022.naacl-main.220. URL https://doi.org/10.18653/v1/2022.naacl-main.220. G. Liu, Z. Feng, Y. Gao, Z. Yang, X. Liang, J. Bao, X. He, S. Cui, Z. Li, and Z. Hu. Composable text controls in latent space with odes, 2022. F. Locatello, B. Poole, G. Rätsch, B. Schölkopf, O. Bachem, and M. Tschannen. Weakly-supervised disentanglement without compromises. In International Conference on Machine Learning, pages 6348 6359. PMLR, 2020. E. Mathieu, T. Rainforth, N. Siddharth, and Y. W. Teh. Disentangling disentanglement in variational autoencoders. In K. Chaudhuri and R. Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, ICML 2019, 9-15 June 2019, Long Beach, California, USA, volume 97 of Proceedings of Machine Learning Research, pages 4402 4412. PMLR, 2019. URL http://proceedings.mlr.press/v97/mathieu19a.html. G. Mita, M. Filippone, and P. Michiardi. An identifiable double vae for disentangled representations. In M. Meila and T. Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 7769 7779. PMLR, 18 24 Jul 2021. URL https://proceedings.mlr.press/v139/mita21a.html. K. Papineni, S. Roukos, T. Ward, and W. Zhu. Bleu: a method for automatic evaluation of machine translation. In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics, July 6-12, 2002, Philadelphia, PA, USA, pages 311 318. ACL, 2002. doi: 10.3115/ 1073083.1073135. URL https://aclanthology.org/P02-1040/. A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, and I. Sutskever. Language models are unsupervised multitask learners. 2019. C. Raffel, N. Shazeer, A. Roberts, K. Lee, S. Narang, M. Matena, Y. Zhou, W. Li, and P. J. Liu. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research, 21(1):5485 5551, 2020. S. Rao and J. R. Tetreault. Dear sir or madam, may i introduce the gyafc dataset: Corpus, benchmarks and metrics for formality style transfer. In North American Chapter of the Association for Computational Linguistics, 2018. P. Riley, N. Constant, M. Guo, G. Kumar, D. C. Uthus, and Z. Parekh. Textsettr: Label-free text style extraction and tunable targeted restyling. Co RR, abs/2010.03802, 2020. URL https: //arxiv.org/abs/2010.03802. A. Ross, T. Wu, H. Peng, M. E. Peters, and M. Gardner. Tailor: Generating and perturbing text with semantic controls. ar Xiv preprint ar Xiv:2107.07150, 2021. M. Shang, P. Li, Z. Fu, L. Bing, D. Zhao, S. Shi, and R. Yan. Semi-supervised text style transfer: Cross projection in latent space. In Conference on Empirical Methods in Natural Language Processing, 2019. T. Shen, T. Lei, R. Barzilay, and T. S. Jaakkola. Style transfer from non-parallel text by crossalignment. Co RR, abs/1705.09655, 2017. URL http://arxiv.org/abs/1705.09655. P. Sorrenson, C. Rother, and U. Köthe. Disentanglement by nonlinear ica with general incompressibleflow networks (gin). ar Xiv preprint ar Xiv:2001.04872, 2020. A. Sudhakar, B. Upadhyay, and A. Maheswaran. "transforming" delete, retrieve, generate approach for controlled text style transfer. In K. Inui, J. Jiang, V. Ng, and X. Wan, editors, Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Hong Kong, China, November 3-7, 2019, pages 3267 3277. Association for Computational Linguistics, 2019. doi: 10.18653/v1/D19-1322. URL https://doi.org/10.18653/v1/D19-1322. J. von Kügelgen, Y. Sharma, L. Gresele, W. Brendel, B. Schölkopf, M. Besserve, and F. Locatello. Self-supervised learning with data augmentations provably isolates content from style. ar Xiv preprint ar Xiv:2106.04619, 2021. H. Wang, C. Zhou, C. Yang, H. Yang, and J. He. Controllable gradient item retrieval. In Proceedings of the Web Conference 2021, pages 768 777, 2021. K. Wang, H. Hua, and X. Wan. Controllable unsupervised text attribute transfer via editing entangled latent representation. Advances in Neural Information Processing Systems, 32, 2019a. Y. Wang, Y. Wu, L. Mou, Z. Li, and W. Chao. Harnessing pre-trained neural networks with rules for formality style transfer. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3573 3578, 2019b. S. Xie, L. Kong, M. Gong, and K. Zhang. Multi-domain image generation and translation with identifiability guarantees. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=U2g8OGONA_V. P. Xu, J. C. K. Cheung, and Y. Cao. On variational learning of controllable representations for text without supervision. In International Conference on Machine Learning, 2019a. P. Xu, J. C. K. Cheung, and Y. Cao. On variational learning of controllable representations for text without supervision. In Proceedings of the 37th International Conference on Machine Learning, ICML 2020, 13-18 July 2020, Virtual Event, volume 119 of Proceedings of Machine Learning Research, pages 10534 10543. PMLR, 2020. URL http://proceedings.mlr.press/v119/ xu20a.html. R. Xu, T. Ge, and F. Wei. Formality style transfer with hybrid textual annotations. Ar Xiv, abs/1903.06353, 2019b. K. Yang and D. Klein. FUDGE: controlled text generation with future discriminators. In K. Toutanova, A. Rumshisky, L. Zettlemoyer, D. Hakkani-Tür, I. Beltagy, S. Bethard, R. Cotterell, T. Chakraborty, and Y. Zhou, editors, Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2021, Online, June 6-11, 2021, pages 3511 3535. Association for Computational Linguistics, 2021. doi: 10.18653/v1/2021.naacl-main.276. URL https://doi.org/10.18653/v1/2021. naacl-main.276. X. Yang, Y. Wang, J. Sun, X. Zhang, S. Zhang, Z. Li, and J. Yan. Nonlinear ICA using volumepreserving transformations. In International Conference on Learning Representations, 2022. URL https://openreview.net/forum?id=AMpki9kp8Cn. Z. Yang, Z. Hu, C. Dyer, E. P. Xing, and T. Berg-Kirkpatrick. Unsupervised text style transfer using language models as discriminators. Advances in Neural Information Processing Systems, 31, 2018. X. Zhang, J. Zhao, and Y. Le Cun. Character-level convolutional networks for text classification. Advances in neural information processing systems, 28, 2015. Y. Zheng, I. Ng, and K. Zhang. On the identifiability of nonlinear ICA: Sparsity and beyond. In A. H. Oh, A. Agarwal, D. Belgrave, and K. Cho, editors, Advances in Neural Information Processing Systems, 2022. URL https://openreview.net/forum?id=Wo1HF2w WNZb. J.-Y. Zhu, T. Park, P. Isola, and A. A. Efros. Unpaired image-to-image translation using cycleconsistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Oct 2017. R. S. Zimmermann, Y. Sharma, S. Schneider, M. Bethge, and W. Brendel. Contrastive learning inverts the data generating process, 2022. A.1 Implementation Details In this section, we first introduce the dataset. We then provide the network architecture details of MATTE. The hyperparameter selection criteria and the training details are summarized. A.1.1 Dataset We use four domain datasets to train our unsupervised model, i.e., Imdb, Yelp, Amazon and Yahoo, and follow the data split provided by Li et al. [2019]. The datasets can be downloaded via https://github.com/cookielee77/DAST. The dataset details can be found in Table 1. We set the sequence length L as 25, which is the 90 percentile of the sentence length of the training dataset. Therefore, shorter sentences are padded and longer sentences are clipped. The vocabulary size is set to 10000. For the sentiment transfer, we collect 100 positive sentences in dev set based on their sentiment labels to derive stransfer to flip the sentiment of the negative sentences in the test set, and vice versa. For the tense transfer, we use stanfordnlp tool 8 to identify the tense for the main verb, then collect 100 sentences of present tense in dev set to derive stransfer to flip the past-tense sentences. A.1.2 Model Architecture We summarize our network architecture below and describe it in detail in Table 7. Encoder: According to Xu et al. [2020], the encoder is fed with a text span x[t1 : t1 +m] extracted from the original sentence x, where t1 is a random word position index, m is set to 10 if t1 + m is smaller than L. Hword is the word embedding dimension, set to 256. Hlstm is the hidden states of LSTM, set to 1024. Hz is the dimension of the latent variable, set to 80. The output of the encoder is the µ, σ and z. All of them are in shape [BS, Hz]. Decoder: Decoder is fed with the input sentence span and the generated latent variable. The final reconstructed sentence span is one timestamp delay compared to the input span, i.e., xt1+1:(t1+1)+m. This is generated by applying beam search to the sequence of output probability over the vocabulary V . The Lrecon is to calculate the cross-entropy loss between the output probability and target sequence span. Content Flow rc: We apply Deep Dense Sigmoid Flow (DDSF) [Huang et al., 2018] to derive the content noise term. To incorporate the domain information, we leverage the domain embedding (after MLP) to parameterize the flow model. Style Flow rs: We apply spline flow [Durkan et al., 2019] to derive the noise term. Similarly, we use the conditional flow 9 with extra input. The conditional input is the combination of content variable and domain embedding. Specially, they are concatenated firstly and the result are fed into a MLP with Tanh activation to derive a attention score α The conditional input is actually the doct Prodcut. A.1.3 Training Training details. The models were implemented in Py Torch 2.0. and Python 3.9. The VAE network is trained for a maximum of 25 epochs and a mini-batch size of 64 is used. We use early stops if the validation reconstruction loss does not decrease for three epochs. For the encoder, we use the Adam optimizer and the learning rate of 0.001. For the decoder, we use SGD with a learning rate of 0.1. For the content and style flow, we use Adam optimizer and the learning rate is 0.001. We set three different random seeds and report the average results. Training objective. The VAE-based model is mainly trained with Lrecon and LVAE. We use a training trick to better jointly train the other three objectives. The Lsparsity could cram the information of s to c, while the Lc-mask is used to prevent the ill-posed situation where s have zero influence. Therefore, 8https://stanfordnlp.github.io/stanfordnlp/ 9The implementation refers to Conditioned Spline in https://docs.pyro.ai/en/stable/_modules/ pyro/distributions/transforms/spline.html Module Description Output 1. Encoder Encoder for Input sentence Input xt1:t1+m random span of sentence Word Emb get word embedding BS m Hword Bi-LSTM Bi-direction, 2layers BS m Hlstm Average Pooling sentencet-level Rep. BS Hlstm MLP µ and σ BS (2 Hz) reparameterization Sampling BS Hz 2. Domain Embedding Embedding Layer Input u number of domain udim BS udim 3. Content Flow rc Input: c, u domain as flow conditional input MLP u conditional context BS |Hrc| DDSF get content noise term c BS cdim 4. Style Flow rs Input: c, u, s content and domain as flow context Concatenate combine c and u BS (cdim + udim) MLP Tanh activation, get attention score α BS cdim Element-wise Multiplication α c BS cdim Spline Flow get style noise term s BS sdim 5. Decoder rs Input: z, xt1:t1+m generate the next token Bi-LSTM Bi-direction, 2layers BS m Hlstm MLP output word probability BS m V Table 7: MATTE overall architecture. DDSF is deep dense sigmoid flow, and Spline Flow is neural spline flow. m is the length of randomly extracted text span from input sentence x. we involve both Lsparsity and Lc-mask at the beginning of the training phrase. For Lpartial, it is used to sparsify the influence intersection but their separate influences change very frequently in the initial training stages. So we involve it after 3 epochs. Computing hardware and running time. We used a machine with the following CPU specifications: AMD EPYC 7282 CPU. We use NVIDIA Ge Force RTX 3090 with 24GB GPU memory. It costs approximately 190ms to run our model on this machine per epoch. A.2 Additional Results This section presents additional results on the hyperparameter sensitivity and the ablation studies on more datasets. A.2.1 Hyperparameter Sensitivity We discuss the effect of the three loss weights λsparsity,λpartial and λc-mask in the training objective. We have performed a grid search of λsparsity [1E-4,1E-3,1E-2], λpartial [3E-5,3E-3,3E-1] and λc-mask [1E-4,1E-3,1E-2]. The best configuration is [λsparisty, λpartial, λc-mask] = [1E-4,3E-3,1E-4]. The model performance is relatively sensitive to λsparsity, so we plot the sentiment accuracy and BLEU as a function of λsparsity in Figure 6. Figure 6: The Sentiment Acc (left) and BLEU (right) with different λsparsity. Among the four datasets, sentiment acc generally decreases as the λsparsity becomes larger; BLEU increases instead. This observation aligns with our content identifiability theory. We determine the λsparsity with the best G-score, i.e., 1E-4. A.2.2 Ablation Results on Amazon and Yahoo Datasets We show the ablation study of the Amazon and Yahoo datasets in Table 8. The full model achieves the best G-score and PPL on the two datasets. Causal Dep improves the BLEU and PPL. λsparsity greatly improves the content preservation at the cost of sentiment acc. After incorporating the Lpartial and Lc-mask, the sentiment acc is recovered. Amazon Yahoo Model Acc( ) BLEU( ) G-score( ) PPL( ) Acc( ) BLEU( ) G-score( ) PPL( ) Backbone 32.60 41.08 30.08 77.61 43.92 25.44 20.28 76.28 Indep [Kong et al., 2022] 48.80 39.50 31.76 77.95 51.70 23.44 21.12 56.95 Causal Dep 33.50 45.25 32.54 66.98 41.50 31.55 21.39 64.29 :w. / Lsparsity 27.10 62.73 34.73 63.37 27.32 48.21 25.16 60.03 :w. / Lpartial 33.10 58.54 34.04 64.42 38.12 41.74 28.03 58.04 :w. / Lc-mask(Full) 34.50 52.25 35.73 63.37 38.45 42.40 29.01 56.12 Table 8: Ablation results on sentiment transfer on two domains. Causal Dep incorporates style flow rs to model dependency of c on s, while Indep assumes the independence between the two variables. marks the improvements over Backbone, while over the Causal Dep. A.3 Semantic Preservation Measurement by CTC score As BLEU has limitations in capturing semantic relatedness beyond literal word-level overlap, we adopt CTC score [Deng et al., 2021] as a complementary evaluation for semantics preservation measurement. For the semantics alignment from a to b, CTC considers the matching embeddings, i.e., maximum cosine similarity of all the tokens in a with the tokens in b, and vice versa. Then, the final semantic preservation is in F1-style definition with one direction result as precision, and the other one as recall. The evaluation results of all the baselines and MATTE are shown in Table 9. The CTC score still favours Optimus and MATTE, with most inferior results on β-VAE, which are similar trends under the BLEU evaluation schema. Admittedly, the CTC score differences are less discriminative than BLEU this phoneme is also observed in Liu et al. [2022]. A.4 Diversity measurements for generated sentences To further demonstrate the generation degradation issue generate oversimplified and repetitious sentences, we use diversity-2 [Li et al., 2016], the ratio of distinct two-grams in all the two-grams in the generated sentences to evaluate the transferred sentences. The diversity-2 for original sentences is also included for better comparison. The results in Table 10 show that all the other methods except for β-VAE generated sentences with similar diversity-2 as the original sentences, but the sentences generated by β-VAE have much lower diversity than the original ones. IMDB Yelp Amazon Yahoo BGST 0.468 0.458 0.472 0.458 β-VAE 0.436 0.433 0.433 0.413 Joint Train 0.456 0.462 0.455 0.437 CPVAE 0.462 0.463 0.461 0.443 GPT2-FT 0.459 0.459 0.458 0.448 Optimus 0.465 0.468 0.465 0.446 Matte 0.465 0.464 0.466 0.452 Table 9: CTC score, a complementary evaluation for semantics preservation. β-VAE displays the least impressive performance, and Optimus and Matte exhibit the overall best results. Dataset IMDB (0.34) Yelp (0.63) Amazon (0.64) Yahoo(0.44) β-VAE 0.11 0.46 0.37 0.22 Joint Train 0.21 0.59 0.56 0.37 CPVAE 0.32 0.59 0.57 0.45 MATTE 0.32 0.62 0.61 0.45 Table 10: Diversity-2 for the transferred sentences. Diversity for the original sentences is included in the bracket for comparison. β-VAE has significantly fewer distinct 2-gram than original datasets. This results are consistent with evaluation results on BLEU. A.5 Proof for Theorem 1 The original Assumption 1 and Theorem 1 are copied below for reference. Assumption 1 (Content identification). i. g is smooth and invertible and its inverse g 1 is also smooth. ii. For all i {1, . . . , dx}, there exist {z(ℓ)}|Gi,:| ℓ=1 and T T , such that span({Jg(z(ℓ))i,:}|Gi,:| ℓ=1 ) = Rdz Gi,: and [Jg(z(ℓ))T]i,: R dz ˆGi,:. iii. For every pair (cjc, sjs) with jc [dc] and js {dc + 1, . . . , dz}, the influence of sjs is sparser than that of cjc, i.e., G:,jc 0 > G:,js 0. Theorem 1. We assume the data-generating process in Equation 1 with Assumption 1. If for given dimensions (dc, ds), a generative model (pˆc, pˆs|ˆc, ˆg) follows the same generating process and achieves the following objective: arg min pˆc,pˆs,ˆg jˆs {dc+1,...,dz} ˆG:,jˆs 0 subject to: pˆx(x) = px(x), x X, (3) then the estimated variable ˆc is an one-to-one mapping of the true variable c. That is, there exists an invertible function hc( ) such that ˆc = hc(c). Proof. We first define the notation z = [c, s] and the indeterminacy function: h := ˆg 1 g, which is an invertible function h : Z ˆZ as g is invertible by Assumption 1-i.. According to the chain rule, we have the following relation among the Jacobian matrices: Jˆg(ˆz) = Jg(z)J 1 h (z). (8) We define the support notations as follows: G := supp(Jg(z)), ˆG := supp(Jˆg(ˆz)), T := supp(J 1 h (z)). In the following, we will show that (jc, jˆs) / T for any jc {1, . . . , dc} and jˆs {dc + 1, . . . , dc + ds}. That is, [J 1 h (z)]jc,jˆs = 0, for any jc {1, . . . , dc} and jˆs {dc + 1, . . . , dc + ds}, which implies that c is not influenced by ˆs. Because of Assumption 1-ii., for any i {1, . . . , dv1 + dv2}, there exists {z(ℓ)}|Gi,:| ℓ=1 , such that span({Jg(z(ℓ))i,:}|Gi,:| ℓ=1 ) = Rdz Gi,:. Since {Jg(z(ℓ))i,:}|Gi,:| ℓ=1 forms a basis of Rdz Gi,:, for any j0 Gi,:, we can express canonical basis vector ej0 Rdz Gi,: as: ℓ Gi,: αℓ Jg(z(ℓ))i,:, (9) where αℓ R is a coefficient. Also, following Assumption 1-ii., there exists a deterministic matrix T where Tj1,j2 = 0 iff (j1, j2) T and Tj0,: = e j0T = X ℓ Gi,: αℓ Jg(z(ℓ))i,:T Rdz ˆGi,:, (10) where is because each element in the summation belongs to Rdz ˆGi,:. j Gi,:, Tj,: Rdz ˆGi,:. Equivalently, we have: (i, j) G, {i} Tj,: ˆG. (11) As both Jg and Jˆg are invertible, Jh(z) is an invertible matrix and thus has a non-zero determinant. Expressing Jh(z) with the Leibniz formulae gives: det(Jh(z)) = X j=1 Jg(z)σ(j),j where Pdz is the set of all dz-permutations. Equation 12 indicates that there exists σ Pdz, such that Qdz j=1 Jg(z)σ(j),j = 0. Equivalently, we have j [dz], (σ(j), j) T . (13) Therefore, for a specific jˆs {dc + 1, . . . , dz}, it follows that (σ(jˆs), jˆs) T . Further, Equation 11 shows that for any ix [dx], s.t., (ix, σ(jˆs)) G, we have {ix} Tσ(jˆs),: ˆG. Together, it follows that (ix, σ(jˆs)) G = (ix, jˆs) ˆG. (14) Equation 14 suggests that the column σ(jˆs) of the true generating function support G is included in the column jˆs of the estimated generating function support ˆG. Together with Assumption 1-iii., it follows that X jˆs {dc+1,...,dz} ˆG:,jˆs 0 X js {dc+1,...,dz} G:,js 0 , (15) where the permutation σ( ) connects the indices of s and those of ˆs. We note that Equation 15 is a lower-bound of the objective Equation 3, which can be attained by a minimizer ˆg = g. In the following, we show by contradiction that the support of J 1 h (z) does not contain (jc, jˆs), for any jc [dc] and any jˆs {dc + 1, . . . , dc}, i.e., (jc, jˆs) / T . We suppose that a specific (j c, j ˆs) T , where j c [dc] and any j ˆs {dc + 1, . . . , dc}. We note that the argument for Equation 14 also applies to (j1, j2) T for any j1, j2 [dz]. Thus, we would have (j c, j ˆs) T = (ix, j ˆs) ˆG, ix {i [dx] : (ix, j c) G}. (16) It would follow that X jˆs {dc+1,...,dz}\{j ˆs} ˆG:,jˆs 0 + ˆG:,j ˆs jˆs {dc+1,...,dz}\{j ˆs} G:,σ(jˆs) 0 + ˆG:,j ˆs jˆs {dc+1,...,dz}\{j ˆs} G:,σ(jˆs) 0 + G:,σ(j ˆs) G:,j c jˆs {dc+1,...,dz}\{j ˆs} G:,σ(jˆs) 0 + G:,j c 0 js {dc+1,...,dz} G:,js 0 , where the inequality (1) is due to Assumption 1-iii. that the influence of c on x is denser than that of s. However, as discussed above, there exists an optimizer that attains the lower-bound Equation 15. Equation 17 contradicts the minimization objective Equation 3. Therefore, (j c, j ˆs) T , for any j c [dc] and any j ˆs {dc + 1, . . . , dc}. As discussed above, this implies that c is not influenced by ˆs. Further, it follows from the invertibility of h( ) that [Jh(z)]jˆc,js = 0, for any jˆc {1, . . . , dc} and js {dc +1, . . . , dc +ds}, which implies that ˆc is not influenced by s. These two conditions and the invertibility of h( ) imply that ˆc and c form a one-to-one mapping. A.6 Proof for Theorem 2 The original Assumption iii. and Theorem 2 are copied below for reference. Assumption 2 (Partially intersecting influence supports). For every pair (cjc, sjs), the supports of their influences on x do not fully intersect, i.e., G:,jc G:,js 0 < min{ G:,jc 0 , G:,js 0}. Theorem 2. We follow the data-generating process Equation 1 and Assumption 1 and Assumption 2. We optimize the objective function in Equation 3 together with (jˆc,jˆs) {1,...,dc} {dc+1,...,dz} ˆG:,jˆc ˆG:,jˆs 0 . (4) The estimated style variable ˆs is a one-to-one mapping to the true variable s. That is, there exists an invertible mapping hs( ) between s and ˆs, i.e., ˆs = hs(s). Proof. As shown in Section A.5, there exists a dz-permutation σ( ) such that j [dz], (σ(j), j) T . Also, we have shown in Theorem 1 that (jc, jˆs) / T for jc [dc] and jˆs {dc + 1, . . . , dz}, which implies that σ(jˆs) {dc + 1, . . . , dz}. Thus, it follows that for any jˆc [dc], σ(jˆc) [dc]. In the following, we show by contradiction that (js, jˆc) T for any js {dc + 1, . . . , dz} and jˆc [dc]. We suppose that (j s, j ˆc) T . Analogous to Equation 16, we would have that (j s, j ˆc) T = (ix, j ˆc) ˆG, ix {i [dx] : (ix, j s) G}. (18) It would follow that ˆG:,j ˆc G:,σ(j ˆc) G:,j s. Also, to attain the objective Equation 3 in Theorem 1, we have j ˆs := σ 1(j s) {dc + 1, . . . , dz}, s.t., ˆG:,j ˆs = G:,j s. It would follow that ˆG:,j ˆc ˆG:,j ˆs. Further, we would have ˆG:,j ˆc ˆG:,j ˆs 0 = ˆG:,j ˆs 0 = G:,j s 0 > |{z} (2) G:,σ(j ˆc) G:,σ(j ˆs) 0 , (19) where (2) is due to Assumption 2. We note that the lower-bound for Equation 4 is X (jˆc,jˆs) {1,...,dc} {dc+1,...,dz} ˆG:,jˆc ˆG:,jˆs 0 X (jˆc,jˆs) {1,...,dc} {dc+1,...,dz} G:,σ(jˆc) G:,σ(jˆs) 0 (jc,js) {1,...,dc} {dc+1,...,dz} G:,jc G:,js 0 , which can be achieved by G = ˆG. Note that the lower-bounds for both Equation 3 and Equation 4 can be attained simultaneously by G = ˆG. Hence, optimizing the sum of the two objectives does not alter the optimal value of either. Applying a similar argument as that in Equation 15, we would have that X (jˆc,jˆs) {1,...,dc} {dc+1,...,dz} ˆG:,jˆc ˆG:,jˆs 0 (jˆc,jˆs) {1,...,dc} {dc+1,...,dz}\{(j ˆc,j ˆs)} ˆG:,jˆc ˆG:,jˆs 0 + ˆG:,j ˆc ˆG:,j ˆs (jˆc,jˆs) {1,...,dc} {dc+1,...,dz}\{(j ˆc,j ˆs)} G:,σ(jˆc) G:,σ(jˆs) 0 + ˆG:,j ˆc ˆG:,j ˆs (jˆc,jˆs) {1,...,dc} {dc+1,...,dz}\{(j ˆc,j ˆs)} G:,σ(jˆc) G:,σ(jˆs) 0 + G:,σ(j ˆc) G:,σ(j ˆs) 0 (jc,js) {1,...,dc} {dc+1,...,dz} G:,jc G:,js 0 , where (3) is due to Equation 19. Hence, this was not the minimizer of Equation 4. By contradiction, we have that (js, jˆc) / T for any js {dc + 1, . . . , dz} and jˆc [dc]. This implies that s is not influenced by ˆc. Further, it follows from the invertibility of h( ) that [Jh(z)]jˆs,jc = 0, for any jˆs {dc+1, . . . , dz} and jc [dc], which implies that ˆs is not influenced by c. These two conditions and the invertibility of h( ) imply that ˆs and s form a one-to-one mapping.