# the_autoencoding_variational_autoencoder__ffb99f53.pdf The Autoencoding Variational Autoencoder A. Taylan Cemgil Sumedh Ghaisas Krishnamurthy Dvijotham Pushmeet Kohli Does a Variational Auto Encoder (VAE) consistently encode typical samples generated from its decoder? This paper shows that the perhaps surprising answer to this question is No ; a (nominally trained) VAE does not necessarily amortize inference for typical samples that it is capable of generating. We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency. Our approach hinges on an alternative construction of the variational approximation distribution to the true posterior of an extended VAE model with a Markov chain alternating between the encoder and the decoder. The method can be used to train a VAE model from scratch or given an already trained VAE, it can be run as a post processing step in an entirely self supervised way without access to the original training data. Our experimental analysis reveals that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks. We provide experimental results on the Color Mnist and Celeb A benchmark datasets that quantify the properties of the learned representations and compare the approach with a baseline that is specifically trained for the desired property. 1 Introduction The variational Auto Encoder (VAE) is a deep generative model [10, 15] where one can simultaneously learn a decoder and an encoder from data. An attractive feature of the VAE is that while it estimates an implicit density model for a given dataset via the decoder, it also provides an amortized inference procedure for computing a latent representation via the encoder. While learning a generative model for data, the decoder is the key object of interest. However, when the goal is extracting useful features from data and learning a good representation, the encoder plays a more central role [20]. In this paper, we will focus primarily on the encoder and its representation capabilities. Learning good representations is one of the fundamental problems in machine learning to facilitate data-efficient learning and to boost the ability to transfer to new tasks. The surprising effectiveness of representation learning in various domains such as natural language processing [7] or computer vision [11] has motivated several research directions, in particular learning representations with desirable properties like adversarial robustness, disentanglement or compactness [1, 3, 4, 5, 12]. In this paper, our starting point is based on the assumption that if the learned decoder can provide a good approximation to the true data distribution, the exact posterior distribution (implied by the decoder) tends to possess many of the mentioned desired properties of a good representation, such as robustness. On a high level, we want to approximate properties of the exact posterior, in a way that supports representation learning. 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. Figure 1: Iteratively encoding and decoding a color MNIST image using a decoder and encoder (a) fitted a VAE with no observation noise (b) with AVAE. We find the drift in the generated images is an indicator of an inconsistency between the encoder and the decoder. One may naturally ask in what extent this goal is different from learning a standard VAE, where the encoder is doing amortized posterior inference and is directly approximating the true posterior. As we are using finite data for fitting within a restricted family of approximation distributions using local optimization, there will be a gap between the exact posterior and the variational approximation. We will illustrate that for finite data, even global minimization of the VAE objective is not sufficient to enforce natural properties which we would like a representation to have. We identify the source of the problem as an inconsistency between the decoder and encoder, attributed to the lack of autoencoding, see also [2]. We argue that, from a probabilistic perspective, autoencoding for a VAE should ideally mean that samples generated by the decoder can be consistently encoded. More precisely, given any typical sample from the decoder model, the approximate conditional posterior over the latents should be concentrated on the values that could be used to generate the sample. In this paper, we show (through analysis and experiments) that this is not the case for a VAE learned with normal training and we propose an additional specification to enforce the model to autoencode, bringing us to the choice of the title of this paper. Our Contributions: The key contributions of our work can be summarized as follows: We uncover that widely used VAE models are not autoencoding - samples generated by the decoder of a VAE are not mapped to the corresponding representations by the encoder, and an additional constraint on the approximating distribution is required. We derive a novel variational approach, Autoencoding VAE (AVAE), that is based on a new lower bound of the true marginal likelihood, and also enables data augmentation and self supervised density estimation in a theoretically principled way. We demonstrate that the learned representations achieve adversarial robustness. We show robustness of the learned representations in downstream classification tasks on two benchmark datasets: color MNIST and Celeb A. Our results suggest that a high performance can be achieved without adversarial training. 2 The Variational Autoencoder The VAE is a latent variable model that has the form Z p(Z) = N(Z; 0, I) X|Z p(X|Z, ) = N(X; g(Z; ), v I) (1) where N( ; µ, ) denotes a Gaussian density with mean and covariance parameters µ and , v is a positive scalar variance parameter and I is an identity matrix of suitable size. The mean function g(Z; ) is parametrized typically by a deep neural network with parameters and the conditional distribution is known as the decoder. We use here a conditionally Gaussian observation model but other choices are possible, such as Bernoulli or Poisson, with mean parameters given by g. To learn this model by maximum likelihood, one can use a variational approach to maximize the evidence lower bound (ELBO) defined as log p(X| ) hlog p(X|Z, )iq(Z|X, ) + hlog p(Z)iq(Z|X, ) + H[q(Z|X, )] B( , ) (2) h(a)q(a)da denotes the expectation of the test function h with respect to the distribution q, and, H is the entropy functional H[q] = hlog qiq. Here, q is an instrumental distribution, also known as the encoder, defined as q(Z|X, ) = N(Z; f µ(X; ), f (X; )) Here, the functions f µ and f are also chosen as deep neural networks with parameters . Using the reparametrization trick [10], it is possible to optimize and jointly by maximization of the ELBO using stochastic gradient descent aided by automatic differentiation. This mechanism is known as amortised inference, as once an encoder-decoder pair is trained, in principle the representation for a new data point can be readily calculated without the need of running a costly iterative inference procedure. The ELBO in (2) is defined for a single data point X. It can be shown that in a batch setting, maximization of the ELBO is equivalent to minimization of the Kullback-Leibler divergence KL(Q|P) = hlog(Q/P)i Q Q = (X)q(Z|X, ) P = p(X|Z, )p(Z) (3) with respect to and . Here, is ideally the true data distribution, but in practice it is replaced by the dataset, i.e., the empirical data distribution ˆ (X) = 1 N i δ(X x(i)). This form has also an intuitive interpretation as the minimization of the divergence between two alternative factorizations of a joint distribution with the desired marginals. See [21] for alternative interpretations of the ELBO including the one in (3). In the sequel, we will build our approach on this interpretation. 2.1 Extended VAE model for a pair of observations In this section, we will justify the VAE from an alternative perspective of learning a variational approximation in an extended model, essentially first arriving at an identical algorithm to the original VAE. This alternative perspective enables us to justify additional specifications that a consistent encoder/decoder pair should satisfy. Imagine the following extended VAE model for a pair of observations X, X0 (Z, Z0) p (Z, Z0) = N X|Z p(X|Z, ) = N(X; g(Z; ), v I) X0|Z0 p(X0|Z0, ) = N(X0; g(Z0; ), v I) (5) where the prior distribution p (Z0, Z) is chosen as symmetric with p(Z0) = p(Z), that is both marginals (here unit Gaussians) are equal in distribution. Here, I is the identity matrix and is a hyperparameter | | 1 that we will refer as the coupling strength. Note that we have p (Z0|Z) = N Z0; Z, (1 2)I . The two border cases = 1 and = 0, correspond to Z0 = Z, and, independence p(Z0, Z) = p(Z0)p(Z) respectively. This model defines the joint distribution P p(X0|Z0; )p(X|Z; )p (Z0, Z) (6) Proposition 2.1. Approximating P with a distribution of form Q ˆ (X)q(Z|X, )p (Z0|Z)p(X0|Z0, ) (7) where ˆ is the empirical data distribution and q and p are the encoder and the decoder models respectively gives the original VAE objective in (3). Proof: See Appendix A.1. Proposition 2.1 shows that we could have derived the VAE from this extended model as well. This derivation is of course completely redundant since we introduce and cancel out terms. However, we will argue that this extended model is actually more relevant from a representation learning perspective where a coupling 1 is preferable as this is the desired behaviour of the encoder. Example 2.1. Conditional Generation: Imagine that we generate a latent z p(Z) and an observation x p(X|Z = z; ) from the decoder of a VAE, but consequently discard z. Clearly, x is a sample from the marginal p(X; ). Now suppose we are told to generate a new sample using the same z as x0 p(X0|Z0 = z; ). As we have discarded z, our best bet would be using the extended model with = 1 and sample from P(X0|X = x; , = 1) instead. Example 2.2. Representation Learning: Imagine as in the previous example, that we generate a latent z and the corresponding observation x, but discard z. Suppose we are told to solve a classification task based on a classifier p(y|z). As we have discarded z, we would ideally compute an expectation under the true posterior p(y|Z) P(Z|X = x; )d Z. If our goal is learning the representation, and the task is unknown at training time, we wish to maintain as much as information about x. One strategy is asking a faithful reconstruction x0 given x, i.e., we would like to be able to sample from P(X0|X = x; , = 1) so the goal is not very different from conditional generation. In practice, P(X0|X = x; , ) is not available and we would be using the approximate transition kernel Q(X0|X; ) R Qd Zd Z0/ˆ (X) obtained from the variational distribution. A natural question is how good the approximation P(X0|X = x; ) Q(X0|X; ) is for different , and in particular 1. Therefore, we will investigate first some properties of this conditional distribution. Proposition 2.2. The marginal P(X0, X; , ) is symmetric in X0 and X, and its marginal does not depend on , i.e., P(X = x; , ) = p(X; ) for any | | 1. Proof: See Appendix A.2. The next proposition shows that if the encoder q is equal to the exact posterior, then Q(X0|X; ) is also exact Proposition 2.3. If the encoder q and decoder p satisfy the consistency condition q(Z|X, )p(X; ) = p(X|Z; )p(Z) (8) then, for all | | 1, we have Q(X0|X; )p(X; ) = P(X0, X; , ). We will say that Q(X0|X; ) is p(X; )-invariant. Proof: See Appendix A.3. The subtle point of Proposition 2.3 is that the exact posterior is valid for any coupling strength parameter . However, we have seen that the approximation computed by the VAE would be completely agnostic to our choice of . Moreover, note that even if the original VAE objective is globally minimized to get KL(ˆ (X)q(Z|X, )||p(X|Z, )p(Z)) = 0, this may not be sufficient to make sure that the transition kernel Q(X0|X; ) to be p(X; )-invariant. The empirical distribution ˆ has a discrete support and we have no control over the encoder out of this support. Intuitively, we need to introduce additional terms to the objective to steer the encoder if we would like to use the model as a conditional generator, or as a representation especially in the regime 1. For this purpose, we will need to make our approximating distribution to match the transition P(Z0|Z; ) = p (Z0|Z) that is by construction p(Z) invariant. In Figure 1, we compare what can happen when a VAE is learned nominally with an example of expected behaviour. Here, we show a sequence of images generated by iteratively sampling from a nominally learned encoder-decoder pair on color MNIST. The drift in the generated images points to a potential inconsistency between the decoder and the encoder. For further insight, we also discuss the special case of probabilistic PCA (Principal Component Analysis) in the appendix Section B, to give an analytically tractable example. 3 Autoencoding Variational Autoencoder (AVAE) Motivated by our analysis, we propose using the following extended model as a target distribution, P = p(X|Z; )p(Z)p (Z0|Z)u( X) (9) where is considered as a fixed hyper-parameter, not to be learned from data. Here X is an additional auxiliary observation (a delusion) that we introduce. We want the target model to be agnostic to its value, hence we choose a flat distribution u( X) = 1. We choose as the approximating distribution QAVAE = q(Z0| X; )p ( X|Z)q(Z|X; ) (X) The central idea of AVAE is making the encoder and the decoder to be consistent both on the training data and on the auxiliary observations generated by the decoder. We use the notation p to highlight that the decoder is considered as constant when used as a factor of QAVAE otherwise the original VAE bound would be invalid. Intuitively, the self generated delusion X should not be changing the true log likelihood as otherwise we would be modifying the original objective. The following proposition justifies our choice: Proposition 3.1. Assume that the consistency condition (8) is true, Then, the transition kernel defined as QAVAE(Z0|Z; , ) q(Z0| X, )p( X|Z; )d X is p(Z) invariant. Moreover, assume that the latent space has a lower dimension than the observation space ( Z 2 RNz and X 2 RNx, and Nz < Nx), and the decoder mean mapping g : RNz ! Xg, Figure 2: Graphical model of the extended target distribution P , and the variational approximation QAVAE. Here X is a sample generated by the decoder that is subsequently encoded by the encoder. is one-to-one, where Xg RNx is the image of g. Then in the limit when the observation noise variance v goes to zero (v ! 0) we have QAVAE(Z0|Z; , ) = p(Z0|Z; = 1). Proof: See Appendix A.4. The proposition shows that our choice of the variational approximation (10) is natural: it forces QAVAE(Z0, Z) to be close to the marginal of the exact posterior P(Z0, Z) = p (Z0|Z)p(Z). The choice QAVAE is also convenient because it uses the original encoder and decoder as building blocks. In the appendix Section C.1, we derive the variational objective BAVAE = KL(QAVAE||P ). The result is 1 BAVAE =+ KL( (X)q(Z|X; )||p(X|Z; )p(Z)) + hlog p (Z0|Z)i q(Z; ) q (Z0|Z; ) log q(Z0| X; ) q(Z0| X; ) q ( X; ) (10) where q(Z; ) q(Z|X; ) (X)d X, q ( X; ) p ( X|Z) q(Z; )d Z and q (Z0|Z; ) R q(Z0| X; )p ( X|Z)d X. In the appendix 1, we provide pseudocode with a stop-gradients primitive to avoid using contributions of dependent terms of q to the gradient computation. The resulting objective (10) is intuitive. The first term is identical to the standard VAE ELBO. The second term is a smoothness term that measures the distance between z (the representation that is used to generate the delusion) and z0 (the encoding of the delusion). The third term is an extra entropy term on auxiliary observations. 3.1 Illustration As a illustration, we show results for a discrete VAE, where both the encoder and the decoder can be represented as parametrized probability tables. Our goal in choosing a discrete model is visualization of the behaviour of the algorithms in a way that is independent from a particular neural network architecture. The details of this model are explained in the appendix Section D. Figure 3: Each Panel is showing (from north east clockwise order) heatmaps of probabilities (darker is higher) i) the decoder p(X0|Z0; ); ii) Q(X0|X; , )p(X; ); iii) the encoder q(Z|X; ); iv) Q(Z|Z0; , )p(Z0). See text for definitions. We can visually compare the learned transition models of the learned encoders and decoders for VAE and AVAE in Figure 3, where we show for each model (starting from north east clockwise order) i) the estimated decoder p(X0|Z0; ) (that is equal in distribution to p(X|Z; ) due to parameter tying), ii) the joint distribution Q(X0|X; , )p(X; ), that should be ideally symmetric, where Q(X0|X; , ) = p(X0|Z0 = z; )q(Z = z|X; ) with iii) the encoder q(Z|X; ) and iv) the joint distribution Q(Z|Z0; , )p(Z0) that should be also ideally symmetric and additionally should be close to identity where Q(Z|Z0; , ) = q(Z|X = x; )p(X0 = x|Z0; ). 1When a and b are proportional, we write log a =+ log b. We see in Figure 3 that the joint distribution Q(Z|Z0; , )p(Z0) is far from an identity mapping, and it is also not symmetrical. In contrast, the distributions learned by AVAE are enforced to be symmetrical, and the joint distribution is concentrated on the diagonal. Intermediate summary: Learning a VAE from data is an highly ill-posed problem and regularization is necessary. Here, instead of modifying the decoder model, we have proposed to constrain the encoder in such a way that it approaches the desired properties of the exact posterior. Our argument started with the extended model in proposition 2.1 that admits the VAE as a marginal. In 2.2, we show that the exact decoder model is by construction independent of the choice of the coupling strength . In plain language this means that, if we had access to the exact decoder and if we were able to do exact inference, any will be acceptable. In this ideal case, the extended model would be actually redundant, and in 2.3, we highlight the properties of an exact encoder. In practice however, we will be learning the decoder from data while doing only approximate amortized inference. Naturally, we want to still retain the properties of the exact posterior. We argue that the extended model is relevant for representation learning (see also example 2.2) and incorporate this explicitly to the encoder by AVAE, and in 3.1 we provide the justification that our choice coincides with the exact target conditional for = 1, the representation learning case, when we encode and decode consistently. We argue that encoder/decoder consistency is related to robustness. The existence of certain surprising adversarial examples [17], where an input image is classified as a very different class after slightly changing input pixels, can be attributed non-smoothness of the representation such as having a large Lipschitz constant, see [5]. The smooth encoder (SE) method proposed in [5] attempted to fix this by data augmentation while training the encoder. In this paper, we introduce a more general framework, (where SE is also a special case) and investigate methods that circumvent the need for computationally costly adversarial attacks during training. The data augmentation is achieved by using the learned generative model itself, as a factor of the approximating QAVAE distribution. The AVAE objective ensures that samples that can be generated by the decoder in the vicinity of the representations corresponding to the training inputs are consistently encoded. In the next experimental section, we will show that the consistency translates to a nontrivial adversarial robustness performance. While our approach does not provide formal guarantees, learning an encoder that retains properties of an exact posterior seems to be central in achieving adversarial robustness. 4 Experimental Results In this section, we will experimentally explore consequences of training with the AVAE objective in (10), implemented as Algorithm 1 in Appendix C. Optimizing the AVAE objective will change the learned encoder and decoder, and we expect that the additional terms will enforce a smooth representation leading to input perturbation robustness in downstream tasks. To test this claim, we will evaluate the encoder in terms of adversarial robustness, using an approach that will be described in the evaluation protocol. To see the effect of the new objective on the decoder and reconstruction quality, we will report Frechet Inception Distance (FID) [9], as well as the test mean square error (MSE). Models: AVAE will be compared to two models, i) VAE trained using the standard ELBO (2), and, ii) Smooth Encoder (SE), a method recently proposed by [5] that uses the same target model as in (6), but uses a different variational approximation. This approximation is computed using adversarial attacks to generate the auxiliary observations X. We provide a self contained derivation of this algorithm in Appendix C.2. We also include two hybrid models in our simulations. iii) AVAE-SS (Self Supervised) a post-training method, where a VAE is first trained normally and then post-trained only by samples generated from the decoder (without training data, see Figure 4 for the graphical model for details see appendix C.4). Finally, we also provide results for a model iv) SE-AVAE, a model that combines both the SE and the AVAE objectives and is concurrently trained using both adversarial attacks and self generated auxiliary observations (for details see appendix C.5). In all models, we use a latent space dimension of 64. Data and Tasks: Experiments are conducted on datasets color MNIST and Celeb A, using both MLP and Convnet architectures in the former, and only Convnet in the latter (for details see Appendix F). The color MNIST dataset has 2 separate downstream classification tasks (color, digit). For Celeb A, we have 17 different binary classification tasks, such as ( has mustache? or wearing hat? ). P ,AVAE-SS QAVAE-SS Figure 4: Graphical model of the AVAESS target distribution P ,AVAE-SS, and the variational approximation QAVAE-SS. Here both X and X is a sample generated by the decoder. The decoder factors (dotted arcs) is fixed (as pretrained by normal VAE). Task digit color Time MSE FID 0.0 0.1 0.2 0.0 0.1 0.2 VAE 93.8 5.8 0.0 100.0 19.9 2.0 1 1369.2 12.44 SE5 0.1 94.3 89.6 1.8 100.0 100.0 21.8 4 1372.5 13.01 SE5 0.2 95.7 92.6 87.3 100.0 99.9 99.9 4 1374.9 11.72 AVAE 97.3 88.1 54.8 100.0 99.8 87.7 1.5 1371.9 15.46 SE0.1-AVAE 97.4 93.6 24.5 100.0 100.0 60.0 4.7 1373.3 13.90 SE0.2-AVAE 97.6 94.2 79.8 100.0 100.0 83.2 4.7 1374.3 13.89 AVAE SS 94.1 72.8 20.8 100.0 99.6 56.8 1.5 1379.3 12.44 Table 1: Adversarial test accuracy (in percentage) of the representations for digit and color classification tasks on color MNIST (Conv Net). Evaluation attack radius is where pixels are normalized between [0, 1]. Time is the ratio of the wall-clock time of method to the time taken by the VAE. We show performance of the decoder in terms of MSE and FID score. For AVAE and SE, the coupling strengths are = 0.975 and SE = 0.95. The subscript of SE 0 is the radius 0 used during adversarial training of SE. Evaluation Protocol: The learned representations will be evaluated by their robustness against adversarial perturbations. For the evaluation of adversarial accuracy we employ a three step process, i) first we train an encoder-decoder pair agnostic to any classification task, and ii) subsequently, we freeze the encoder parameters and use the mean mapping f µ as a representation to train a linear classifier on top. Thus, each task specific classifier will share the common representation learned by the encoder. This linear classifier is trained normally, without any adversarial attacks. Finally, iii) we evaluate the adversarial accuracy of the resulting classifiers. For this, we compute an adversarial perturbation δ such that kδk1 using projected gradient descent (PGD). Here, is the attack radius and the optimization goal is changing the classification decision to a different class. The adversarial accuracy is reported in terms of percentage of examples where the attack is not able to find an adversarial example. Results: In Table 1, we show a comparison of the models on color MNIST, where we evaluate adversarial accuracy for different attack radii (0.0 means no attack). Our key observation is that AVAE increases the nominal accuracy and achieves high adversarial robustness, that extends well to strong attacks with a large radius. The surprising fact is that the method is trained completely agnostic to the particular attack type (e.g, PGD), or the attack radius , and it is able to achieve comparable performance to SE. Due to the computational burden of adversarial attacks, a single iteration of SE takes almost 2.7 times more computation time than a single AVAE iteration, making AVAE practical for large networks. We also observe that for small , both objectives can be combined to provide improved adversarial accuracy (digit, = 0.1), however, for higher radius ( = 0.2), the advantage seems to disappear. Another notable algorithm in Table 1 is AVAE-SS, where a pretrained VAE model can be further trained to significantly improve the robustness of the encoder. While this approach seems to be not competitive in terms of the final adversarial accuracy, the fact that it can improve robustness entirely in a self supervised way (Section C.4) is an attractive property. In Figure 5, we show the summary of several experiments with different architectures, models and various choices of the coupling strength hyper-parameter , on the color MNIST dataset. The first column shows the effect of varying , and as expected from our analysis, the adversarial accuracy is high for close to one. The middle column compares different architectures in terms of training dynamics. For MLP, we see that adversarial accuracy slowly increases, while for Convnet, the improvement is very rapid. We also observe that AVAE-SS can significantly improve the robustness of a pretrained VAE, but not to the level of the other methods (results with further are in Appendix G). The third column highlights the behaviour of the algorithms with increasing attack radius. We see Figure 5: Adversarial accuracy of digit classification task of color MNIST. (Top row) All results use MLP architecture (Bottom row) Convnet architecture. (Left Column), Adversarial accuracy as a function of the coupling strength for attack radius . (Middle column) Comparison of algorithms and test adversarial accuracy for attack radius = 0.1, for = 0.975 and SE = 0.95. (Right column) Adversarial accuracy as a function of the evaluation attack radius, SE model is trained with 0 = 0.1 (SE0.1). AVAE SE5 SE20 [5] SE5 AVAE AVAE SS Task / 0.0 0.1 0.0 0.1 0.0 0.1 0.0 0.1 0.0 0.1 Bald 97.9 85.2 97.9 72.0 97.4 86.5 97.9 87.0 97.8 70.0 Mustache 96.1 91.5 95.0 69.5 95.7 84.4 96.0 92.3 94.9 74.3 Necklace 86.1 78.4 87.8 56.7 88.0 78.9 86.1 80.3 88.0 59.7 Eyeglasses 95.4 68.9 95.9 20.3 95.7 33.0 95.4 67.5 94.4 57.1 Smiling 77.7 3.6 87.0 3.10 85.7 1.1 77.9 6.3 81.4 0.9 Lipstick 81.0 7.3 83.9 2.0 80.3 0.6 80.2 11.5 80.7 0.9 Time 2.2 3.1 7.8 4.3 2.2 MSE 7276.6 7208.8 N/A 7269.2 7347.3 FID 97.92 98.00 N/A 109.4 99.8 Table 2: Adversarial test accuracy (in percentage) of the representations for subset of classification tasks on Celeb A. For SE methods, superscript L in SEL denotes the number of PGD iterations used during training of the model. that for SE, the robustness does not extend beyond the radius that the model was trained for, whereas the accuracy of the AVAE trained model degrades gracefully. In Table 2, we show that the robustness of a representation learned with AVAE extends to a complex dataset such as Celeb A. The VAE results are omitted as they achieve around 0.0 adversarial accuracy and can be found in Table 3 in Appendix F, along with complete results on all tasks. We see that AVAE performs robustly in downstream tasks, and even surpasses robustness of SE5 and even challenges SE20 (reported by [5]) which is substantially requires more computation. We observe that AVAE-SS also achieves non trivial adversarial accuracy across tasks, in some cases even beating SE5. Finally we also report results for SE-AVAE model, showing non-trivial improvements over both SE and AVAE model in most of the downstream tasks. Yet, the increased robustness seems to come with a cost: in all the experiments, we observe an increase in FID (lower is better) and MSE for AVAE models, indicating a tendency of reduced decoder quality and test set reconstruction. We conjecture that this tradeoff may be the result of the encoder being much more constrained than in the VAE case. 5 Conclusions We show that VAE models derived from the canonical formulation in (1) are not unique. We used an alternative model (5) to show that the standard model is unable to capture some desired properties of the exact posterior that are important for representation learning. We proposed a principled alternative, the AVAE, and observed that it gives rise to robust representations without requiring adversarial training. In addition, we present a self supervised variant (AVAE-SS) that also exhibits robustness. The paper justifies, theoretically and experimentally, the modelling choices of AVAE. Using two benchmark datasets, we demonstrate that the approach can achieve surprisingly high adversarial accuracy without adversarial training. An important feature of the AVAE is that the likelihood function is still equal to the standard VAE likelihood. In doing so, we also provide a principled justification for data augmentation in a density estimation context, where naive data augmentation is not valid as it alters the data distribution. In our framework, the generated data points act like nuisance parameters of the approximating distribution and do not have an effect on the estimated decoder. Although we are not modifying the original likelihood function, currently, we observe a tradeoff: while the learnt encoder becomes more robust, the corresponding decoder seems to slightly suffer in quality, as measured by FID and MSE. It remains to be seen if there is actually a fundamental reason behind this and we believe that with more carefully designed training methods, both the decoder and the encoder could be in principle improved. The autoencoding specification is related loosely to the concept of cycle consistency, that is often enforced between two data modalities [22]. In [8], authors propose a supervised VAE model for pairs of data points with the same labels to enforce disentanglement by using extra relational information (similar relational models include [18, 13]). In contrast, our approach does only change the encoder. Our method is a VAE based representation learning approach, and there is a large literature on the subject, see, e.g. [20], however models are often not evaluated on their robustness properties. Another fruitful approach in representation learning, especially for the image modality, is contrastive learning [14, 6]. These approaches can challenge and surpass supervised approaches in terms of label efficiency and accuracy, and it remains a future work to investigate the links with our work and whether or not these models can also be used for learning robust representations. 6 Societal Impact General Research Direction. In recent years, researchers have trained deep generative models that can generate synthetic examples, often indistinguishable from natural data. The high quality of these samples suggest that these models may be able to learn latent representations useful for other downstream tasks. Learning such representations without task specific supervision facilitates transfer to yet unseen, future tasks. This also fosters label efficiency, and interpretability. Unsupervised representation learning would have a high societal impact as it could enable learning representations from data that can be shared with a wider community of researchers who do not have the computational resources for training such a representations, or do not have direct access to training data due to privacy/security/commercial considerations. However, the properties of these representations in terms of test accuracy, robustness, privacy preservation must be carefully studied before their release, especially if systems will be deployed in the real world. In the current work, we have taken a step towards learning representations in an unsupervised way, that exhibit robustness against transformations of the input. Ethical Considerations. The current work studies representations learned by a specific generative model, the VAE and shares a finding that training a VAE by enforcing an additional natural autoencoding specification is able to provide significant robustness on the learned representation without adversarial training. The study is not proposing a particular system for a specific application. The human face dataset Celeb A is choosen as a standard benchmark dataset with several attributes to illustrate the viability of the approach. Still, we have decided to exclude potentially sensitive and subjective attributes from the original dataset, such as big-nose or Asian , and use only 17 neutral attributes that we have selected using our own judgment. Acknowledgments and Disclosure of Funding We would like to thank Andriy Minh for the fruitful discussions and excellent feedback. [1] Alessandro Achille and Stefano Soatto. Emergence of invariance and disentanglement in deep representations. The Journal of Machine Learning Research, 19(1):1947 1980, 2018. [2] Guillaume Alain and Yoshua Bengio. What regularized auto-encoders learn from the data- generating distribution. J. Mach. Learn. Res., 15(1):3563 3593, January 2014. [3] Yoshua Bengio. Deep learning of representations for unsupervised and transfer learning. In Proceedings of ICML workshop on unsupervised and transfer learning, pages 17 36, 2012. [4] Christopher P. Burgess, Irina Higgins, Arka Pal, Loic Matthey, Nick Watters, Guillaume Desjardins, and Alexander Lerchner. Understanding disentangling in β-VAE. ar Xiv e-prints, page ar Xiv:1804.03599, Apr 2018. [5] A. T. Cemgil, S. Ghaisas, K. Dvijotham, and P. Kohli. Adversarially robust representations with smooth encoders. In Proceedings of the Eighth International Conference on Learning Representations, ICLR 2020, 2020. [6] Ting Chen, Simon Kornblith, Mohammad Norouzi, and Geoffrey Hinton. A simple framework for contrastive learning of visual representations. In International Conference on Machine Learning, 2020. [7] Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. Bert: Pre-training of deep bidirectional transformers for language understanding, 2018. [8] Ananya Harsh Jha, Saket Anand, Maneesh Singh, and VSR Veeravasarapu. Disentangling factors of variation with cycle-consistent variational auto-encoders. In Proceedings of the European Conference on Computer Vision (ECCV), pages 805 820, 2018. [9] Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS 17, page 6629 6640, Red Hook, NY, USA, 2017. Curran Associates Inc. [10] Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. In Yoshua Bengio and Yann Le Cun, editors, 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014. [11] Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolutional neural networks. In F. Pereira, C. J. C. Burges, L. Bottou, and K. Q. Weinberger, editors, Advances in Neural Information Processing Systems 25, pages 1097 1105. Curran Associates, Inc., 2012. [12] Francesco Locatello, Stefan Bauer, Mario Lucic, Gunnar Raetsch, Sylvain Gelly, Bernhard Schölkopf, and Olivier Bachem. Challenging common assumptions in the unsupervised learning of disentangled representations. In Kamalika Chaudhuri and Ruslan Salakhutdinov, editors, Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pages 4114 4124, Long Beach, California, USA, 09 15 Jun 2019. PMLR. [13] Christos Louizos, Xiahan Shi, Klamer Schutte, and Max Welling. The Functional Neural Process. ar Xiv e-prints, page ar Xiv:1906.08324, Jun 2019. [14] Aaron van den Oord, Yazhe Li, and Oriol Vinyals. Representation learning with contrastive predictive coding. ar Xiv preprint ar Xiv:1807.03748, 2018. [15] Danilo Jimenez Rezende, Shakir Mohamed, and Daan Wierstra. Stochastic backpropagation and approximate inference in deep generative models. In Eric P. Xing and Tony Jebara, editors, Proceedings of the 31st International Conference on Machine Learning, volume 32 of Proceedings of Machine Learning Research, pages 1278 1286, Bejing, China, 22 24 Jun 2014. PMLR. [16] Sam T Roweis. Em algorithms for pca and spca. In Advances in neural information processing systems, pages 626 632, 1998. [17] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfel- low, and Rob Fergus. Intriguing properties of neural networks. ar Xiv preprint ar Xiv:1312.6199, 2013. [18] Da Tang, Dawen Liang, Tony Jebara, and Nicholas Ruozzi. Correlated Variational Auto- Encoders. ar Xiv e-prints, page ar Xiv:1905.05335, May 2019. [19] Michael E. Tipping and Christopher M. Bishop. Probabilistic principal component analysis. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 61(3):611 622, 1999. [20] Michael Tschannen, Olivier Bachem, and Mario Lucic. Recent advances in autoencoder-based representation learning. In Third workshop on Bayesian Deep Learning (Neur IPS 2018), 2018. [21] Shengjia Zhao, Jiaming Song, and Stefano Ermon. Infovae: Balancing learning and inference in variational autoencoders. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 5885 5892, 2019. [22] Jun-Yan Zhu, Taesung Park, Phillip Isola, and Alexei A Efros. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE international conference on computer vision, pages 2223 2232, 2017.