# sparse_communication_via_mixed_distributions__6fdb8de7.pdf Published as a conference paper at ICLR 2022 SPARSE COMMUNICATION VIA MIXED DISTRIBUTIONS António Farinhas 1, Wilker Aziz 2, Vlad Niculae 3, André F. T. Martins 1,4 1Instituto de Telecomunicações, Instituto Superior Técnico (Lisbon ELLIS Unit), 2ILLC, University of Amsterdam, 3Iv I, University of Amsterdam, 4Unbabel {antonio.farinhas,andre.t.martins}@tecnico.ulisboa.pt, {w.aziz,v.niculae}@uva.nl Neural networks and other machine learning models compute continuous representations, while humans communicate mostly through discrete symbols. Reconciling these two forms of communication is desirable for generating human-readable interpretations or learning discrete latent variable models, while maintaining endto-end differentiability. Some existing approaches (such as the Gumbel-Softmax transformation) build continuous relaxations that are discrete approximations in the zero-temperature limit, while others (such as sparsemax transformations and the Hard Concrete distribution) produce discrete/continuous hybrids. In this paper, we build rigorous theoretical foundations for these hybrids, which we call mixed random variables. Our starting point is a new direct sum base measure defined on the face lattice of the probability simplex. From this measure, we introduce new entropy and Kullback-Leibler divergence functions that subsume the discrete and differential cases and have interpretations in terms of code optimality. Our framework suggests two strategies for representing and sampling mixed random variables, an extrinsic ( sample-and-project ) and an intrinsic one (based on face stratification). We experiment with both approaches on an emergent communication benchmark and on modeling MNIST and Fashion-MNIST data with variational auto-encoders with mixed latent variables. Our code is publicly available. 1 INTRODUCTION Historically, discrete and continuous domains have been considered separately in machine learning, information theory, and engineering applications: random variables (r.v.) and information sources are chosen to be either discrete or continuous, but not both (Shannon, 1948). In signal processing, one needs to opt between discrete (digital) and continuous (analog) communication, whereas analog signals can be converted into digital ones by means of sampling and quantization. Discrete latent variable models are appealing to facilitate learning with less supervision, to leverage prior knowledge, and to build more compact and interpretable models. However, training such models is challenging due to the need to evaluate a large or combinatorial expectation. Existing strategies include the score function estimator (Williams, 1992; Mnih & Gregor, 2014), pathwise gradients (Kingma & Welling, 2014) combined with a continuous relaxation of the latent variables (such as the Concrete distribution, Maddison et al. (2017); Jang et al. (2017)), and sparse parametrizations (Correia et al., 2020). Pathwise gradients, in particular, require continuous approximations of quantities that are inherently discrete, sometimes requiring proxy gradients (Jang et al., 2017; Maddison et al., 2017), sometimes giving the r.v. different treatment in different terms of the same objective (Jang et al., 2017), sometimes creating a discrete-continuous hybrid (Louizos et al., 2018). Since discrete variables and their continuous relaxations are so prevalent, they deserve a rigorous mathematical study. Throughout, we will use the name mixed variable to denote a hybrid variable that takes on both discrete and continuous outcomes. This work takes a first step into a rigorous study of mixed variables and their properties. We will call communication through mixed variables sparse communication: its goal is to retain the advantages of differentiable computation but still be able to represent and approximate discrete symbols. Our main contributions are: We provide a direct sum measure as an alternative to the Lebesgue and counting measures used for continuous and discrete variables, respectively (Halmos, 2013). The direct sum measure hinges Published as a conference paper at ICLR 2022 Table 1: Discrete, continuous, and mixed distributions considered in this work, all involving the probability simplex K 1. For each distribution we indicate if it assigns probability mass to all faces of the simplex or only some, if it is multivariate (K 2), and, for mixed distributions, if it is characterized extrinsically (sample-and-project) or intrinsically (based on face stratification). Distribution All faces? Multivariate? Intrinsic? Categorical Discrete % Yes ! Dirichlet, Logistic-Normal, Concrete Continuous % Yes ! Hard Concrete, Rectified Gaussian Mixed ! No % Extrinsic % K-D Hard Concrete, Gaussian-Sparsemax (this paper) Mixed ! Yes ! Extrinsic % Mixed Dirichlet (this paper) Mixed ! Yes ! Intrinsic ! on a face lattice stratification of polytopes, including the probability simplex, avoiding the need for Dirac densities when expressing densities with point masses in the boundary of the simplex ( 3). We use the direct sum measure to formally define Kth-dimensional mixed random variables. We provide extrinsic ( sample-and-project ) and intrinsic (based on face stratification) characterizations of these variables, leading to several new distributions: the K-D Hard Concrete, the Gaussian Sparsemax, and the Mixed Dirichlet (summarized in Table 1). See Figure 1 for an illustration. We propose a new direct sum entropy and Kullback-Leibler divergence, which decompose as a sum of discrete and continuous (differential) entropies/divergences. We provide an interpretation in terms of optimal code length, and we derive an expression for the maximum entropy ( 4). We illustrate the usefulness of our framework by learning mixed latent variable models in an emergent communication task and with VAEs to model Fashion-MNIST and MNIST data ( 5). 2 BACKGROUND We assume throughout an alphabet with K 2 symbols, denoted [K] = {1, . . . , K}. Symbols can be encoded as one-hot vectors ek. RK denotes the K-dimensional Euclidean space, RK >0 its strictly positive orthant, and K 1 RK the probability simplex, K 1 := {y RK | y 0, 1 y = 1}, with vertices {e1, . . . , e K}. Each y K 1 can be seen as a vector of probabilities for the K symbols, parametrizing a categorical distribution over [K]. The support of y K 1 is the set of nonzero-probability symbols supp(y) := {k [K] | yk > 0}. The set of full-support categoricals corresponds to the relative interior of the simplex, ri( K 1) := {y K 1 | supp(y) = [K]}. 2.1 TRANSFORMATIONS FROM RK TO K 1 In many situations, there is a need to convert a vector of real numbers z RK (scores for the several symbols, often called logits) into a probability vector y K 1. The most common choice is the softmax transformation (Bridle, 1990): y = softmax(z) exp(z). Since the exponential function is strictly positive, softmax reaches only the relative interior ri( K 1), that is, it never returns a sparse probability vector. To encourage more peaked distributions (but never sparse) it is common to use a temperature parameter β > 0, by defining softmaxβ(z) := softmax(β 1z). The limit case β 0+ corresponds to the indicator vector for the argmax, which returns a one-hot distribution indicating the symbol with the largest score. While the softmax transformation is differentiable (hence permitting end-to-end training with the gradient backpropagation algorithm), the argmax function has zero gradients almost everywhere. With small temperatures, numerical issues are common. A direct sparse probability mapping is sparsemax (Martins & Astudillo, 2016), the Euclidean projection onto the simplex: sparsemax(z) := arg miny K 1 y z . Unlike softmax, sparsemax reaches the full simplex K 1, including the boundary, often returning a sparse vector y, without sacrificing differentiability almost everywhere. With K = 2 and parametrizing z = (z, 1 z), sparsemax becomes a hard sigmoid, sparsemax (z, 1 z) 1 = max{0, min{1, z}}. We will come back to this point in 3.3. Other sparse transformations include α-entmax (Peters et al., 2019; Blondel et al., 2020), top-k softmax (Fan et al., 2018; Radford et al., 2019), and others (Laha et al., 2018; Sensoy et al., 2018; Kong et al., 2020; Itkina et al., 2020). Published as a conference paper at ICLR 2022 (1, 0, 0) (0, 1, 0) (1, 0, 0) (0, 1, 0) 0.0 0.2 0.4 0.6 0.8 1.0 0.0 Figure 1: Multivariate distributions over K 1. Standard distributions, like the Logistic-Normal (left), assign zero probability to all faces but ri( K 1). Our mixed distributions support assigning probability to the full simplex, including its boundary: the Gaussian-Sparsemax (right) induces a distribution over the 1-dimensional edges (shown as a histogram), and assigns Pr{(1, 0, 0)} = .022. 2.2 DENSITIES OVER THE SIMPLEX Let us now switch from deterministic to stochastic maps. Denote by Y a r.v. taking on values in the simplex K 1 with probability density function p Y (y). The density of a Dirichlet r.v. Y Dir(α), with α RK >0 is p Y (y; α) QK k=1 yαk 1 k . Sampling from a Dirichlet produces a point in ri( K 1), and, although a Dirichlet can assign high density to y close to the boundary of the simplex when α < 1, a Dirichlet sample can never be sparse. A Logistic-Normal r.v. (Atchison & Shen, 1980), also known as Gaussian-Softmax by analogy to other distributions to be presented, is given by the softmax-projection of a multivariate Gaussian r.v. with mean z and covariance Σ: Y = softmax(z + Σ 1 2 N) with Nk N(0, 1). Since the softmax is strictly positive, the Logistic-Normal places no probability mass to points in the boundary of K 1. A Concrete (Maddison et al., 2017), or Gumbel-Softmax (Jang et al., 2017), r.v. is given by the softmax-projection of K independent Gumbel r.vs., each with mean zk: Y = softmaxβ(z + G) with Gk Gumbel(0, 1). Like in the previous cases, a Concrete draw is a point in ri( K 1). When the temperature β approaches zero, the softmax approaches the indicator for argmax and Y becomes closer to a categorical r.v. (Luce, 1959; Papandreou & Yuille, 2011). Thus, a Concrete r.v. can be seen as a continuous relaxation of a categorical. 2.3 TRUNCATED UNIVARIATE DENSITIES Binary Hard Concrete. For K = 2, a point in the simplex can be represented as y = (y, 1 y) and the simplex is isomorphic to the unit interval, 1 [0, 1]. For this binary case, Louizos et al. (2018) proposed a Hard Concrete distribution which stretches the Concrete and applies a hard sigmoid transformation (which equals the sparsemax with K = 2, per 2.1) as a way of placing point masses at 0 and 1. These stretch-and-rectify techniques enable assigning probability mass to the boundary of 1 and are similar in spirit to the spike-and-slab feature selection method (Mitchell & Beauchamp, 1988; Ishwaran et al., 2005) and for sparse codes in variational auto-encoders (Rolfe, 2017; Vahdat et al., 2018). We propose in 3.3 a more general extension to K 2. Rectified Gaussian. Rectification can be applied to other continuous distributions. A simple choice is the Gaussian distribution, to which one-sided (Hinton & Ghahramani, 1997) and two-sided rectifications (Palmer et al., 2017) have been proposed. Two-sided rectification yields a mixed r.v. in [0, 1]. Writing y = (y, 1 y) and z = (z, 1 z), this distribution has the following density: p Y (y) = N(y; z, σ2) + 1 erf(z/( 2σ)) 2 δ0(y) + 1 + erf((z 1)/( 2σ)) 2 δ1(y), (1) where δs(y) is a Dirac delta density. Extending such distributions to the multivariate case is nontrivial. For K > 2, a density expression with Diracs would be cumbersome, since it would require a combinatorial number of Diracs of several orders, depending on whether they are placed at a vertex, edge, face, etc. Another annoyance is that Dirac deltas have differential entropy, which prevents information-theoretic treatment. The next section shows how we can obtain densities that assign mass to the full simplex while avoiding Diracs, by making use of the face lattice and a new base measure. Published as a conference paper at ICLR 2022 3 FACE STRATIFICATION AND MIXED RANDOM VARIABLES 3.1 THE FACE LATTICE Let P be a convex polytope whose vertices are bit vectors (i.e., elements of {0, 1}K). Examples are the probability simplex K 1, the hypercube [0, 1]K, and marginal polytopes of structured variables (Wainwright & Jordan, 2008). The combinatorial structure of a polytope is determined by its face lattice (Ziegler, 1995, 2.2), which we now describe. A face of P is any intersection of P with a closed halfspace such that none of the interior points of P lie on the boundary of the halfspace; we denote by F(P) the set of all faces of P and by F(P) := F(P) \ { } the set of proper faces. We denote by dim(f) the dimension of a face f F(P). Thus, the vertices of P are 0-dimensional faces, and P itself is a face of the same dimension as P, called the maximal face . Any other face of P can be regarded as a lower-dimensional polytope. The set F(P) has a partial order induced by set inclusion, that is, it is a partially ordered set (poset), and more specifically a lattice. The full polytope P can be decomposed uniquely as the disjoint union of the relative interior of its faces, which we call face stratification: P = F f F(P) ri(f). For example, the simplex 2 is composed of its face ri( 2) (i.e., excluding the boundary), three edges (excluding the vertices in the corners), and three vertices (the corners). This is represented schematically in Figure 2. Likewise, the square [0, 1]2 is composed of its maximal face (0, 1)2, four edges (excluding the corners) and four vertices (the corners). The partition above implies that any subset A P can be represented as a tuple A = (Af)f F(P), where Af = A ri(f); and the sets Af are all disjoint. Figure 2: Left: Decomposition of a simplex as the disjoint union of the relative interior of its faces. Right: Hasse diagram of the face lattice a DAG where points represent faces, and a directed path from a point to another represents face inclusion. Simplex and hypercube. If P is the simplex K 1, each face corresponds to an index set I [K], i.e., it can be expressed as f I = {p K 1 | supp(p) I}, with dimension dim(f I) = |I| 1: the set of distributions assigning zero probability mass outside I. The set F( K 1) has 2K 1 elements. Since 1 [0, 1], the hypercube [0, 1]K can be regarded as a product of K binary probability simplices. It has 3K nonempty faces for each dimension we choose between {0}, {1}, and [0, 1]. We experiment with P { K 1, [0, 1]K} in 5. 3.2 MIXED RANDOM VARIABLES Categorical distributions assign probability only to the vertices of K 1. In the opposite extreme, the densities listed in 2.2 assign probability mass to the maximal face only, that is, Pr{y f I} = R f I p Y (y) = 0 for any I = [K]. Any proper density (without Diracs) has this limitation, since nonmaximal faces have zero Lebesgue measure in RK 1. While for K = 2 it is possible to circumvent this by defining densities that contain Dirac functions (as in 2.3), this becomes cumbersome for K > 2. Fortunately, there is a more elegant construction that does not require generalized functions. The key is to replace the Lebesgue measure by a measure inspired by face stratification. Definition 1 (Direct sum measure). The direct sum measure on a polytope P is f F(P) µf(A ri(f)), (2) where µf is the dim(f)-dimensional Lebesgue measure for dim(f) > 0, and the counting measure for dim(f) = 0. We show in App. A that µ is a valid measure on P under the product σ-algebra of its faces. We can then define probability densities p Y (y) w.r.t. this base measure and use them to compute probabilities of measurable subsets of P. Such distributions can equivalently be defined as follows: (i) define a probability mass function PF (f) on F(P), and (ii) for each face f F(P), define a probability Published as a conference paper at ICLR 2022 density p Y |F (y | f) over ri(f). Random variables with a distribution of this form have a discrete part and a continuous part, thus we call them mixed random variables. This is formalized as follows. Definition 2 (Mixed random variable). A mixed r.v. is a r.v. Y over a polytope P, including the boundary. Let F be the corresponding discrete r.v. over F(P), with PF (f) = Pr{y ri(f)}. Since the mapping Y F is deterministic, we have p Y (y) = PF (f)p Y |F (y | f) for y ri(f). The probability of a set A P is given by: Pr{y A} = Z A p Y (y)dµ = X f F(P) PF (f) Z A ri(f) p Y |F (y | f). (3) Equation (3) may be regarded as a manifestation of the law of total probability mixing discrete and continuous variables. Using (3), we can write expectations over a mixed r.v. as Ep Y [g(Y )] = EPF Ep Y |F [g(Y ) | F = f] , where g : P R. (4) Both discrete and continuous distributions are recovered with our definition: If PF (f) = 0 for dim(f) > 0, we have a categorical distribution, which only assigns probability to the 0-faces. In the other extreme, if PF (P) = 1, we have a continuous distribution confined to ri(P). That is, mixed random variables include purely discrete and purely continuous r.vs. as particular cases. To parametrize distributions of high-dimensional mixed r.vs., it is not efficient to consider all degrees of freedom suggested in Definition 2, since there can be exponentially many faces. Instead, we need to derive parametrizations that exploit the lattice structure of the faces. We next build upon this idea. 3.3 EXTRINSIC AND INTRINSIC CHARACTERIZATIONS There are two possible characterizations of mixed random variables: an extrinsic one, where one starts with a distribution over the ambient space (e.g. RK) and then applies a deterministic, noninvertible, transformation that projects it to P; and an intrinsic one, where one specifies a mixture of distributions directly over the faces of P, by specifying PF and p Y |F for each f F(P). We next provide constructions for both cases: We extend the Hard Concrete and Rectified Gaussian distributions reviewed in 2.3, which are instances of the extrinsic characterization, to K 2; and we present a new Mixed Dirichlet distribution which is an instance of the intrinsic characterization. K-D Hard Concrete. We define the K-D Hard Concrete as the following generative story: Y Concrete(z, β), Y = sparsemax(λY ), with λ 1. (5) When K = 2, sparsemax becomes a hard sigmoid and we recover the binary Hard Concrete ( 2.3). For K 2 this is a projection of a stretched Concrete r.v. onto the simplex the larger λ, the higher the tendency of this projection to hit a non-maximal face of the simplex and induce sparsity. Gaussian-Sparsemax. A similar idea (but without any stretching required) can be used to obtain a sparsemax counterpart of the Logistic-Normal in 2.2, which we call Gaussian-Sparsemax : N N(0, I), Y = sparsemax(z + Σ1/2N). (6) Unlike the Logistic-Normal, the Gaussian-Sparsemax can assign nonzero probability mass to the boundary of the simplex. When K = 2, we recover the double-sided rectified Gaussian described in 2.3. In that case, using Dirac deltas, the density with respect to the Lebesgue measure in R has the form in (1). With θ0 = 1 erf(z/( 2σ)) 2 , θ1 = 1+erf((z 1)/( 2σ)) 2 and θc = 1 θ0 θ1, the same distribution can be expressed intrinsically via the density p Y (y) = PF (f)p Y |F (y | f) as PF ({0}) = θ0, PF ({1}) = θ1, PF ([0, 1]) = θc, p Y |F (y | F = [0, 1]) = N(y;z,σ2)/θc. (7) For K > 2, expressions for PF and p Y |F (i.e., an intrinsic representation) are less direct; we express those distributions as a function of the orthant probability of multivariate Gaussians in App. B. Mixed Dirichlet. We now propose an intrinsic mixed distribution over K 1, the Mixed Dirichlet, whose generative story is as follows. First, a face F = f I is sampled with probability PF (f I; w) = exp(w φ(f I) log Z(w)), φk(f I) = ( 1)1 [k I] , (8) Published as a conference paper at ICLR 2022 where w RK is the natural parameter (a.k.a. log-potentials), φ(f I) { 1, 1}K is the sufficient statistic, and log Z(w) is the log-normalizer. We then parametrize a Dirichlet distribution over the relative interior of f I, that is, Y |F = f I Dir(α(I)), where α(I) R|I| >0. For a compact parametrization, we have a single K-dimensional vector α of concentration parameters, one parameter per vertex, and α(I) gathers the coordinates of α associated with the vertices in f. The normalizer of (8) can be evaluated in time O(K) via the forward algorithm (Baum & Eagon, 1967) on a directed acyclic graph (DAG) that encodes each non-empty corner I [K] of the face lattice as a path. Similarly, we can draw independent samples by stochastic traversals through this DAG. The graph needed for this construct and the associated algorithms are detailed in App. C. 4 INFORMATION THEORY FOR MIXED RANDOM VARIABLES Now that we have the theoretical pillars for mixed random variables, we proceed to defining information theoretic quantities for them: their entropy and Kullback-Leibler divergence. Direct sum entropy and KL divergence. The entropy of a r.v. X with respect to a measure µ is: X p X(x) log p X(x)dµ(x), (9) where p X(x) is a probability density satisfying R X p X(x)dµ(x) = 1. When X is finite and µ is the counting measure, the integral becomes a sum and we recover Shannon s discrete entropy, which is non-negative and upper bounded by log |X|. When X Rk is continuous and µ is the Lebesgue measure, we recover the differential entropy, which can be negative and, for compact X, is upper bounded by the logarithm of the volume of X. When relaxing a discrete r.v. to a continuous one in a variational model (e.g., using the Concrete distribution), correct variational lower bounds require switching to differential entropy (Maddison et al., 2017) (although discrete entropy is sometimes used (Jang et al., 2017)). This is problematic since the differential entropy is not a limit case of the discrete entropy (Cover & Thomas, 2012). Our direct sum entropy, defined below, obviates this. The key idea is to plug in (9) the direct sum measure (2). Since Y F is deterministic, we have H(Y, F) = H(Y ). This leads to: Definition 3 (Direct sum entropy and KL divergence). The direct sum entropy of a mixed r.v. Y is H (Y ) := H(F) + H(Y | F) (10) f F(P) PF (f) log PF (f) | {z } discrete entropy f F(P) PF (f) Z f p Y |F (y | f) log p Y |F (y | f) | {z } differential entropy The KL divergence between distributions p Y (PF , p Y |F ) and q Y (QF , q Y |F ) is: D KL(p Y q Y ) := DKL(PF QF ) + Ef PF DKL(p Y |F ( | F = f) q Y |F ( | F = f) (11) f F(P) PF (f) log PF (f) QF (f) | {z } discrete KL f F(P) PF (f) Z f p Y |F (y | f) log p Y |F (y | f) q Y |F (y | f) | {z } continuous KL As shown in Definition 3, the direct sum entropy and the KL divergence have two components: a discrete one over faces and an expectation of a continuous one over each face. The KL divergence is always non-negative and it becomes + if supp(PF ) supp(QF ) or if there is some face where supp(p Y |F =f) supp(q Y |F =f).1 App. D provides more information theoretic extensions. Relation to optimal codes. The direct sum entropy and KL divergence have an interpretation in terms of optimal coding, described in Proposition 1 for the case P = K 1 and proven in App. D. In words, the direct sum entropy is the average length of the optimal code where the sparsity pattern of y K 1 must be encoded losslessly and where there is a predefined bit precision for the fractional 1In particular, this means that mixed distributions shall not be used as a relaxation in VAEs with purely discrete priors using the ELBO rather, the prior should be also mixed. Published as a conference paper at ICLR 2022 entries of y. On the other hand, the KL divergence between p Y and q Y expresses the additional average code length if we encode variable Y p Y (y) with a code that is optimal for distribution q Y (y), and it is independent of the required bit precision. Proposition 1. Let Y be a mixed r.v. in K 1. In order to encode the face of Y losslessly and to ensure an N-bit precise encoding of Y in that face we need the following bits on average: H N(Y ) = H (Y ) + N k=1 (k 1) X f F( K 1) dim(f)=k 1 PF (f). (12) Entropy of Gaussian-Sparsemax. Revisiting the Gaussian-Sparsemax with K = 2 ( 3.3) and using the intrinsic representation (7), the direct sum entropy becomes H(Y ) = H(F) + H(Y |F) = P0 log P0 P1 log P1 Z 1 0 N(y; z, σ2) log N(y; z, σ2)dy. (13) This leads to simple expressions for the entropy and KL divergence, detailed in App. E. We use these expressions in our experiments with mixed bit-vector VAEs in 5. Entropy of Mixed Dirichlet. The intrinsic representation of the Mixed Dirichlet ( 3.3) allows for O(K) computation of H(F) and DKL(PF ||QF ) via dynamic programming, necessary for H (Y ) and D KL(p Y ||q Y ) (see App. C for details). The continuous parts H(Y |F) and EPF [DKL(p Y |F =f||q Y |F =f)] require computing an expectation with an exponential number of terms (one per proper face) and can be approximated with an MC estimate by sampling from PF and assessing closed-form the differential entropy of Dirichlet distributions over the sampled faces. Maximum entropy density in the full simplex. An important question is to characterize maximum entropy mixed distributions. If we consider only continuous distributions, confined to the maximal face ri( K 1), the answer is the flat distribution, with entropy log (K 1)! , corresponding to a deterministic F which puts all probability mass in this maximal face. But constraining ourselves to a single face is quite a limitation, and in particular knowing this constraint provides valuable information that intuitively should reduce entropy.2 What if we consider densities that assign probability to the boundaries? This is answered by the next proposition, proved in App. F. Proposition 2 (Maxent mixed distribution on simplex). Let Y be a mixed r.v. on K 1 with PF (f) 2N(dim(f)) dim(f)! and p Y |F (y | f) uniform for each f F( K 1). Then, Y has maximal direct sum entropy H N,max(Y ). The value of the entropy is: H N,max(Y ) = log K k 2N(k 1) (k 1)! = log L(1) K 1( 2N). (14) where L(α) n (x) denotes the generalized Laguerre polynomial (Sonine, 1880). For example, for K = 2, the maximal entropy distribution is PF ({0}) = PF ({1}) = 1/(2 + 2N), PF ([0, 1]) = 2N/(2 + 2N), with H N,max(Y ) = log(2 + 2N). Therefore, in the worst case, we need log2(2 + 2N) bits on average to encode Y 1 with bit precision N. For K = 3, H N,max(Y ) = log(3 + 3 2N + 22N 1). A plot of the entropy as a function of K is shown in Figure 4, App. F. 5 EXPERIMENTS We report experiments in three representation learning tasks:3 an emergent communication game, where we assess the ability of mixed latent variables over K 1 to induce sparse communication 2At the opposite extreme, if we only assign probability to pure vertices, i.e., if we are constrained to minimal faces, the maximal discrete entropy is log K. We will see that looking at all faces further increases entropy. 3Appendix G.4 contains a fourth experiment regression towards voting proportions where we use our Mixed Dirichlet as a likelihood function in a generalized linear model. Published as a conference paper at ICLR 2022 Table 2: Test results. Left: Emergent communication success average and standard error over 10 runs. Random guess baseline: 6.25%. Right: Fashion-MNIST bit-vector VAE NLL (bits/dim, lower is better). Entropy column legend: continuous/discrete/mixed, : estimated, =: exact. Method Success (%) Nonzeros Gumbel-Softmax 78.84 8.07 256 Gumbel-Softmax ST 49.96 9.51 1 K-D Hard Concrete 76.07 7.76 21.43 17.56 Gaussian-Sparsemax 80.88 0.50 1.57 0.02 Method Entropy NLL Sparsity (%) Binary Concrete C 3.60 0 Gumbel-Softmax D = 3.49 0 Gumbel-Softmax ST D = 3.57 100 Hard Concrete X 3.57 45.64 Gaussian-Sparsemax X 3.53 82.82 Gaussian-Sparsemax X = 3.49 73.83 between two agents; a bit-vector VAE modeling Fashion-MNIST images (Xiao et al., 2017), where we compare several mixed distributions over [0, 1]K and study the impact of the direct sum entropy ( 4); and a mixed-latent VAE, where we experiment with the Mixed Dirichlet over K 1 to model MNIST images (Le Cun et al., 2010). Full details about each task and the experimental setup are reported in App. G. Throughout, we report our three mixed distributions described in 3.3 alongside the following distributions: Concrete (Maddison et al. (2017); a purely continuous density), Gumbel Softmax (Jang et al. (2017); like above, but using a discrete KL term in a VAE, which introduces inconsistency), Gumbel-Softmax ST (Jang et al. (2017); categorical latent variable, but using Concrete straight-through gradients), and Dirichlet and Gaussian, when applicable. Emergent communication. This is a cooperative game between two agents: a sender sees an image and emits a single-symbol message from a fixed vocabulary [K]; a receiver reads the symbol and tries to identify the correct image out of a set. We follow the architecture in Lazaridou & Baroni (2020); Havrylov & Titov (2017), choosing a vocabulary size of 256 and 16 candidate images. For the mixed variable models, the message is a mixed symbol , i.e., a (sparse) point in K 1. Table 2 reports communication success (accuracy of the receiver), along with the average number of nonzero entries in the samples at test time. Gaussian-Sparsemax learns to become very sparse, attaining the best overall communication success, exhibiting in general a better trade-off than the K-D Hard Concrete. Bit-Vector VAE. We model Fashion-MNIST following Correia et al. (2020), with 128 binary latent bits, maximizing the ELBO. Each latent bit has a uniform prior for purely discrete and continuous models, and the maxent prior (Prop. 2), which assigns probability 1/3 to each face of 1, in the mixed case. For Hard Concrete (Louizos et al., 2018) and Gaussian-Sparsemax, we use mixed latent variables and our direct sum entropy, yielding a coherent objective and unbiased gradients. For the Hard Concrete, we estimate H(Y | X = x) via MC; for Gaussian-Sparsemax, we consider both cases: using an MC estimate and computing the entropy exactly using the expression derived in 3.3. In Table 2 (right) we report an importance sampling estimate (1024 samples) of negative log-likelihood (NLL) on test data, normalized by the number of pixels. For Gaussian-Sparsemax, exact entropy computation improves performance, matching Gumbel-Softmax while being sparse. Table 3: MNIST test results (avg. of 5 runs). Categorical is marginalized exactly. Method D R NLL Gaussian (R10) 76.67 19.94 91.12 Dirichlet 78.62 19.94 93.81 Categorical 164.72 2.28 166.95 Gumbel-Softmax ST 171.76 1.70 168.50 Mixed Dirichlet 90.34 19.39 106.59 Mixed-Latent VAE on MNIST. We model the binarized MNIST dataset using a mixed-latent VAE, with y 10 1, a maximum-entropy prior on Y , and a factorized decoder. We use a feed-forward encoder with a single-hidden-layer to predict the variational distributions (e.g., log-potentials w and concentrations α, for Mixed Dirichlet). For Mixed Dirichlet, gradients are estimated via a combination of implicit reparametrization (Figurnov et al., 2018) and the score function estimator (Mnih & Gregor, 2014), see App. C for details. We report a single-sample MC estimate of distortion (D) and rate (R) as well as a 1000-samples importance sampling estimate of NLL. The Gaussian prior has access to all of the R10, whereas all other priors are constrained to the simplex 10 1, namely, its relative interior (Dirichlet), its vertices (Gumbel-Softmax ST, and Categorical), or all of it (Mixed Dirichlet). The Dirichlet model clusters digits as well as a Gaussian VAE does, while purely discrete models struggle to discover structure. App. G lists qualitative evidence. Compared to a Dirichlet, the Mixed Dirichlet makes some plausible confusions (Fig. 3, Published as a conference paper at ICLR 2022 left), but, crucially, it solves part of the problem by allocating digits to specific faces (Fig. 3, right). Even though we employed a sampled self-critic, Mixed Dirichlet seems to suffer from variance of SFE, which may explain why Mixed Dirichlet under-performs in terms of NLL. 6 RELATED WORK Figure 3: t SNE plots of: posterior samples y K 1 (left), predicted log-potentials w RK (right). Colors encode digit label (not available to models). Clusters are formed in latent space and in how digits are assigned to faces (recall that w parameterizes a Gibbs distribution over F( K 1)). Existing strategies to learn discrete latent variable models include the score function estimator (Williams, 1992; Mnih & Gregor, 2014) and pathwise gradients combined with a Concrete relaxation (Maddison et al., 2017; Jang et al., 2017). The latter is often used with straightthrough gradients and by combining a continuous latent variable with a discrete entropy, which is theoretically not sound. Besides the Concrete, continuous versions of the Bernoulli and Categorical have also been proposed (Loaiza-Ganem & Cunningham, 2019; Gordon-Rodriguez et al., 2020), but they are also purely continuous densities, assigning zero probability mass to the boundary of the simplex. Our approach is inspired by discrete-continuous hybrids based on truncation and rectification (Hinton & Ghahramani, 1997; Palmer et al., 2017; Louizos et al., 2018), which have been proposed for univariate distributions. We generalize this idea to arbitrary dimensions replacing truncation by sparse projections to the simplex. The direct sum measure and our proposed intrinsic sampling strategies are related to the concept of manifold stratification proposed in the statistical physics literature (Holmes-Cerfon, 2020). Other mixed r.vs. have also been recently considered (for a few special cases) in (Bastings et al., 2019; Murady, 2020; van der Wel, 2020). For instance, Burkhardt & Kramer (2019) induce sparsity in hierarchical models by deterministically masking the concentration parameters of a Dirichlet distribution yielding a special case of Mixed Dirichlet with degenerate PF , trained with the straight-through estimator. Our paper generalizes these attempts, providing solid theoretical support for manipulating mixed distributions in higher dimensions. Discrete and continuous representations in the context of emergent communication have been discussed in Foerster et al. (2016); Lazaridou & Baroni (2020). Discrete communication is computationally more challenging because it prevents direct gradient backpropagation (Foerster et al., 2016; Havrylov & Titov, 2017), but it is hypothesized that this discrete bottleneck forces the emergence of symbolic protocols. Our mixed variables bring a new perspective into this problem, leading to sparse communication, which lies in between discrete and continuous communication and supports gradient backpropagation without the need for straight-through gradients. 7 CONCLUSIONS We presented a mathematical framework for handling mixed random variables, which are discrete/continuous hybrids. Key to our framework is the use of a direct sum measure as an alternative to the Lebesgue-Borel and the counting measures, which considers all faces of the simplex. We developed generalizations of information theoretic concepts for mixed symbols, and we experimented on emergent communication and on variational modeling of MNIST and Fashion-MNIST images. We believe the framework described here is only scratching the surface. For example, the Mixed Dirichlet is just one example of an intrinsic mixed distribution; more effective intrinsic parametrizations may exist and are a promising avenue. While our main focus was on the probability simplex and hypercube, mixed structured variables are another promising direction, enabled by our theoretical characterization of direct sum measures, which can be defined for any polytope via their face lattice (Ziegler, 1995; Grünbaum, 2003). Current methods for structured variables include perturbations (continuous, Corro & Titov (2019); Berthet et al. (2020); Paulus et al. (2020)), and sparsemax extensions (discrete, Niculae et al. (2018); Correia et al. (2020)), lacking tractable densities. Published as a conference paper at ICLR 2022 Ethics statement. We highlight indirect impact of our work through applications such as generation and explainable AI, where improved performance must be carefully scrutinized. Our proof-of-concept emergent communication, like previous work, uses Image Net, whose construction exhibits societal biases, including racism and sexism (Crawford & Paglen, 2019) that communicating agents may learn. Even in unsupervised settings, biases may be learned by communicating agents. Reproducibility statement. We now discuss the efforts that have been made to ensure reproducibility of our work. We state the full set of assumptions of our theoretical results and include complete proofs in App. A, App. B, App. C, App. D, App. E, and App. F. Additionally, code and instructions to reproduce our experiments are available at https://github.com/deep-spin/ sparse-communication. We report the standard error over 10 runs for the emergent communication experiment due to the high variance of results across seeds. We include the type of computing resources used in our experiments in App. G.5. ACKNOWLEDGMENTS We would like to thank Mário Figueiredo, Gonçalo Correia, and the Deep SPIN team for helpful discussions, Tim Vieira, who answered several questions about order statistics, Sam Power, who pointed out to manifold stratification, and Juan Bello-Rivas, who suggested the name mixed random variables. This work was built on open-source software; we acknowledge Van Rossum & Drake (2009); Oliphant (2006); Virtanen et al. (2020); Walt et al. (2011); Pedregosa et al. (2011), and Paszke et al. (2019). AF and AM are supported by the P2020 program MAIA (LISBOA-01-0247FEDER045909), the European Research Council (ERC St G Deep SPIN 758969), and by the Fundação para a Ciência e Tecnologia through project PTDC/CCI-INF/4703/2021 (PRELUNA) and contract UIDB/50008/2020. WA received funding from the European Union s Horizon 2020 research and innovation programme under grant agreement No 825299 (Go URMET). VN is partially supported by the Hybrid Intelligence Centre, a 10-year programme funded by the Dutch Ministry of Education, Culture and Science through the Netherlands Organisation for Scientific Research (https:// hybrid-intelligence-centre.nl). J Atchison and Sheng M Shen. Logistic-normal distributions: Some properties and uses. Biometrika, 67(2):261 272, 1980. Jasmijn Bastings, Wilker Aziz, and Ivan Titov. Interpretable neural predictions with differentiable binary variables. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 2963 2977, Florence, Italy, July 2019. Association for Computational Linguistics. doi: 10.18653/v1/P19-1284. URL https://aclanthology.org/P19-1284. Leonard E. Baum and J. A. Eagon. An inequality with applications to statistical estimation for probabilistic functions of Markov processes and to a model for ecology. Bulletin of the American Mathematical Society, 73(3):360 363, 1967. doi: bams/1183528841. Quentin Berthet, Mathieu Blondel, Olivier Teboul, Marco Cuturi, Jean-Philippe Vert, and Francis Bach. Learning with differentiable pertubed optimizers. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 9508 9519. Curran Associates, Inc., 2020. URL https://proceedings.neurips. cc/paper/2020/file/6bb56208f672af0dd65451f869fedfd9-Paper.pdf. Mathieu Blondel, André F.T. Martins, and Vlad Niculae. Learning with fenchel-young losses. Journal of Machine Learning Research, 21(35):1 69, 2020. URL http://jmlr.org/papers/v21/ 19-021.html. Diane Bouchacourt and Marco Baroni. How agents see things: On visual representations in an emergent language game. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 981 985, Brussels, Belgium, October-November 2018. Association for Computational Linguistics. doi: 10.18653/v1/D18-1119. URL https://aclanthology. org/D18-1119. Published as a conference paper at ICLR 2022 John S. Bridle. Probabilistic interpretation of feedforward classification network outputs, with relationships to statistical pattern recognition. In Françoise Fogelman Soulié and Jeanny Hérault (eds.), Neurocomputing, pp. 227 236, Berlin, Heidelberg, 1990. Springer Berlin Heidelberg. ISBN 978-3-642-76153-9. Sophie Burkhardt and Stefan Kramer. Decoupling sparsity and smoothness in the dirichlet variational autoencoder topic model. Journal of Machine Learning Research, 20(131):1 27, 2019. URL http://jmlr.org/papers/v20/18-569.html. John B Conway. A course in functional analysis, volume 96. Springer, 2019. Gonçalo Correia, Vlad Niculae, Wilker Aziz, and André Martins. Efficient marginalization of discrete and structured latent variables via sparsity. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 11789 11802. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/ paper/2020/file/887caadc3642e304ede659b734f79b00-Paper.pdf. Caio Corro and Ivan Titov. Differentiable perturb-and-parse: Semi-supervised parsing with a structured variational autoencoder. In International Conference on Learning Representations, 2019. URL https://openreview.net/forum?id=BJlg Nh0q KQ. Thomas M Cover and Joy A Thomas. Elements of Information Theory. John Wiley & Sons, 2012. Kate Crawford and Trevor Paglen. Excavating AI: The politics of images in machine learning training sets. Accessed 28 May 2021, , 2019. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scale hierarchical image database. In 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 248 255, 2009. doi: 10.1109/CVPR.2009.5206848. Angela Fan, Mike Lewis, and Yann Dauphin. Hierarchical neural story generation. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 889 898, Melbourne, Australia, July 2018. Association for Computational Linguistics. doi: 10.18653/v1/P18-1082. URL https://aclanthology.org/P18-1082. Mikhail Figurnov, Shakir Mohamed, and Andriy Mnih. Implicit reparameterization gradients. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL https://proceedings.neurips.cc/paper/2018/file/ 92c8c96e4c37100777c7190b76d28233-Paper.pdf. Jakob Foerster, Ioannis Alexandros Assael, Nando de Freitas, and Shimon Whiteson. Learning to communicate with deep multi-agent reinforcement learning. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016. URL https://proceedings.neurips.cc/ paper/2016/file/c7635bfd99248a2cdef8249ef7bfbef4-Paper.pdf. Alan Genz. Numerical computation of multivariate normal probabilities. Journal of computational and graphical statistics, 1(2):141 149, 1992. Elliott Gordon-Rodriguez, Gabriel Loaiza-Ganem, and John Cunningham. The continuous categorical: a novel simplex-valued exponential family. In Hal Daumé III and Aarti Singh (eds.), Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pp. 3637 3647. PMLR, 13 18 Jul 2020. URL https://proceedings. mlr.press/v119/gordon-rodriguez20a.html. Branko Grünbaum. Convex polytopes, volume 221. Springer, Graduate Texts in Mathematics, 2003. Paul R Halmos. Measure Theory, volume 18. Springer, 2013. Serhii Havrylov and Ivan Titov. Emergence of language with multi-agent games: Learning to communicate with sequences of symbols. In I. Guyon, U. V. Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. URL https://proceedings.neurips. cc/paper/2017/file/70222949cc0db89ab32c9969754d4758-Paper.pdf. Published as a conference paper at ICLR 2022 Geoffrey E Hinton and Zoubin Ghahramani. Generative models for discovering sparse distributed representations. Philosophical Transactions of the Royal Society of London. Series B: Biological Sciences, 352(1358):1177 1190, 1997. Miranda Holmes-Cerfon. Simulating sticky particles: A Monte Carlo method to sample a stratification. The Journal of Chemical Physics, 153(16):164112, 2020. Hemant Ishwaran, J Sunil Rao, et al. Spike and slab variable selection: frequentist and bayesian strategies. Annals of Statistics, 33(2):730 773, 2005. Masha Itkina, Boris Ivanovic, Ransalu Senanayake, Mykel J Kochenderfer, and Marco Pavone. Evidential sparsification of multimodal latent spaces in conditional variational autoencoders. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 10235 10246. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper/2020/file/ 73f95ee473881dea4afd89c06165fa66-Paper.pdf. Eric Jang, Shixiang Gu, and Ben Poole. Categorical reparameterization with gumbel-softmax. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. Open Review.net, 2017. URL https://openreview. net/forum?id=rk E3y85ee. Diederik P. Kingma and Max Welling. Auto-encoding variational bayes. In 2nd International Conference on Learning Representations, ICLR 2014, Banff, AB, Canada, April 14-16, 2014, Conference Track Proceedings, 2014. URL https://openreview.net/forum?id= 33X9fd2-9Fy Zd. Weiwei Kong, Walid Krichene, Nicolas Mayoraz, Steffen Rendle, and Li Zhang. Rankmax: An adaptive projection alternative to the softmax function. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 633 643. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/ paper/2020/file/070dbb6024b5ef93784428afc71f2146-Paper.pdf. Anirban Laha, Saneem Ahmed Chemmengath, Priyanka Agrawal, Mitesh Khapra, Karthik Sankaranarayanan, and Harish G Ramaswamy. On controllable sparse alternatives to softmax. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL https://proceedings.neurips.cc/paper/2018/file/ 6a4d5952d4c018a1c1af9fa590a10dda-Paper.pdf. Angeliki Lazaridou and Marco Baroni. Emergent multi-agent communication in the deep learning era. preprint ar Xiv:2006.02419, 2020. Yann Le Cun, Corinna Cortes, and CJ Burges. MNIST handwritten digit database. Available online: , 2010. Gabriel Loaiza-Ganem and John P Cunningham. The continuous bernoulli: fixing a pervasive error in variational autoencoders. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 32. Curran Associates, Inc., 2019. URL https://proceedings.neurips.cc/paper/2019/file/ f82798ec8909d23e55679ee26bb26437-Paper.pdf. Christos Louizos, Max Welling, and Diederik P. Kingma. Learning sparse neural networks through l0 regularization. In International Conference on Learning Representations, 2018. URL https: //openreview.net/forum?id=H1Y8hhg0b. R Duncan Luce. Individual choice behavior: A theoretical analysis. New York: Wiley, 1959, 1959. Chris J. Maddison, Andriy Mnih, and Yee Whye Teh. The concrete distribution: A continuous relaxation of discrete random variables. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. Open Review.net, 2017. URL https://openreview.net/forum?id=S1j E5L5gl. Published as a conference paper at ICLR 2022 Andre Martins and Ramon Astudillo. From softmax to sparsemax: A sparse model of attention and multi-label classification. In Maria Florina Balcan and Kilian Q. Weinberger (eds.), Proceedings of The 33rd International Conference on Machine Learning, volume 48 of Proceedings of Machine Learning Research, pp. 1614 1623, New York, New York, USA, 20 22 Jun 2016. PMLR. URL https://proceedings.mlr.press/v48/martins16.html. Toby J Mitchell and John J Beauchamp. Bayesian variable selection in linear regression. Journal of the American Statistical Association, 83(404):1023 1032, 1988. Andriy Mnih and Karol Gregor. Neural variational inference and learning in belief networks. In Proceedings of the 31st International Conference on International Conference on Machine Learning - Volume 32, ICML 14, pp. II 1791 II 1799. JMLR.org, 2014. Lina Murady. Probabilistic models for joint classification and rationale extraction. Master s thesis, University of Amsterdam, 2020. Vlad Niculae, Andre Martins, Mathieu Blondel, and Claire Cardie. Sparse MAP: Differentiable sparse structured inference. In Jennifer Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pp. 3799 3808. PMLR, 10 15 Jul 2018. URL https://proceedings.mlr. press/v80/niculae18a.html. Frank Nielsen and Richard Nock. Entropies and cross-entropies of exponential families. In 2010 IEEE International Conference on Image Processing, pp. 3621 3624, 2010. doi: 10.1109/ICIP. 2010.5652054. Travis E Oliphant. A guide to Num Py, volume 1. Trelgol Publishing USA, 2006. Andrew W Palmer, Andrew J Hill, and Steven J Scheding. Methods for stochastic collection and replenishment (scar) optimisation for persistent autonomy. Robotics and Autonomous Systems, 87: 51 65, 2017. George Papandreou and Alan L. Yuille. Perturb-and-map random fields: Using discrete optimization to learn and sample from energy models. In 2011 International Conference on Computer Vision, pp. 193 200, 2011. doi: 10.1109/ICCV.2011.6126242. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary De Vito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. Pytorch: An imperative style, high-performance deep learning library. In H. Wallach, H. Larochelle, A. Beygelzimer, F. d'Alché-Buc, E. Fox, and R. Garnett (eds.), Advances in Neural Information Processing Systems 32, pp. 8024 8035. Curran Associates, Inc., 2019. Max Paulus, Dami Choi, Daniel Tarlow, Andreas Krause, and Chris J Maddison. Gradient estimation with stochastic softmax tricks. In H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, and H. Lin (eds.), Advances in Neural Information Processing Systems, volume 33, pp. 5691 5704. Curran Associates, Inc., 2020. URL https://proceedings.neurips.cc/paper/2020/file/ 3df80af53dce8435cf9ad6c3e7a403fd-Paper.pdf. F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, V. Dubourg, J. Vanderplas, A. Passos, D. Cournapeau, M. Brucher, M. Perrot, and E. Duchesnay. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825 2830, 2011. Ben Peters, Vlad Niculae, and André F. T. Martins. Sparse sequence-to-sequence models. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 1504 1519, Florence, Italy, July 2019. Association for Computational Linguistics. doi: 10.18653/ v1/P19-1146. URL https://aclanthology.org/P19-1146. Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, and Ilya Sutskever. Language models are unsupervised multitask learners. Open AI blog, 1(8):9, 2019. Published as a conference paper at ICLR 2022 Jason Tyler Rolfe. Discrete variational autoencoders. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. Open Review.net, 2017. URL https://openreview.net/forum?id=ry Mx XPFex. Murat Sensoy, Lance Kaplan, and Melih Kandemir. Evidential deep learning to quantify classification uncertainty. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL https://proceedings.neurips.cc/paper/2018/file/ a981f2b708044d6fb4a71a1463242520-Paper.pdf. Claude E. Shannon. A mathematical theory of communication. The Bell system technical journal, 27 (3):379 423, 1948. Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. In Yoshua Bengio and Yann Le Cun (eds.), 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, 2015. URL http://arxiv.org/abs/1409.1556. N Sonine. Recherches sur les fonctions cylindriques et le développement des fonctions continues en séries. Mathematische Annalen, 16(1):1 80, 1880. Arash Vahdat, William Macready, Zhengbing Bian, Amir Khoshaman, and Evgeny Andriyash. DVAE++: Discrete variational autoencoders with overlapping transformations. In Jennifer Dy and Andreas Krause (eds.), Proceedings of the 35th International Conference on Machine Learning, volume 80 of Proceedings of Machine Learning Research, pp. 5035 5044. PMLR, 10 15 Jul 2018. URL https://proceedings.mlr.press/v80/vahdat18a.html. Eelco van der Wel. Improving controllable generation with semi-supervised deep generative models. Master s thesis, University of Amsterdam, 2020. Guido Van Rossum and Fred L. Drake. Python 3 Reference Manual. Create Space, Scotts Valley, CA, 2009. ISBN 1441412697. Pauli Virtanen, Ralf Gommers, Travis E. Oliphant, Matt Haberland, Tyler Reddy, David Cournapeau, Evgeni Burovski, Pearu Peterson, Warren Weckesser, Jonathan Bright, Stéfan J. van der Walt, Matthew Brett, Joshua Wilson, K. Jarrod Millman, Nikolay Mayorov, Andrew R. J. Nelson, Eric Jones, Robert Kern, Eric Larson, CJ Carey, Ilhan Polat, Yu Feng, Eric W. Moore, Jake Vand er Plas, Denis Laxalde, Josef Perktold, Robert Cimrman, Ian Henriksen, E. A. Quintero, Charles R Harris, Anne M. Archibald, Antônio H. Ribeiro, Fabian Pedregosa, Paul van Mulbregt, and Sci Py 1. 0 Contributors. Sci Py 1.0: Fundamental Algorithms for Scientific Computing in Python. Nature Methods, 2020. doi: https://doi.org/10.1038/s41592-019-0686-2. Martin J Wainwright and Michael I Jordan. Graphical models, exponential families, and variational inference. Foundations and Trends in Machine Learning, 1(1 2):1 305, 2008. Stéfan van der Walt, S Chris Colbert, and Gael Varoquaux. The Num Py array: a structure for efficient numerical computation. Computing in Science & Engineering, 13(2):22 30, 2011. Ronald J. Williams. Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning, 8(3):229 256, 1992. doi: 10.1007/BF00992696. URL https: //doi.org/10.1007/BF00992696. Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-MNIST: a novel image dataset for benchmarking machine learning algorithms. preprint ar Xiv:1708.07747, 2017. Günter M Ziegler. Lectures on polytopes, volume 152. Springer, Graduate Texts in Mathematics, 1995. Published as a conference paper at ICLR 2022 A PROOF OF WELL-DEFINEDNESS OF DIRECT SUM MEASURE We start by recalling the definitions of σ-algebras, measures, and measure spaces. A σ-algebra on a set X is a collection of subsets, Ω 2X, which is closed under complements and under countable unions. A measure µ on (X, Ω) is a function from Ωto R { } satisfying (i) µ(A) 0 for all A Ω, (ii) µ( ) = 0, and (iii) the σ-additivity property: µ( j NAj) = P j N µ(Aj) for every countable collections {Aj}j N Ωof pairwise disjoint sets in Ω. A measure space is a triple (X, Ω, µ) where X is a set, A is a σ-algebra on X and µ is a measure on (X, A). An example is the Euclidean space X = RK endowed with the Lebesgue measure, where Ωis the Borel algebra generated by the open sets (i.e. the set Ωwhich contains these open sets and countably many Boolean operations over them). The well-definedness of the direct sum measure µ comes from the following more general result, which appears (without proof) as exercise I.6 in Conway (2019). Lemma 1. Let (Xk, Ωk, µk) be measure spaces for k = 1, . . . , K. Then, (X, Ω, µ) is also a measure space, with X = LK k=1 Xk = QK k=1 Xk (the direct sum or Cartesian product of sets Xk), Ω= {A X | A Xk Ωk, k [K]}, and µ(A) = PK k=1 µk(A Xk). Proof. First, we show that Ωis a σ-algebra. We need to show that (i) if A Ω, then A Ω, and (ii) if Ai Ωfor each i N then S i N Ai Ω. For (i), we have that, if A Ω, then we must have A Xk Ωk for every k, and therefore A Xk = Xk \ A = Xk \ (A Xk) Ωk, since Ωk is a σ-algebra on Xk. This implies that A Ω. For (ii), we have that, if Ai Ω, then we must have Ai Xk Ωk for every i N and k [K], and therefore S i N Ai Xk = S i N(Ai Xk) Ωk, since Ωk is closed under countable unions. This implies that S i N Ai Ω. Second, we show that µ is a measure. We clearly have µ(A) = PK k=1 µk(A Xk) 0, since each µk is a measure itself, and hence it is non-negative. We also have µ( ) = PK k=1 µk( Xk) = PK k=1 µk( ) = 0. Finally, if {Aj}j N Ωis a countable collection of disjoint sets, we have µ( j NAj) = PK k=1 µk( j N(Aj Xk)) = PK k=1 P j N µk(Aj Xk) = P j N PK k=1 µk(Aj Xk) = P We have seen in 3 that the simplex K 1 can be decomposed as a disjoint union of the relative interior of its faces. Each of these relative interiors is an open subset of an affine subspace isomorphic to Rk 1, for k [K], which is equipped with the Lebesgue measure for k > 1 and the counting measure for k = 1. Lemma 1 then guarantees that we can take the direct sum of all these affine spaces as a measure space with the direct sum measure µ = µ of Definition 1. B K-DIMENSIONAL GAUSSIAN-SPARSEMAX We derive expressions for the density p Y,F (y, f) for the sample-and-project case (stochastic sparsemax). We assume without loss of generality that f = {1, 2, . . . , s} and that we want to compute the density p Y2:K(y2, . . . , ys, 0, . . . , 0) for given y2, . . . , ys > 0 such that y1 := 1 Ps j=2 yj > 0. The process that generates the data is as follows: first, zi p Zi(zi) independently for i [K]. Then, y is obtained deterministically from z1:K as y = sparsemax(z1:K). We assume here that each p Zi(zi) is a univariate Gaussian distribution with mean µi and variance σ2 i . We make use of the following well-known properties of multivariate Gaussians: Lemma 2. Let Z N(µ, Σ). Then, for any matrix A, not necessarily square we have AZ N(Aµ, AΣA ). Furthermore, splitting , µ = µx µy , Σ = Σxx Σxy Σ xy Σyy we have the following expression for the marginal distribution PX(x): X N(µx, Σxx) (16) Published as a conference paper at ICLR 2022 and the following expression for the conditional distribution PX|Y (x | Y = y): X | Y = y N(µx + ΣxyΣ 1 yy (y µy), Σxx ΣxyΣ 1 yy Σ xy). (17) We start by picking one index in the support of y we assume without loss of generality that this pivot index is 1 and that the support set is {1, . . . , s} as stated above and introducing new random variables Ui := Zi Z1 for i {2, . . . , K}. Note that these new random variables are not independent since they all depend on Z1. Then, from the change of variable formula for non-invertible transformations, we have, for given y2, . . . , ys > 0 such that y1 := 1 Ps j=2 yj > 0: p Y2:K(y2, . . . , ys, 0, . . . , 0) = p U2:s(u2:s) |B| Z y1 p U(s+1):K|U1:s(u(s+1):K | U1:s = u1:s)dus+1 du K. where ui = yi y1 for i 2, . . . , s, which can written as u2:s = By2:s+c with B = Is 1 1s 11 s 1 and c = 1s 1. The determinant of B is simply |B| = s. Note that U2:K = AZ1:K, where A = [ 1K 1, IK 1] R(K 1) K, that is 1 1 0 0 1 0 1 0 ... ... 0 1 0 0 1 From Lemma 2, we have that p U(u) = N(u; Aµ, ADiag(σ2)A ) = N(u; µ2:K µ11K 1, Diag(σ2 2:K) + σ2 11K 11 K 1), (20) and the marginal distribution p U2:s(u2:s) is p U2:s(u2:s) = N(u2:s; µ2:s µ11s 1, Diag(σ2 2:s) + σ2 11s 11 s 1) = N(y2:s y11s 1; µ2:s µ11s 1, Diag(σ2 2:s) + σ2 11s 11 s 1). (21) Note that this is a multivariate distribution whose covariance matrix is the sum of a diagonal matrix with a constant matrix. We now calculate the conditional distribution of the variables which are not in the support conditioned on the ones which are in the support. We have according to the notation in Lemma 2: Σxx = Diag(σ2 (s+1):K) + σ2 11K s1 K s Σyy = Diag(σ2 2:s) + σ2 11s 11 s 1 Σxy = σ2 11K s1 s 1. (22) Using the Sherman-Morrison formula, we obtain Σ 1 yy = Diag(σ 2 2:s) σ 2 2:s(σ 2 2:s) σ 2 1 + Ps i=2 σ 2 i , (23) from which we get µ := µx + ΣxyΣ 1 yy (y µy) = µ(s+1):K µ11K s + 1 σ 2 1 + Ps i=2 σ 2 i 1K s(σ 2 2:s) (u2:s µ2:s + µ11s 1) = µ(s+1):K µ11K s + 1 Ps i=1 σ 2 i 1K s(σ 2 2:s) (y2:s y11s 1 µ2:s + µ11s 1) = µ(s+1):K y11K s + Ps i=1 σ 2 i (yi µi) Ps i=1 σ 2 i 1K s. (24) Published as a conference paper at ICLR 2022 Σ := Σxx ΣxyΣ 1 yy Σ xy = Diag(σ2 (s+1):K) + σ2 11K s1 K s σ2 11K s(σ 2 2:s) 1s 11 K s σ 2 1 + Ps i=2 σ 2 i = Diag(σ2 (s+1):K) + σ2 11K s1 K s σ2 1 Ps i=2 σ 2 i σ 2 1 + Ps i=2 σ 2 i 1K s1 K s = Diag(σ2 (s+1):K) + σ2 1 1 Ps i=2 σ 2 i σ 2 1 + Ps i=2 σ 2 i = Diag(σ2 (s+1):K) + 1 σ 2 1 + Ps i=2 σ 2 i 1K s1 K s = Diag(σ2 (s+1):K) + 1 Ps i=1 σ 2 i 1K s1 K s. (25) Finally, also from Lemma 2, we get p U(s+1):K|U1:s(u(s+1):K | U1:s = u1:s) = N(u(s+1):K; µ, Σ). (26) Note that this is again a multivariate Gaussian distribution with a covariance matrix Σ which is the sum of a diagonal and a constant matrix. Putting everything together, we get p Y2:K(y2, . . . , ys, 0, . . . , 0) = s N(y2:s y11s 1; µ2:s µ11s 1, Diag(σ2 2:s) + σ2 11s 11 s 1) F µ(s+1):K + Ps i=1 σ 2 i (yi µi) Ps i=1 σ 2 i 1K s; Σ , (27) where F(v; Σ) = R 0 R 0 N(0; v, Σ) is the negative orthant cumulative distribution of a multivariate Gaussian with mean v and covariance Σ. Efficient Monte Carlo approximations of this integral have been proposed by Genz (1992). Using again Lemma 2 but now in the reverse direction, the function F(v; Σ), setting v = µ(s+1):K + Ps i=1 σ 2 i (yi µi) Ps i=1 σ 2 i 1K s, can be reduced to a unidimensional integral: F(v; Σ) = Z dt N t; Ps i=1 σ 2 i (µi yi) Ps i=1 σ 2 i , 1 Ps i=1 σ 2 i dr N(r; µi, σ2 i ) Ps i=1 σ 2 i (t + yi µi) q Ps i=1 σ 2 i i=s+1 Φ t µi σj q Ps i=1 σ 2 i Ps i=1 σ 2 i (yi µi + µj) σj Ps i=1 σ 2 i where φ(x) and Φ(x) are the p.d.f. and c.d.f. of a standard Gaussian distribution, respectively, and we applied the change of variables formula in the last line, with u = Φ Ps i=1 σ 2 i (t + yi µi)/ q Ps i=1 σ 2 i , whose inverse is t = Φ 1(u) Ps i=1 σ 2 i Ps i=1 σ 2 i (yi µi) Ps i=1 σ 2 i ; the Jacobian of this transformation cancels with the φ-term. We can compute the quantities above via the erf function and its inverse erf 1, as Φ(x) = 1 2 1 + erf x and Φ 1(u) = 2erf 1(2u 1). Published as a conference paper at ICLR 2022 When the variance is constant, σi = σ i [K], and since Ps i=1 yi = 1, the expressions above simplify to F(v; Σ) = Z dt N t; 1 + Ps i=1 µi |s| , σ2 dr N(r; µi, σ2) 1 + |s|t Ps i=1 µi σ p i=s+1 Φ t µi µj + 1 Ps i=1 µi |s| C SAMPLING FROM THE MIXED DIRICHLET DISTRIBUTION Py Torch code for a batched implementation of the procedures discussed in this section will be made available online. Distribution over Face Lattice. The distribution PF is a discrete exponential family whose support is the set of all 2K 1 proper faces of the simplex. The computation of the natural parameter in Equation (8) factorizes over vertices, but, because the empty face is not in the support of PF , the computation of the log-normalizer requires special attention. The obvious strategy is to enumerate all assignments to (FI, w φ(FI)), one per proper face. This strategy is not feasible for even moderately large K. We can, however, exploit a compact representation of the set of proper faces by encoding the face lattice in a directed acyclic graph (DAG) where each proper face is associated with a complete path from the DAG s source to the DAG s sink. This DAG G = (Q, A) is a collection of states u Q {0, . . . , K + 1} {0, 1} {0, 1} and arcs (u, v) A Q Q. A path that ends in (k, b, s) Q corresponds to a non-empty subset of vertices if, and only if, s = 1. A path that contains (k, b, s) Q corresponds to a face that includes the kth vertex if, and only if, b = 1. If u = (k1, b1, s1) Q and v = (k2, b2, s2) Q, then u v iff k1 k2. The state (0, 0, 0) is the DAG s unique source, and the state (K + 1, 0, 1) is the DAG s unique sink. Because = 1 for the sink, no complete path (i.e., from source to sink) will correspond to an empty face. An arc is a pair of states (u, v), where u is the origin and v is the destination. For every state (k, b, s) such that k < K, we have arcs ((k, b, s), (k + 1, 1, 1)) and ((k, b, s), (k + 1, 0, s)). For every state (K, b, 1), we have an arc to the final state (K + 1, 0, 1). By construction, the number of states (and arcs) in the DAG is proportional to K. A complete path (from source to sink) has length K + 1, and it uniquely identifies a proper face. To compute the log-normalizer of Equation (8), we run the forward algorithm through G, with an arc s weight given by ( 11 b)wk if the arc s destination is the state (k, b, s) or 0 if the arc s destination is the DAG s sink. Running the backward algorithm through G (that is, running the forward algorithm from sink to source) evaluates the marginal probabilities required for ancestral sampling. Entropy of Mixed Dirichlet. The intrinsic representation of a Mixed Dirichlet distribution allows for efficient computation of H(F) and DKL(PF ||QF ) necessary for H (Y ) and D KL(p Y ||q Y ). The continuous parts H(Y |F) and EPF [DKL(p Y |F =f||q Y |F =f)] require solving an expectation with an exponential number of terms (one per proper face). The distribution PF is a discrete exponential family indexed by the natural parameter w RK, its discrete entropy is H(F) = log Z(w) w, w log Z(w) , where w log Z(w) = E[φ(F)]. KL divergence from a member QF of the same family but with parameter v is given by DKL(PF ||QF ) = log Z(v) log Z(w) v w, w log Z(w) . (30) The forward-backward algorithm computes both the log-normalizer and its gradient (the expected sufficient statistic) in a single pass through a DAG of size O(K). For results concerning entropy, cross-entropy, and relative entropy of exponential families see for example Nielsen & Nock (2010). The continuous part H(Y |F) is the expectation of the differential entropy of Y |F = f, each a Dirichlet distribution over the face f, under PF . For small K we can enumerate the 2K 1 terms in this expectation, since the entropy of each Dirichlet is known in closed-form. In general, for large K, Published as a conference paper at ICLR 2022 we can obtain an MC estimate by sampling independently from PF and assessing the differential entropy of Y |F = f only for sampled faces. For EPF [DKL(p Y |F =f||q Y |F =f)] the situation is similar, since KL for two Dirichlet distributions on the same face f is known in closed-form. Gradient Estimation. Parametrizing a Mixed Dirichlet variable takes two parameter vectors, namely, w and α. In a VAE, those are predicted from a given data point (e.g., an MNIST digit) by an inference network. To update the parameters of the inference network we need a Monte Carlo estimate of the gradient w,αEF |w[EY |F =f,α[ℓ(y)]] with respect to w and α of a loss function ℓ computed in expectation under the Mixed Dirichlet distribution. By chain rule, we can rewrite the gradient as follows: EF |w EY |F =f,α[ℓ(y)] w log PF (f|w) + αEY |F =f,α[ℓ(y)] . (31) Given a sampled face F = f, we can MC estimate both contributions to the gradient, the first takes MC estimation of EY |F =f ,α[ℓ(y)], which is straightforward, the second can be done via implicit reprametrization (Figurnov et al., 2018). To reduce the variance of the score function estimator (first term), we employ a sampled self-critic baseline, i.e., an MC estimate of EY |F =f ,α[ℓ(y)] given an independently sampled face f . D INFORMATION THEORY FOR MIXED RANDOM VARIABLES D.1 MUTUAL INFORMATION FOR MIXED RANDOM VARIABLES Besides the direct sum entropy and the Kullback-Leibler divergence for mixed distributions, we can also define a mutual information for mixed random variables as follows: Definition 4 (Mutual information). For mixed random variables Y and Z, the mutual information between Y and Z is I (Y ; Z) = H (Y ) H (Y | Z) = H(F) + H(Y | F) H(F | Z) + H(Y | F, Z) = I(F; Z) + I(Y ; Z | F) 0. (32) With these ingredients it is possible to provide counterparts for channel coding theorems by combining Shannon s discrete and continuous channel theorems (Shannon, 1948). D.2 PROOF OF PROPOSITION 1 (CODE OPTIMALITY) Proposition 1 is a consequence of the following facts (Shannon, 1948): The discrete entropy of a random variable representing an alphabet symbol corresponds to the average length of the optimal code for the symbols in the alphabet, in a lossless compression setting. Besides, it is known (Cover & Thomas, 2012) that the optimal number of bits to encode a D-dimensional continuous random variable with N bits of precision equals its differential entropy (in bits) plus ND.4 Therefore, the direct sum entropy (10) is the average length of the optimal code where the sparsity pattern of y K 1 must be encoded losslessly and where there is a predefined bit precision for the fractional entries of y. Eq. 12 follows from these facts and the definition of direct sum entropy (Definition 3). 4Cover & Thomas (2012, Theorem 9.3.1) provide an informal proof for D = 1, but it is straightforward to extend the same argument for D > 1. Published as a conference paper at ICLR 2022 E DIRECT SUM ENTROPY AND KULLBACK-LEIBLER DIVERGENCE OF 2D GAUSSIAN-SPARSEMAX From (1) and Definition 3, the direct sum entropy of the 2D Gaussian sparsemax is: H (Y ) = H(F) + H(Y | F) = H([P0, P1, 1 P0 P1]) (1 P0 P1) Z 1 N(y; z, σ2) 1 P0 P1 log N(y; z, σ2) = P0 log P0 P1 log P1 Z 1 0 N(y; z, σ2) log N(y; z, σ2)dy, (33) where we have P0 = 1 erf(z/( 2σ)) 2 , P1 = 1+erf((z 1)/( 2σ)) 2 , and 0 N(y; z, σ2) log N(y; z, σ2)dy = 0 N(y; z, σ2) log( 2πσ2) + (y z)2 = (1 P0 P1) log( σ N(t; 0, 1)t2dt = (1 P0 P1) log( which leads to a closed form for the entropy. As for the KL divergence, we have: D KL(p Y q Y ) := DKL(PF QF ) + Ef PF DKL(p Y |F ( | F = f) q Y |F ( | F = f) f F(P) PF (f) log PF (f) f F(P) PF (f) Z f p Y |F (y | f) log p Y |F (y | f) q Y |F (y | f) = P0 log P0 Q0 + P1 log P1 0 N(y; z P , σ2 P ) log N(y; z P , σ2 P ) N(y; z Q, σ2 Q), (35) Published as a conference paper at ICLR 2022 where we have P0 = 1 erf(z P /( 2σP )) 2 , P1 = 1+erf((z P 1)/( 2σP )) 2 , Q0 = 1 erf(z Q/( Q1 = 1+erf((z Q 1)/( 2σQ)) 2 , and 0 N(y; z P , σ2) log N(y; z P , σ2 P ) N(y; z Q, σ2 Q)dy = 0 N(y; z P , σ2 P ) σQ (y z P )2 2σ2 P + (y z Q)2 = (1 P0 P1) log σP σP N(t; 0, 1)t2dt + σQ σQ N(t; 0, 1)t2dt = (1 P0 P1) log σP σP ; 0, 1 z P σQ ; 0, 1 z Q F MAXIMUM DIRECT SUM ENTROPY OF MIXED DISTRIBUTIONS Start by noting that the differential entropy of a Dirichlet random variable Y Dir(α) is H(Y ) = log B(α) + (α0 K)ψ(α0) k=1 (αk 1)ψ(αk), (37) where α0 = PK k αk and ψ is the digamma function. When α = 1, this becomes a flat (uniform) density and the entropy attains its maximum value: H(Y ) = log B(α) = log(K 1)!. (38) This value is negative for K > 2; it follows that the differential entropy of any distribution in the simplex is negative. We next determine the distribution p Y (y) with the largest direct sum entropy. Considering only the maximal face, which corresponds to ri( K 1), the distribution with the largest entropy is the flat distribution, whose entropy is given in (38). In our definition of entropy in (10) this corresponds to a deterministic F which puts all probability mass in this maximal face. At the opposite extreme, if we only assign probability to pure vertices, i.e., if we constrain to minimal faces, a uniform choice leads to a (Shannon) entropy of log K. We will show that looking at all faces further increases entropy. Looking at (10), we see that the differential entropy term H(Y | F = f) can be maximized separately for each f, the solution being the flat distribution on face f k 1, which has entropy log(k 1)!, where 1 k K. By symmetry, all faces of the same dimension k 1 look the same, and there are K k of them. Therefore the maximal entropy distribution is attained with PF of the form PF (f) = g(k)/ K k where g : [K] R+ is a function satisfying PK k=1 g(k) = 1 (which can be regarded as a categorical probability mass function). If we choose a precision of N bits, this Published as a conference paper at ICLR 2022 2 3 4 5 6 7 8 9 K mixed N=0 mixed N=1 mixed N=2 mixed N=3 mixed N=4 discrete continuous N=4 Figure 4: Maximum entropies for mixed distributions for several values of bit precision N, as a function of the simplex dimensionality K 1. Shown are also the maximum entropies for the corresponding discrete and continuous cases, for comparison. H N(Y ) = X f F( K 1) PF (f) log PF (f) + X f F( K 1) PF (f) Z f p Y |F (y | f) log p Y |F (y | f) k=1 (k 1)g(k) (39) k=1 g(k) log g(k) K k k=1 g(k)(log(k 1)! N(k 1)) k=1 g(k) log g(k) + k=1 g(k) log K k 2N(k 1) (k 1)! . (40) Note that the first term is the entropy of g( ) and the second term is a linear function of g( ), that is, (40) is a entropy-regularized argmax problem, hence the g( ) that maximizes this objective is the softmax transformation of the vector with components log ( K k)2N(k 1) (k 1)! , that is: g(k) = ( K k)2N(k 1) (k 1)! PK j=1 ( K j )2N(j 1) and the maximum entropy value is H N,max(Y ) = log K k 2N(k 1) (k 1)! = log L(1) K 1( 2N), (42) where L(α) n (x) denotes the generalized Laguerre polynomial (Sonine, 1880), as stated in Proposition 2. Figure 4 shows how the largest direct sum entropy varies with K, compared to the discrete and continuous entropies. For example, for K = 2, we obtain g(1) = 2 2+2N , g(2) = 2N 2+2N , and H N(Y ) = log(2 + 2N), therefore, in the worst case, we need at most log2(2 + 2N) bits to encode Published as a conference paper at ICLR 2022 Y 1 with bit precision N. This is intuitive: the faces of 1 are the two vertices {(0, 1)} and {(1, 0)} and the line segment [0, 1]. The first two faces have a probability of 1 2+2N and the last one have a probability 2N 2+2N . To encode a point in the simplex we first need to indicate which of these three faces it belongs to (which requires 2 2+2N log2(2 + 2N) + 2N 2+2N log2 2+2N 2N = log2(2 + 2N) N2N 2+2N bits), and with 2N 2+2N probability we need to encode a point uniformly distributed in the segment [0, 1] with N bit precision, which requires extra N2N 2+2N bits on average. Putting this all together, the total number of bits is log2(2 + 2N), as expected. G EXPERIMENTAL DETAILS In this section, we give further details on the architectures and hyperparameters used in our main experiments (App. G.1, App. G.2, and App.G.3), which are focused on learning sparse representations, in App. G.4 we experiment with the Mixed Dirichlet as a likelihood function to model simplex-valued data. Finally, in App G.5, we describe our computing infrastructure. G.1 EMERGENT COMMUNICATION GAME Let V = {v1, . . . , v|V|} be the collection of images from which the sender sees a single image vj. We consider a latent variable model with observed variables x = (V, j) X and latent stochastic variables y chosen from a vocabulary set Y. We let the sender correspond to the probability model π(y|x, θ) = p(y | vj, θπ) of the latent variable and the receiver be modeled as p(j | V, y, θℓ). The overall fit to the dataset D is P x D Lx(θ), where we marginalize the latent variable y to compute the loss of each observation, that is Lx(θ) = Eπ(y|x,θ) [ℓ(x, y; θ)] = X y Y π(y|x, θ) ℓ(x, y; θ), (43) where the receiver is used to define the downstream loss, ℓ(x, y; θ) := log p(j | V, y, θℓ). Notably, we do not add a (discrete) entropy regularization term of π(y | x, θ) to (43) with a coefficient as an hyperparameter as in Correia et al. (2020). Data. The dataset consists of a subset of Image Net (Deng et al., 2009)5containing 463,000 images that are then passed through a pretrained VGG (Simonyan & Zisserman, 2015), from which the representations at the second-to-last fully connected layers are saved and used as input to the sender and the receiver. To get the dataset visit https://github.com/Diane Bouchacourt/ Signaling Game (Bouchacourt & Baroni, 2018). Architecture and hyperparameters. We follow the experimental procedure described in Correia et al. (2020): the architecture of both the sender and the receiver are identical to theirs, we set the size of the collection of images |V| to 16, the size of the vocabulary of the sender to 256, the hidden size to 512, and the embedding size to 256. We choose the best hyperparameter configuration by doing a grid search on the learning rate (0.01, 0.005, 0.001) and, for each configuration, evaluating the communication success on the validation set. For the Gumbel-Softmax models, the temperature is annealed using the schedule τ = max(0.5, exp rt), where r = 1e 5 and t is updated every N = 1000 steps. For the K-D Hard Concrete we use a scaling constant λ = 1.1 and for Gaussian Sparsemax we set Σ = I. All models were trained for 500 epochs using the Adam optimizer with a batch size of 64. G.2 BIT-VECTOR VAE In this experiment, the training objective is to minimize the negative ELBO: we rewrite (43) with ℓ(x, y; θℓ) = log p(x,y|φ) q(y|x,λ), where we decompose the approximate posterior as q(y | x, λ) = QD i=1 q(yi | x, λ), with D = 128 being the number of binary latent variables yi. For each bit, we 5Available for research under the terms and conditions described in https://www.image-net.org/ download. Published as a conference paper at ICLR 2022 consider a uniform prior p(y) when the approximate posterior is purely discrete or continuous, and the maxent mixed distribution (Prop. 2) for the mixed cases. Data. We use Fashion-MNIST (Xiao et al., 2017)6, comprising of 28 28 grayscale images of fashion products from different categories. Architecture and hyperparameters. We follow the architecture and experimental procedure described in Correia et al. (2020), where the inference and generative network consist of one 128-node hidden layer with Re LU activation functions. We choose the best hyperparameter configuration by doing a grid search on the learning rate (0.0005, 0.001, 0.002) and choosing the best model based on the value of the negative ELBO on the validation set. For the Gumbel-Softmax models, the temperature is annealed using the schedule τ = max(0.5, exp rt), where we search r {1e 5, 1e 4} and t is updated every N = 1000 steps; also, for these models, we relax the sample into the continuous space but assume a discrete distribution when computing the entropy of π(y | x, θ), leading to an incoherent evaluation for Gumbel-Softmax without ST. For the Hard Concrete distribution we follow Louizos et al. (2018) and stretch the concrete distribution to the ( 0.1, 1.1) interval and then apply a hard-sigmoid on its random samples. For the Gaussian-Sparsemax, we use σ2 = 1. All models were trained for 100 epochs using the Adam optimizer with a batch size of 64. G.3 MIXED-LATENT VAE ON MNIST Data. We use stochastically binarized MNIST (Le Cun et al., 2010).7The first 55,000 instances are used for training, the next 5,000 instances for development and the remaining 10,000 for test. Architecture and hyperparameters. The model is a VAE with a K-dimensional latent code (with K = 10 throughout), both the decoder (which parametrizes the observation model) and the encoder (which parametrizes the inference model) are based on feed-forward neural networks. The decoder maps from a sampled latent code z to a collection of D = 28 28 Bernoulli distributions. We use a feed-forward decoder with two Re LU-activated hidden layers, each with 500 units. The encoder maps from a D-dimensional data point x to the parameters of a variational approximation to the model s posterior distribution. For Gaussian models we predict K locations and K scales (with softplus activation). For Dirichlet models we predict K concentrations using softplus clamped to [10 3, 103]. For Mixed Dirichlet models we predict K log-potentials clamped to [ 10, 10], and K concentrations using softplus clamped to [10 3, 103]. For regularization we employ dropout and L2 penalty. We use two Adam optimizers, one for the parameters of the decoder and another for the parameters of the encoder, each with its own learning rate. We search for the best configuration of hyperparameters using importance-sampled estimates of negative log-likelihood of the model given a development set. The parameters we consider are: learning rates (in {10 5, 10 4, 10 3}, and their halves), regularization strength (in {0, 10 7, 10 6, 10 5}), and dropout rate (in {0, 0.1, 0.2, 0.3}). All models are trained for 500 epochs with stochastic mini batches of size 100. During training, for gradient estimation, we sample each latent variable exactly once. For importance sampling, we use 100 samples for model selection, and 1000 samples for evaluation on the test set. Gradient estimation. We use single-sample reparametrized gradient estimates for most models: Gaussian (unbiased), Dirichlet (unbiased), Gumbel-Softmax ST (biased due to straight-trough), and Mixed Dirichlet (unbiased). A gradient estimate for Mixed Dirichlet models, in particular, has two sources of stochasticity, one from the implicit reparametrization for Dirichlet samples, another from score function estimation for Gibbs samples. To reduce variance due to score function estimation, the Mixed Dirichlet model we report employs an additional sample, used to compute a self-critic baseline. We have also looked into a simpler variant, which dispenses with this self-critic baseline, and instead employs a simple running average baseline, this model performed very close to that with a self-critic. The purely discrete latent variable model (Categorical) uses the exact gradient of the ELBO, given a mini batch, which we obtain by exactly marginalizing the K = 10 assignments of the latent variable. 6License available at https://github.com/zalandoresearch/fashion-mnist/blob/ master/LICENSE. Published as a conference paper at ICLR 2022 0 1 2 3 4 5 6 7 8 9 Figure 5: t SNE plots of posterior samples, from left-to-right: Gaussian, Dirichlet, Mixed Dirichlet, Categorical, Gumbel-Softmax ST. Figure 6: Pixel-wise average of 100 model samples generated by conditioning on each of the ten vertices of the simplex. From left-to-right: Mixed Dirichlet, Categorical, Gumbel-Softmax ST, and Dirichlet (as the Dirichlet does not support the vertices of the simplex, we add uniform noise (between 0 and 0.1) to each coordinate of a vertex and renormalize). Prior, posterior, and KL term in the ELBO. The Gaussian model employs a standard Gaussian prior and a parametrized Gaussian posterior approximation, the KL term in the ELBO is computed in closed-form. The Dirichlet model employs a uniform Dirichlet prior (concentration 1) and a parametrized Dirichlet posterior approximation, the KL term in the ELBO is also computed in closed-form. Both the Categorical and the Gumbel-Softmax ST models employ a uniform Categorical prior and a Categorical approximate posterior (albeit parametrized via Gumbel-Softmax ST for the latter model), the KL term in the ELBO is between two Categorical distributions and thus computed in closed form. The Mixed Dirichlet model employs a maximum entropy prior with bit-precision parameter N = 0 and a parametrized Mixed Dirichlet approximate posterior, the KL term in the ELBO is partly exact and partly estimated, in particular, the contribution of the Gibbs distributions over faces is computed in closed form. Additional plots. Figure 5 shows t SNE plots where each validation digit is encoded by a sample from the approximate posterior obtained by conditioning on that digit. The continuous (Gaussian and Dirichlet) and mixed (Mixed Dirichlet) models learn to represent the digit rather well, with some plausible confusions by the Mixed Dirichlet model (e.g., 4, 7, and 9). The models that can only encode data using the vertices of the simplex (Categorical and Gumbel-Softmax ST) typically map multiple digits to the same vertex of the simplex. For models whose priors support the vertices of the simplex (Mixed Dirichlet, Categorical, and Gumbel-Softmax ST), we can inspect what the vertices of the simplex typically map to (in data space). Figure 6 shows 100 such samples per vertex. We also include samples from the Dirichlet model, though note that the Dirichlet prior does not support the vertices of the simplex, thus we sample from points very close to the vertices (but in the relative interior of the simplex). We also report conditional and unconditional generations from each model in Figures 7 and 8. G.4 SIMPLEX-VALUED REGRESSION Simplex-valued data are observations in the form of probability vectors, they appear in statistics in contexts such as time series (e.g., modeling polling data) and in machine learning in contexts such as knowledge distillation (e.g., teaching a compact network to predict the categorical distributions predicted by a larger network). 7Available under the terms of the Creative Commons Attribution-Share Alike 3.0 license (https:// creativecommons.org/licenses/by-sa/3.0/). Published as a conference paper at ICLR 2022 Figure 7: Conditional generation. For each instance of each class in the validation set (note that none of the models has access to the label), we sample a latent code conditioned on the digit, and re-sample a digit from the model. The illustration displays the pixel-wise average across all validation instances of the same class. From left-to-right: Gaussian, Dirichlet, Mixed Dirichlet, Categorical, Gumbel-Softmax ST. Figure 8: Unconditional generation. For each model, we sample a digit via ancestral sampling (i.e., z is sampled from the prior, then, given z, x is sampled from the observation model). We gather 5000 such samples and group them by class as predicted by a 5-nearest neighbour classifier trained on the MNIST training set (we use kd tree from scikit-learn), the classifier achieves 95% F1 on the MNIST test set. Each image in the illustration is a pixel-wise average of the samples in the cluster, we also report the percentage of digits in each cluster. From left-to-right: Gaussian, Dirichlet, Mixed Dirichlet, Categorical, Gumbel-Softmax ST. Data and task. We experiment with the UK election data setup by Gordon-Rodriguez et al. (2020, Section 5.2).8 The UK electorate is partitioned into 650 constituencies, each electing one member of parliament in a winner-takes-all vote. Hence the data are 650 observed vectors of proportions over the four major parties plus a remainder category (i.e., each observation is a point in the 5 1 simplex, including its faces). Modeling simplex-valued data with the Dirichlet distribution is tricky for the Dirichlet does not support sparse outcomes. While pre-processing the data into the relative interior of the simplex is a simple strategy (e.g., add small positive noise to coordinates and renormalize), it is ineffective for the Dirichlet pdf either diverges or vanishes at extrema (neighbourhoods of the lower-dimensional faces of the simplex). Gordon-Rodriguez et al. (2020) document this and other difficulties in modeling with the Dirichlet likelihood function. To address these limitations they develop the continuous categorical (CC) distribution, an exponential family that supports the entire simplex (i.e., it assigns non-zero density to any point in the simplex), and does not diverge at the extrema. The CC enjoys various analytical properties, but it still cannot assign non-zero mass to the lower-dimensional faces of the simplex, thus while it is a better choice of likelihood function than the Dirichlet, CC samples are never truly sparse (thus test-time predictions are always dense). Architecture and hyperparameters. For Dirichlet and CC, we use a linear layer to map from the input predictors to 5 log-concentration parameters. For Mixed Dirichlet we use 2 linear layers: one maps from the input predictors to 5 scores (clamped to [ 10, 10]) which parametrize PF , the other maps to 5 strictly positive concentrations (we use softplus activations, with pre activations constrained to [ 10, 10]) which parametrize PY |F =f. We train all models using Adam with learning rate 0.1 and no weight decay for exactly 400 steps without mini-batching with 20% of the available data used for training. Following Gordon-Rodriguez et al. (2020), we pre-process the data (add 10 3 to each coordinate and renormalize) for Dirichlet and CC, but this is not done for Mixed Dirichlet. Results. Figure 9 compares the three choices of likelihood function. We report two prediction rules for our Mixed Dirichlet model. Sample mean: we predict stochastically by drawing 100 samples and outputting the sample mean. Most probable mean: we predict deterministically by outputting the mean of the Dirichlet on the face that is assigned highest probability by the model (finding this most probable face is an operation that takes time O(K), and recall K = 5 in this task). We can see that Mixed Dirichlet does not suffer from the pathologies of the Dirichlet, due to face stratification, and lowers test error a bit more than CC (see Table 9), likely due to the ability to sample actual zeros. 8https://commonslibrary.parliament.uk/research-briefings/cbp-8749/ Published as a conference paper at ICLR 2022 0 50 100 150 200 250 300 350 400 Training step 0 50 100 150 200 250 300 350 400 Training step Dir CC Mixed Dir (sample mean) Mixed Dir (most probable mean) Figure 9: Test error (root-mean-square error on the left, mean absolute error on the right) of generalized linear regression towards 5-dimensional vote proportions (UK election data). We compare 3 likelihood functions: Dirichlet, Continuous Categorical (CC), and Mixed Dirichlet. Model RMSE MAE CC 0.1124 0.0847 Mixed Dirichlet sample mean 0.1030 0.0774 most probable mean 0.0987 0.0740 Table 4: Test root-mean-square error and mean absolute error as a function of choice of likelihood function and prediction rule. The Mixed Dirichlet uses twice more parameters (we need to parametrize two components), but training time is barely affected (sampling and density assessments are all linear in K), the training loss and its gradients are stable, and the algorithm converges just as early as CC s. As the Mixed Dirichlet produces sparse samples, it is interesting to inspect how often it succeeds to predict whether an output coordinate is zero or not (i.e., whether yk > 0, which is true for 77.3% of the targets in the text set). The sample mean predicts whether yk > 0 with macro F1 0.92, whereas the most probable mean achieves macro F1 0.94. G.5 COMPUTING INFRASTRUCTURE Our infrastructure consists of 5 machines with the specifications shown in Table 5. The machines were used interchangeably, and all experiments were executed in a single GPU. Despite having machines with different specifications, we did not observe large differences in the execution time of our models across different machines. Table 5: Computing infrastructure. 1. 4 Titan Xp - 12GB 16 AMD Ryzen 1950X @ 3.40GHz - 128GB 2. 4 GTX 1080 Ti - 12GB 8 Intel i7-9800X @ 3.80GHz - 128GB 3. 3 RTX 2080 Ti - 12GB 12 AMD Ryzen 2920X @ 3.50GHz - 128GB 4. 3 RTX 2080 Ti - 12GB 12 AMD Ryzen 2920X @ 3.50GHz - 128GB 5. 2 GTX Titan X - 12GB 12 Intel Xeon E5-1650 v3 @ 3.50GHz - 64 GB