# multilayer_state_evolution_under_random_convolutional_design__6fd74be1.pdf Multi-layer State Evolution Under Random Convolutional Design Max Daniels Northeastern University daniels.g@northeastern.edu Cédric Gerbelot ENS Paris cedric.gerbelot@ens.fr Florent Krzakala Ide PHIcs Laboratory, EPFL florent.krzakala@epfl.ch Lenka Zdeborová SPOC Laboratory, EPFL lenka.zdeborova@epfl.ch Signal recovery under generative neural network priors has emerged as a promising direction in statistical inference and computational imaging. Theoretical analysis of reconstruction algorithms under generative priors is, however, challenging. For generative priors with fully connected layers and Gaussian i.i.d. weights, this was achieved by the multi-layer approximate message (ML-AMP) algorithm via a rigorous state evolution. However, practical generative priors are typically convolutional, allowing for computational benefits and inductive biases, and so the Gaussian i.i.d. weight assumption is very limiting. In this paper, we overcome this limitation and establish the state evolution of ML-AMP for random convolutional layers. We prove in particular that random convolutional layers belong to the same universality class as Gaussian matrices. Our proof technique is of an independent interest as it establishes a mapping between convolutional matrices and spatially coupled sensing matrices used in coding theory. 1 Introduction In a typical signal recovery problem, one seeks to recover a data signal x0 given access to measurements y0 = G (x0), where the parameters of the signal model are known. In many problems, it is natural to view the measurement generation process as a composition of simple forward operators, or layers. In this work, we are concerned with multi-layer signal models of the form G (x0) = '(1)(W (1)'(2)(W (2) . . . '(L)(W (L)x0; (L)) . . .). (1) where W (l) 2 Rnl 1 nl are linear sensing matrices and where '(l)(z; ) are separable channel functions that may be non-linear and may depend on unknown channel noise . In the multi-layer case L > 1, this signal model can be viewed as a simple instance of recovery under a Generative Neural Network (GNN) prior, a technique which has recently shown promise as a generalization of sparsity priors for signal processing applications Bora et al. [2017]. For example, Gaussian compressive sensing y0 = AG (x0)+ under a prior with random Gaussian weights can be naturally viewed as an instance of the multilayer signal model, and given an estimate ˆx, one can recover the target signal ˆs = G (ˆx) from the weights {W (l)} and noise-independent channel functions {'(l)(z)}. In practical settings GNNs often use structured convolutional weight matrices, preventing the direct application of estimators that require W (l) to have unstructured entries. Motivated by this, we take interest in a variant of the recovery problem (1) in which some of the sensing matrices W (l) may be 36th Conference on Neural Information Processing Systems (Neur IPS 2022). 0.0 0.2 0.4 0.6 0.8 1.0 β - Measurement Ratio Sparse Compressive Sensing (q = P = 1024) ρ = 0.25 ρ = 0.5 ρ = 0.75 MCC AMP Dense AMP Dense SE 0 20 40 60 AMP Iteration AMP Iterates for ρ = 0.25 (q = P = 1024) β = 0.4 β = 0.45 β = 0.5 β = 0.55 Figure 1: Agreement between the performance of the AMP algorithm run with random multichannel convolutional matrices and its state evolution as proven in this paper. (left) Compressive sensing y0 = Wx0 + for noise i N(0, 10 4) and signal prior x0 N(0, 1) + (1 )δ(x), where W 2 RDq P q has varying aspect ratio β = D/P. Crosses correspond to AMP evaluations for W MCC(D, P, q, k) according to Definition 3.2, averaged over 10 independent trials. Dots correspond to AMP evaluations for W 2 RD P with i.i.d. Gaussian entries Wij N(0, 1/P). Lines show the state evolution predictions when Wij N(0, 1/Pq). The system size is P = 1024, q = 1024, k = 3, where β and D = βP vary. While our theorem treats the limit P, D ! 1, q, k = O(1), we observe strong empirical agreement even when q P. In Appendix C.1 we give the same figure for q = 10 P. (right) AMP iterates at = 0.25 and β near the recovery transition. Rather than showing these models have equivalent fixed points, we show a stronger result: the state evolution equations are equivalent at each iteration. multi-channel convolutional (MCC) matrices, having a certain block-sparse circulant structure which captures the convolutional layers used by many modern generative neural network architectures Karras et al. [2018, 2019]. In this work, we develop an asymptotic analysis of the performance of an Approximate Message Passing (AMP) algorithm Donoho et al. [2009] for recovery from multichannel convolutional signal models. This family of algorithms originates in statistical physics Mézard and Montanari [2009], Zdeborová and Krzakala [2016] and allows to compute the marginals of an elaborate posterior distribution defined by inference problems involving dense random matrices. A number of AMP iterations have been proposed for various inference problems, such as compressed sensing Donoho et al. [2009], low-rank matrix recovery Rangan and Fletcher [2012] or generalized linear modeling Rangan [2011]. More recently, composite AMP iterations (ML-AMP) have been proposed to study multilayer inference problems Manoel et al. [2017], Aubin et al. [2019]. Here we consider the ML-AMP proposed in Manoel et al. [2017] to compute marginals of a multilayer generalized linear model with unstructured random weights, but we replace these weights random convolutional weight matrices. A major benefit of AMP lies in the fact that the asymptotic distribution of their iterates can be exactly determined by a low-dimensional recursion: the state evolution equations. This enables to obtain precise theoretical results for the reconstruction performance of the proposed algorithm. Another benefit of such iterations is their low computational complexity, as they only involve matrix-multiplication and, in the separable case, pointwise non-linearities. Previous works on AMP suggest that the state evolution is not readily applicable to our setting because its derivation requires strong independence assumptions on the coordinates of the {W (l)} which are violated by structured multi-channel convolution matrices. Despite this, we use AMP for our setting and rigorously prove its state evolution. Our main contributions are: 1. We rigorously prove state evolution equations for models of the form (1), where weights are allowed to be either i.i.d. Gaussian or random structured MCC matrices, as in Definition 3.2. 2. For separable channel functions '(l) and separable signal priors, we show that the original ML-AMP of Manoel et al. [2017] used with sensing matrices that are either dense Gaussian matrices or random convolutional ones admits the same state evolution equations, up to a rescaling. Multi-layer MCC signal models can therefore simulate dense signal models while making use of fast structured matrix operations for convolutions. 3. The core of our proof shows how an AMP iteration involving random convolutional matrices may be reduced to another one with dense Gaussian matrices. We first show that random convolutional matrices are equivalent, through permutation matrices, to dense Gaussian ones with a (sparse) block-circulant structure. We then show how the block-circulant structure can be embedded in a new, matrix-valued, multilayer AMP with dense Gaussian matrices, the state evolution equations of which are proven using the results of Gerbelot and Berthier [2021], with techniques involving spatially coupled matrices Krzakala et al. [2012], Javanmard and Montanari [2013]. 4. We validate our theory numerically and observe close agreement between convolutional AMP iterations and its state evolution predictions, as shown in Figure 1 and in Section 5. Our code can be used as a general purpose library to build compositional models and evaluate AMP and its state evolution. We make this code available at https://github. com/mdnls/conv-ml-amp.git. 2 Related Work AMP-type algorithms arose independently in the contexts of signal recovery and spin-glass theory. In the former case, Donoho et al. [2009] derives AMP for Gaussian compressive sensing. This approach was later generalized by Rangan [2011] to recovery problems with componentwise channel functions that may be stochastic and/or nonlinear, and generalized further by Manoel et al. [2017] to multi-layer or compositional models. Due to the versatility of this approach, a wide variety of general purpose frameworks for designing AMP variants have since been popularized Fletcher et al. [2018], Baker et al. [2020], Gerbelot and Berthier [2021]. Proof techniques to show the concentration of AMP iterates to the state evolution prediction developed alongside new variants of the algorithm. In the context of spin-glass theory, Bolthausen s seminal work Bolthausen [2009] introduces a Gaussian conditioning technique used widely to prove AMP concentration. Following this approach, Bayati and Montanari [2011], Javanmard and Montanari [2013], Berthier et al. [2020] treat signal models with dense couplings and generalized channel functions. More recently, a proof framework adaptable to composite inference problems was proposed in Gerbelot and Berthier [2021], which we use in our proof. There has also been significant interest in relaxing the strong independence assumptions required by AMP algorithms on sensing matrix coordinates. In one direction, Vector AMP (VAMP) algorithms target signal models whose sensing matrices are drawn from right orthogonally invariant distributions. The development of VAMP algorithms followed a similar trajectory to that of vanilla AMP Schniter et al. [2016], Fletcher et al. [2018], Rangan et al. [2019], Baker et al. [2020]. The MCC ensemble considered in this work is not right orthogonally invariant, but we observe strong empirical evidence that an analogous version of Theorem 4.2 holds for VAMP as well, as described in Appendix C.2. In a second direction, there has been much interest in spatial coupling sensing matrices, which were used to achieve the information-theoretically optimal performance in sparse compressive sensing Donoho et al. [2013], Barbier et al. [2015], Krzakala et al. [2012], with complementary state evolution guarantees Javanmard and Montanari [2013]. The concept of spatial coupling and proofs of its performance originated in the literature of error correcting codes Kudekar et al. [2011, 2013], where it developed from the so-called convolutional codes Felstrom and Zigangirov [1999]. The connection between spatial coupling and convolution layers of neural networks, that we establish in this paper, is as far as we know novel. Another direction of related work is the design of generative neural network architectures, and correspondingly, the design of signal recovery procedures that can make use of new generative prior models. Bora et al. [2017], one of the original works on compressive sensing with generative models, studies signal recovery under a fully connected VAE prior and a convolutional DC-GAN prior. More advanced convolutional architectures such as PG-GAN Karras et al. [2018] and Style-GAN Karras et al. [2019] have been studied in followup work as priors for a variety of signal recovery problems Daras et al. [2021], Gu et al. [2020]. For simplicity and theoretical tractibility, we do not consider some fine-grained practical modifications used by these architetures, like batch normalization Ioffe and Szegedy [2015] or strided convolution layers Radford et al. [2015], focusing instead on the essential elements of simple convolutional networks. Lastly, while our focus is on feedforward convolutional priors such as GAN/VAE networks, there is growing interest in alternative approaches to signal recovery under neural network priors, such as normalizing flows Rezende and Mohamed Figure 2: MCC matrices operate on Pq dimensional input data, composed of q-dimensional signals for each of P separate channels. The i-th output channel is a linear combination of convolutional features extracted from input channels, where k is the convolutional filter size: y(i) = P j=1...P Cijx(j). Each row of W is a concatenation of P different rows from convolutional matrices, each row hvaing k nonzero entries, for a total of k P nonzero entries. Blue boxes show which k P input entries contribute to a single output entry. [2015], Dinh et al. [2015, 2017], Kingma and Dhariwal [2018], Asim et al. [2020] and score-based generative models Song and Ermon [2020], Jalal et al. [2021]. These approaches fall outside the scope of our work and may be interesting directions for future investigation. 3 Definition of the problem 3.1 Multi-channel Convolutional Matrices We focus our attention on multichannel convolution matrices that have localized convolutional filters. In this section, we introduce our notation and define the random matrix ensembles which are relevant to our result. We consider block structured signal vectors x 2 RP q of the form x = [x(i)]P i=1, and we refer to the blocks x(i) 2 Rq as channels. For any vector of dimension d, we denote by Pd 2 Rd d the cyclic coordinate permutation matrix of order d, whose coordinates are hei, Pdeji = 1[i = j + 1]. For a block-structured vector x 2 RP q, we denote by PP,q 2 RP q P q the block cyclic permutation matrix satisfying (PP,qx)(i) = x(i+1) for 1 i < P, and (PP,qx)(P ) = x(1). Similarly, we denote by Si,j 2 RP q P q the swap permutation matrix which exchanges blocks i, j: [Si,jx](i) = x(j), [Si,jx](j) = x(i), and [Si,jx](k) = x(k) for k 6= i, j. Last, given a vector ! 2 Rk for k q, denote by Zero-Padq,k(!) the vector whose first k coordinates are !, and whose other coordinates are zero. Zero-Padq,k(!) = [!1 !2 . . . !k 0 . . . 0] 2 Rq. We define the following ensemble for random multi-channel convolution matrices. Definition 3.1 (Gaussian i.i.d. Convolution). Let q k be integers. The convolutional ensemble C(q, k) contains random circulant matrices C 2 Rq q whose first row is given by C1 = Zero-padq,k[!] where ! 2 Rk has i.i.d. Gaussian coordinates !i N(0, 1/k). The remaining rows Ci are determined by circulant structure, ie. Ci = Pi 1 q Zero-padq,k[!]. Random multi-channel convolutions are block-dense matrices with independent C(q, k) blocks. Definition 3.2 (Multi-channel Gaussian i.i.d. Convolution). Let D, P 1 and q k 1 be integers. The random multi-channel convolution ensemble M(D, P, k, q) contains random block matrices M 2 RDq P q of the form C1,1 C1,2 . . . C1,P ... ... ... CD,1 . . . CD,P where each Ci,j C(q, k) is sampled independently. Fig. 2 gives a graphical explanation of the link between these matrices and the convolutional layers. The parameter P (D) is the number of input (output) channels, q is the dimension of the input and k the filter size. Layer D P q k 1 ! 2 1024 512 42 1 2 ! 3 512 256 82 5 3 ! 4 256 128 162 5 4 ! 5 128 3 642 5 Figure 3: System sizes for convolutional layers in a DC-GAN architecture used to generate LSUN images [Radford et al., 2015, Figure 1]. These are not directly comparable to MCC matrices, as DCGAN uses fractionally strided convolutions, which can be thought of as a composition of an MCC matrix with superresolution. However, they give a reasonable picture of the sizes of typical layers in convolutional neural networks. 3.2 Thermodynamic-like Limit and Finite-size Regimes We prove our main result in a thermodynamic-like limit D, P ! 1 while β = D/P is fixed and q, k = O(1). From a practical perspective, convolutional layers in deep neural networks often use large channel dimensions (D, P 1), large signal dimensions (q 1), and a small filter size (k = O(1)). As an example, we show in Figure 3 the sizes of convolutional layers used by the DC-GAN architecture to generate LSUN images [Radford et al., 2015, Figure 1]. Interestingly, our theoretical predictions do not depend explicitly on the relative sizes of q and (D, P). We observe empirically that these predictions become accurate at finite sizes of (D, P) which may seem small relative to q, and which are realistic from a practical neural network perspective. For example, in Figure 1, we observe strong empirical agreement with predictions for q = P = 1024 as β and D = βP vary. 3.3 Multi-layer AMP In this section, we define a class of probabilistic graphical models (PGMs) that captures the inference problems of interest, and we state the Multi-layer Approximate Message Passing (ML-AMP) Manoel et al. [2017] iterations, which can be used for inference on these PGMs. We consider the following signal model. Definition 3.3 (Multi-layer Signal Model). Let {W (l)}1 l L be matrices of dimension W (l) 2 Rnl 1 nl. Let {'(l)(z; )}1 l L be scalar channel functions for which z is the estimation quantity and represents channel noise. For vectors z, 2 Rnl 1, we write '(l)(z; ) to indicate the coordinatewise application of '(l). The multi-layer GLM signal model is given by y = '(1)(W (1)'(2)(W (2)(. . . '(L)(W (L)x; (L)) . . .). We assume x 2 Rn L follows a known separable prior, xi PX(x) i.i.d., and that the channel noise (l) has i.i.d. N(0, (σ(l))2) coordinates of known variance (σ(l))2 0. The full estimation quantities of the model are the coordinates of the vectors {h(l)}1 l L, {z(l)}1 l L, which are related by yµ = '(1)(z(1); (1)) z(1) i = '(2)(z(2); (2)) z(2) i = '(L)(z(L); (L)) z(L) where we take h(L) = x0. The corresponding conditional probabilities, which define the factor nodes of the underlying PGM, are given by P (l)(h | z) = 2 2δ(h '(z; )) , To compute the posterior marginals, ML-AMP iteratively updates the parameters of independent 1D Gaussian approximations to each marginal. Each coordinate h(l) i (t) has corresponding parameters {A(l) i (t), B(l) i (t)} and each z(l) µ (t) has corresponding {V (l) µ (t), !(l) µ (t)}, where t 1 indexes the ML-AMP iterations. The recursive relationship between these parameters is defined in terms of scalar denoising functions, ˆh(l) and g(l), which compute posterior averages of the estimation quantities given their prior parameters. In general, these denoising functions can be chosen (up to regularity assumptions) to adjust MLAMP s performance in applied settings, such as in Metzler et al. [2015], and in these cases the denoisers may be nonseparable vector valued functions. However, in the separable, Bayes-optimal regime where Px(x) and P (l)(h | z) are known, the optimal denoisers are given by, i (t + 1) := @B log Z(l+1)(A(l) i , V (l+1) i (t + 1) := @Bˆh(l) µ (t) := @! log Z(l)(A(l 1) µ (t) := @!g(l) Z(l)(A, B, V, !) := 1 p P (l)(h | z) exp 2Ah2 (z !)2 where 2 L L 1, t 2 and the prior parameters on the right hand side are taken at iteration t 2. The corresponding ML-AMP iterations are given by, i (t) V (l) µ (t 1) (4) µ (t) + A(l) For the boundary cases t = 1, l = 1, and l = L, the iterations (3), (4) are modified as follows. 1. At t = 1, we initialize B(l) B0 and !(l) !0 , where P (l) !0 are the distributions of the signal model parameters (2) when xi PX. We take (A(l) i ) 1 = Var(B(l) i ) and V (l) µ = Var(!(l) 2. At l = 1, the denoiser g(1) µ (t) = @! log Z(1)(y, V (1) Z(1)(y, V (1) P (1)(y | z) exp 3. At l = L, the denoiser ˆh(L)(t) = @B log Z(L)(A(L) 3.3.1 Computational Savings of MCC Matrices As ML-AMP requires only matrix-vector products, its computational burden can be significantly reduced when using structured and/or sparse sensing matrices. In our setting, multi-channel convolutions M MCC(D, P, q, k) have DPk nonzero coordinates, compared to DPq2 nonzero coordinates of a Gaussian i.i.d. matrix. Typically, k represents the size of a localized filter applied to a larger image, with k q [Gonzalez and Woods, 2008, Section 3.4], leading to significant space savings by a factor k/q2. This same is true in convolutional neural networks, where the use of localized convolutional filters represents an inductive bias towards localized features that is considered a key aspect of their practical success Krizhevsky et al. [2012], Zeiler and Fergus [2014]. In addition to space savings, specialized matrix-vector product implementations can reduce the time complexity of ML-AMP with MCC sensing matrices. Simple routines for sparse matrix-vector products run in time proportional to the number of nonzero entries, resulting in the same k/q2 constant factor speed up for MCC matrix-vector products. Alternatively, if k log q, then a simple algorithm using a fast Fourier transform for convolution-vector products yields time complexity O(DPq log q). Such an algorithm is sketched Appendix B. 4 Main result We now state our main technical result, starting with the set of required assumptions. (A1) for any 1 6 l 6 L, the function '(l)(z; ) is continuous and there exists a polynomial b(l) of finite order such that, for any x 2 R, |'(l)(x; )| 6 |b(l)(x, )|. (A2) for any 1 6 l 6 L, the matrix W(l) is sampled from the ensemble M(Dl, P l, kl, ql) where P lql = Dl 1ql 1 (A3) the iteration (4) is initialized with a random vector independent of the mixing matrices verifying 2 < +1 almost surely, and the prior distribution PX is sub Gaussian (A4) for any 1 6 l 6 L, Dl, Pl ! 1 with constant ratio βl = Dl/Pl, with finite ql. Under these assumptions, we may define the following state evolution recursion Definition 4.1 (State Evolution). Consider the following recursion, ˆ (l)(t) = β(l)E(l)[@!g(ˆ (l 1), b, 1 (l), h)] (5) (l 1)(t + 1) = E(l)[h ˆh(l 1)(ˆ (l 1), b, 1 (l), h)], (6) where (l) is the second moment of P (l) B0, where the right hand side parameters are taken at time t, and the expectations E(l) are taken with respect to P (l)(w, z, h, b) = P (l) out (h | z) N(z; w, (l) (l)) N(w; 0, (l)) N(b; ˆ (l 1)h, ˆ (l 1)). At t = 1, the state evolution is initialized at (l) = 0 and (ˆ (l)) 1 = (l). At the boundaries l = 1, L, the expectations are modified analogously to the ML-AMP iterations as described by Manoel et al. [2017]. We then have the following asymptotic characterization of the iterates from the convolutional ML-AMP algorithm Theorem 4.2. Under the set of assumptions (A1)-(A4), for any sequences of uniformly pseudo Lipschitz functions N 2 of order k, for any 1 6 l 6 L and any t 2 N, the following holds i (t), B(l 1) Z(l)(t), ˆZ(l 1)(t) i (t), B(l) Z(l+1)(t), ˆZ(l)(t) where Zl(t) N(0, l(t)), ˆZl(t) N(0, ˆ l(t)) are independent random variables. 4.1 Proof Sketch The proof of Theorem 4.2, which is given in Appendix A, has two key steps. First, we construct permutation matrices U, U such that for W MCC(D, P, q, k), the matrix W = UW U T is a block matrix whose blocks either have i.i.d. Gaussian elements or are zero valued, and has a block-circulant structure. The effect of the permutation is that entries of W which are correlated due to circulant structure of W are relocated to different blocks. Once these permutation matrices are defined, we define a new, matrix-valued AMP iteration involving the dense Gaussian matrices obtained from the permutations, and whose non-linearities account for the block-circulant structures and the permutation matrices. The state evolution of this new iteration is proven using the results of Gerbelot and Berthier [2021]. This provides an explicit example of how the aforementioned results can be used to obtain rigorous, non Bayes-optimal SE equations on a composite AMP iteration. The separability assumption is key in showing that the AMP iterates obtained with the convolutional matrices can be exactly embedded in a larger one. Note that this is a stronger result than proving SE equations for an algorithm that computes marginals of a random convolutional posterior: we show the SE equations are the same as in the dense case. We finally invoke the Nishimori conditions, see e.g. Krzakala et al. [2012], to simplify the generic, non Bayes-optimal SE equations to the Bayes-optimal ones. z11 w11 z12 w12 z13 w13 z11 w11 z12 w12 z13 w13 w11 z11 w12 z12 w13 z13 z21 w21 z22 w22 z23 w23 z21 w21 z22 w22 z23 w23 w21 z21 w22 z22 w23 z23 z31 w31 z32 w32 z33 w33 z31 w31 z32 w32 z33 w33 w31 z31 w32 z32 w33 z33 z41 w41 z42 w42 z43 w43 z41 w41 z42 w42 z43 w43 w41 z41 w42 z42 w43 z43 z11 z12 z13 w11 w12 w13 z21 z22 z23 w21 w22 w23 z31 z32 z33 w31 w32 w33 z41 z42 z43 w41 w42 w43 z11 z12 z13 w11 w12 w13 z21 z22 z23 w21 w22 w23 z31 z32 z33 w31 w32 w33 z41 z42 z43 w41 w42 w43 w11 w12 w13 z11 z12 z13 w21 w22 w23 z21 z22 z23 w31 w32 w33 z31 z32 z33 w41 w42 w43 z41 z42 z43 Figure 4: A sketch of the permutation lemma applied to matrix W MCC(4, 3, 3, 2). Left: W before permutation. Right: after permutation, UW U T . The idea of embedding a non-separable effect such as a block-circulant structure or different variances in a mixing matrix is the core idea in the proofs of SE equations for spatially coupled systems, notably as done in Javanmard and Montanari [2013], Donoho et al. [2013]. We note that in the numerical experiments shown at Figure 1, the parameter q, considered finite in the proof, is actually comparable to the number of channel, considered to be extensive. Empirically we observe that this does not hinder the validity of the result, something that was also observed in the spatial coupling literature, e.g. Krzakala et al. [2012], where large number of different blocks in spatially coupled matrices were considered, with convincing numerical agreement. The existence of permutations matrices verifying the property described above is formalized in the following lemma: Lemma 4.3 (Permutation Lemma). Let W M(D, P, k, q) be a multi-channel convolution matrix. There exist row and column permutation matrices U 2 RDq Dq, U 2 RP q P q such that W = UW U T is a block-convolutional matrix with dense, Gaussian i.i.d. blocks. That is, A(1) A(2) . . . A(k) A(1) A(2) . . . A(k) ... A(2) . . . A(k) ... ... A(2) A(3) . . . A(k) A(1) where each A(s) 2 RD,P , 1 s k has i.i.d. N(0, 1/P) coordinates. Proof. Consider the elements of the matrix M which are non-zero and sampled i.i.d. as opposed to exact copies of other variables. They are positioned on the first row of each block of size q q, and thus the indexing for their rows and columns can be written as Maq+1,bq+c where a, b, c are integers such that 0 6 a 6 D 1, 0 6 b 6 P 1 and 1 6 c 6 k. The integers a, b describe the position of the q q block the variable is in, and c describes, for each block, the position in the initial random Gaussian vector of size k that is zero-padded and circulated to generate the block. The goal is to find the mapping that groups these variables into k dense blocks of extensive size D P. To do so, one can use the following bijection Mγ, P +β = Maq+1,bq+c where γ = a + 1, = c 1 and β = b + 1. By doing this, c becomes the block index and a, b become the position in the dense block. This mapping can be represented by left and right permutation matrices which also prescribe the permutation for the rest of the elements of M. Figure 4 shows a sketch of this permutation. The state evolution recursion given in Definition 4.1 is equivalent to the state evolution for unstructured matrices proposed by Manoel et al. [2017], which may seem surprising relative to existing literature on spatial coupling, where spatially coupled sensing matrices are used to achieve improved performance over unstructured sensing matrices. This equivalence is a consequence of the fact that W has dense Gaussian blocks which each have the same variance, whereas in the literature, spatially coupled matrices typically have Gaussian blocks with inhomogeneous variances that can be tuned to improve recovery performance. 5 Numerical Experiments In this section, we compare state evolution predictions from Theorem 4.2 with a numerical implementation of the ML-AMP algorithm described in Section 3.3. Our first experiment, shown in Figure 1, is a noisy compressive sensing task under a sparsity prior PX(x) = N(x; 0, 1) + (1 )δ(x), where is the expected fraction of nonzero components of x0. Measuremements are generated y0 = Wx0 + for noise N(0, 10 4), where W MCC(D, P, q, k). We show recovery performance at sparsity levels 2 {0.25, 0.5, 0.75} as the measurement ratio β = D/P varies, averaged over 10 independent AMP iterates. Additionally, we show convergence of the (averaged) AMP iterates for sparsity = 0.25 at a range of β near the recovery threshold. We observe strong agreement between AMP empirical performance and the state evolution prediction. The system sizes are P = 1024, q = 1024, with D = βP varying. In Figure 5, we consider two examples of L = 2, 3, 4 layer models following Equation (2). In both, the output channel l = 1 generates noisy, compressive linear measurements y = z(1) + for i N(0, σ2) and for dense couplings W (1) ij N(0, 1/n(1)). Layers 2 l 4 use MCC couplings W (l) MCC(Dl, Pl, q, k), where q Pl = nl and Dl = βPl = qnl 1. Channel functions {'(l)} vary across the two experiments. The input prior is PX(x) = N(x; 0, 1) and the problem parameters are q = 10 channels, filter size k = 3, noise level σ2 = 10 4, input dimension n(L) = 5000, and layerwise aspect ratios β(L) = 2 and β(l) = 1 for 2 l < L. Finally, the channel aspect ratio β(1) varies in each experiment. 0.0 0.2 0.4 0.6 0.8 1.0 β(1) - Measurement Ratio Linear Activations L = 1 L = 2 L = 3 MCC AMP Dense AMP Dense SE 0.0 0.2 0.4 0.6 0.8 1.0 β(1) - Measurement Ratio Re LU Activations L = 1 L = 2 L = 3 MCC AMP Dense AMP Dense SE Figure 5: ML-AMP compressive sensing recovery under multichannel convolutional designs (crossed) and the state evolution for the corresponding fully connected model (lined). For comparison, we also plot the corresponding fully connected AMP iterations (dotted), in which W (l) 2 RDl Pl with Wij N(0, 1/Pl), with the dimensions of the prior and output channel adjusted appropriately. Left: For 2 l L, the channel functions are '(l)(z; ) = z + where i N(0, σ2). Right: For 2 l L, the channel functions are '(l)(z; ) = max(z, 0) where the maximum is applied coordinatewise. This channel function is the popular Re LU activation function used by generative convolutional neural networks such as in Radford et al. [2015], Bora et al. [2017]. We compare the state evolution equations to empirical AMP results in two cases. In the left panel, we show multilayer models with identity channel functions, and in the right panel, we show models with Re LU channel functions. The latter captures a simple but accurate example of a convolutional GNN. 6 Discussion and Future Work We have proven state evolution recursions for the ML-AMP algorithm for signal recovery from multilayer convolutional networks. We consider networks whose weight matrices are drawn either i.i.d. Gaussian or from an ensemble of random multi-channel convolution matrices. Interestingly, under a separable prior and separable channel functions, these two matrix ensembles yield the same state evolution (up to a rescaling). These predictions closely match empirical observations in compressive sensing under a sparsity prior (Figure 1) and under multi-layer priors (Figure 5). Based on the discussion in Section 4.1, the equivalence property is expected to break down if W is replaced by a block matrix Wsp with dense Gaussian blocks organized in the same block-circulant structure, with inhomogeneous variances. As we discuss in Appendix D, by inverting the permutation lemma, Wsp corresponds to a multi-channel convolution Wsp with inhomogeneous convolutional filters, which can be taken as a simple model for structured convolutions. Inhomogeneous blocks can also easily be incorporated in the Graph-AMP framework using nonseparable denoisers, so the key elements of our proof also apply to the inhomogeneous case. This generalized model for MCC matrices represents an interesting direction for future exploration, whose state evolution is expected to diverge from that of the dense Gaussian ensemble. Acknowledgments and Disclosure of Funding M.D. acknowledges funding from Northeastern University s Undergraduate Research and Fellowships office and the Goldwater Award. We also acknowledge funding from the ERC under the European Union s Horizon 2020 Research and Innovation Program Grant Agreement 714608SMi Le, and by the Swiss National Science Foundation grant SNFS Opera GOST, 200021_200390. Muhammad Asim, Max Daniels, Oscar Leong, Ali Ahmed, and Paul Hand. Invertible generative models for inverse problems: mitigating representation error and dataset bias. In Hal Daumé III and Aarti Singh, editors, Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 399 409. PMLR, 13 18 Jul 2020. URL https://proceedings.mlr.press/v119/asim20a.html. Benjamin Aubin, Bruno Loureiro, Antoine Maillard, Florent Krzakala, and Lenka Zdeborová. The spiked matrix model with generative priors. Advances in Neural Information Processing Systems, 32, 2019. Antoine Baker, Benjamin Aubin, Florent Krzakala, and Lenka Zdeborová. Tramp: Compositional inference with tree approximate message passing. ar Xiv preprint ar Xiv:2004.01571, 2020. Jean Barbier, Christophe Schülke, and Florent Krzakala. Approximate message-passing with spatially coupled structured operators, with applications to compressed sensing and sparse superposition codes. Journal of Statistical Mechanics: Theory and Experiment, 2015(5):P05013, 2015. Mohsen Bayati and Andrea Montanari. The dynamics of message passing on dense graphs, with applications to compressed sensing. IEEE Transactions on Information Theory, 57(2):764 785, 2011. Raphael Berthier, Andrea Montanari, and Phan-Minh Nguyen. State evolution for approximate message passing with non-separable functions. Information and Inference: A Journal of the IMA, 9(1):33 79, 2020. Erwin Bolthausen. On the high-temperature phase of the sherrington-kirkpatrick model, Sep 2009. Ashish Bora, Ajil Jalal, Eric Price, and Alexandros G Dimakis. Compressed sensing using generative models. In International Conference on Machine Learning, pages 537 546. PMLR, 2017. Giannis Daras, Joseph Dean, Ajil Jalal, and Alex Dimakis. Intermediate layer optimization for inverse problems using deep generative models. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 2421 2432. PMLR, 18 24 Jul 2021. URL https://proceedings. mlr.press/v139/daras21a.html. Laurent Dinh, David Krueger, and Yoshua Bengio. NICE: non-linear independent components estimation. In Yoshua Bengio and Yann Le Cun, editors, 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Workshop Track Proceedings, 2015. URL http://arxiv.org/abs/1410.8516. Laurent Dinh, Jascha Sohl-Dickstein, and Samy Bengio. Density estimation using real NVP. In International Conference on Learning Representations, 2017. URL https://openreview. net/forum?id=Hkpbn H9lx. David L Donoho, Arian Maleki, and Andrea Montanari. Message-passing algorithms for compressed sensing. Proceedings of the National Academy of Sciences, 106(45):18914 18919, 2009. David L Donoho, Adel Javanmard, and Andrea Montanari. Information-theoretically optimal compressed sensing via spatial coupling and approximate message passing. IEEE transactions on information theory, 59(11):7434 7464, 2013. A Jimenez Felstrom and Kamil Sh Zigangirov. Time-varying periodic convolutional codes with low-density parity-check matrix. IEEE Transactions on Information Theory, 45(6):2181 2191, 1999. Alyson K Fletcher, Sundeep Rangan, and Philip Schniter. Inference in deep networks in high dimensions. In 2018 IEEE International Symposium on Information Theory (ISIT), pages 1884 1888. IEEE, 2018. Cédric Gerbelot and Raphaël Berthier. Graph-based approximate message passing iterations. ar Xiv preprint ar Xiv:2109.11905, 2021. Rafael C. Gonzalez and Richard E. Woods. Digital image processing, 3rd Edition. Pearson Education, 2008. ISBN 9780135052679. URL https://www.worldcat.org/oclc/241057034. Jinjin Gu, Yujun Shen, and Bolei Zhou. Image processing using multi-code gan prior. In CVPR, Sergey Ioffe and Christian Szegedy. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Francis Bach and David Blei, editors, Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pages 448 456, Lille, France, 07 09 Jul 2015. PMLR. URL https:// proceedings.mlr.press/v37/ioffe15.html. Ajil Jalal, Sushrut Karmalkar, Alex Dimakis, and Eric Price. Instance-optimal compressed sensing via posterior sampling. In Marina Meila and Tong Zhang, editors, Proceedings of the 38th International Conference on Machine Learning, volume 139 of Proceedings of Machine Learning Research, pages 4709 4720. PMLR, 18 24 Jul 2021. URL https://proceedings.mlr.press/ v139/jalal21a.html. Adel Javanmard and Andrea Montanari. State evolution for general approximate message passing algorithms, with applications to spatial coupling. Information and Inference: A Journal of the IMA, 2(2):115 144, 2013. Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. Progressive growing of GANs for improved quality, stability, and variation. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=Hk99z Ce Ab. Tero Karras, Samuli Laine, and Timo Aila. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 4401 4410, 2019. Durk P Kingma and Prafulla Dhariwal. Glow: Generative flow with invertible 1x1 convolutions. In S. Bengio, H. Wallach, H. Larochelle, K. Grauman, N. Cesa-Bianchi, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 31. Curran Associates, Inc., 2018. URL https://proceedings.neurips.cc/paper/2018/file/ d139db6a236200b21cc7f752979132d0-Paper.pdf. Alex Krizhevsky, Ilya Sutskever, and Geoffrey E Hinton. Imagenet classification with deep convolu- tional neural networks. Advances in neural information processing systems, 25, 2012. Florent Krzakala, Marc Mézard, François Sausset, YF Sun, and Lenka Zdeborová. Statistical-physics- based reconstruction in compressed sensing. Physical Review X, 2(2):021005, 2012. Shrinivas Kudekar, Thomas J Richardson, and Rüdiger L Urbanke. Threshold saturation via spatial coupling: Why convolutional ldpc ensembles perform so well over the bec. IEEE Transactions on Information Theory, 57(2):803 834, 2011. Shrinivas Kudekar, Tom Richardson, and Rüdiger L Urbanke. Spatially coupled ensembles universally achieve capacity under belief propagation. IEEE Transactions on Information Theory, 59(12): 7761 7813, 2013. Andre Manoel, Florent Krzakala, Marc Mézard, and Lenka Zdeborová. Multi-layer generalized linear estimation. In 2017 IEEE International Symposium on Information Theory (ISIT), pages 2098 2102. IEEE, 2017. Christopher A Metzler, Arian Maleki, and Richard G Baraniuk. Bm3d-amp: A new image recovery algorithm based on bm3d denoising. In 2015 IEEE International Conference on Image Processing (ICIP), pages 3116 3120. IEEE, 2015. Marc Mézard and Andrea Montanari. Information, physics, and computation. Oxford University Press, 2009. Alec Radford, Luke Metz, and Soumith Chintala. Unsupervised representation learning with deep convolutional generative adversarial networks. ar Xiv preprint ar Xiv:1511.06434, 2015. Sundeep Rangan. Generalized approximate message passing for estimation with random linear mixing. In 2011 IEEE International Symposium on Information Theory Proceedings, pages 2168 2172. IEEE, 2011. Sundeep Rangan and Alyson K Fletcher. Iterative estimation of constrained rank-one matrices in noise. In 2012 IEEE International Symposium on Information Theory Proceedings, pages 1246 1250. IEEE, 2012. Sundeep Rangan, Philip Schniter, and Alyson K Fletcher. Vector approximate message passing. IEEE Transactions on Information Theory, 65(10):6664 6684, 2019. Danilo Rezende and Shakir Mohamed. Variational inference with normalizing flows. In Francis Bach and David Blei, editors, Proceedings of the 32nd International Conference on Machine Learning, volume 37 of Proceedings of Machine Learning Research, pages 1530 1538, Lille, France, 07 09 Jul 2015. PMLR. URL https://proceedings.mlr.press/v37/rezende15.html. Philip Schniter, Sundeep Rangan, and Alyson K Fletcher. Vector approximate message passing for the generalized linear model. In 2016 50th Asilomar Conference on Signals, Systems and Computers, pages 1525 1529. IEEE, 2016. Yang Song and Stefano Ermon. Improved techniques for training score-based generative models. In Hugo Larochelle, Marc Aurelio Ranzato, Raia Hadsell, Maria-Florina Balcan, and Hsuan-Tien Lin, editors, Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, Neur IPS 2020, December 6-12, 2020, virtual, 2020. Lenka Zdeborová and Florent Krzakala. Statistical physics of inference: Thresholds and algorithms. Advances in Physics, 65(5):453 552, 2016. Matthew D Zeiler and Rob Fergus. Visualizing and understanding convolutional networks. In European conference on computer vision, pages 818 833. Springer, 2014. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] See Section 3.2 (c) Did you discuss any potential negative societal impacts of your work? [No] We believe that because of the theoretical nature of this work, any potential negative societal impacts would be highly indirect. We are not aware of any potential direct negative societal impacts to discuss. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] See Section 4 (b) Did you include complete proofs of all theoretical results? [Yes] See Section 4 and the Appendix 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experi- mental results (either in the supplemental material or as a URL)? [Yes] The code to reproduce our results is included in supplemental material and will be made publically available on Github. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [N/A] Our results do not require training (c) Did you report error bars (e.g., with respect to the random seed after running experi- ments multiple times)? [No] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [No] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [N/A] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [N/A] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]