# multilinear_operator_networks__80f5de9e.pdf Published as a conference paper at ICLR 2024 MULTILINEAR OPERATOR NETWORKS Yixin Cheng1, Grigorios G. Chrysos2, Markos Georgopoulos, Volkan Cevher1 1LIONS - Ecole Polytechnique F ed erale de Lausanne 2University of Wisconsin-Madison Despite the remarkable capabilities of deep neural networks in image recognition, the dependence on activation functions remains a largely unexplored area and has yet to be eliminated. On the other hand, Polynomial Networks is a class of models that does not require activation functions, but have yet to perform on par with modern architectures. In this work, we aim close this gap and propose MONet, which relies solely on multilinear operators. The core layer of MONet, called Mu-Layer, captures multiplicative interactions of the elements of the input token. MONet captures high-degree interactions of the input elements and we demonstrate the efficacy of our approach on a series of image recognition and scientific computing benchmarks. The proposed model outperforms prior polynomial networks and performs on par with modern architectures. We believe that MONet can inspire further research on models that use entirely multilinear operations. The source code is available at MONet. 1 INTRODUCTION Image recognition has long served as a crucial benchmark for evaluating architecture designs, including the seminal Res Net (He et al., 2016) and MLP-Mixer (Tolstikhin et al., 2021). As architectures are applied to new applications, there are additional requirements for the architecture design. For instance, encryption is a key requirements in safety-critical applications (Caruana et al., 2015). Concretely, the Leveled Fully Homomorphic Encryption (LFHE) (Brakerski et al., 2014), can provide a high level of security for sensitive information. The core limitation of FHE (and especially LFHE) is that they support only addition and multiplication as operations. That means that traditional neural networks cannot fit into such privacy constraints owing to their dependence on elementwise activation functions, making developments in MLP-Mixer and similar models invalid for many real-world applications. Therefore, new designs that can satisfy those constraints and still achieve high accuracy on image recognition are required. A core advantage of Polynomial Nets (PNs), that express the output as high-degree interactions between the input elements (Ivakhnenko, 1971; Shin & Ghosh, 1991; Chrysos et al., 2020), is that they can satisfy constraints, such as encryption or interpretability (Dubey et al., 2022). However, a major drawback of PNs so far is that they fall short of the performance of modern architectures on standard machine learning benchmarks, such as image recognition. This is precisely the gap we aim to close in this work. We introduce a class of PNs, dubbed Multilinear Operator Network (MONet), which is based solely on multilinear operations1. The core layer captures multiplicative interactions within the token elements2. The multiplicative interaction is captured using two parallel branches, each of which assumes a different rank to enable different information to flow. By composing sequentially such layers, the model captures high-degree interactions between the input elements and can predict the target signal, e.g., class in the case of image recognition. Concretely, our contributions can be summarized as: 1The terminology on multilinear operations arises from the multilinear algebra. Concretely, we follow the terminology of the seminal paper of Kolda & Bader (2009). 2Consistent with the recent literature of MLP-based models and transformers (Dosovitskiy et al., 2021), we consider sequences of tokens as inputs. In the case of images, a token refers to a (vectorized) patch of the input image. Published as a conference paper at ICLR 2024 We propose Mu-Layer, a new module which uses purely multilinear operations to capture multiplicative interactions. We showcase how this module can serve as a plug-in replacement to standard MLP. We introduce MONet, which captures high-degree interactions between the input elements. To our knowledge, this is the first network that obtains a strong performance on challenging benchmarks. We conduct a thorough evaluation of the proposed model across standard image recognition benchmarks to show the efficiency and effectiveness of our method. MONet significantly improves the performance of the prior art on polynomial networks, while it is on par with other recent strong-performing architectures. The source code is available at MONet. We hope the code can enable further improvement of models relying on linear projections. 2 RELATED WORK We present a brief overview of the most closely related categories of MLP-based and polynomialbased architectures from the vast literature of deep learning architectures. For a detailed overview, the interested reader can find further information on dedicated surveys on image recognition (Lu & Weng, 2007; Plested & Gedeon, 2022; Peng et al., 2022). MLP models: The resurgence of MLP models in image recognition is an attempt to reduce the computational overhead of transformers (Vaswani et al., 2017). MLP-based models rely on linear layers, instead of convolutional layers or the self-attention block. The MLP-Mixer (Tolstikhin et al., 2021) is among the first networks that demonstrate a high accuracy on challenging benchmarks. The MLP-Mixer uses tokens (i.e., vectorized patches of the image) and captures both interand intra-token correlations. Follow-up works improve upon the simple idea of token-mixing MLP structure (Touvron et al., 2022; Liu et al., 2021). Concretely, Vi P (Hou et al., 2022), cycle MLP (Chen et al., 2022), S2-MLPv2 (Yu et al., 2022a) design strategies to improve feature aggregation across spatial positions. Our work differs from MLP-based models, as it is inspired by the idea of capturing high-order interactions using polynomial expansions. Polynomial models: Polynomial expansions establish connections between input variables and learnable coefficients through addition and multiplication. Polynomial Nets (PNs) express the output variable (e.g., class) as a high-degree expansion of the input elements (e.g., pixels of an image) using learnable parameters. Even though extracting such polynomial features is not a new idea (Shin & Ghosh, 1991; Li, 2003; Dauphin et al., 2017), it has been largely sidelined from architecture design. The last few years PNs have demonstrated promising performance in standard benchmarks in various vision tasks including image recognition (Chrysos et al., 2022). In particular, PNs augmented with activation functions, which are referred to as hybrid models in this work, can achieve state-of-the-art performance (Hu et al., 2018; Li et al., 2019; Yin et al., 2020; Babiloni et al., 2021; Yang et al., 2022; Georgopoulos et al., 2020; Chrysos et al., 2021; Georgopoulos et al., 2021). The work of Chrysos et al. (2022) introduces a taxonomy for a single block of PNs based on the degree of interactions captured. This taxonomy allows the comparison of various approaches based on a specific degree of interactions. Building upon this work, researchers have explored ways to modify the degree or type of interactions captured to improve the performance. For example, Babiloni et al. (2021) reduce computational cost of the popular non-local block (Wang et al., 2018) by capturing exactly the same third degree interactions, while Chrysos et al. (2022) investigate modifications to the degree of expansion. Arguably, the works most closely related to ours are Chrysos et al. (2020; 2023). In Π-Nets (Chrysos et al., 2020), three different models are instantiated based on separate manifestations of three tensor decompositions. That is, the paper establishes a link between a concrete tensor decomposition and its assumptions to a concrete architecture that can be implemented in standard frameworks. The authors evaluate those different architectures and notice that they perform well, but do not surpass standard networks, such as Res Net. Chrysos et al. (2023) improve upon the Π-Nets by introducing carefully designed (implicit and explicit) regularization techniques, which reduce the overfitting of high-degree polynomial expansions. Published as a conference paper at ICLR 2024 Figure 1: The architecture of the proposed Mu-Layer (on the left) and MONet (on the right). In the left figure, the grey box represents layer normalization. The color solid line boxes represent channel projection in different dimensions, all projection operations are linear. The box denotes an elementwise (Hadamard) product. The red dash box represents the spatial aggregation module. Contrary to the aforementioned PNs, we are motivated to introduce a new architecture using PN that is comparable to modern networks. To achieve that, we are inspired by the modern setup of considering the input as a sequence of tokens. The token-based input is widely used across a range of domains and modalities that last few years. The token-based input also departs from the design of previous PNs that utilize convolutional layers instead of simple matrix multiplications. Arguably, our choice results in a weaker inductive bias as pointed out in the related works of MLP-based models. This can be particularly useful in domains outside of images, e.g., in the ODE experiments. Let us now introduce MONet, which expresses the output as a polynomial expansion of the input using multilinear operations. Firstly, we introduce the core layer, called Mu-Layer, in Section 3.1. Then, in Section 3.2, we design the whole architecture with a schematic visualized in Fig. 1. Tokens: Following the conventions of recent MLP-based models, we consider an image as a sequence of tokens. In practice, each token is a patch of the image. We denote the sequence of tokens as X Rd s, where d N is the length of a token and s N is the number of tokens. As such, our method below assumes an input sequence X is provided. MLP-based models capture linear interactions between the elements of a token. That operation would be denoted as ΓX, where Γ Ro d is a learnable matrix. 3.1 MU-LAYER Our goal is to capture rich interactions among the input elements of a token. We can achieve that using multiplicative interactions. Specifically, we use two branches, and each captures a set of linear interactions inside the token. An elementwise product is then used to capture the multiplicative interactions between the branches. Notation-wise, the output is expressed as (AX) (ΛX), where denotes the elementwise product and A, Λ are learnable matrices. We perform the following three modifications to the aforementioned block. Firstly, we add a shortcut connection to capture the first-degree interactions. Secondly, we decompose Λ into two matrices as Λ = BD. This rank factorization of Λ enables us to capture low-rank representations in this branch. Thirdly, we add one matrix C to capture the linear interactions of the whole expression. Then, the complete layer is expressed as follows: Y = C [(AX) (BDX) + AX] , (1) Published as a conference paper at ICLR 2024 Table 1: Configurations of different MONet models. We present four models: two variants with different parameter-size (Tiny, Small) and two variants with different hidden-size. The original model use the same hidden-size across the architecture. The Multi-stage model adopts a different hidden sizes across the architecture. The number in stages represents the number of blocks that each stage includes, while the hidden size list numbers correspond to each stage s hidden size. We provide further details along with a schematic in Appendix B. We only use Multi-stage models for high-resolution image classification benchmarks, e.g., on Image Net1K. Specification Tiny Small Multi-stage-Tiny Multi-stage-Small Numbers of Blocks 32 45 32 32 Embedding Method Pyramid Patch Embedding Pyramid Patch Embedding Pyramid Patch Embedding Pyramid Patch Embedding Hidden size 192 320 64,128,192,192 128,192,256,384 Stages - - 4,8,12,10 4,6,12,14 Expansion Ratio 3 3 3 3 Shrinkage Ratio 4 4 8 8 Parameters 14.0M 56.8M 10.3M 32.9M FLOPs 3.6 11.2 2.8 6.8 where A Rm d, B Rm l, C Ro m, D Rl d are learnable parameters and symbolizes an elementwise product. The block has two hyperparameters we can choose, i.e., the rank m N and the value l N. In practice, we utilize a shrinkage ratio m l > 1 to encourage different information flow in the branches. Prop. 1, the proof of which is in Appendix A, indicates the interactions captured. The schematic of this layer, called Mu-Layer, is depicted in Appendix C. Proposition 1. The Mu-Layer captures multiplicative interactions between elements of each token. Poly-Block: The Poly-Block, which is the core operation of MONet architecture, stacks Mu-Layers sequentially. Concretely, we connect two Mu-Layers sequentially and add a a layer normalization (Ba et al., 2016) between them3. Additionally, we add a shortcut connection to skip the first block. The two Mu-Layers are similar except for two differences: (a) a spatial shift operation (Yu et al., 2022b) is added to the first block, (b) the hidden dimension is m in the first Mu-Layer, and m ϵ in the second, where ϵ is an expansion factor. The Poly-Block is illustrated in Fig. 1. 3.2 NETWORK ARCHITECTURE Our final architecture is composed of N Poly-Blocks. Each block captures up to 4th degree interactions, which results in the final architecture capturing up to 4N interactions, with N > 10 in practice4. Beyond the Poly-Block, one critical detail in our final architecture is the patch embedding. Patch embeddings: In deep learning architectures, the choice of patch size plays a crucial role in capturing fine-grained features. Smaller patches prove more effective in this regard. However, a trade-off arises, as reducing the patch size leads to an increase in floating-point operations, which scales approximately quadratically. To overcome this challenge, we introduce a novel scheme called pyramid patch embedding. This scheme operates by considering embeddings at smaller scales and subsequently extracting new patch embeddings on top. By adopting this approach, we can leverage the finer details of features without introducing additional parameters or computational complexity. Our experiments validate the effectiveness of incorporating multi-level patches, enhancing the overall performance. For the Image Net1K dataset, we employ a two-level pyramid scheme, although it is worth noting that higher-resolution images may benefit from additional levels. We present the schematic and further details of this approach in Appendix D. 4 EXPERIMENTS In this section, we conduct a thorough experimental validation of MONet. We conduct experiments on large-scale image classification in Section 4.1, fine-grained and small-scale image classification 3We also conduct experiments with batch normalization and obtain similar performance; results and analysis on Section 4.1. 4A theoretical proof on the degree of expansion required to tackle a specific task is an interesting problem that is left as future work. Published as a conference paper at ICLR 2024 in Section 4.2. In addition, we exhibit a unique advantage of models without activation functions to learn dynamic systems in scientific computing in Section 4.3. Lastly, we validate the robustness of our model to diverse perturbations in Section 4.4. We summarize four configurations of the proposed MONet in Table 1 with different versions of MONet. We present a schematic for our configuration in Appendix B. Due to the limited space, we conduct additional experiments in the Appendix. Concretely, we experiment on semantic segmentation in Appendix H, while we add additional ablations and experimental details on Appendices K to M. 4.1 IMAGENET1K CLASSIFICATION Image Net1K, which is the standard benchmark for image classification, contains 1.2M images with 1,000 categories annotated. We consider a host of baseline models to compare the performance of MONet. Concretely, we include strong-performing polynomial models5 (Chrysos et al., 2020; 2023), MLP-like models (Tolstikhin et al., 2021; Touvron et al., 2022; Yu et al., 2022b), models based on vanilla Transformer (Vaswani et al., 2017; Touvron et al., 2021) and several classic convolutional networks (Bello et al., 2019; Chen et al., 2018b; Simonyan & Zisserman, 2015; He et al., 2016). Training Setup: We train our model using Adam W optimizer (Loshchilov & Hutter, 2019). We use a batch size of 448 per GPU to fully leverage the memory capacity of the GPU. We use a linear warmup and cosine decay schedule learning rate, while the initial learning rate is 1e-4, linear increase to 1e-3 in 10 epochs and then gradually drops to 1e-5 in 300 epochs. We use label smoothing (Szegedy et al., 2016), standard data augmentation strategies, such as Cut-Mix (Yun et al., 2019), Mix-up (Zhang et al., 2018) and auto-augment (Cubuk et al., 2019), which are used in similar methods (Tolstikhin et al., 2021; Trockman & Kolter, 2023; Touvron et al., 2022). Our data augmentation recipe follows the one used in MLP-Mixer (Tolstikhin et al., 2021). We do not use any external data for training. We train our model using native Py Torch training on 4 NVIDIA A100 GPUs. We exhibit the performance of the compared networks in Table 2. Notice that our smaller model achieves a 10% improvement over previous PNs. Our larger model further strengthens this performance gain and manages for the first time to close the gap between PNs and be on par with other recent models. 4.2 ADDITIONAL BENCHMARKS IN IMAGE RECOGNITION Beyond Image Net1K, we experiment with a number of additional benchmarks to further assess MONet. We use the standard datasets of CIFAR10 (Krizhevsky et al., 2009), SVHN (Netzer et al., 2011) and Tiny Image Net1K (Le & Yang, 2015) for image recognition. A fine-grained classification experiment on Oxford Flower102 (Nilsback & Zisserman, 2008) is conducted. Beyond the different distributions, those datasets offer insights into the performance of the proposed model in datasets with limited data6. The results in Table 3 compare the proposed method with other strong-performing methods. MONet outperforms the compared methods with the second strongest being the R-Poly Nets. MLP-models usually do not perform well in small datasets as indicated in Tolstikhin et al. (2021) due to their weaker inductive bias. Notably, MONet still performs better than other CNN-based polynomial models. 4.3 POLY NEURAL ODE SOLVER An additional advantage of our method is the ability to model functions that have a polynomial form (e.g, in ordinary differential equations) in an interpretable manner (Fronk & Petzold, 2023). A neural ODE solver is a computational technique that models dynamic systems relying on ODEs using neural networks. Chen et al. (2018a); Dupont et al. (2019) illustrate the potential of neural ODEs to capture the behavior of complex systems that evolve over time. However, an evident limitation is that those models remain black-box which means the learned equation remains unknown after training. On the 5We note that we report results of Π nets without activation functions from Chrysos et al. (2020) for a fair comparison. The models with activation functions are reported as hubrid . 6Previous MLP-based models have demonstrated weaker performance than CNN-based models in such datasets. The original S2MLP only reports two large models on these datasets. We use an open source S2MLP code design two models with 12-14 M parameters for a fair comparison. Those models, noted as S2MLP-Wide-S (hidden size 512, depth 8) and S2MLP-Deep-S (hidden size 384, depth 12) are used for the comparisons. Published as a conference paper at ICLR 2024 Table 2: Image Net1K classification accuracy (%) on the validation set for different models. In the Activation Function column, G denotes Gelu, R denotes Relu, T denotes Tanh, and P denotes RPRe LU. The models with are special polynomial models with activation functions. We mark model with up to 15M params in red color, models with 15-40M parameters with green color and model with over 40M parameters with blue . Beyond our models with Layer Normalization, we conduct an experiment with Batch Normalization and notice that the performance is similar (i.e., the model Multi-stage MONet-T-Batch Norm below). Extra Data Top-1(%) Top-5(%) FLOPs (B) Params (M) Activation Attention Res Net-18 (He et al., 2016) 69.7 89.0 1.8 11.0 R Res Net-50 (He et al., 2016) 77.2 92.9 4.1 25.0 R A2Net (Chen et al., 2018b) 77.0 93.5 31.3 33.4 R AA-Res Net-152 (Bello et al., 2019) 79.1 94.6 23.8 61.6 R VGG-16 (Simonyan & Zisserman, 2015) 71.5 92.7 15.5 138.0 R Transformer-based Dei T-S/16 (Touvron et al., 2021) 81.2 - 5.0 24.0 G Vi T-B/16 (Dosovitskiy et al., 2021) 77.9 - 55.5 86.0 G Bi MLP-S (Xu et al., 2022) 70.0 - 1.21 - P Bi MLP-B (Xu et al., 2022) 72.7 - 1.21 - P Res MLP-12 (Touvron et al., 2022) 76.6 - 3.0 15.0 G Hire-MLP-Tiny (Guo et al., 2022) 79.7 - 2.1 18.0 G Cycle MLP-T (Chen et al., 2022) 81.3 - 4.4 28.8 G Res MLP-24 (Touvron et al., 2022) 79.4 - 6 30.0 G MLP-Mixer-B/16 (Tolstikhin et al., 2021) 76.4 - 11.6 59.0 G MLP-Mixer-L/16 (Tolstikhin et al., 2021) 71.8 - 44.6 207.0 G S2MLP-Wide (Yu et al., 2022b) 80.0 94.8 14.0 71.0 G S2MLP-Deep (Yu et al., 2022b) 80.7 95.4 10.5 51.0 G FF (Melas-Kyriazi, 2021) 74.9 - 7.21 59.0 G Polynomial-based Π-Nets (Chrysos et al., 2020) 65.2 85.9 1.9 12.3 None Hybrid Π-Nets (Chrysos et al., 2020) 70.7 89.5 1.9 11.9 R,T PDC-comp (Chrysos et al., 2022) 69.8 89.9 1.3 7.5 R,T PDC (Chrysos et al., 2022) 71.0 89.9 1.6 10.7 R,T R-Poly Nets (Chrysos et al., 2023) 70.2 89.3 1.9 12.3 None D-Poly Nets (Chrysos et al., 2023) 70.0 89.4 1.9 11.3 None Multi-stage MONet-T-Layer Norm (Ours) 77.0 93.4 2.8 10.3 None Multi-stage MONet-T-Batch Norm (Ours) 77.7 93.8 2.8 10.3 None Multi-stage MONet-S-Layer Norm (Ours) 81.3 95.5 6.8 32.9 None Table 3: Experimental validation of different polynomial architectures on smaller datasets. The best are marked in bold. D-Poly Nets perform favorably to most of the other baseline polynomial networks and the rest baselines. MONet outperforms all the baselines and all polynomial networks in two of the four datasets, and shares the best accuracy with R-Poly Nets in another dataset. Model CIFAR10 SVHN Oxford Flower Tiny Imagenet Resnet18 94.4 97.3 87.8 61.5 MLP-Mixer 90.6 96.8 80.2 45.6 Res-MLP 92.3 97.1 83.2 58.9 S2MLP-Deep-S 92.6 96.7 93.0 59.3 S2MLP-Wide-S 92.0 96.6 91.5 52.7 Hybrid Pi-Nets 94.4 - 88.9 61.1 Π-Nets 90.7 96.1 82.6 50.2 PDC 90.9 96.0 88.5 45.2 R-Poly Nets 94.5 97.6 94.9 61.5 D-Poly Nets 94.7 97.5 94.1 61.8 MONet-T 94.8 97.6 95.0 61.5 contrary, the proposed model can recover the equations behind the dynamic system and explicitly restore the symbolic representation. Published as a conference paper at ICLR 2024 Figure 2: The training loss change with epochs trained(Left). The ground truth and model predicted trajectory. (Right) Our model achieves low loss in 20 epochs and successfully predicts real trajectory. Table 4: Robustness on Image Net-C (Hendrycks & Dietterich, 2019). The corruption error (and mean corruption error (m CE)) is used as the robustness metric, where a lower score indicates a better performance. The best performance per category is highlighted in bold. Notice that the proposed method is robust to a range of corruptions. In fact, in many categories, such as Weather and Digital , the proposed model outperforms the compared models in all designated corruptions. Noise Blur Weather Digital Network m CE( ) Gauss Shot Impulse Defocus Glass Motion Zoom Snow Frost Fog Bright Contrast Elastic Pixel JPEG Res Net50 76.7 79.8 81.6 82.6 74.7 88.6 78 79.9 77.8 74.8 66.1 56.6 71.4 84.7 76.9 76.8 Dei T 54.6 46.3 47.7 46.4 61.6 71.9 57.9 71.9 49.9 46.2 46 44.9 42.3 66.6 59.1 60.4 Swin 62.0 52.2 53.7 53.6 67.9 78.6 64.1 75.3 55.8 52.8 51.3 48.1 45.1 75.7 76.3 79.1 MLP-Mixer 78.8 80.9 82.6 84.2 86.9 92.1 79.1 93.6 78.3 67.4 64.6 59.5 57.1 90.5 72.7 92.2 Res MLP 66.0 57.6 58.2 57.8 72.6 83.2 67.9 76.5 61.4 57.8 63.8 53.9 52.1 78.3 72.9 75.3 g MLP 64 52.1 53.2 52.5 73.1 77.6 64.6 79.9 77.7 78.8 54.3 55.3 43.6 70.6 58.6 67.5 Cycle MLP 53.7 42.1 43.4 43.2 61.5 76.7 56.0 66.4 51.5 47.2 50.8 41.2 39.5 72.3 57.5 56.1 Hire MLP 51.9 52.4 55.3 55.3 60.3 71.6 59.6 57.8 54.3 52.5 43.5 29.1 42.2 54.9 49.7 40.2 R-Poly Nets 73.8 84.0 84.4 88.0 81.9 83.6 75.0 77.9 71.8 72.2 69.4 45.7 67.4 65.8 74.8 65.4 MONet-S 49.7 51.3 52.3 53.0 57.8 72.1 55.0 60.6 50.4 46.0 42.0 27.6 38.1 53.7 47.7 39.0 We provide an experiment of a polynomial neural ODE approximating the Lotka-Volterra ODE model. This model captures the dynamics of predator-prey population within a biological system. The equations representing the Lotka-Volterra formula are in the form of dx dt = αx βxy, dy dt = δy +γxy . Given α = 1.56, β = 1.12, δ = 3.10, γ = 1.21, the ground truth equation are presented below: dx dt = 1.56x 1.12xy , dy dt = 3.10y + 1.21xy . (2) We generate N = 100 discrete points between time 0 and 10 for training. The training loss with epochs and predicted trajectory are shown in Fig. 2. Notice that our model can directly recover the right hand side of the Lotka-Volterra formula. Indeed, the learned model in our case recovers the following formula: dt = 1.56x 1.12001xy , dy dt = 3.10y + 1.21001xy . (3) Examining the results, it is evident that our Poly Neural ODE achieves high accuracy in recovering the ground truth formula and demonstrates rapid convergence. A more comprehensive comparison with Neural ODE and Augmented Neural ODE is conducted in Appendix G. 4.4 ROBUSTNESS We further conduct experiments on Image Net-C (Hendrycks & Dietterich, 2019) to analyze the robustness of our model. Image Net-C contains 75 types of corruptions, which makes it a good testbed for our evaluation. We follow the setting of Cycle MLP (Chen et al., 2022) and compare MONet with existing models using the corruption error as the metric (lower value indicates a better performance). Table 4 illustrates that our models exhibits the strongest ability among recent models. Notice that in important categories, such as the weather or the digital category, MONet outperforms the Published as a conference paper at ICLR 2024 Table 5: The influence of the different module in Block. Based on depth 16, hidden size 384 model on Image Net100. We incorporate spatial shift since this operation is multilinear and brings a performance gain. However, the primary performance improvement comes from the Mu-Layer we proposed as shown in the following table. Notice all models mentioned below are without activation functions. Block Type Layer 1 Layer 2 Paras. (M) Top-1 Acc(%) Poly-Block Mu-Layer+Spatial-Shift Mu-Layer 1.70 82.94% Poly-Block* Mu-Layer Mu-Layer 1.70 81.50% Mix Block Mu-Layer+Spatial-Shift MLP Layer 1.26 67.61% Linear Block MLP Layer+Spatial-Shift MLP Layer 1.21 55.11% Table 6: The influence of the patch embedding approach on a model with depth 16, hidden size 384 trained on Image Net100. Notice that the pyramid patch embedding can improve the performance with a minor difference in the parameters and maintain small FLOPs when using small patch. Method Patch size Top-1 Acc(%) Paras (M) FLOPs Patch Embedding 7 83.08 27.77 28.14 Patch Embedding 14 79.54 27.94 7.07 Pyramid Patch Embedding 7 82.04 28.02 7.23 Pyramid Patch Embedding 14,7 82.94 28.54 7.28 compared methods in all types of corruption. Those categories can indeed be corruptions met in realistic scenaria, making us more confident on the robustness of the proposed model in such cases. 4.5 ABLATION STUDY We conduct an ablation study using Image Net100, a subset of Image Net1K with 100 classes selected from the original ones. Previous studies have reported that Image Net100 serves as a representative subset of Image Net1K (Yan et al., 2021; Yu et al., 2022a; Douillard et al., 2022). Therefore, Image Net100 enables us to make efficient use of computational resources while still obtaining representative results for the self-evaluation of the model. In every experiment below, we assess one module or hyperparameter, while we assume that the rest remain the same as in the final model. Module ablation: In this study, we investigate the influence of various modules on the final network performance by removing and reassembling modules within the block. Please note that the following blocks do not contain any activation functions. In Table 5, we report the result using a hidden size of 384 and depth 16 on the Image Net100. We observe that spatial shift contributes to the performance improvement in our model but it is not crucial. The network still achieves high accuracy even in the absence of this module. At the same time, the removal of the Mu-Layer from the model leads to a significant drop in performance. The results validate that the spatial shift module alone cannot replace the proposed multilinear operations. Patch Embedding: In this study, we validate the effectiveness of our proposed pyramid patch embedding. We replace the input embedding layer of a model with depth 16 and hidden size 384 on the Image Net100 dataset. As depicted in Table 6, the model utilizing multi-level embedding achieves an Multi-stage performance with a Top-1 accuracy of 82.94%. The model utilizing a single-level embedding achieves a Top-1 accuracy of 81.78%, which is slightly lower than the multi-level embedding approach. For normal patch embedding, the computational and parameter overhead increases quadratically with a decrease in patch size. The result highlights that our approach significantly improves performance while adding only a small number of parameters and computational overhead. We conduct the same study for MLP-Mixer in Appendix D with similar outcomes, indicating the efficacy of our core Poly-Block independently of the patch embedding module. Hidden size: We evaluate the impact of different hidden size m. We fix the depth of the network to 16 and vary the hidden size. We observe a sharp decrease when MLP-Mixer uses a small hidden size under 128. The results in Table 7 validate that our proposed method is more robust in hidden size change compared to the normal MLP layer. Published as a conference paper at ICLR 2024 Table 7: The influence of the hidden size, m.The models result are based on depth 16 model with different hidden size, the number in the second and their row indicates Top-1 Accuracy(%) on Image Net100. The MLP-Mixer shows a sharp decline when hidden size change from 128 to 96. m 96 128 192 256 384 512 MLP-Mixer 48.3% 70.05% 72.39% 76.88% 78.24% 77.93% MONet 67.5% 75.82% 79.3% 81.94% 82.94% 82.24% Table 8: The influence of the depth, N. The model is based on depth 16, hidden size 384 and is trained on Image Net100. N Top-1(%) Top-5(%) Paras (M) 1 44.25 71.18 2.96 3 70.58 78.30 6.37 6 78.30 93.20 11.49 12 80.03 94.12 21.72 16 82.94 95.04 28.54 Table 9: The influence of the shrinkage ratio, r. The model is based on depth 16, hidden size m = 384, trained on Image Net100. r Top-1(%) Top-5(%) Paras (M) 1 83.18 94.86 52.98 2 83.88 95.54 36.46 4 82.94 95.04 28.54 8 82.28 95.12 24.06 16 82.34 94.62 21.99 Depth: The backbone of our proposed model, MONet, consists of N blocks. We evaluate the influence of the number of blocks (depth N) on Top-1 accuracy and parameter numbers, using a model with a fixed hidden size and shrinkage ratio. We observe that for our network, the depth holds a more significance role than the hidden size. Our experiments result in Table 8 validate our theoretical conjecture, as increasing depth yields a more pronounced improvement in performance. Shrinkage ratio: As mentioned in Section 3.1 the shrinkage ratio is defined as r = m l . The results of different shrinkage ratios based on a hidden size of m = 384 and a depth of 16 in our model are presented in Table 9. We observe that: i) A larger shrinkage ratio effectively reduces the number of parameters while having a relatively minor impact on performance; ii) a lower effective rank in the weights is sufficient for performance gain. 5 CONCLUSION In this work, we introduce a model, called MONet that expresses the output as a polynomial expansion of the input elements. MONet leverages solely linear and multilinear operations, which avoids the requirement for activation functions. At the core of our model lies the Mu-Layer, which captures multiplicative interactions inside the token elements (i.e. input). Through a comprehensive evaluation, we demonstrate that MONet surpasses recent polynomial networks, showcasing performance levels outperforms modern transformers models across a range of challenging benchmarks in image recognition. We anticipate that our work will further encourage the community to reconsider the role of activation functions and actively explore alternative classes of functions that do not require them. Lastly, we encourage the community to extend our illustration of the polynomial ODE solver in order to tackle scientific applications with PNs. Limitation: A theoretical characterization of the polynomial expansions that can be expressed with MONet remains elusive. In our future work, we will conduct further theoretical analysis of our model. We believe that such an analysis would further shed light on the inductive bias of the block and its potential outside of image recognition. Published as a conference paper at ICLR 2024 REPRODUCIBILITY STATEMENT Throughout this study, we exclusively utilize publicly accessible benchmarks, guaranteeing that other researchers can replicate our experiments. Additionally, we provide comprehensive information about the hyperparameters employed in our study and strive to offer in-depth explanations of all the techniques employed. Our plan is to make the source code of our model open source once our work gets accepted. ACKNOWLEDGMENTS We are thankful to Dr. Giorgos Bouritsas and the ICLR reviewers for their feedback and constructive comments. We are also thankful to Colby Fronk for his help in the Neural ODE symbolic representation restoration. We thank Zulip7 for their project organization tool. This work was supported by the Hasler Foundation Program: Hasler Responsible AI (project number 21043). Research was sponsored by the Army Research Office and was accomplished under Grant Number W911NF-24-1-0048. This work was supported by the Swiss National Science Foundation (SNSF) under grant number 200021 205011. Jimmy Lei Ba, Jamie Ryan Kiros, and Geoffrey E Hinton. Layer normalization. ar Xiv preprint ar Xiv:1607.06450, 2016. Francesca Babiloni, Ioannis Marras, Filippos Kokkinos, Jiankang Deng, Grigorios G Chrysos, and Stefanos Zafeiriou. Poly-nl: Linear complexity non-local layers with polynomials. In International Conference on Computer Vision (ICCV), 2021. Irwan Bello, Barret Zoph, Ashish Vaswani, Jonathon Shlens, and Quoc V Le. Attention augmented convolutional networks. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3286 3295, 2019. Zvika Brakerski, Craig Gentry, and Vinod Vaikuntanathan. (leveled) fully homomorphic encryption without bootstrapping. ACM Transactions on Computation Theory (TOCT), 6(3):1 36, 2014. Rich Caruana, Yin Lou, Johannes Gehrke, Paul Koch, Marc Sturm, and Noemie Elhadad. Intelligible models for healthcare: Predicting pneumonia risk and hospital 30-day readmission. In Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, pp. 1721 1730, 2015. Ricky TQ Chen, Yulia Rubanova, Jesse Bettencourt, and David K Duvenaud. Neural ordinary differential equations. Advances in neural information processing systems (Neur IPS), 31, 2018a. Shoufa Chen, Enze Xie, Chongjian Ge, Runjian Chen, Ding Liang, and Ping Luo. Cyclemlp: A mlp-like architecture for dense prediction. International Conference on Learning Representations (ICLR), 2022. Yunpeng Chen, Yannis Kalantidis, Jianshu Li, Shuicheng Yan, and Jiashi Feng. Aˆ 2-nets: Double attention networks. In Advances in neural information processing systems (Neur IPS), volume 31, 2018b. Grigorios G Chrysos, Stylianos Moschoglou, Giorgos Bouritsas, Yannis Panagakis, Jiankang Deng, and Stefanos Zafeiriou. P-nets: Deep polynomial neural networks. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7325 7335, 2020. Grigorios G Chrysos, Markos Georgopoulos, and Yannis Panagakis. Conditional generation using polynomial expansions. In Advances in neural information processing systems (Neur IPS), volume 34, pp. 28390 28404, 2021. 7https://zulip.com Published as a conference paper at ICLR 2024 Grigorios G Chrysos, Markos Georgopoulos, Jiankang Deng, Jean Kossaifi, Yannis Panagakis, and Anima Anandkumar. Augmenting deep classifiers with polynomial neural networks. In European Conference on Computer Vision (ECCV), pp. 692 716. Springer, 2022. Grigorios G Chrysos, Bohan Wang, Jiankang Deng, and Volkan Cevher. Regularization of polynomial networks for image recognition. In Conference on Computer Vision and Pattern Recognition (CVPR), 2023. Ekin D Cubuk, Barret Zoph, Dandelion Mane, Vijay Vasudevan, and Quoc V Le. Autoaugment: Learning augmentation policies from data. Conference on Computer Vision and Pattern Recognition (CVPR), 2019. Yann N Dauphin, Angela Fan, Michael Auli, and David Grangier. Language modeling with gated convolutional networks. In International Conference on Machine Learning (ICML), pp. 933 941, 2017. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. In International Conference on Learning Representations (ICLR), 2021. Arthur Douillard, Alexandre Ram e, Guillaume Couairon, and Matthieu Cord. Dytox: Transformers for continual learning with dynamic token expansion. In Conference on Computer Vision and Pattern Recognition (CVPR), 2022. Abhimanyu Dubey, Filip Radenovic, and Dhruv Mahajan. Scalable interpretability via polynomials. Advances in neural information processing systems (Neur IPS), 2022. Emilien Dupont, Arnaud Doucet, and Yee Whye Teh. Augmented neural odes. Advances in neural information processing systems (Neur IPS), 32, 2019. Colby Fronk and Linda Petzold. Interpretable polynomial neural ordinary differential equations. Chaos: An Interdisciplinary Journal of Nonlinear Science, 33(4), 2023. Markos Georgopoulos, Grigorios G Chrysos, Maja Pantic, and Yannis Panagakis. Multilinear latent conditioning for generating unseen attribute combinations. In International Conference on Machine Learning (ICML), pp. 3442 3451, 2020. Markos Georgopoulos, James Oldfield, Mihalis A Nicolaou, Yannis Panagakis, and Maja Pantic. Mitigating demographic bias in facial datasets with style-based multi-attribute transfer. International Journal of Computer Vision (IJCV), 129(7):2288 2307, 2021. Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward neural networks. In International Conference on Artificial Intelligence and Statistics (AISTATS), pp. 249 256, 2010. Jianyuan Guo, Yehui Tang, Kai Han, Xinghao Chen, Han Wu, Chao Xu, Chang Xu, and Yunhe Wang. Hire-mlp: Vision mlp via hierarchical rearrangement. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 826 836, 2022. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In International Conference on Computer Vision (ICCV), pp. 1026 1034, 2015. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770 778, 2016. Dan Hendrycks and Thomas G Dietterich. Benchmarking neural network robustness to common corruptions and surface variations. In International Conference on Learning Representations (ICLR), 2019. Published as a conference paper at ICLR 2024 Qibin Hou, Zihang Jiang, Li Yuan, Ming-Ming Cheng, Shuicheng Yan, and Jiashi Feng. Vision permutator: A permutable mlp-like architecture for visual recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(1):1328 1334, 2022. Jie Hu, Li Shen, and Gang Sun. Squeeze-and-excitation networks. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7132 7141, 2018. Alexey Grigorevich Ivakhnenko. Polynomial theory of complex systems. IEEE transactions on Systems, Man, and Cybernetics, 1(4):364 378, 1971. Tamara G Kolda and Brett W Bader. Tensor decompositions and applications. SIAM review, 51(3): 455 500, 2009. Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. University of Toronto, 2009. Ya Le and Xuan Yang. Tiny imagenet visual recognition challenge. CS 231N, 7(7):3, 2015. Yann Le Cun, L eon Bottou, Genevieve B Orr, and Klaus-Robert M uller. Efficient backprop. In Neural networks: Tricks of the trade, pp. 9 50. Springer, 2002. Chien-Kuo Li. A sigma-pi-sigma neural network (spsnn). Neural Processing Letters, 17:1 19, 2003. Xiang Li, Wenhai Wang, Xiaolin Hu, and Jian Yang. Selective kernel networks. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 510 519, 2019. Tsung-Yi Lin, Piotr Doll ar, Ross Girshick, Kaiming He, Bharath Hariharan, and Serge Belongie. Feature pyramid networks for object detection. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2117 2125, 2017. Hanxiao Liu, Zihang Dai, David So, and Quoc V Le. Pay attention to mlps. Advances in neural information processing systems (Neur IPS), 34:9204 9215, 2021. Jonathan Long, Evan Shelhamer, and Trevor Darrell. Fully convolutional networks for semantic segmentation. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3431 3440, 2015. Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. International Conference on Learning Representations (ICLR), 2019. Dengsheng Lu and Qihao Weng. A survey of image classification methods and techniques for improving classification performance. International journal of Remote sensing, 28(5):823 870, 2007. James Martens et al. Deep learning via hessian-free optimization. In International Conference on Machine Learning (ICML), volume 27, pp. 735 742, 2010. Luke Melas-Kyriazi. Do you even need attention? a stack of feed-forward layers does surprisingly well on imagenet. ar Xiv preprint ar Xiv:2105.02723, 2021. Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Bo Wu, and Andrew Y Ng. Reading digits in natural images with unsupervised feature learning. Advances in neural information processing systems (Neur IPS), 2011. Maria-Elena Nilsback and Andrew Zisserman. Automated flower classification over a large number of classes. In 2008 Sixth Indian Conference on Computer Vision, Graphics and Image Processing, pp. 722 729, 2008. doi: 10.1109/ICVGIP.2008.47. Luzhou Peng, Bowen Qiang, and Jiacheng Wu. A survey: Image classification models based on convolutional neural networks. In 2022 14th International Conference on Computer Research and Development (ICCRD), pp. 291 298. IEEE, 2022. Jo Plested and Tom Gedeon. Deep transfer learning for image classification: a survey, 2022. Published as a conference paper at ICLR 2024 Olaf Ronneberger, Philipp Fischer, and Thomas Brox. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18, pp. 234 241. Springer, 2015. Yoan Shin and Joydeep Ghosh. The pi-sigma network: An efficient higher-order neural network for pattern classification and function approximation. In IJCNN-91-Seattle international joint conference on neural networks, volume 1, pp. 13 18. IEEE, 1991. Karen Simonyan and Andrew Zisserman. Very deep convolutional networks for large-scale image recognition. International Conference on Learning Representations (ICLR), 2015. Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinking the inception architecture for computer vision. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2818 2826, 2016. Ilya O Tolstikhin, Neil Houlsby, Alexander Kolesnikov, Lucas Beyer, Xiaohua Zhai, Thomas Unterthiner, Jessica Yung, Andreas Steiner, Daniel Keysers, Jakob Uszkoreit, et al. Mlp-mixer: An all-mlp architecture for vision. Advances in neural information processing systems (Neur IPS), 34: 24261 24272, 2021. Hugo Touvron, Matthieu Cord, Matthijs Douze, Francisco Massa, Alexandre Sablayrolles, and Herv e J egou. Training data-efficient image transformers & distillation through attention. In International Conference on Machine Learning (ICML), pp. 10347 10357, 2021. Hugo Touvron, Piotr Bojanowski, Mathilde Caron, Matthieu Cord, Alaaeldin El-Nouby, Edouard Grave, Gautier Izacard, Armand Joulin, Gabriel Synnaeve, Jakob Verbeek, et al. Resmlp: Feedforward networks for image classification with data-efficient training. IEEE Transactions on Pattern Analysis and Machine Intelligence (T-PAMI), 2022. Asher Trockman and J Zico Kolter. Patches are all you need? Transactions on Machine Learning Research, 2023. ISSN 2835-8856. Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. In Advances in neural information processing systems (Neur IPS), volume 30, 2017. Wenhai Wang, Enze Xie, Xiang Li, Deng-Ping Fan, Kaitao Song, Ding Liang, Tong Lu, Ping Luo, and Ling Shao. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 568 578, 2021. Xiaolong Wang, Ross Girshick, Abhinav Gupta, and Kaiming He. Non-local neural networks. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 7794 7803, 2018. Enze Xie, Wenhai Wang, Zhiding Yu, Anima Anandkumar, Jose M Alvarez, and Ping Luo. Segformer: Simple and efficient design for semantic segmentation with transformers. Advances in Neural Information Processing Systems, 34:12077 12090, 2021. Yixing Xu, Xinghao Chen, and Yunhe Wang. Bimlp: Compact binary architectures for vision multi-layer perceptrons. Advances in neural information processing systems (Neur IPS), 2022. Shipeng Yan, Jiangwei Xie, and Xuming He. Der: Dynamically expandable representation for class incremental learning. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3014 3023, 2021. Guandao Yang, Sagie Benaim, Varun Jampani, Kyle Genova, Jonathan Barron, Thomas Funkhouser, Bharath Hariharan, and Serge Belongie. Polynomial neural fields for subband decomposition and manipulation. Advances in neural information processing systems (Neur IPS), 35:4401 4415, 2022. Jiancheng Yang, Rui Shi, and Bingbing Ni. Medmnist classification decathlon: A lightweight automl benchmark for medical image analysis. In IEEE 18th International Symposium on Biomedical Imaging (ISBI), pp. 191 195, 2021. Published as a conference paper at ICLR 2024 Minghao Yin, Zhuliang Yao, Yue Cao, Xiu Li, Zheng Zhang, Stephen Lin, and Han Hu. Disentangled non-local neural networks. In European Conference on Computer Vision (ECCV), pp. 191 207. Springer, 2020. Tan Yu, Xu Li, Yunfeng Cai, Mingming Sun, and Ping Li. S2-mlpv2: Improved spatial-shift mlp architecture for vision. In Winter Conference on Applications of Computer Vision (WACV), 2022a. Tan Yu, Xu Li, Yunfeng Cai, Mingming Sun, and Ping Li. S2-mlp: Spatial-shift mlp architecture for vision. In Conference on Computer Vision and Pattern Recognition (CVPR), pp. 297 306, 2022b. Sangdoo Yun, Dongyoon Han, Seong Joon Oh, Sanghyuk Chun, Junsuk Choe, and Youngjoon Yoo. Cutmix: Regularization strategy to train strong classifiers with localizable features. Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6023 6032, 2019. Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and David Lopez-Paz. mixup: Beyond empirical risk minimization. International Conference on Learning Representations (ICLR), 2018. Published as a conference paper at ICLR 2024 CONTENTS OF THE APPENDIX The contents of the supplementary material are organized as follows: In Appendix A, we provide a technical proofs regarding the interactions learned by Mu-Layer and Poly-Block. In Appendix B, we exhibit the details of the architecture design for our model, including the design of single-stage and multi-stage architecture. In Appendix C, we present the schematic of the Mu-Layer to provide a more intuitive understanding of its structure. Additionally, we provide detailed explanations of the Mu Layer and Poly-Block components. In Appendix D, we elaborate on the pyramid patch embedding method mentioned in Section 3.2. We provide illustrative comparisons between the conventional patch embedding approach and our method, showcasing the differences in the representation of input data between the two approaches. In Appendix E, we extend our method beyond natural images to the medical domain. In Appendix F, we visualize the effective receptive field of our Image Net-1K pre-trained model. We also visualize related models with same input image as a comparison. In Appendix G, we follow the original evaluation experiment to further compare our method with existing Neural ODE Solver (Dupont et al., 2019; Chen et al., 2018a). In Appendix H, we scrutinize our model further beyond image recognition. We conduct a semantic segmentation evaluation on ADE20K dataset. The result shows our proposed model also greatly surpasses previous PN models in semantic segmentation. In Appendix I, we measure computation cost using FLOPs. We notice that the tool impacts the number of FLOPs and indeed we observe that this resulted in less accurate measurements of previous PN models. We utilize open-source code to reevaluate the computational costs of existing PN models. In this section, we present precise FLOPs (Floating Point Operations) values for a range of PNs, along with their corresponding performance on Image Net1K. In Appendix J, we conduct complexity analysis of our proposed model. In Appendix K, we list the training setting and hyperparameter of our model used for Image Net-1K training. We also conduct an error analysis using saliency maps in Section Appendix K.2 to investigate the weaknesses of our model further. In Appendix L, we list the training settings and hyperparameters used for the medical image classification experiment and the fine-grained classification experiment. Additionally, we include the model settings table. In Appendix M, we perform experiments to examine the effects of various parameter initialization methods on the final performance of our model. Published as a conference paper at ICLR 2024 In this section, we derive the proof of Prop. 1 and also include a new proposition for the Poly-Block, along with the associated proof. A.1 PROOF OF PROP. 1 In this paper, we use a bold capital letter to represent a matrix, e.g., X, F , while a plain letter to represent a scalar. For instance, Xj,ρ represents the element in the jth row and the ρth column. The matrix multiplication of two matrices F , K results in the following j, ρ element: [F K]j,ρ = P q Fj,q Kq,ρ. As a reminder, the input is X Rd s, where d N is the length of a token and s N is the number of tokens. Thus, to prove the proposition, we need to show that products of the form Xτ,ρXψ,ρ exist between elements of each token. Eq. (1) relies on matrix multiplications and a Hadamard (elementwise) product. Concretely, if we express the j, ρ element of (AX) (BDX), we obtain: [(AX) (BDX)]j,ρ = ω Aj,τXτ,ρBj,ωDω,ψXψ,ρ . (4) Then, we can add the additive term and the matrix C. If we express Eq. (1) elementwise, we obtain the following expression: ω Aj,τXτ,ρBj,ωDω,ψXψ,ρ + τ=1 Aj,τXτ,ρ That last expression indeed contains sums of products Xτ,ρXψ,ρ for every output element, which concludes our proof. A.2 INTERACTIONS OF THE POLY-BLOCK As we mention in the main paper, a Poly-Block comprises of two Mu-Layers, so one reasonable question is how the multiplicative interactions of the Mu-Layer extend in the Poly-Block. We prove below that up to fourth degree interactions are captured in such a block. To simplify the derivation, we focus on two consecutive Mu-Layers as the Poly-Block. Proposition 2. The Poly-Block captures up to fourth degree interactions between elements of each token. Proof. All we need to show is that products of the form Xτ,ρXψ,ρXγ,ρXϵ,ρ appear. Using the insights from Appendix A.1, we expect the fourth degree interactions to appear in the Hadamard product, so we will focus on the term A(2)X(2) B(2)D(2)X(2) , where the (2) declares the weights and input of the second Mu-Layer. The elementwise expression for X(2) is directly obtained from Eq. (5). To simplify the expression, we ignore the additive term, since exhibiting the fourth degree interactions is enough. The q, w element of the expression is the following: h A(2)X(2) B(2)D(2)X(2) i α Cq,ξCδ,θA(2) w,q B(2) w,αD(2) α,δ {f1} {f2} , (6) where the expression {f1} {f2} is the following: τ=1,ψ=1,γ=1,ϵ=1 ω,ω Aξ,τXτ,ρBξ,ωDω,ψXψ,ρAθ,γXγ,ρBθ,ω Dω ,ϵXϵ,ρ . (7) Notice that the last expression indeed contains products of the form Xτ,ρXψ,ρXγ,ρXϵ,ρ, which concludes the proof. Published as a conference paper at ICLR 2024 B MULTI-STAGE MONET Figure 3: The Schematic of (simple) MONet and Multi-stage MONet. PPE represents our pyramid patch embedding. In deep architectures often a different number of channels or hidden size is used for different blocks (He et al., 2016). Our preliminary experiments indicate that MONet is amenable to different hidden size as well. Following recent works (Chen et al., 2022; Guo et al., 2022), we set 4 different hidden sizes across the network (referred to as stages) and we refer to this variant as the Multi-stage MONet . This variant is mostly used for large-scale experiments, since for datasets such as CIFAR10, and CIFAR100 a single hidden size is sufficient. C DETAILS OF MU-LAYER AND POLY-BLOCK Figure 4: The Schematic of Mu-Layer. Blue boxes correspond to learnable parameters. Green and red boxes denote input and output, respectively. The denotes the Hadamard product, the + denotes element-wise addition. The gray box denotes the spatial aggregation module, the dotted line represents it as an optional module. In our design the first Mu-Layer of each Poly-Block includes a spatial aggregation unit, while the second Mu-Layer does not. In Fig. 4, we present the schematic of the Mu-Layer. The matrices A, B, C, and D in Fig. 4 are described in Section 3.1. As mentioned in Section 3.1, the Poly-Block serves as a larger unit module in our proposed model. In our implementation, the Poly-Block consists of two Mu-Layer modules. Only the first Mu-Layer in each block incorporates a spatial aggregation module, where we utilize a spatial shift operation (Yu et al., 2022b). D PYRAMID EMBEDDING A Patch Embedding, which is typically used in the input space of the model, converts an image into a sequence of N non-overlapping patches and projects to dimension d. If the original image has Published as a conference paper at ICLR 2024 Input Image 7*7Conv,stride=7 2*2Conv,stride=2 14*14Conv,stride=14 7*7Conv,stride=7 Pyramid Patch Embedding Patch Embedding Input Image Figure 5: Pyramid Patch Embedding Table 10: Influence of the patch embedding on MONet. Method Patch size Top-1 Acc(%) Paras (M) FLOPs Patch Embedding 7 83.08 27.77 28.14 Patch Embedding 14 79.54 27.94 7.07 Pyramid Patch Embedding 7 82.04 28.02 7.23 Pyramid Patch Embedding 14,7 82.94 28.54 7.28 Table 11: The influence of the patch embedding approach on MLP-Mixer. Method Patch size Top-1 Acc(%) Paras (M) FLOPs Patch Embedding 7 77.92 26.60 19.44 Patch Embedding 14 77.93 23.68 4.91 Pyramid Patch Embedding 7 79.64 24.50 5.18 Pyramid Patch Embedding 14,7 79.96 24.80 5.26 (H, W) resolution, each patch Xi has resolution (P, P). Then, we obtain S = H W/P 2 patches, where Xi RP P d. For models based on MLP structure, smaller patches can capture finer-grained features better. However, due to the quadratic growth in computation caused by reducing the patch size, most models (Guo et al., 2022; Chen et al., 2022) do not follow this design. To reduce the computational burden by smaller patches, these networks assume multiple hidden sizes across the network, see Appendix B for further details. Instead, we utilize an additional method for reducing the computational cost. After performing the first patch embedding on the original image using a small patch size, we can further reduce the input size by performing 2 2 convolution with stride 2. With this approach, we can achieve performance comparable to directly using a small patch size while significantly reducing computational costs and allowing us to adopt a more elegant and simple structure. Inspired by FPN (Lin et al., 2017), we introduce a different level of patch embedding to further enhance the performance. Each level of the pyramid represents an image at a different scale, with the lower levels representing images with higher resolution and the higher levels representing images with lower resolution. We use lower resolution image patch merged with downsampled higher-resolution image patch by element-wise addition. In Section 4.5, we compare our method with normal patch embedding. The embedding method we propose not only reduces network computational complexity but also improves the performance when compared to larger patch sizes. We conduct 8 comparisons to prove the soundness of our method. We employ different patch embedding and patch sizes for the MLP-Mixer and our model. We train those models on Image Net100 dataset and report Top-1 Accuracy, their parameters and computation costs (FLOPs) in Table 10 and Table 11. Appendix D indicates that our pyramid patch embedding improves the performance compared to normal patch embedding. By leveraging our proposed method, we can utilize small patch size to achieve better results while preventing quadratic growth of computation costs. This could be used as an ad-hoc module in other MLP models. At the same time, even with a normal patch embedding, our model still outperforms MLP-Mixer. In the context of merging different levels of features, we initially adopt an elementwise addition approach. Additionally, we explore a U-Net-style approach (Ronneberger et al., 2015). In this approach, we concatenate the features along the channel dimension and then reduced the dimension using a 1x1 depthwise convolution. However, after conducting experiments, we observe that this approach had limited improvement and was considerably less effective compared to the elementwise addition method. Published as a conference paper at ICLR 2024 Table 12: MONet performance compared to Res Net18 on the Med MNIST benchmark, with detailed numbers. The best results are marked in bold. Dataset MONet-T PDC Π-Nets MLP-Mixer Res MLP S2MLP-D-S S2MLP-W-S Res Net Auto Sklearn Path 90.8 92.1 90.8 89.2 89.6 90.1 90.7 90.7 83.4 Derma 77.5 77.5 72.9 76.1 76.5 76.2 76.9 73.5 74.9 Oct 80.1 77.2 79.8 78.2 79.7 79.1 79.5 74.3 60.1 Pneumonia 93.4 92.5 89.1 93.1 89.1 91.9 93.1 85.4 87.8 Retina 55.7 53.6 52.2 54.5 54.7 53.7 54.0 52.4 51.5 Blood 96.7 95.5 94.5 94.7 95.3 95.8 95.2 93.0 96.1 Tissue 67.7 67.5 - 65.9 66.8 65.9 66.0 67.7 53.2 Organ A 93.6 93.5 92.5 90.5 91.4 92.3 92.7 93.5 76.2 Organ C 88.9 93.0 89.3 90.6 87.5 88.9 89.4 90.0 87.9 Organ S 78.5 77.9 75.0 76.9 75.7 77.5 78.4 78.2 67.2 E MEDICAL IMAGE CLASSIFICATION To assess the efficacy of our model beyond natural images, we conduct an experiment on the Med MNIST challenge across ten datasets (Yang et al., 2021). The dataset encompasses diverse medical domains, such as chest X-rays, brain MRI scans, retinal fundus images, and more. Med MNIST serves as a benchmark dataset for evaluating models in the field of medical image analysis, enabling us to evaluate the performance of our model across various medical imaging domains. In our experiments, we train the variant of MONet-T with 14.0M parameters and 0.68G FLOPs. The results are shown in Table 12 exhibit that our model outperforms other models in eight datasets. F VISUALIZATION OF REPRESENTATIONS FROM LEARNED MODELS In this section, we use visualizations to understand the difference between the learned models of MLP-Mixer, Res MLP and CNN architectures. The effective receptive field refers to the region in the input data that a neural network s output is influenced by, which has been proved as an effective approach to understand where a learned model focuses on. Specifically, we visualize the hidden unit effective receptive field of our model trained on Image Net1k, i.e., the output of the last layer, and compare it with pretrained MLP-Mixer, Res MLP and Resnet in Fig. 6. We adopt a random image from Image Net1K as input and visualize the effective receptive field. We can notably observe that due to the flattening of input tokens into one-dimensional structures by MLP-Mixer and Res MLP, the features they learn exhibit a distinct grid effect due to the loss of 2D information. They also tend to emphasize low-level texture information to a greater extent. Res Net s effective receptive field is more discrete, encompassing both background and foreground elements to some extent, focusing on the global context. On the contrary, MONet focuses on the semantic parts (e.g. on the dog s face) which achieves a balance between global and local context. G NEURALODE SOLVER COMPARSION Following the evaluation setup of the Neural-ODE network, we implement a Poly-Neural ODE based on our method, which can be used to solve ordinary differential equations for simulating physics problems. We conduct scientific computing experiments using simulated data, with the aim of predicting and interpreting the dynamics of physical systems The specific task is to simulate moving the randomly generated inner sphere particles out of the annulus by pushing against and stretching the barrier. Eventually, since we are mapping discrete points and not a continuous region, the flow is able to break apart the annulus to let the flow through. The initial particle states are shown in Figure Fig. 7. We set the model with hidden dimension 32 and train on simulated data for 12 epoches. The revolution of the first 10 epochs are shown in Fig. 8. We implement two models based on Neural ODE (Chen et al., 2018a) and Augmented Nerual ODE (Dupont et al., 2019), called Poly Nerual ODE(Poly NODE) and Poly Augmented Neural ODE(Poly Published as a conference paper at ICLR 2024 Figure 6: Input image(Left). The ERF of model pretrained on Image Net-1K(Right). We visualize the effective receptive field of MLP-Mixer, Res MLP, Res Net and MONet. Our model more focus on the dog s face and the outline shape of dog, while the rest MLP models more focus on the low-level texture feature. Figure 7: The initial particles states. The task is to move the inner sphere of blue particles out of the red particles. Figure 8: The move of particles. From left to right it s feature evolution with iterations increase. Figure 9: The predicted trajectory of particles. The lines are their history trajectory. Published as a conference paper at ICLR 2024 ANODE). For Poly Neural ODE, higher-order model could achieve better accuracy with the cost of higher computation cost, the following experiment are based on minimal order 2 Poly Neural ODE. We first compare NODE with Poly NODE, the loss plots shown in Fig. 10. We can observe that with 40 epochs training, The Poly NODE approximates the functions faster and more accurately. We then Figure 10: Loss plots for NODEs and ANODEs trained on 2d data. Poly NODEs easily approximate the functions and are consistently faster than NODEs. Figure 11: Loss plots for Poly ANODEs and ANODEs trained on 2d data. Poly NODEs easily approximate the functions and are consistently faster than NODEs. compare ANODE with Poly ANODE. The loss plot trained 30 epoches is shown in Fig. 11. ANODEs augment the space on which the ODE is solved, allowing the model to use the additional dimensions to learn more complex functions using simpler flows. Our Poly ANODE inherent its advantage while converge faster. H SEMANTIC SEGMENTATION Settings. To further explore the performance of MONet in downstream task. We conduct semantic segmentation experiments on ADE20K dataset and present the final performance compared to previous models. The Appendix H shows the result. Previous PNN models exhibited poor performance on downstream tasks. Our model has overcome this issue and achieved results comparable to some state-of-the-art models. I COMPUTATION COST COMPARED TO PREVIOUS POLYNOMIAL MODELS FLOPs (Floating-Point Operations Per Second) is commonly used metric to measure the computational cost of a model. However, tools like flops-counters and mmcv, which use Py Torch s module forward hooks, have significant limitations. They are accurate only if every custom module includes a corresponding flop counter. In this paper, we adopt fvcore Flop Counter8 developed by Meta FAIR, which provides the first operator-level FLOP counters. This tool observes all operator calls and collects operator-level FLOP counts. For models like polynomial models that involve numerous custom operations, an operatorlevel counter will give more accurate results. We reproduced previous polynomial models according 8fvcore flops counter Published as a conference paper at ICLR 2024 Table 13: Semantic segmentation on ADE20K val,All models use Semantic FPN head. The transformer and Res Net models result from the original paper. The polynomial networks result from the reproduced experiment. Backbone Semantic FPN m Io U(%) Res Net18 (He et al., 2016) 32.9 FCN (Long et al., 2015) 29.3 Pv T-Tiny (Wang et al., 2021) 35.7 Seg-Former (Xie et al., 2021) 37.4 R-PDC (Chrysos et al., 2022) 20.7 R-Poly Nets(Chrysos et al., 2023) 19.4 Multi-stage MONet-S 37.5 to their open-source code and re-measure their computation cost with fvcore flops counter. This causes a slight difference from the FLOPs reported in previous PNs. For instance, operations such as custom normalization modules and Hadamard product computations were often overlooked due to former tool limitations. We adopt the number taken from original papers in Table 2 and we have corrected and updated the FLOPs report for recent polynomial models in the table below. In comparison to previous work, our model has significant performance benefits while only incurring 20% of the computational cost of the previous state-of-the-art work. Simultaneously, our performance on Image Net-1K surpasses the previous state-of-the-art by 10%. Table 14: The re-measure computational costs for polynomial models, along with their corresponding Top-1 accuracy on Image Net-1K Model FLOPs(B) Accuracy(%) PDC 30.78 71.0 R-PDC 39.56 - R-Poly Nets 29.34 70.2 D-Poly Nets 26.41 70.0 MONet-T (Ours) 3.6 76.0 MONet-S(Ours) 11.2 79.6 Multi-stage MONet-T(Ours) 2.8 77.0 Multi-stage MONet-S(Ours) 6.8 81.3 J COMPLEXITY ANALYSIS To simplify the process, we conduct a complexity analysis of MONet adopted the single-stage design as follows. Pyramid Patch Embedding Layer (PPEL) crops the input raw image I RH W 3 where W is the width and H is the height into several non-overlapping patches of dimension P Rp p 3. Consider 1-level embedding,Our method conducts compression to the patches with a 2 2 convolution layer, resulting in the token W0 Rp p c, token W1 R p 2 p 2 c,where W0 is the output of the first convolution layer, W1 is the output of the second convolution layer. Noted that the final number of patches are Np = W 2p which is 25% of normal patch embedding. Thus, the total number of parameters of the Pyramid embedding is: Params P P EL = (3p2 + 1)c + c(4c + 1) = (3p2 + 4c + 2)c (8) In summary the floating operations in PPEL is FLOPs P P EL = 3 W p cp2 + 4 c2 W 2p = 4c Np(3p2 + c) (9) Poly Blocks(PB): Our model consists of N identical Poly Blocks. Each blocks contains 2 Poly Layer. The first Poly Layers(PL1) contains two Spatial Shift module in two branches, and the Published as a conference paper at ICLR 2024 second Poly Layers doesn t contain Spatial Shift module. We denotes that first Poly Layer four fully connected layer as W1, b1, W2, b2, W3, b3, W4, b4 and shrinkage ratio as s, where W1 Rc c, W2 Rc c s , W3 R c s c, W4 Rc c. Therefore, the total parameter numbers of the first Poly Layers is Params P L1 = c (2c + 2c s) + 3c + c s = c2(2 + 2 s) + c(3 + 1 The second Poly Layer is similar, while we have a expansion rate r similar to other MLP models.We denotes that second Poly Layer four fully connected layer as W5, b5, W6, b6, W7, b7, W8, b8 and shrinkage ratio as s, where W5 Rc rc, W2 Rc rc s , W3 R rc s rc, W4 Rrc c. We get the total parameter numbers of the second Poly Layers is Params P L2 = c (2rc + rc s ) + 2rc + c + rc s = c2(2r + r + r2 s ) + c(2r + 1 s + 1) (11) Hence, the total number of a Poly Block is Params P B = Params P L1 + Params P L2 = c2(2r + 2 + r2 + r + 2 s ) + c(4 + 2r + 2 Besides the fully connected layer, the hadamard product bring extra computation where the flops of hadamard product of two matrices in shape n m is n m. The overall hadamard product flops is FLOPs P rod = Npc + Nprc = Npc(1 + r) (13) and the flops becomes FLOPs P B = Npc2(2r+2+r2 + r + 2 s )+FLOPs P rod = Npc2(2r+2+r2 + r + 2 s )+Npc(1+r) (14) Fully-connected classification layer(FCL): takes input c-dimensional vector and feedforward to average pooling layer. The output vectore is in k dimension where k is the number of classes. In summary, the number of parameters of FCL is: Params F CL = k(c + 1) (15) The flops of FCL is: FLOPs F CL = Npck (16) Overall architecture: We conclude that the summary of parameters of overall of architecture. The total numbers of parameters of our architecture is: Params = Params P P EL + N Params P B + Params F CL (17) The flops of our architecture is: FLOPs = FLOPs P P EL + N FLOPs P B + FLOPs F CL (18) Published as a conference paper at ICLR 2024 K IMAGENET1K CLASSIFICATION K.1 IMAGENET1K TRAINING SETTING The Appendix K.1 shows our experiment setting for training MONet on Image Net1K dataset. Our implementation for the optimizer and data augmentation follow the popular timm library9. We train 300 epochs in Image Net1K on 4 GPUs, we enable exponential moving average on 200th epochs. Table 15: Image Net1K Training Settings in Section 4.1 Image Net1K Training Setting optimizer Adam W base learning rate 1e-3 weight decay 0.01 (Multi-stage MONet-T) 0.02 (Multi-stage MONet-S) batch size 448 training epochs 300 learning rate schedule cosine warmup label smoothing 0.1 auto augmentation random erase 0.1 cutmix 0.5 mixup 0.5 K.2 ERROR ANALYSIS We utilize our best-performing MONet model to compute per-class accuracy rates for all 1000 classes on the validation dataset of Image Net1K. In Table 16, we present the top 10 least accurate and misclassified classes. Additionally, in Fig. 12 we showcase the images of the most misclassified class (laptop). In Fig. 13, we show a failure case of the screen. In Fig. 14, we show a failure case of tennis. In Fig. 15, we show a successful case of small object sunglasses. Table 16: The top-10 least accurate classes and their labels Class Name Accuracy (%) Class Name Accuracy (%) tiger Cat 20 laptop,laptop computer 24 screen 24 chiffionier,commode 28 sunglasses 28 cassette player 30 letter opener 30 malliot 32 projectile,misslle 32 spotlight 32 Our analysis dictates that the majority of classification failures occur when the model fails to focus on the main object. This is even more pronounced in cases where the classification label is provided by an item occupying a small portion of the image. We believe this is an interesting phenomenon that is shared with similar models relying on linear operations and not using convolutions. 9Hugging Face timm Published as a conference paper at ICLR 2024 Figure 12: Failure case of MONet as explained from the saliency map. The red figure in the middle denotes the saliency map, which is the area where the model focuses on. The correct class for the image is laptop, while the saliency map shows the model focuses on the person. This is not an unreasonable error though, since the main figure lies in the center of the image. Figure 13: Failure case of MONet as explained from the saliency map. The red figure in the middle denotes the saliency map, which is the area where the model focuses on. The correct class for the image is screen, while the saliency map shows the model focusing on the wall. Figure 14: Failure case of MONet as explained from the saliency map. The red figure in the middle denotes the saliency map, which is the area where the model focuses on. The correct class for the image is tennis, while the saliency map shows the model focusing on the background wall, instead of on the small object. Figure 15: Success case of MONet as explained from the saliency map. The correct class for the image is sunglasses, The saliency map shows the model focusing on the face of the woman. In some cases, successful classification of a category can be achieved based on global information, even if the object occupies only a small portion of the image. Published as a conference paper at ICLR 2024 K.3 SOURCES FOR BASELINES To further increase the transparency of our comparisons, we clarify the details on the main results on Image Net1K. We aim for reporting the accuracy and the other metrics using the respective papers or public implementations. Concretely: The paper of Chrysos et al. (2022) is used as the source for PDC. Chrysos et al. (2023) is used as the source for Res Net18,Pi-Nets, Hybrid Pi-Nets, RPoly Nets and D-Poly Nets. We clarify here once more that Π-Nets refers to the model without activation functions, while Hybrid Π-Nets refers to the model with activation functions. Yu et al. (2022b) is used as the source of Res Net50 performance. We updated the number to 77.2% (on Image Net1K) to align with He et al. (2016). The performance data for the other models in the Table 2 are sourced from their respective papers. L ADDITIONAL BENCHMARKS AND MEDICAL IMAGE CLASSIFICATION We introduce dataset details including dataset information for various datasets used for the empirical validation of our method. In this section, we give a brief introduction to the dataset we use and the overview statistics of datasets are shown in Appendix L. CIFAR10: CIFAR-10 (Krizhevsky et al., 2009) is a well-known dataset in the field of computer vision. It consists of 60,000 labeled images that are divided into ten different classes. Each image in the CIFAR-10 dataset has a resolution of 32x32 pixels and is categorized into one of the following classes: airplane, automobile, bird, cat, deer, dog, frog, horse, ship, or truck. The images in the CIFAR10 dataset are of size 32 32 pixels. We keep its original resolution for training. SVHN:SVHN (Netzer et al., 2011) stands for Street View House Numbers. It is a widely used dataset for object recognition tasks in computer vision. The SVHN dataset consists of images of house numbers captured by Google s Street View vehicles. These images contain digit sequences that represent the house numbers displayed on buildings. It contains over 600,000 labeled images, which are split into three subsets: a training set with approximately 73,257 images, a validation set with around 26,032 images, and a test set containing approximately 130,884 images. The images in the SVHN dataset are of size 32 32 pixels, we keep its original resolution for training. Oxford Flower: The Oxford Flower (Nilsback & Zisserman, 2008) dataset, also known as the Oxford 102 Flower dataset, is a widely used collection of images for fine-grained image classification tasks in computer vision. It consists of 102 different categories of flowers, with each category containing a varying number of images. The Oxford Flower dataset provides a challenging testbed for researchers and practitioners due to the high intra-class variability among different flower species. This variability arises from variations in petal colors, shapes, and overall appearances across different species. The images in the Oxford Flower dataset are of size 256 256 pixels, in our training, we use bicubic interpolation to resize all images to 224 224 pixels. Tiny-Imagenet: Tiny Image Net(Le & Yang, 2015) is a dataset derived from the larger Image Net dataset, which is a popular benchmark for object recognition tasks in computer vision. The Tiny Image Net dataset is a downsized version of Image Net, specifically designed for research purposes and computational constraints. While the original Image Net dataset consists of millions of images spanning thousands of categories, the Tiny Image Net dataset contains 200 different classes, each having 500 training images and 50 validation and test images. This results in a total of 100,000 labeled images in the dataset. The images in the Tiny Image Net dataset are of size 64 64 pixels, we keep its original resolution for training. Med MNIST: Med MNIST (Yang et al., 2021) is a specialized dataset designed for medical image analysis and machine learning tasks. It is inspired by the popular MNIST dataset, but instead of handwritten digits, Med MNIST focuses on medical imaging data. The Med MNIST dataset includes several sub-datasets, each corresponding to a different medical imaging modality or task. Some Published as a conference paper at ICLR 2024 examples of these sub-datasets are Chest X-ray, Dermatology, OCT (Optical Coherence Tomography), and Retinal Fundus. Each sub-dataset contains labeled images that are typically 28x28 pixels in size, resembling the format of the original MNIST dataset. We keep the original size for training. Due to the presence of both the RGB sub-dataset and grayscale sub-dataset in Med MNIST, we employed different model configurations to accommodate this variation. The Path MNIST, Derma MNIST, Blood MNIST, and Retina MNIST are RGB images, the rest 7 dataset are greyscale images. Image Net-C: Image Net-C (Hendrycks & Dietterich, 2019) is a dataset of 75 common visual corruptions. This dataset serves as a benchmark for evaluating the resilience of machine learning models to different forms of visual noise and distortions. This benchmark provides a more comprehensive understanding of a model s performance and can lead to the development of more robust algorithms that perform well under a wider range of scenarios. The metric used in Image Net-C is mean corruption error, which is the average error rate of 75 common visual corruptions of Imange Net validation set. The lower the m CE, the better the result. To ensure a fair comparison with the original paper that presented Res Net18, we followed the training protocol outlined in the Med MNIST paper. We trained our model for 100 epochs without any data augmentation, using early stopping as the stopping criterion. Table 17: The overview of the dataset we use in Appendix E and Section 4.2. The * in the below table since Med MNIST dataset consists of multiple sub-datasets, each containing medical images from specific categories. These sub-datasets have varying numbers of training samples, testing samples, and classes. Dataset CIFAR10 SVHN Oxford Flower Tiny-Imagenet Med MNIST Classes 10 10 102 200 * Train samples 50000 73257 1020 100000 * Test samples 10000 26032 6149 10000 * Resolution 32 32 256 64 28 Attribute Natural Images Numbers Flowers Natural Images Medical Images Image Type RGB RGB RGB RGB RGB+Greyscale M INITIALIZATION We evaluate various initialization methods for the parameters. The results in Table 18 indicate that the Xavier Normal initialization method yields the most favorable results. By adopting the Xavier Normal initialization, we observe improvements in the performance of the model trained on CIFAR10. Our result in the main paper use Xavier Normal as our default initialzation. Previous works on PNs (Chrysos et al., 2022; 2023) have demonstrated the crucial role of appropriate parameter initialization in the final performance. In this section, we explore different parameter initialization methods for the linear layers in the model and train them on the CIFAR-10 dataset from scratch. To minimize the impact of confounding factors, all experiments are conducted without any data augmentation techniques or regularization methods. The results are shown in Table 18. N ABLATION STUDY ON THE DESIGN OF MU-LAYER Our original design in Eq. (1) factorizes Λ, but not A. One interesting question would be to explore factorizing also A as a low-rank decomposition. We follow the same idea as in Λ to factorize A as A = R Q. The new design is visualized in Fig. 16. We conduct an experiment on CIFAR10 to assess the performance of the model when both branches assume a low-rank decomposition. To provide a fair comparison with the main model, we leave the secondary branch as mentioned in the main paper. The results in Table 19 indicate that the original design performs favorably when compared to having low-rank decompositions in both branches. Published as a conference paper at ICLR 2024 Table 18: Comparison of Initialization Methods. The best performance is marked in bold. Note that * in the table uses Py Torch default initialization, it depends on the layer type. For linear layer of shape (out, in), the values are initialized from U( k), where k = 1 in. Initialization Method Top-1 Acc (%) Top-5 Acc (%) Xavier Uniform (Glorot & Bengio, 2010) 88.85 99.38 Xavier Normal (Glorot & Bengio, 2010) 89.13 99.85 Kaiming Uniform (He et al., 2015) 88.56 98.70 Kaiming Normal (He et al., 2015) 88.72 99.46 Lecun Normal (Le Cun et al., 2002) 88.95 99.51 Normal 88.06 99.34 Sparse (Martens et al., 2010) 88.18 99.48 Pytorch default* 88.37 99.49 Table 19: Ablation study on the Mu-Layer design in Appendix N. Although considering the second branch as a low-rank decomposition performs generally worse, in some cases the difference is not large. Baseline (ours) Design 1 Design 2 Design 3 Design 4 Design 5 Design 6 The rank of Matrix RQ - 64 64 64 64 32 32 The rank of Matrix BD 64 32 64 16 8 16 8 Acc 88.1 87.1 85.2 86.4 86.2 86.1 85.6 Figure 16: Different designs of Mu-Layer as studied in Appendix N.