# efficient_learning_with_sineactivated_lowrank_matrices__3b1a5a2f.pdf Published as a conference paper at ICLR 2025 EFFICIENT LEARNING WITH SINE-ACTIVATED LOW-RANK MATRICES Yiping Ji Australian Institute for Machine Learning University of Adelaide DATA61, CSIRO Hemanth Saratchandran* Australian Institute for Machine Learning University of Adelaide Cameron Gordon Australian Institute for Machine Learning University of Adelaide Zeyu Zhang Australian National University Simon Lucey Australian Institute for Machine Learning University of Adelaide Low-rank decomposition has emerged as a vital tool for enhancing parameter efficiency in neural network architectures, gaining traction across diverse applications in machine learning. These techniques significantly lower the number of parameters, striking a balance between compactness and performance. However, a common challenge has been the compromise between parameter efficiency and the accuracy of the model, where reduced parameters often lead to diminished accuracy compared to their full-rank counterparts. In this work, we propose a novel theoretical framework that integrates a sinusoidal function within the lowrank decomposition. This approach not only preserves the benefits of the parameter efficiency of low-rank methods but also increases the decomposition s rank, thereby enhancing model performance. Our method proves to be a plug-in enhancement for existing low-rank methods, as evidenced by its successful application in Vision Transformers (Vi T), Large Language Models (LLMs), Neural Radiance Fields (Ne RF) and 3D shape modelling. The code is publicly available at https://yipingji.github.io/sine_activated_PEL/. 1 INTRODUCTION In the last few years, large-scale machine learning models have shown remarkable capabilities across various domains, achieving groundbreaking results in tasks related to computer vision and natural language processing (Vaswani et al., 2017; Dosovitskiy et al., 2021). However, these models come with a significant drawback: their training necessitates an extensive memory footprint. This challenge has spurred the demand for more compact, parameter-efficient architectures. A prominent solution that has emerged is the use of low-rank techniques (Hu et al., 2022; Chen et al., 2024; Liu et al., 2024; Kopiczko et al., 2024), which involve substituting the large, dense matrices in large scale models with smaller, low-rank matrices. This substitution not only simplifies the models but also shifts the computational complexity from quadratic to linear, making a significant impact on efficiency. In the context of high-capacity models like Vision Transformers (Vi Ts) and Large Language Models (LLMs) that utilize millions to billions of parameters, transitioning from dense to low-rank matrices can result in considerable cost savings. Nonetheless, adopting low-rank architectures does introduce a trade-off, as they typically do not achieve the same level of accuracy as their full-rank counterparts, presenting a balance between parameter efficiency and model performance. Equal contribution. Correspondence to Yiping Ji and Hemanth Saratchandran . Published as a conference paper at ICLR 2025 In this paper, we tackle the challenge of balancing parameter efficiency and model performance by introducing a novel technique that enhances the representational capacity of low-rank methods. Our approach builds on the insight that augmenting low-rank matrices with high-frequency sinusoidal functions can increase their rank without adding parameters. We provide a theoretical framework that explains how this modulation increases rank and demonstrate how incorporating this non-linearity into low-rank decompositions enables compact architectures that preserve efficiency while achieving good accuracy across various machine learning tasks. We direct the reader s attention to figure 1, which showcases our method across various machine learning applications. Our comparisons with standard low-rank methods consistently demonstrate superior performance across these diverse tasks. Sine Low-Rank Low-Rank PSNR: 19.81 PSNR: 23.27 (a) Ne RF (k=5), 4.7% Full Rank Params Sine Low-Rank Low-Rank (b) 3D Occupancy (k=1), 2.1% Full Rank Params 30 35 40 45 50 Parameters (M) Top-1 Accuracy (%) Low-Rank Sine Low-Rank Full-Rank (88M) (c) Vi T Classification (d) Lo RA (k=32) Figure 1: Applying a drop-in sine-activation increases the rank of low-rank matrix methods, leading to improved parameter efficiency and performance on a variety of tasks including: a) Ne RF, b) 3D Occupancy, c) Vi T image classification, and d) Fine-tuning Large Language Models (Lo RA). Our approach s advantages are corroborated across a range of machine learning applications, including low rank methods for Vi T (He et al., 2022b), LLMs (Hu et al., 2022; Liu et al., 2024), Ne RF for novel view synthesis (Mildenhall et al., 2020), and 3D shape modeling via Binary Occupancy Fields (Mescheder et al., 2019). Across the board, our approach not only matches the parameter savings offered by low-rank methods but also results in an improvement in accuracy, showing its broad applicability and superior performance among diverse machine learning tasks. The main contributions of our paper are: 1. A theoretical framework demonstrating that applying sinusoidal non-linearities to low-rank matrices can effectively increase their rank without introducing additional parameters. 2. Demonstrating that our theoretical framework leads to a drop-in component that can be applied to various low-rank architectures, resulting in improved accuracy while maintaining computational and parameter efficiency. 3. A comprehensive validation of our method across a range of diverse machine learning tasks, including computer vision, 3D shape modeling, and natural language processing. Published as a conference paper at ICLR 2025 2 RELATED WORK 2.1 LOW-RANK DECOMPOSITION: Low-rank decomposition stands as a crucial method across disciplines such as information theory, optimization, and machine learning, providing an approach to reduce memory costs (Strang, 2019; Xinwei et al., 2023). Notably, (Cand es et al., 2011) uncovered that matrices can precisely separate low-rank and sparse components through convex programming, linking to matrix completion and recovery. Expanding its application, (Yu et al., 2017) devised a low-rank learning framework for convolutional neural networks, enhancing compression while maintaining accuracy. (Sharma et al., 2023) found that performance improvements in large language models could be achieved by eliminating higher-order weight matrix components without extra parameters or data. In neural radiance fields, (Tang et al., 2022) introduced a rank-residual learning strategy for optimal low-rank approximations, facilitating model size adjustments. Additional contributions include (Shi & Guillemot, 2023) with rank-constrained distillation, (Chen et al., 2022) applying vector-matrix decomposition, and (Schwarz et al., 2023) using soft-gated low-rank decompositions for compression. 2.2 PARAMETER-EFFICIENT LEARNING: Parameter-efficient learning is an important research area in deep learning, merging various techniques to enhance model adaptability with minimal resource demands (Menghani, 2023). Techniques like parameter-efficient fine-tuning (PEFT) allow pretrained models to adjust to new tasks efficiently, addressing the challenges of fine-tuning large models due to high hardware and storage costs. Among these, Visual Prompt Tuning (VPT) stands out for its minimal parameter alteration less than 1% in the input space, effectively refining large Transformer models while keeping the core architecture unchanged (Jia et al., 2022). Similarly, Bit Fit offers a sparse-finetuning approach, tweaking only the model s bias terms for cost-effective adaptations (Zaken et al., 2022). Moreover, Lo RA introduces a low-rank adaptation that maintains model quality without additional inference latency or altering input sequence lengths, by embedding trainable rank matrices within the Transformer layers (Hu et al., 2022). Recent studies also combine Lo RA with other efficiency strategies like quantization, pruning, and random projections for further model compression (Dettmers et al., 2024; Li et al., 2024; Zhang et al., 2024; Kopiczko et al., 2024; Liu et al., 2024). 3 METHODOLOGY We introduce our main technique that we term a sine-activated low-rank approximation. The purpose of this technique is to increase the rank of a low-rank matrix without adding any extra parameters. 3.1 PRELIMINARIES 3.1.1 FEED-FORWARD LAYER Our technique is defined for feed-forward layers of a neural architecture. In this section, we fix the notation for such layers following Prince (2023). We express a feed-forward layer as: y = Wx + b (1) where W Rm n is a dense weight matrix, b Rm 1 is the bias of the layer, and x is the input from the previous layer. The output y is often activated by a non-linearity σ producing σ(y). The weight matrix W and bias b are trainable parameters of the layer. In contemporary deep learning models, the feed-forward layers weight matrices, W, are often large and dense yielding a high rank matrix. While the high-rank property of the weight matrix helps in representing complex signals, it significantly adds to the overall parameter count within the network yielding the need for a trade-off between the rank of the weight matrix and overall architecture capacity. 3.1.2 LOW-RANK DECOMPOSITION A full-rank weight matrix W can be replaced by low-rank matrices UVT , such that W = UVT , where U Rm k, V Rn k and k min(m, n). y = Wx + b = (UVT )x + b (2) Published as a conference paper at ICLR 2025 This is the most common way to reduce the parameter count in a feed-forward layer. During the training process, this method performs optimization on U and V alternatively. Low-rank multiplication then reduces the learnable parameter count and memory footprint from O(mn) to O(k (m+n)). Although UVT has the same matrix shape as the full-rank matrix W, the rank of UVT is constrained and rank(UVT ) k. Thus while we have significantly decreased the number of trainable weights in such a layer, we have paid the price by obtaining a matrix of much smaller rank. In the next section, we address this trade-off by developing a technique that can raise back the rank of a low-rank decomposition while keeping its low parameter count. 3.2 THEORETICAL FRAMEWORK 3.2.1 NON-LINEAR LOW-RANK DECOMPOSITION We introduce a non-linearity transformation into low-rank matrices as follows y = ϕ(ω UVT ) g x + b (3) where ϕ( ) is the non-linearity function, ω is a non-learnable frequency parameter, and g is a nonlearnable parameter to control the gain of the transformation. Our theoretical work will show that ϕ(ωx) = sin(ωx) is an optimal choice to increase the rank of UVT . 3.2.2 MAIN RESULT In this section, we provide a theoretical framework that clearly shows how to increase the rank of a low-rank decomposition using a non-linearity without adding any parameters. We will show that if we choose the non-linearity, in the decomposition defined in section 3.2.1, to be a sine function then provided the frequency ω is chosen high enough, the rank of the matrix ϕ(ω UVT ) will be larger than that of UVT . The proofs of the theorems are given in the appendix A.1. To begin with, we fix ω > 0 and let sin(ω A) denote the matrix obtained from a fixed m n matrix A by applying the function sin(ω x) component-wise to A. Assuming A = 0 we define A0 min as: A0 min = min i,j s.t.Aij =0 |Aij|. (4) Note that such a quantity is well defined precisely because A has a finite number of entries and all such entries cannot be zero from the assumption that A = 0. The following theorem relates the rank of sin(ω A) to the frequency ω and the quantity A0 min. Proposition 1. Fix an m n matrix A s.t. A = 0. Then Rank(sin(ω A)) ω if 0 ω π 3A0 min . (5) Proposition 1 shows that if we modulate the matrix sin(ω A) by increasing ω > 0 then the rank of the matrix sin(ω A) can be increased provided ω < π 3A0 min . We can apply proposition 1 to the context of a low-rank decomposition as defined in section 3.1.2. Given a low-rank decomposition UVT with U Rm k and V Rn k with k min{m, n} the following theorem shows how we can increase the rank of the decomposition by applying a sin(ω ) function. Theorem 1. Let U Rm k and V Rn k with k min{m, n}. Assume both U and V are initialized according to a uniform distribution U( 1/N, 1/N) where N > k. Then there exists an ω0 such that the matrix sin(ω A) will satisfy the inequality Rank(sin(ω UVT )) > Rank(UVT ) (6) provided ω ω0. Published as a conference paper at ICLR 2025 Theorem 1 also holds for the case we initialize U and V by a normal distribution of variance N. Weight matrices within feed-forward layers are typically initialized using a distribution that is contingent upon the layer s neuron count. When considering low-rank decompositions characterized by matrices U Rm k and VT Rk n, where k min{m, n}, the variance of this initialization distribution is influenced by m and n. These dimensions are significantly larger than k, ensuring that the condition specified in theorem 1 that N > k is always met, making this theorem especially relevant for low-rank decompositions in feed-forward layers. For example the most common initialization schemes such as Kaiming (He et al., 2015) and Xavier (Glorot & Bengio, 2010) satisfy the requirements of our theorem. Theorem 1 offers a viable strategy for maintaining a high-rank characteristic in feed-forward layers while simultaneously minimizing the parameter count. By introducing a sinusoidal non-linearity with a sufficiently high frequency ω into a low-rank decomposition, it s possible to increase the rank of the layer without altering the quantity of trainable parameters. Full Rank !!" Low Rank !#" = #$$ Sine Activated Low Rank !%&'(#") = sin 100 * #$$ Sine Activated Low Rank !%&'(#") = sin 2000 * #$$ Sine Low-Rank Full-Rank Sine Low-Rank Low-Rank Figure 2: These figures display weight magnitudes for matrices with dimension 128 128. The first figure shows a heatmap of a full-rank matrix initialized by Kaiming uniform, highlighting linear independence among rows. The second shows a low-rank matrix Wlr = UVT R128 128, with U, V R128 1 illustrating minimal linear independence. The final pair of figures reveal how applying a sine function element-wise, sin(ω UVT ), with varying ω, affects linear independence in low-rank matrices; specifically, ω = 100 and ω = 2000 progressively increase linear independence. In figure 2 we give a visualization of our method in action. We consider a full-rank matrix, a lowrank matrix, and two sine-activated low-rank matrices with different frequencies. By visualizing the weight magnitudes in each matrix via a heatmap, we can clearly see how the sine-activated low-rank matrix increases rank and furthermore how increasing the frequency of the sine function increases the rank in accord with theorem 1. Building upon equation 3, we explore the application of various non-linear functions to a low-rank decomposition, with a particular focus on the sine function. This choice is inspired by theorem 1, which theoretically demonstrates that applying a sine function effectively increases the matrix rank. In figure 3, we present a comparative analysis of the sine function against other common non-linear functions in machine learning, such as the sigmoid and Re LU. The results clearly show that the sine function increases the rank, making it an optimal non-linearity to apply to a low-rank decomposition. Further, theorem 1 suggests that augmenting the frequency of the sine function applied to a low-rank decomposition contributes to a further increase in rank. To empirically validate this, we conducted experiments applying sine functions of various frequencies to a constant low-rank matrix. The outcomes, depicted in figure 3 (right), corroborate the theorem s prediction, showcasing a positive correlation between the frequency of the sine function and the resultant rank increase. 4 EXPERIMENTS This section is dedicated to validating and analyzing the efficacy of our proposed low-rank methods across a spectrum of neural network architectures. To demonstrate the broad applicability and versatility of our approach, we examine its performance in three distinct contemporary applications. Published as a conference paper at ICLR 2025 0 100 200 Index Singular Values Full-Rank Low-Rank Sine LR Sigmoid LR Re LU LR 0 100 200 Index Singular Values Full-Rank Sine LR =100 Sine LR =200 Sine LR =400 Figure 3: In this figure we depict the singular value spectrum of a Kaiming uniform initialized matrix Wfr R256 256 and a low-rank k = 5 approximation matrix Wlr = UVT . All singular values are normalized to 1. Left: the spectral advantages of applying a non-linear function ϕ(ω UVT ) where ω is a hyper-parameter. Here we see the natural advantages of the sine function such that ϕ(x) = sin(ω x). Right: manipulating ω within the sine function changes these spectral properties. Specifically, we explore its integration into the fine-tuning of large language models through Lo RA (Hu et al., 2022), the pretraining of Vi T (Dosovitskiy et al., 2021), the reconstruction of scenes using Ne RF (Mildenhall et al., 2020), and 3D shape modeling (Mescheder et al., 2019). This collectively underscores our model s adaptability to a diverse array of low-rank frameworks, highlighting its potential to significantly impact various domains within the field of computer vision. 4.1 LARGE LANGUAGE MODEL Lo RA is a highly effective strategy for finetuning large pretrained models, as described in (Hu et al., 2022). Lo RA targets the adaptation of pretrained weight matrices W0 Rm n by limiting updates to a low-rank representation, expressed as W0x + Wx = W0x + UVT x, where U Rm k and V Rn k with the rank k min{m, n}. This method does not introduce additional inference latency or necessitate reducing the input sequence length, thus preserving the model quality. We conduct thorough experiments to evaluate the performance of our novel approach, termed sine Lo RA, against the standard Lo RA framework, demonstrating the effectiveness of our method. Dataset. We evaluate the natural language understanding (NLU) task performance on the Ro BERTa V3 base model (Reimers & Gurevych, 2019). Specifically, we adopt the widely recognized GLUE benchmark (Wang et al., 2018), including Co LA (Warstadt et al., 2018), MRPC (Dolan & Brockett, 2005), QQP, STS-B(Cer et al., 2017), MNLI (Williams et al., 2018), QNLI (Rajpurkar et al., 2016), and RTE (Dagan et al., 2006; Haim et al., 2006; Giampiccolo et al., 2007; Bentivogli et al., 2009). Furthermore, we evaluate sine Lo RA by fine-tuning large scale language models LLa MA 3-8B on commonsense reasoning tasks, which includes Bool Q (Clark et al., 2019), PIQA (Bisk et al., 2019), SIQA (Sap et al., 2019), Hella Swag (HS) (Zellers et al., 2019), Wino Grande (WG) (Sakaguchi et al., 2021), ARC-c, ARC-e (Clark et al., 2018) and OBQA (Mihaylov et al., 2018). Setting. In the Transformer architecture, there are four weight matrices in the self-attention module (Wq, Wk, Wv, Wo) and two in the MLP module(Wup, Wdown). To evaluate Ro BERTA V3, we follow up the Lo RA architecture and implement low-rank adaptation only on Wq and Wv. We study the performance of Lo RA and sine Lo RA in terms of different rank k = 1, 2, 4, 8. To evaluate LLa MA 3-8B, we implement low-rank adaptations on five modules (Wq, Wk, Wv, Wup, Wdown) with different rank k = 4, 8, 16, 32. For further implementation details see Appendix A.2.1. Results: We replicated the experimental framework of naive Lo RA to establish a baseline, and then evaluated our sine Lo RA, as detailed in table 1 and table 2. Our results reveal that sine Lo RA con- Published as a conference paper at ICLR 2025 Table 1: Performance and parameter count of the Ro BERTa V3 model fine-tuned using the Lo RA and sine Lo RA methods across varying kmax settings on the GLUE benchmark. Method Params COLA MRPC STSB SST2 RTE QNLI MNLI QQP Avg. Lo RAk=1 36.9K 66.31 90.15 90.15 94.70 78.80 93.06 88.18 87.61 85.63 0.32 x Sine Lo RAk=1 67.99 90.44 90.85 94.79 78.05 92.76 88.35 87.90 85.95 Lo RAk=2 73.7K 68.38 89.42 89.19 95.02 78.27 93.32 89.15 88.57 85.99 0.45 x Sine Lo RAk=2 68.93 90.79 90.94 94.81 79.10 93.29 88.26 88.70 86.44 Lo RAk=4 147.5K 68.56 89.69 88.79 95.23 80.39 93.34 89.78 88.70 86.41 0.73 x Sine Lo RAk=4 68.93 90.86 90.87 95.25 82.00 93.53 89.68 89.18 87.14 Lo RAk=8 294.9K 68.62 89.82 89.50 95.25 80.37 93.56 89.86 88.83 86.57 0.42 x Sine Lo RAk=8 68.54 90.22 90.85 95.11 81.82 93.58 89.69 89.38 86.99 Table 2: Performance and parameter count of the LLa MA 3-8B model fine-tuned using the Lo RA and sine Lo RA methods across varying kmax settings on the commonsense reasoning benchmark. Method Params Bool Q PIQA SIQA HS WG ARC-e ARC-c OBQA Avg. Lo RAk=4 7.1M 73.58 86.29 79.99 94.92 79.95 63.91 78.7 83 80.04 3.57 x Sine Lo RAk=4 72.69 87.38 79.32 94.39 85.32 75.01 88.64 86.2 83.61 Lo RAk=8 14.2M 72.97 87.43 78.81 72.18 85.80 77.47 88.38 83.20 80.79 2.87 x Sine Lo RAk=8 73.42 86.51 80.3 94.16 85.87 76.36 88.05 84.6 83.66 Lo RAk=16 28.3M 73.57 85.58 79.27 93.97 85.71 75.42 86.44 83.2 82.9 2.45 x Sine Lo RAk=16 73.7 87.65 80.76 94.93 84.45 79.1 89.77 84.4 84.35 Lo RAk=32 56.6M 70.64 86.13 78.25 91.48 83.19 69.71 85.73 81.4 80.82 2.74 x Sine Lo RAk=32 72.42 86.51 79.78 93.96 85.16 78.07 87.58 85 83.56 sistently surpasses the performance of the standard Lo RA at different rank levels (k), highlighting the effectiveness of the sine function in enhancing the representation capabilities of low-rank matrices. Notably, sine Lo RA at k = 4 not only exceeds Lo RA s performance at k = 8 by 0.57 but also halves the parameter count, illustrating significant efficiency and parameter savings. Surprisingly, with the LLa MA 3-8B model, our method with rank 4 already outperforms Lo RA with all higher ranks, achieving 83.61 compared to 82.9. Analysis. Within the Lo RA framework, featuring a low-rank multiplication component W = UVT, we enhance this low-rank component with a sine function and assess the efficacy of our method. This adaptation amplifies the update significance due to the intrinsic rank increase, facilitated by the sine-activation. Consequently, our approach attains superior performance at reduced rank levels k, compared to Lo RA, effectively decreasing the count of learnable parameters. Further results. For comparisons to the recent DORA paper (Liu et al., 2024) see Appendix A.2.1. 4.2 PRETRAINING VISION TRANSFORMERS Vision Transformers have risen to prominence as powerful models in the field of computer vision, demonstrating remarkable performance across a variety of tasks. When pretrained on large-scale datasets such as Image Net-21K and JFT-300M, Vi Ts serve as robust foundational architectures, particularly excelling in feature extraction tasks (Deng et al., 2009; Sun et al., 2017). A critical observation regarding the architecture of Vi Ts is that the two feed-forward layers in each block dedicated to channel mixing contribute to nearly 66% of the total model parameter count. In light of this, focused experiments on these specific layers have been conducted to rigorously assess the effectiveness of our proposed method, facilitating a direct comparison with the baseline model. Experimental setup. We trained the Vi T-Small and Vi T-Base models from scratch, utilizing the CIFAR-100 and Image Net-1k datasets, respectively, to establish our baseline performance metrics (Deng et al., 2009; Krizhevsky, 2012). The Vi T-Small model, characterized by its two MLP layers with input/output dimensions of 384 and hidden dimensions of 1536, was modified by replacing the full-rank weight matrices with low-rank matrices across a range of ranks (k). Similarly, the Vi T-Base model, which features two MLP layers with input/output dimensions of 768 and hidden dimensions of 3072, underwent a parallel modification, where its full-rank weight matrices were Published as a conference paper at ICLR 2025 30 35 40 45 50 Parameters (M) Top-1 Accuracy (%) Low-Rank Sine Low-Rank Full-Rank (88M) (a) Vi T-Base 7.5 8.0 8.5 9.0 9.5 10.0 Parameters (M) Top-1 Accuracy (%) Low-Rank Sine Low-Rank Full-Rank (22M) (b) Vi T-Small Figure 4: Low Rank Vi T classification performance. Use of the sine-activation improves performance of the low-rank models, and even enables improvement relative to the Full Rank model. substituted with low-rank matrices for a range of ranks. For the training of the Vi T-Base model, we follow the training methodology described in Masked Autoencoders (MAE) (He et al., 2022a), implementing a batch size of 1024. This structured approach allows us to rigorously evaluate the impact of introducing low-rank matrices to these model architectures. For further implementation details see Appendix A.2.2. Results. Figure 4 shows the outcomes of training Vi T models from scratch on the Image Net-1k and CIFAR100 datasets, respectively. These findings are compared with those of conventional baseline training of Vi T models, which demonstrate that employing aggressive low-rank levels (k) compromises accuracy. Remarkably, the Vi T-Base model, even when operating at a rank of 250 with only 50% of its parameters in comparison to the baseline, attains the performance metrics of the baseline on the Image Net-1k dataset, albeit at the cost of increased training loss. The use of sine low-rank matrices consistently yields substantial improvements in test accuracy across all examined rank levels for both datasets. This suggests that the sine function significantly bolsters the representational capacity of low-rank weight matrices, as suggested by the theory in section 3.2. Analysis. Large models, such as Vi T-Base, with an excessively large number of parameters, are prone to overfitting, where they perform well on training data but poorly on unseen data, especially when trained on relatively small datasets like Image Net-1k (Xu et al., 2024). Low-rank learning techniques can help in designing models that generalize better to new data by encouraging the model to learn more compact and generalizable representations to reduce overfitting. Additionally, while Vi T architectures often underperform on smaller datasets, this method introduces a novel approach for efficiently training Vi T models using small data collections. For frequency ablations in the rank 1 case, see Appendix A.2.2. Further results: Conv Ne Xt. In order to show that our method works on convolutional only architectures we implemented sine low-rank on Conv Ne Xt (Liu et al., 2022), a leading convolutional architecture for image classification. For implementation details and results see Appendix A.2.3. Neural Radiance Fields (Ne RFs) represent 3D scene signals by utilizing a set of 2D sparse images (Mildenhall et al., 2020). The 3D reconstruction is obtained by a forward pass fθ(x, y, z, θ, ϕ), involving position (x, y, z) and viewing direction (θ, ϕ). We evaluate our methods by training a Ne RF model on the standard benchmarks LLFF dataset, which consists of 8 real-world scenes captured by hand-held cameras (Mildenhall et al., 2019). To evaluate our method on Ne RF we substitute each fully dense layer with low-rank decomposition and use a range of rank levels (k). Results. Table 3 provides results on the LLFF dataset (Mildenhall et al., 2019; 2020). We report the peak signal-to-noise ratio (PSNR) with the compression rate representing the percentage of parameters used in comparison to the parameter count of the Full Rank Ne RF model. Employing lowrank matrices in Ne RF learning reduces parameter count while significantly enhancing compression. However, performance dips with very low-rank levels (k), where models capture minimal informa- Published as a conference paper at ICLR 2025 Table 3: Quantitative results for Ne RF evaluated on the LLFF dataset. Fern Flower Fortress Horn Leaves Orchids Room Trex Average Compression Rate Full-Rank 26.38 27.54 30.93 28.20 21.79 21.33 30.96 27.68 26.85 - 100% Low-Rankk=1 15.03 14.60 14.74 13.66 12.89 12.50 15.04 13.54 14.00 5.77 x 1.3% Sine Low-Rankk=1 20.77 20.14 24.13 19.00 15.92 16.25 25.53 16.42 19.77 Low-Rankk=5 20.64 19.81 24.90 20.40 15.74 16.07 22.74 19.79 20.01 3.10 x 4.7% Sine Low-Rankk=5 23.50 23.27 26.78 23.99 18.49 18.90 27.05 22.96 23.11 Low-Rankk=10 22.83 22.18 25.96 22.76 17.36 18.12 26.12 21.69 22.12 2.27 x 8.7% Sine Low-Rankk=10 24.56 24.61 28.01 25.39 19.62 20.02 28.70 24.21 24.39 Low-Rankk=30 24.48 24.68 28.10 25.54 19.36 20.04 38.92 24.24 24.42 1.45 x 24.6% Sine Low-Rankk=30 25.71 26.01 29.46 27.16 20.95 21.17 30.18 26.27 25.86 Low-Rankk=60 25.26 26.16 29.50 26.74 20.39 20.85 30.00 25.81 25.59 0.77 x 48.6% Sine Low-Rankk=60 26.09 26.70 29.75 27.78 21.56 21.37 30.54 27.16 26.36 tion. Our methods, nevertheless, substantially elevate performance. For instance, with k = 1, our sine low-rank approach yields an average PSNR of 19.77, outperforming the naive low-rank by 5.77 and achieving a compression rate of merely 1.3%. Even at a 48% compression rate, it surpasses the basic low-rank model by 0.77 PSNR, narrowly trailing the baseline by just 0.49 PSNR, as shown in Figure 5b. Our rate-distortion analysis, applying Akima interpolation for Bjøntegaard Delta calculation, reveals a BD-Rate of 64.72% and BD-PSNR of 2.72d B, signifying marked improvements in compression efficiency (Bjøntegaard, 2001; Herglotz et al., 2022). Visualization for k = 1 results are shown in Figure 5a and more results are shown in Figure 8 in the Appendix. Sine Low-Rank Low-Rank (a) Qualitative Ne RF results for LLFF datasets (k = 1). 0 20 40 Compression Rate (%) Sine Low-Rank Low-Rank Full-Rank (b) Rate-Distortion curve (LLFF average). Figure 5: (a) Using a non-transformed Low-Rank model leads to a complete loss of signal at extreme (rank k = 1). In contrast, applying a sine-activation function is able to reconstruct details even at 1.3% of the Full-Rank parameters. (b) The Sine Low-Rank Ne RF models show significant improvements across the rate-distortion curve relative to the Low-Rank models. Analysis. Ne RF models fit entire 3D scenes, and a high training PSNR leads to a high testing PSNR (Mildenhall et al., 2020). Employing structured weight matrices could result in a drop in performance due to the inherent constraints imposed by their structural design. Increasing the matrices rank enhances their memorization abilities significantly, especially when using a very low k. Starting from a low frequency, there is a rapid and consistent increase in PSNR. Consequently, as we elevate the rank level k, our results gradually align with the baseline Ne RFs, which serve as the upper bound. For frequency ablations in the case of rank 1 and rank 5 see Appendix A.2.4. Published as a conference paper at ICLR 2025 4.4 3D SHAPE MODELING For this experiment, we evaluate performance on binary occupancy field reconstruction, which involves determining whether a given coordinate is occupied (Mescheder et al., 2019). Following (Saragadam et al., 2023), we sampled over a 512 512 512 grid with each voxel within the volume assigned a 1, and voxels outside the volume assigned a 0. We use the Thai Statue, Dragon and Lucy instance from the Stanford Scanning Repository.1 We evaluate intersection over union (Io U) for the occupancy volumes. We used a coordinate-based MLP that includes two hidden layers, each with a width of 256 neurons, and employed the Gaussian activation function. The full-rank model achieves an accuracy of 97 (Io U). Figure 6 shows the 3D mesh representation of the Thai Statue, visualized using the low-rank method and the sine low-rank method for k = 1, 2, 5. Applying the sine function to the low-rank matrix resulted in a significant enhancement and more precise shape delineation. Results on Dragon and Lucy are given in Table 13 in Appendix A.2.5. A frequency ablation in the rank 1 case is given in Appendix A.2.5. k = 1 k = 2 Low-Rank Low-Rank Sine Low-Rank (Ours) Sine Low-Rank (Ours) (Io U : 84.3) (Io U : 90.8) (Io U : 88) (Io U : 92) k = 5 Low-Rank Sine Low-Rank (Ours) (Io U : 93.4) (Io U : 94.4) Figure 6: Binary occupancy field reconstruction on the Thai Statue. Note that without a sine function, the low-rank model is unable reconstruct any finer details for the k = 1 case; however, even at that level the sine low-rank model is able to reconstruct fine structural details of the statue, including the trunks of the elephants. The k = 1, k = 2 and k = 5 model utilizes only 2.1%, 2.9% and 5.2%, respectively, of the parameters of the full-rank model. 5 LIMITATIONS Our exploration into sine low-rank matrices illuminates their promising capabilities, yet it also has a limitation: notably, while these matrices can reach rank levels comparable to their full-rank counterparts upon the application of a sine function, their accuracy falls short. This highlights an ongoing challenge in finding the optimal balance between the need for sufficient parameterization to ensure high accuracy and the preferable rank of matrices. Overparameterization is widely recognized in the literature as vital for deep learning models to achieve strong generalization and memorization. Moving forward, developing strategies that not only increase rank but also clearly define the necessary degree of overparameterization will be crucial for creating cost-effective deep learning architectures, presenting an intriguing avenue for future research. 6 CONCLUSION In this work we have demonstrated that applying a sinusoidal non-linearity improves the accuracy of low-rank approximations by increasing their rank. While simple, this method is highly applicable to parameter-constrained models such as Lo RA, as it improves approximation without adding capacity, by overcoming representation limits of the matrix rank. We have fully justified this approach from theoretical first principles. When applied as a drop-in component we showed that this method leads to surprisingly large improvements across a range of tasks involving low-rank models, including language tasks, image classification, neural radiance fields and 3D shape modelling. 1Available at https://graphics.stanford.edu/data/3Dscanrep/ Published as a conference paper at ICLR 2025 ACKNOWLEDGEMENTS Hemanth Saratchandran and Simon Lucey acknowledge support from Commonwealth Bank of Australia through the Comm Bank Centre for Foundational AI Research. This funding was essential for the completion of the research described in this publication. Luisa Bentivogli, Peter Clark, Ido Dagan, and Danilo Giampiccolo. The fifth PASCAL recognizing textual entailment challenge. TAC, 7(8):1, 2009. Yonatan Bisk, Rowan Zellers, Ronan Le Bras, Jianfeng Gao, and Yejin Choi. PIQA: Reasoning about physical commonsense in natural language, 2019. URL https://arxiv.org/abs/ 1911.11641. Gisle Bjøntegaard. Calculation of average PSNR differences between RD-curves. Technical report, VCEG-M33, Austin, TX, USA, April 2001. Emmanuel J Cand es, Xiaodong Li, Yi Ma, and John Wright. Robust principal component analysis? Journal of the ACM (JACM), 58(3):1 37, 2011. Daniel Cer, Mona Diab, Eneko Agirre, I nigo Lopez-Gazpio, and Lucia Specia. Sem Eval-2017 task 1: Semantic textual similarity multilingual and crosslingual focused evaluation. In Steven Bethard, Marine Carpuat, Marianna Apidianaki, Saif M. Mohammad, Daniel Cer, and David Jurgens (eds.), Proceedings of the 11th International Workshop on Semantic Evaluation (Sem Eval2017), pp. 1 14, Vancouver, Canada, August 2017. Association for Computational Linguistics. doi: 10.18653/v1/S17-2001. URL https://aclanthology.org/S17-2001. Anpei Chen, Zexiang Xu, Andreas Geiger, Jingyi Yu, and Hao Su. Tenso RF: Tensorial radiance fields. In Computer Vision ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23 27, 2022, Proceedings, Part XXXII, pp. 333 350, Berlin, Heidelberg, 2022. Aochuan Chen, Ziqi Gao, Zijing Liu, Yu Li, and Jia Li. Parameter-efficient fine-tuning via circular convolution. ar Xiv preprint ar Xiv:2407.19342, 2024. Christopher Clark, Kenton Lee, Ming-Wei Chang, Tom Kwiatkowski, Michael Collins, and Kristina Toutanova. Bool Q: Exploring the surprising difficulty of natural yes/no questions. ar Xiv preprint ar Xiv:1905.10044, 2019. Peter Clark, Isaac Cowhey, Oren Etzioni, Tushar Khot, Ashish Sabharwal, Carissa Schoenick, and Oyvind Tafjord. Think you have solved question answering? try ARC, the AI2 reasoning challenge. ar Xiv preprint ar Xiv:1803.05457, 2018. Ido Dagan, Oren Glickman, and Bernardo Magnini. The PASCAL recognising textual entailment challenge. In Machine learning challenges. evaluating predictive uncertainty, visual object classification, and recognising tectual entailment, pp. 177 190. Springer, 2006. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Image Net: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 248 255, 2009. doi: 10.1109/CVPR.2009.5206848. Tim Dettmers, Artidoro Pagnoni, Ari Holtzman, and Luke Zettlemoyer. QLo RA: Efficient finetuning of quantized LLMs. Advances in Neural Information Processing Systems, 36, 2024. Ning Ding, Xingtai Lv, Qiaosen Wang, Yulin Chen, Bowen Zhou, Zhiyuan Liu, and Maosong Sun. Sparse low-rank adaptation of pre-trained language models, 2023. URL https://arxiv. org/abs/2311.11696. William B Dolan and Chris Brockett. Automatically constructing a corpus of sentential paraphrases. In Proceedings of the International Workshop on Paraphrasing, 2005. Published as a conference paper at ICLR 2025 Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16 words: Transformers for image recognition at scale. ICLR, 2021. Danilo Giampiccolo, Bernardo Magnini, Ido Dagan, and Bill Dolan. The third PASCAL recognizing textual entailment challenge. In Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing, pp. 1 9. Association for Computational Linguistics, 2007. Xavier Glorot and Yoshua Bengio. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 249 256. JMLR Workshop and Conference Proceedings, 2010. R Bar Haim, Ido Dagan, Bill Dolan, Lisa Ferro, Danilo Giampiccolo, Bernardo Magnini, and Idan Szpektor. The second PASCAL recognising textual entailment challenge. In Proceedings of the Second PASCAL Challenges Workshop on Recognising Textual Entailment, volume 7, pp. 785 794, 2006. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Delving deep into rectifiers: Surpassing human-level performance on Image Net classification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 1026 1034, 2015. doi: 10.1109/ICCV.2015.123. Kaiming He, Xinlei Chen, Saining Xie, Yanghao Li, Piotr Doll ar, and Ross Girshick. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16000 16009, 2022a. Xuehai He, Chunyuan Li, Pengchuan Zhang, Jianwei Yang, and Xin Eric Wang. Parameter-efficient model adaptation for vision transformers. ar Xiv preprint ar Xiv:2203.16329, 2022b. Christian Herglotz, Matthias Kr anzler, Rubens Mons, and Andr e Kaup. Beyond Bjøntegaard: Limits of video compression performance comparisons. In International Conference on Image Processing (ICIP), 2022. doi: 10.1109/icip46576.2022.9897912. Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. Lo RA: Low-rank adaptation of large language models. In International Conference on Learning Representations, 2022. Menglin Jia, Luming Tang, Bor-Chun Chen, Claire Cardie, Serge Belongie, Bharath Hariharan, and Ser-Nam Lim. Visual prompt tuning. In European Conference on Computer Vision, pp. 709 727. Springer, 2022. Dawid J. Kopiczko, Tijmen Blankevoort, and Yuki M. Asano. Ve RA: Vector-based random matrix adaptation. In International Conference on Learning Representations, 2024. Alex Krizhevsky. Learning multiple layers of features from tiny images. University of Toronto, 05 2012. Yixiao Li, Yifan Yu, Chen Liang, Nikos Karampatziakis, Pengcheng He, Weizhu Chen, and Tuo Zhao. Loft Q: Lo RA-fine-tuning-aware quantization for large language models. In The Twelfth International Conference on Learning Representations, 2024. URL https://openreview. net/forum?id=Lz PWWPAd Y4. Shih-Yang Liu, Chien-Yi Wang, Hongxu Yin, Pavlo Molchanov, Yu-Chiang Frank Wang, Kwang Ting Cheng, and Min-Hung Chen. Do RA: Weight-decomposed low-rank adaptation. ar Xiv preprint ar Xiv:2402.09353, 2024. Zhuang Liu, Hanzi Mao, Chao-Yuan Wu, Christoph Feichtenhofer, Trevor Darrell, and Saining Xie. A convnet for the 2020s, 2022. URL https://arxiv.org/abs/2201.03545. Gaurav Menghani. Efficient deep learning: A survey on making deep learning models smaller, faster, and better. ACM Computing Surveys, 55(12), mar 2023. ISSN 0360-0300. doi: 10.1145/3578938. URL https://doi.org/10.1145/3578938. Published as a conference paper at ICLR 2025 Lars Mescheder, Michael Oechsle, Michael Niemeyer, Sebastian Nowozin, and Andreas Geiger. Occupancy networks: Learning 3D reconstruction in function space. In Proceedings IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2019. Todor Mihaylov, Peter Clark, Tushar Khot, and Ashish Sabharwal. Can a suit of armor conduct electricity? a new dataset for open book question answering. ar Xiv preprint ar Xiv:1809.02789, 2018. Ben Mildenhall, Pratul P. Srinivasan, Rodrigo Ortiz-Cayon, Nima Khademi Kalantari, Ravi Ramamoorthi, Ren Ng, and Abhishek Kar. Local light field fusion: Practical view synthesis with prescriptive sampling guidelines. ACM Transactions on Graphics (TOG), 2019. Ben Mildenhall, Pratul P. Srinivasan, Matthew Tancik, Jonathan T. Barron, Ravi Ramamoorthi, and Ren Ng. Nerf: Representing scenes as neural radiance fields for view synthesis. In ECCV, 2020. Simon JD Prince. Understanding deep learning. MIT press, 2023. Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. SQu AD: 100,000+ questions for machine comprehension of text. In Proceedings of EMNLP, pp. 2383 2392. Association for Computational Linguistics, 2016. Sameera Ramasinghe and Simon Lucey. Beyond periodicity: Towards a unifying framework for activations in coordinate-mlps, 2022. URL https://arxiv.org/abs/2111.15135. Nils Reimers and Iryna Gurevych. Sentence-bert: Sentence embeddings using siamese bertnetworks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing. Association for Computational Linguistics, 11 2019. URL http://arxiv.org/ abs/1908.10084. Keisuke Sakaguchi, Ronan Le Bras, Chandra Bhagavatula, and Yejin Choi. Wino Grande: An adversarial Winograd schema challenge at scale. Communications of the ACM, 64(9):99 106, 2021. Maarten Sap, Hannah Rashkin, Derek Chen, Ronan Le Bras, and Yejin Choi. Social IQA: Commonsense reasoning about social interactions. ar Xiv preprint ar Xiv:1904.09728, 2019. Vishwanath Saragadam, Daniel Le Jeune, Jasper Tan, Guha Balakrishnan, Ashok Veeraraghavan, and Richard G. Baraniuk. WIRE: Wavelet implicit neural representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 18507 18516, June 2023. Jonathan Richard Schwarz, Jihoon Tack, Yee Whye Teh, Jaeho Lee, and Jinwoo Shin. Modalityagnostic variational compression of implicit neural representations. In Andreas Krause, Emma Brunskill, Kyunghyun Cho, Barbara Engelhardt, Sivan Sabato, and Jonathan Scarlett (eds.), Proceedings of the 40th International Conference on Machine Learning, volume 202 of Proceedings of Machine Learning Research, pp. 30342 30364. PMLR, 23 29 Jul 2023. URL https://proceedings.mlr.press/v202/schwarz23a.html. Pratyusha Sharma, Jordan T Ash, and Dipendra Misra. The truth is in there: Improving reasoning in language models with layer-selective rank reduction. ar Xiv preprint ar Xiv:2312.13558, 2023. Jinglei Shi and Christine Guillemot. Light field compression via compact neural scene representation. In ICASSP 2023 - 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1 5, 2023. doi: 10.1109/ICASSP49357.2023.10095668. Gilbert Strang. Linear Algebra and Learning from Data. Wellesley-Cambridge Press, 2019. ISBN 978-0-692-19638-0. Chen Sun, Abhinav Shrivastava, Saurabh Singh, and Abhinav Gupta. Revisiting unreasonable effectiveness of data in deep learning era. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Oct 2017. Jiaxiang Tang, Xiaokang Chen, Jingbo Wang, and Gang Zeng. Compressible-composable Ne RF via rank-residual decomposition. ar Xiv preprint ar Xiv:2205.14870, 2022. Published as a conference paper at ICLR 2025 Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Ł ukasz Kaiser, and Illia Polosukhin. Attention is all you need. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett (eds.), Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc., 2017. Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel R Bowman. GLUE: A multi-task benchmark and analysis platform for natural language understanding. ar Xiv preprint ar Xiv:1804.07461, 2018. Alex Warstadt, Amanpreet Singh, and Samuel R Bowman. Neural network acceptability judgments. ar Xiv preprint ar Xiv:1805.12471, 2018. Adina Williams, Nikita Nangia, and Samuel R. Bowman. A broad-coverage challenge corpus for sentence understanding through inference. In Proceedings of NAACL-HLT, 2018. Ou Xinwei, Chen Zhangxin, Zhu Ce, and Liu Yipeng. Low rank optimization for efficient deep learning: Making a balance between compact architecture and fast training. Journal of Systems Engineering and Electronics, PP:1 23, 01 2023. doi: 10.23919/JSEE.2023.000159. Zhiqiu Xu, Yanjie Chen, Kirill Vishniakov, Yida Yin, Zhiqiang Shen, Trevor Darrell, Lingjie Liu, and Zhuang Liu. Initializing models with larger ones. In International Conference on Learning Representations (ICLR), 2024. Xiyu Yu, Tongliang Liu, Xinchao Wang, and Dacheng Tao. On compressing deep models by low rank and sparse decomposition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7370 7379, 2017. Elad Ben Zaken, Yoav Goldberg, and Shauli Ravfogel. Bit Fit: Simple parameter-efficient finetuning for transformer-based masked language-models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pp. 1 9, Dublin, Ireland, 2022. Association for Computational Linguistics. Rowan Zellers, Ari Holtzman, Yonatan Bisk, Ali Farhadi, and Yejin Choi. Hella Swag: Can a machine really finish your sentence? ar Xiv preprint ar Xiv:1905.07830, 2019. Mingyang Zhang, Hao Chen, Chunhua Shen, Zhen Yang, Linlin Ou, Xinyi Yu, and Bohan Zhuang. Lo RAPrune: Pruning meets low-rank parameter-efficient fine-tuning, 2024. URL https:// openreview.net/forum?id=9KVT1e1qf7. Published as a conference paper at ICLR 2025 A.1 THEORETICAL FRAMEWORK In this section we give the proof of proposition 1 and theorem 1 from section 3.2 of the paper. We recall from section 3.2 the following notation: For fixed ω > 0, let sin(ω A) denote the matrix obtained from a fixed m n matrix A by applying the function sin(ω x) component-wise to A. Assuming A = 0 we define A0 min as: A0 min = min i,js.t.Aij =0 |Aij|. (7) Note that such a quantity is well defined precisely because A has a finite number of entries and all such entries cannot be zero from the assumption that A = 0. Before we give the proof of proposition 1 from section 3.2 of the main paper, we will prove two lemmas. Lemma 1. For a fixed m n matrix A. We have ||sin(ωA)||2 F ω2(A0 min) if 0 < ω < π 3A0 min (8) where A0 min is defined as follows: A0 min = min i,js.t.Aij =0 |Aij| (9) for 1 i m and 1 j n. Proof. Observe by definition of the Frobenius norm that ||sin(ωA)||2 F = j=1 sin(ωAij)2. (10) We then find that ||sin(ωA)||2 F sin(ωA0 min)2. (11) The goal is to now find a lower bound on sin(ωA0 min)2. In order to do this consider the function f(ω) = sin(ωx) ωx 2 , where x R is fixed and positive. Differentiating this function we have f (ω) = xcos(ωx) x To find a critical point we solve the equation f (ω) = 0 to find cos(ωx) = 1 We see that equation 13 has the solution ωx = π 3 . In order to check what type of critical point ωx = π 3 we need to look at f (ω) f (ω) = x2sin(ωx) < 0 (14) when ω = π 3x implying that the critical point ω = π 3x is a maximum point. Observe that f(0) = 0 it thus follows that f(ω) 0 on the interval [0, π Applying this to the function sin(ωA0 min) we obtain that sin(ωA0 min) ωA0 min 2 if 0 ω π 3A0 min . (15) Substituting the lower bound in equation 15 into equation 11 we obtain the proposition. Published as a conference paper at ICLR 2025 The next lemma establishes an upper bound on the operator norm of sin(ωA). We remind the reader that the operator norm of A is denoted by ||A||op and is defined by ||A||op = sup ||x||2=1 ||Ax||2 (16) where x is a vector and || ||2 represents the vector 2-norm. It can be shown that the operator norm of A is also given by the largest singular value of A see (Strang, 2019). Lemma 2. Let A be a fixed m n matrix. Then ||sin(ωA)||2 op |A| denotes the matrix obtained from A by taking the absolute value and then square root component wise. Proof. By definition we have ||sin(ωA)||2 op = sup ||x||2=1 ||sin(ωA)x||2 2 (18) where || ||2 denotes the 2-norm of a vector. For any fixed unit vector x we will show how to upper bound the quantity ||sin(ωA)x||2 2. In order to do this we will use the fact that for x 0, we have the bound sin(x) p ||sin(ωA)x||2 2 = j=1 sin(ωAij)xj |A| x||2 2. (21) It follows that sup ||x||=1 ||sin(ωA)x||2 2 sup ||x||=1 || ω p |A|x||2 2 (22) which implies ||sin(ωA)||2 op We can now give the proof of proposition 1 from section 3.2 of the main paper. In order to do so we will need the definition of the stable rank of a matrix. Assume A is a non-zero m n matrix. We define the stable rank of A by SR(A) := ||A||2 F ||A||2op . (24) It is easy to see from the definition that the stable rank is continuous, unlike the rank, and is bounded above by the rank SR(A) Rank(A). (25) Remark 1. We observe that lemmas 1 and 2 give the main reasons why we chose a sine function as the non-linearty to apply on a weight matrix A. The periodic nature of a sine function that can be controlled by a frequency parameter ω > 0 is what allows us to obtain a proof of lemmas 1 and 2. When we give a proof of Thm. 1 we will see that these two lemmas are crucial. of proposition 3.1 from section 3.2 of main paper. Observe that from equation 25 it suffices to prove the lower bound on SR(A). This is immediate from lemma 1 and lemma 2. Published as a conference paper at ICLR 2025 Proof of theorem 1 from section 3.2 of main paper. From the assumption of the theorem 1 we have that N >> k. Further, we are assuming that both U and V have entries sampled from U( 1/N, 1/N). This means if we let A = UVT , then there exists a C > 0 such that Furthermore, observe that || p |A|||op || p |A|||F ||A||F (27) which implies ω A0 min || p Now observe that from proposition 1 in section 3.2 from the main paper we have that Rank(sin(ωA)) ω A0 min || p |A|||op (29) if 0 ω π 3A0 min . We can rewrite this last condition to say that equation 29 holds if 0 ω πN2 In particular, by using equation 28 we find that there exists ω0 within the interval 0 ω0 πN2 Rank(sin(ω0A)) ω0 A0 min || p |A|||op k Rank(A). (30) This completes the proof. Remark 2. Observe that if ω is very small, then sin(ωx) ωx and thus applying a sin simply scales the matrix by ω which cannot change rank. It is only when ω is sufficiently large that we see that the rank increases. If ω = 0 then applying sin with frequency ω to the matrix produces the zero matrix which has zero rank and in general will thus have rank less than A. Remark 3. In general low-rank matrices inherently have fewer degrees of freedom compared to high-rank matrices, which limits their representational capacity. For complex datasets, we hypothesize that high-rank weight matrices within a neural model provide additional degrees of freedom, enabling the model to capture and learn key features more effectively from the input data. Our empirical results on all tasks seem to validate this hypothesis. A.2 EXPERIMENTS For the experiments we observed that the gain factor g in Equation (3) should be chosen analogously to how weights are initialized in (He et al., 2015). In particular we chose g = n, where n was the number of rows of the weight matrix. During backpropagation the frequency parameter ω scales the gradients which can cause gradient exploding. To mitigate this we found the choice of g = n worked best. We also point out that for the experiments there is no principled way to set ω. Therefore, we will obtain ω by treating it as a hyperparameter and tuning it according to best results. A.2.1 FINE TUNING LARGE LANGUAGE MODELS Implementation details for Roberta V3 We followed the settings in (Hu et al., 2022) and (Ding et al., 2023). In the Transformer architecture, there are four weight matrices in the self-attention module (Wq, Wk, Wv, Wo) and two in the MLP module(Wup, Wdown). To evaluate Ro BERTA V3, we follow up the Lo RA architecture and implement low-rank adaptation only on Wq and Wv. We study the performance of Lo RA and sine Lo RA in terms of different rank k = 1, 2, 4, 8. For sine Lo RA, we use frequency = 200 across all the ranks. We use different learning rate and epoch for different datasets as shown in Table 4. Implementation details for Llama3-8B We followed the settings in (Liu et al., 2024). In the Transformer architecture, there are four weight matrices in the self-attention module (Wq, Wk, Wv, Wo) and two in the MLP module(Wup, Wdown). To evaluate Llama3-8B, we implement low-rank adaptation only on Wq, Wk, Wv, Wup, and Wdown. We study the performance of Lo RA and sine Lo RA for different rank k = 4, 8, 16, 32 and configurations are as shown in 5. Published as a conference paper at ICLR 2025 Dataset lr epoch Co LA 8e-5 20 SST-2 1e-4 10 MRPC 1e-4 20 QQP 3e-4 10 STS-B 1e-4 20 MNLI 3e-4 10 QNLI 3e-4 10 RTE 1.2e-3 50 Table 4: Learning rate and epoch for each dataset for the Roberta V3 model. rank frequency lr epoch 4 800 1e-4 3 8 600 1e-4 3 16 200 1e-4 3 32 200 1e-4 3 Table 5: Sine Lo RA. Frequency, learning rate and epochs for the Llama3-8B model. Do RA In order to compare our method to the current state-of-the-art we apply our sine-Lo RA method to Weight-Decomposed Low-Rank Adaptation (Do RA) (Liu et al., 2024). Do RA decomposes pre-trained weights W Rm n into two components: a magnitude vector q R1 n and a direction matrix D Rm n, normalized by the column vector-wise norm c such that each column remains a unit vector. Do RA decompose a low-rank change in the direction matrix into the decomposition D = UVT, where U Rm k and V Rn k. To implement our sine-Do RA, we apply the sine function to this decomposition as per Equation (3). We evaluate the performance of Do RA and sine-Do RA in terms of different rank k = 8, 16, 32 in Table 7 and configurations are shown in Table 8. We implement k = 8 directly as this is not a setting used in (Liu et al., 2024), and compare against reported results for k = 16 and k = 32. By introducing our methods on top of Do RA, our model (Sine Do RA k = 8) achieve state-of-the-art results while utilizing only 25% of the parameters required by Do RA(k = 32). Computational cost. Finetuning Llama3-8B takes roughly 6 hours using Lo RA, 7 hours using Sine Lo RA, 11 hours using Do RA, and 11 hours using Sine Do RA using a NVIDIA H100 GPU with 96GB of memory. Training memory cost is shown in Table 6. A.2.2 VISION TRANSFORMERS Implementation details for Vi T-Small on CIFAR100: The Vi T-Small model, characterized by its two MLP layers with input/output dimensions of 384 and hidden dimensions of 1536, was modified by replacing the full-rank weight matrices with low-rank matrices across a range of ranks (k). We use learning rate 1e-3, batch size 512 and train for 200 epochs. Choices of frequency for different ranks are shown in Table 9. Table 6: Training memory (GB) cost for Lo RA, Sine Lo RA, Do RA, Sine Dora on finetuning Llama3-8B as reported by Nvidia-SMI. Method Rank 8 Rank 16 Rank 32 Lo RA 54.9 55.2 55.3 Sine Lo RA 72.0 72.0 72.2 Do RA 75.6 75.9 76.0 Sine Do RA 90.4 90.7 90.8 Published as a conference paper at ICLR 2025 Table 7: Performance and parameter count of the LLa MA 3-8B model fine-tuned using the Do RA and sine Do RA methods across varying kmax settings on the commonsense reasoning benchmark. * Results reported in the paper. Method Params Bool Q PIQA SIQA HS WG ARC-e ARC-c OBQA Avg. Do RAk=8 14.9M 73.2 87.7 79.9 94.7 84.5 89.3 78.0 83.2 83.8 1.4 x Sine Do RAk=8 73.9 89.0 81.0 95.3 86.1 90.1 79.0 87.0 85.2 Do RA*k=16 29.1M 74.5 88.8 80.3 95.5 84.7 90.1 79.1 87.2 85.0 0.3 x Sine Do RAk=16 75.1 89.0 81.0 95.3 86.1 90.0 79.3 86.2 85.3 Do RA*k=32 57.4M 74.6 89.3 79.9 95.5 85.6 90.5 80.4 85.8 85.2 0.1 x Sine Do RAk=32 75.8 89.3 80.3 95.9 86.1 90.2 79.4 85.4 85.3 Table 8: Sine Do RA. Frequency, learning rate and epochs for the Llama3-8B model. Rank Frequency Learning Rate Epoch 8 300 6e-5 3 16 150 6e-5 3 32 100 6e-5 3 Implementation details for Vi T-Base on Image Net-1k: We followed the settings in (He et al., 2022a). We use batch size of 1024, learning rate of 3e-4 and we train for 300 epochs. Choices of frequency for different ranks are shown in Table 9. Table 9: Frequencies used for different ranks for Vi T-Base and Vi T-Small. rank 1 5 10 30 60 100 150 250 Vi T-small (CIFAR100) ω 500 500 300 300 300 - - - Vi T-Base (Image Net-1k) ω 1000 800 800 400 200 200 150 150 Ablation on frequencies: In table 10, we examine the performance of our method on training the Vi T-Small model from scratch on the CIFAR100 dataset using different frequencies, when k = 1. Table 10: Top-1 Accuracy of Vi T-Small (k = 1) on CIFAR100 with varying frequencies ω. Frequency ω 100 200 300 400 500 600 700 PSNR 55.0 55.8 56.8 58.0 58.1 57.6 57.5 A.2.3 CONVNEXT ON CIFAR100 Conv Ne Xt is a family of convolutional neural networks (CNNs) models introduced in (Liu et al., 2022). These models are designed to modernize traditional CNNs architectures by incorporating design elements inspired by Vision Transformers (Vi Ts) to enhance performance in image recognition. Implementation details: We employ our methods on Conv Ne Xt-Tiny model using CIFAR100 datasets and the Timm codebase. Conv Ne Xt-Tiny consists of 4 stages with block numbers [3, 3, 9, 3] and feature dimensions with [96, 192, 384, 768]. The majority of parameters (50%) in Conv Ne Xt are used in the last stage, therefore we apply a low rank decomposition only to the linear feature layer in this stage. We use a batch size of 512, learning rate of 5e-3, and we train for 150 epochs. Results: In Table 11, we present our performance and configurations. We demonstrate that our method consistently outperform the naive low rank method, and even outperforms the baseline fullrank method (Conv Next-Tiny) with approximately 50% fewer parameters. Published as a conference paper at ICLR 2025 Table 11: Performance and compression rate of Conv Ne Xt-Tiny model trained on CIFAR100 datasets. We use frequency [400, 300, 300] for rank [1, 5, 20] respectively. Method # Params Acc % Compression Rate % Conv Ne Xt-Tiny 27.9M 62.3 100 LR-Conv Ne Xt-Tinyk=1 13.8M 59.5 49.5 Sine LR-Conv Ne Xt-Tinyk=1,ω=400 61.5 LR-Conv Ne Xt-Tinyk=5 13.9M 59.5 50.0 Sine LR-Conv Ne Xt-Tinyk=5,ω=300 62.0 LR-Conv Ne Xt-Tinyk=20 14.2M 62.1 50.9 Sine LR-Conv Ne Xt-Tinyk=20,ω=300 62.8 Implementation details: We followed the settings in (Ramasinghe & Lucey, 2022). We use 8 fully connected layers each with 256 neurons, a learning rate of 5e-4 and train for 500k iterations. We evaluate the performance of our method by experimenting with different ranks [1, 5, 10, 30, 60], corresponding to frequencies [1400, 800, 600, 400, 300] respectively. Results: The full qualitative results on Ne RF are given in figure 8. Ablations: In figure 7, we illustrate the impact of varying frequency on PSNR for cases where k=1 (shown on the left) and k=5 (shown on the right). Computational cost: In Table 12, we present the training memory usage (MB) on Ne RF experiments across different ranks. Figure 7: Ablation Ne RF results for the LLFF dataset. These two figures show PSNR of Ne RF using different frequencies, when k = 1 (on the left) and k = 5 (on the right) Table 12: Training memory usage(MB) on Ne RF experiments Method rank 1 rank 5 rank 10 rank 30 rank 60 Low-Rank 5620 5622 5622 5626 5630 Sine Low-Rank 5630 5630 5632 5634 5638 Published as a conference paper at ICLR 2025 Figure 8: Qualitative Ne RF results for the LLFF dataset (Mildenhall et al., 2019; 2020) using rank k = 1 and k = 5. Using a low-rank model leads to a complete loss of signal for k = 1, however, applying sine is able to reconstruct details even at the extreme low-rank case. At k = 5 the sine low-rank model is noticeably sharper and clearer than using the low-rank. Low-Rank Low-Rank sine-Low-Rank (Ours) Fern Flower Fortress Leaves Horns Orchids Trex Room sine-Low-Rank (Ours) k = 5 Published as a conference paper at ICLR 2025 Figure 9: Ablation binary occupancy results for Thai Statue. This figure shows Io U accuracy of 3D shape modeling using different frequencies, when k=1. A.2.5 3D SHAPE MODELLING Implementation details: We use 2 fully connected layers each with 256 neurons, a learning rate of 1e-3 and train for 200 epochs. Results: Table 13 reports the Intersection over Union (Io U) and Compression Rate of the binary occupancy task using different rank levels (k). Our sine low-rank methods. Table 13: This table illustrates Intersection over Union for 3D shape modeling (Thai Statue, Lucy and Dragon) across different rank levels (k). It also includes the compression rate, indicating the proportion of parameters utilized relative to the total parameter count of the baseline model, thereby detailing the parameter usage versus model performance at different levels of model complexity. We use frequency [200, 100, 50, 20] for ranks [1, 2, 5, 20], respectively Io U Compression # Params Thai Lucy Dragon Rate Full-Rank 132K 97.2 97.8 98.7 100% Low-Rankk=1 2.8K 84.3 79.3 90.4 2.1% Sine Low-Rankk=1,ω=200 90.8 90.7 94.6 Low-Rankk=2 3.8K 88.0 89.4 90.9 2.9% Sine Low-Rankk=2,ω=100 92.0 93.2 96.6 Low-Rankk=5 6.9K 93.4 94.8 96.9 5.2% Sine Low-Rankk=5,ω=50 94.3 95.3 97.4 Low-Rankk=20 22.8K 95.4 96.2 98.0 16.8% Sine Low-Rankk=20,ω=20 95.4 96.3 98.1