# accelerated_training_on_lowpower_edge_devices__1235ddb6.pdf Published in Transactions on Machine Learning Research (10/2025) Accelerated Training on Low-Power Edge Devices Mohamed Aboelenien Ahmed mohamed.ahmed3@kit.edu Karlsruhe Institute of Technology Kilian Pfeiffer kilian.pfeiffer@kit.edu Karlsruhe Institute of Technology Osama Abboud osama.abboud@huawei.com Huawei Research Center Munich Ramin Khalili ramin.khalili@huawei.com Huawei Research Center Munich Heba Khdr heba.khdr@kit.edu Karlsruhe Institute of Technology Jörg Henkel henkel@kit.edu Karlsruhe Institute of Technology Reviewed on Open Review: https: // openreview. net/ forum? id= c Gj Q41j BEn Training on edge devices poses several challenges as these devices are generally resourceconstrained, especially in terms of power. State-of-the-art techniques at the device level reduce the GPU frequency to enforce power constraints, leading to a significant increase in training time. To accelerate training, we propose to jointly adjust the system and application parameters (in our case, the GPU frequency and the batch size of the training task) while adhering to the power constraints on devices. We introduce a novel cross-layer methodology that combines predictions of batch size efficiency and device profiling to achieve the desired optimization. Our evaluation on real hardware shows that our method outperforms the current baselines that depend on state of the art techniques, reducing the training time by up to 2.3 with results very close to optimal. Our measurements also indicate a substantial reduction in the overall energy used for the training process. These gains are achieved without reduction in the performance of the trained model. 1 Introduction Recent trends are moving toward the deployment of deep learning models on edge devices due to the need for real-time decision-making, reduced communication cost, and privacy compared to offloading computation to cloud-based servers Wang et al. (2018); He et al. (2018); Chen & Ran (2019); Ganesh et al. (2022). Low-latency inference is essential in applications such as autonomous vehicles, robotics, surveillance, and smart agriculture. To meet these requirements, edge devices are equipped with compact GPUs optimized for computer vision and deep learning inference tasks Maghazeh et al. (2013); Barba-Guaman et al. (2020); Saddik et al. (2021); Fernández-Sanjurjo et al. (2021); Zhang et al. (2018); Ganesh et al. (2022). Such devices are often power constrained, running on batteries with limited power outputs. Many applications on edge devices involve sensitive or private information that cannot be shared due to regulatory or ethical concerns, making it challenging to maintain accurate models without access to this data. Training on the edge offers adaptability to the data on these devices without transmitting it, reducing exposure risk. Transfer learning is thereby applied for on-device training, where a model is pre-trained with Published in Transactions on Machine Learning Research (10/2025) Power Limit = 5 W (128, 307 MHz) (64, 460 MHz) 4 8 16 32 64 128 153 GPU frequency [MHz] 4 8 16 32 64 128 153 GPU frequency [MHz] Time (Ts)[s] Figure 1: Peak power and time for a single training pass over a set of samples across different batch size and GPU frequency combinations. The gray plane represents the power limit in the left figure, and the black dots in both figures are the feasible combinations that can be utilized under that constraint. The black circle represents the operating point with maximum feasible frequencies for the batch sizes of 128, which will be selected by the state-of-the-art techniques. The green circle represents an operating point at batch size 64, that could be selected when the frequency and the batch size are jointly selected to accelerate training under the power constraint. Selecting this operating point accelerates training by 31.9% (see Time curves). a large (public) dataset and fine-tuned on-device on the target domain Cai et al. (2020); Chiang et al. (2023); Lin et al. (2022); Yang et al. (2023). While this offers less training time compared to training from scratch, transfer learning is still a computationally demanding task utilizing higher power compared to inference. To accelerate training, usually large batch sizes for mini-batch gradient descent are used. However, increasing the batch size increases the power consumption which is limited at edge devices. To enforce the power constraint, the circuit-level control will select the frequency level that satisfies this constraint considering the worst-case computation scenario. Intuitively, this will decelerate training: in the state of the art, the system-level and application-level (training) parameters are set independently from each other, and this is sub-optimal as will be demonstrated in the following example. Figure 1 shows the power and time required to train a Res Net18 He et al. (2016) model on Nvidia Jetson Nano with 4096 of CIFAR10 s training samples, while using different combinations of the GPU s frequency and batch size. Following the state of the art, and for a power constraint of 5 W, the system would select the GPU s frequency f = 307 MHz and the batch size b = 1281. When we look at the training time, this selection is sub-optimal. In particular, we could choose a smaller batch size (b = 64), allowing the GPU to operate at a higher frequency (f = 460 MHz), and thus reducing the training time by 31.9%. This shows that to accelerate the training on low-power edge devices, both the GPU frequency (a system parameter) and the batch size (an application parameter) should be jointly selected. Motivated by this observation, we introduce, for the first time, a cross-layer approach that simultaneously optimizes both the GPU frequency and the batch size used for training, aiming to accelerate the process while adhering to device power restrictions. Note that the example in Fig. 1 illustrates the impact of batch size on a single training pass. However, modifying the batch size in training (in contrast to the inference process) influences the number of iterations needed to reach the targeted accuracy, and consequently the overall training time. This presents an additional dimension to the problem, which we will consider in our approach as detailed in Section 3. Furthermore, our cross-layer optimization is capable of leveraging operating points within the design space that previous works have not considered. This approach facilitates a marked decrease in the necessary training time to accuracy. 1We select the maximum batch size which is allowed by the memory at the device. Published in Transactions on Machine Learning Research (10/2025) Our solution incorporates offline profiling of Neural Network (NN) training on the device, where we assess power and time demands for different combinations of batch sizes and GPU frequencies. Furthermore, we estimate how batch size correlates to the number of iterations required to attain a desired accuracy by training on a proxy dataset on a server. By merging these insights, we identify the combination that minimizes overall training time on the device while respecting its specified power constraints. In summary, we make the following novel contributions: We motivate and demonstrate the importance of the joint optimization of GPU frequency and batch size to accelerate on-device training under a power constraint, as shown across representative models (CNNs and Transformers) and tasks (image classification and next character prediction). We introduce a new cross-layer approach that performs the discussed joint optimization, taking into account the effect of batch size on the overall training time, aiming to accelerate on-device training while complying with the device s power limitations. Our evaluation on real hardware under different power constraints for image classification task on SVHN Netzer et al. (2011) and CINIC Darlow et al. (2018) datasets, using models Res Net18, Mobile Net V2 Sandler et al. (2018), and Efficient VIT Liu et al. (2023)2, and next character prediction task using a transformer network shows a significant improvement in the training speed by an average gain of 1.5 and peak of 2.3 compared to baselines. We observe a reduction in the total energy used for the training at the device. We provide a comprehensive sensitivity analysis that demonstrates the robustness of our solution towards the proxy dataset selection over five image classification datasets and three next character prediction datasets, validating the practicality of our approach and ensuring its effectiveness. 2 Background and Related Work This section discusses the control parameters (i.e., GPU frequency and batch size) used in this work and the state-of-the-art techniques. We do not consider training on the CPU for the GPU-equipped devices, as the GPU is an order of magnitude faster (and energy efficient) while consuming nearly the same average power (see Appendix A.1 for some related discussion). 2.1 Voltage Frequency Scaling (VFS) VFS is a system-level technique used for power management in processors Cochran et al. (2011); Lee et al. (2011). It adjusts the voltage and frequency at runtime, providing a trade-off between performance and energy efficiency Le Sueur & Heiser (2010); Guerreiro et al. (2019). Reducing each of the processor s frequency f and voltage V values can result in a cubic reduction in the dynamic power, i.e., Pdynamic f V 2. Within a device, the chip manufacturer provides a discrete set of frequencies to operate with, where fi, i {1, 2, . . . , max}. The voltage is a function of the operating frequency f, and will be updated automatically for any given frequency. To enforce power constraints on processors, the industry standard technique is to have an upper bound f (and V ) for the device to use under a power limit. VFS is also applied to the CPU s frequency, but this is beyond the scope of our work as we train on the GPUs. VFS has been studied in Nabavinejad et al. (2019); Liu & Karanth (2021) for improving performance and energy efficiency for inference only. Tang et al. (2019) conducted a study on the impact of GPU VFS on performance and energy for both training and inference. However, the study mainly focused on the computational perspective without addressing accuracy and training speed till convergence. In contrast, we show how changing the frequency and batch size due to power limits, accompanied by the difference in iterations to reach accuracy, can lead to different optimal configurations in training. 2The selected models and tasks align with existing literature for training on edge devices Ponzina et al. (2023); Choi & Sobelman (2022); Jia et al. (2024); Pfeiffer et al. (2023), and Jetson Nano and TX2NX represent typical edge GPU platforms. Published in Transactions on Machine Learning Research (10/2025) 2.2 Batch Size One of the state-of-the-art techniques for training NNs is the minibatch gradient descent, where gradients are computed using the samples drawn in mini-batch to approximate the overall gradient, enabling iterative updates of the model parameters. The latency for processing a batch is influenced by the GPU s ability to parallelize computations and its memory bandwidth. Increasing the batch size generally decreases the latency of processing data samples by parallelizing more computations on multiple processing units. Nevertheless, this increases the number of computational operations on the GPU, and hence the power consumption. The impact of batch size on accuracy and convergence speed is explored from multiple aspects. Goyal et al. (2017); Krizhevsky (2014); Zhang et al. (2019) studied the interdependencies between batch size and other hyperparameters such as learning rate, weight decay, and optimizers. Smith et al. (2018) proposed to increase the batch size during training rather than decreasing the learning rate to leverage the regularization effect employed by larger batch sizes. Mc Candlish et al. (2018) proposed a gradient noise scale metric to predict the most efficient batch size that maximizes the throughput in the next training step. This approach is tailored for distributed learning, where a batch of data is split across multiple devices. These aforementioned works have not considered the power efficiency of batching on the device. In contrast, the works in Nabavinejad et al. (2021; 2022) consider the impact of the batch size on power consumption of inference operations. In particular, a binary search approach is employed in Nabavinejad et al. (2022) to find an appropriate batch size that accelerates inference and then the GPU frequency is adjusted accordingly to satisfy power constraints on GPU servers. Most importantly, this approach only considers inference where the statistical impact of the batch size does not exist. You et al. (2023) aimed at optimizing the energy efficiency of periodic training jobs (continuously re-executed for incoming data flow) on powerful GPU clusters. To achieve their optimization (which is different from the one we target in this paper), they propose to adjust the batch size, set a power limit on the server, and depend on circuit-level control to select the frequency that satisfies that power limit. In particular, they obtain Pareto-optimal combinations of batch sizes and power limits that optimize for energy and performance through profiling the whole training job on the server, since this job will be repeated for new data flows. However, this solution can be applied on powerful GPU clusters and not on edge devices with limited resources. In summary, none of the state of the art has addressed the joint selection of the batch size and GPU frequency to accelerate training at edge devices under power constraints. 3 Problem Statement We consider the following scenario: for a specific training task, an edge device requests a pre-trained NN model M with its weights θ from a server in order to fine-tune it on local data D till reaching a given accuracy threshold. Importantly, the edge device has a power constraint Pmax, which should not be exceeded during the training process. Our goal in this paper is to minimize the training (fine-tuning) time at edge devices under their given power constraints. We introduce Ts as the training time required to apply training using a fixed number of samples s. As shown in Fig. 1, the joint selection of b and f will help reduce Ts under a power constraint. However, the ultimate optimization goal is to minimize the total training time required to reach a target accuracy, which we label TTacc. A set of parameters, i.e., frequency and batch size (f, b), that are optimal for training a fixed number of samples (Ts) might not be necessarily optimal for the training to accuracy (TTacc). We display in Fig. 2 the training time to reach an accuracy of 78% ( TTacc) for Res Net18 using two batch sizes of b1 = 8 and b2 = 32 under three different power constraints (i.e., P1 = 4.5 W, P2 = 5 W, and P3 = 7 W). We notice that for the three power constraints, selecting b1 allows to utilize a higher frequency than b2. For each batch size, we select the highest frequency that satisfies the power constraint, and measure Ts and TTacc. We observe that using b2 (the higher batch size) always leads to a lower Ts. However, selecting the same batch size over b1 leads to a longer TTacc for P1 and P2, and shorter TTacc for P3. This shows the complexity of the targeted problem. In particular, TTacc does not only depend on Ts, but it also depends on the number of times of processing s to reach target accuracy (Nsacc). In this example, Nsacc Published in Transactions on Machine Learning Research (10/2025) GPU Frequency [MHz] 153 307 460 614 768 921 GPU Frequency [MHz] 153 307 460 614 768 921 TTacc (b1) TTacc (b2) Ts Feasible Operating Points(b, f) for the power constraints Figure 2: The training time on fixed number of samples Ts and the total training time TTacc to reach an accuracy threshold of 78% using two batch sizes 8 and 32, while considering the maximum feasible GPU frequencies under three power constraints; P1 = 4.5 W, P2 = 5 W, and P3 = 7 W. We observe that for P1 and P2, selecting b = 8 will lead to lower TTacc, while for P3 selecting b = 32 is better. This is in contrast with our observation for Ts, where selecting b = 32 is the best option in all cases. for b2 is equal to 15 while Nsacc for b1 is equal to 10. These values and the effect of power constraint on the feasible frequency highly influence the optimal batch size to minimize TTacc. In summary, there is no clear indication on how to select the optimal operating points (f, b) to achieve the target goal. We formulate our optimization problem as follows: min b B,f F TTacc(b, f, M, D) subject to P(b, f, M) Pmax (1) where B is the set of feasible batch sizes, F is the set of available GPU s frequencies, and P(b, f, M) is the required power to training M using b and f. We rewrite TTacc as the multiplication of Ts and Nsacc, we thus have: TTacc(b, f, M, D) = Ts(b, f, M) Nsacc(b, M, D) (2) s is selected, s.t. bmax s |D|, where bmax is the largest batch size that can fit into the memory of the devices. This detached formulation enables our proposed optimization method, presented in Section 4. In particular, the first factor Ts does not depend on the training data D, nor on the accuracy threshold. The second factor Nsacc is independent of the GPU frequency of the device. 4 Power-Aware Training We propose an optimization approach that co-selects b and f to minimize TTacc on an edge device under Pmax. Following the problem split that we propose in Equation (2), our solution consists of two main parts. The first one considers device specifics, and involves measuring time and power for training a given model M, i.e., Ts(b, f, M) and P(b, f, M) for every b B and f F. The second part is responsible for estimating the efficiency of batch sizes, i.e., their impact on Nsacc. This part does not depend on device characteristics and it is computationally expensive, therefore we consider it to be performed on the server that is responsible for pre-training and sending of the model. The workflow for our approach is as follows: The device first sends a request to the server for a pre-trained neural network model (M) with a specific architecture and input size. The server responds with the requested model and its pre-trained weights. Our proposed power-aware training then begins. The device profiles the model in terms of time and power as will be described in Section 4.1, while the server estimates Nsacc for various batch sizes, as will be discussed in Section 4.2, and sends these estimations to the device. Based on the profiling and estimations, the device selects the best combination of batch size and frequency to minimize TTacc. An overview is presented in Figure 3. Published in Transactions on Machine Learning Research (10/2025) Train with different batch sizes on proxy dataset and approximate Ns( ) Train for few iterations and record P(b, f), Ts(b, f) Store in lookup tables LUTpower[b, f] = P(b, f) LUTtime[b, f] = Ts(b, f) Configuration Selection (Section 4.3) Set GPU freq Train GPU Server Edge Device Proxy Dataset Trainer Figure 3: Overview of our proposed cross-layer approach that accelerate training under power constraint though the joint selection of batch size b and GPU frequency f. 4.1 Profiling The power and time for training on a device also depend on the NN s architecture M, in addition to the frequency and batch size. Besides, changing the input shape, such as the image size in vision tasks, can lead to different time and power requirements even when using the same NN architecture. Furthermore, the time and power are independent of the actual training data. Therefore, profiling Ts(b, f, M) and P(b, f, M) given the task s input shape can be performed before full data acquisition, as long as the input dimensions of the data samples are known. Processing a few mini-batches is adequate to obtain accurate profiling. For each batch size and frequency combination, we set the GPU frequency to f and process a few mini-batches m of size b. The power sensor values are monitored to extract the peak power. Furthermore, the processing time is recorded, and then the average processing time for a mini-batch is calculated. This average processing time is then scaled to the processing time for s samples, denoted as Ts. Ts(b, f, M) and P(b, f, M) are then stored in two lookup tables, denoted as LUTM Time and LUTM Power. The proposed profiling strategy considers hardware and NN structures. Also, the time and power values are not affected by the network weight updates; thus, profiling is applicable before receiving pre-trained weights if M is available at the device and is needed to be perdomed only once. In a future and more practical setting, we could assume such profiling to be provided by the device manufacturer and for a pre-defined set of tasks and models. This can be further supported by periodic validation of current values with those recorded in the LUTs. If a deviation is detected, specifically relative performance change between operating points (e.g., due to memory or cache contention), a re-profiling procedure is initiated to accommodate the change, requiring only a few batches per operating point, that is much cheaper than full training. This process can be further optimized by interpolating the deviations across operating points, thereby enabling efficient updates to the LUTs with minimal computational overhead. Furthermore, if Pmax is known at the device, the profiling can be performed more efficiently as follows. Given that B and F are sorted, and power consumption increases with both b and f, we begin by profiling the largest b with the minimum f, incrementing f until the maximum feasible value under Pmax is reached. Next, we move to the second largest b, starting from the highest feasible frequency found for the previous batch size, and repeat the process until the smallest batch size is profiled. This profiling is equivalent at most to 1.9% and 6% of the training times performed on Res Net18 and Mobile Net V2 in Section 5.2. Published in Transactions on Machine Learning Research (10/2025) 4.2 Estimation of Nsacc for Batch Sizes As discussed in Section 3, batch sizes have different efficiency in terms of Nsacc(bi, M, D). Estimating Nsacc(bi, M, D) for every bi B depends also on the training data that is only available at the device and M s pre-trained weights. To solve Equation (1), the exact number of samples processed to reach target accuracy for each batch size can be replaced with the relative ratio between batch size rbi (i.e., normalized to the maximum Nsacc) as follows: rbi = Nsacc(bi, M, D) max b B Nsacc(b, M, D) (3) With this simplification, we adjust our focus to estimate the relation vector between batch sizes such that r = (rb1, rb2, . . . , rbmax), where rbi (0, 1]. However, this is still a complex task to solve given the non-linear training dynamics of deep learning; especially as the convergence speed of every batch size changes across training, making it impossible to estimate with few probes of multiple batch sizes given the training state. Obviously, training till convergence for multiple batch sizes is computationally-expensive task and cannot be conducted on the device. These all make on-device estimation of r inaccurate, if not infeasible. We thus propose to estimate r on a powerful GPU server. Particularly, M is trained on the server with multiple batch sizes until convergence (reaching the target accuracy) using a proxy dataset DS since the server does not have access to D. Each batch size run starts from the same pre-trained weights and yields its corresponding Nsacc on DS. This systematic exploration allows us to comprehensively assess the longhorizon impact of the batch size on the model s convergence while leveraging the computational capabilities of the server, along with the datasets available on it and augmentation techniques. Since M, which would be used on the edge device, is already designed for a specific task type and pre-trained on a public dataset, a proxy dataset DS should share the same task type (e.g., image object classification) and similar input shapes. Training network M on DS, despite having different objectives, allows us to estimate the relationship between batch sizes and their relative examples to accuracy on D. Thus, we can finally have a mapping such that r DS r D. By estimating the batch size relation vector r on the server, any edge device aiming to train M to utilize this vector. In Section 5.2, we provide evaluation for two different devices, namely Nvidia Jetson Nano and Nvidia Jetson TX2NX, utilizing r. 4.3 Batch Size and Frequency Selection The device profiling and estimation of batch size efficiently are performed in an offline manner and at the design time. In contrast, the configuration selection is performed at runtime, as described below. Given a power constraint Pmax, and the power measurements stored for training a specific M in LUTM Power, we construct a set of feasible combinations C consisting of every feasible batch size with its corresponding highest (and fastest) frequency satisfying Pmax as follows: C {(i, ji))|i [1, . . . , |B|]}, ji max{j|j [1, . . . , |F|], LUTM Power(i, j) < Pmax} (4) The processing time for the feasible combinations is then extracted from LUTM Time. Following this, an approximate training time is computed by multiplying the time for every bi (and frequency) by the corresponding ri element from the relation vector r (i.e., estimated at design time). The configuration (b , f ) that minimizes the approximate total training time is selected. Finally, we set the GPU frequency to f and then start the training using batch size b . We provide the configuration selection in Algorithm 1. The selection part is O(n2) in the worse case scenario (O(|B| |F|)). Since B and F are sorted, and for each b, the power/time over F are also sorted, the selection reduces to O(n log n) via binary search and is negligible to training time. In this section, we evaluate the training time and energy consumption for finetuning vision and text tasks on Nvidia Jetson Nano and TX2NX. Additionally, we assess the effectiveness of our approach by conducting a sensitivity analysis on the selection of the proxy dataset. Published in Transactions on Machine Learning Research (10/2025) Algorithm 1 Batch size and GPU frequency selection Require: Power Limit Pmax, Neural Network M, List of feasible batch sizes B, List of GPU frequencies F, Dataset at server DS 1: Server: 2: r Proxy(B, M, DS) See section (Section 4.2) 3: 4: Device: 5: Let LUTM Time R|B| |F| 6: Let LUTM Power R|B| |F| 7: for i 0, 1, ..., |B| do 8: b B[i] 9: for j 0, 1, ..., |F| do 10: f F[j] 11: time, power Profile(b, f, M) On Device (Section 4.1) 12: LUTM Time[i, j] time 13: LUTM Power[i, j] power 14: // On-device Configuration Selection: 15: TTacc = LUTM Time r Element-wise multiplication 16: C Get Combinations(B, F, LUTM Power, Pmax) Based on Equation (4) 17: TT p acc TTacc[C] 18: (b_idx, f_idx) argmin(TT p acc) 19: b , f B[b_idx], F[f_idx] Selected batch size and GPU frequency 5.1 Experiment Setup Datasets and models: We evaluate our approach for the image object classification and next-word prediction tasks. For image classification, a model is pre-trained with the full CIFAR100 Krizhevsky et al. (2009) and to be trained on subsets (i.e., quarter) of SVHN and CINIC datasets on the device. We use a subset of CIFAR10 Krizhevsky et al. (2009) as the proxy dataset on the server. For all datasets, the input images are of size 3 32 32. Each dataset (subset) is divided into training and testing sets with an 80/20 split, with target accuracy evaluated on the test split. We evaluate Res Net18, Mobile Net V2 and Efficient Vi T-(M1) models, trained with Adam optimizer Kingma & Ba (2015). A summary of the experimental settings is provided in Table 1. For the next character prediction task, we evaluate a 6 layer transformers with 6 attention heads per attention block, 256 embedding dimensions, and a sequence length of 64. We use Adam W Loshchilov & Hutter (2017) as an optimizer. We pre-train the model on Wiki Text-2 dataset Merity et al. (2016), utilize tiny shakespeare Karpathy (2015) as a proxy dataset, and train it on some Jane Austin and Charles dickens novels3. We use 90% of a dataset for training and the rest for testing. For all datasets, we fix the vocabulary to include only words with English letters, digits, punctuation, spaces, and new lines. We set the target character level accuracy for Austin and Dickens at 62% and 61%, respectively. Batch size and learning rate: The choice of appropriate learning rate and batch size are often intertwined, as they impact each other s effectiveness. Larger batch sizes provide more stable gradient estimates, potentially permitting the use of higher learning rates. Therefore, to preserve the performance of deep models with different batch sizes, we apply learning rate scaling (i.e., square root scaling Krizhevsky (2014) for Adam and Adam W). For Res Net18, the batch sizes ranged from 4 to 128, consisting exclusively of powers of two. The initial learning rate of 5 10 4 is used for the largest batch size of 128 (with learning rates scaled for other batch sizes). The same setup was also applied to Mobile Net V2 and Efficient Vi T; however, the batch size of 128 was omitted due to memory constraints. For transformers, we similarly consider batch sizes of 4 to 128 with a learning rate of 1 10 3 for the batch size of 128. 3We used the text for works of Jane Austin s from nltk package Bird et al. (2009) and downloaded Dickens works from project Gutenburg. More details are provided in Appendix A.3 Published in Transactions on Machine Learning Research (10/2025) Table 1: Experimental setup for image classification datasets and models used. Dataset Dataset size Target Acc. Res Net18 Mobile Net V2 Efficient Vi T Cifar10 12500 87% 84.5% 84% SVHN 18315 92% 91.5% 90% CINIC 22500 76% 74% 72% Hardware and power limits: We evaluate our method on Nvidia Jetson Nano with 4 GB memory on three scenarios. In the first and second scenarios, the power limits at the device are set to P 1 max = 4.5 W and P 2 max = 7 W. In the third, the device operates without any power limits, denoted as N/A. To show that our solution and our results are not device specific, we also provide evaluation on another device (i.e., Nvidia Jetson TX2NX). We use Py Torch 1.10 Paszke et al. (2019) for Jetson Nano and TX2NX. For pretraining and proxy datasets training we use NVIDIA A6000 GPU with Pytorch 2.1. Comparison baselines: To highlight the impact of joint selection of GPU frequency and batch size, we compare our approach to the following baselines that depend on state of the art techniques: Baseline 1: The state of practice is to use the largest b that can fit into memory Goyal et al. (2017); Camelo & Cretu (2023), where the latter use edge GPUs. For the three power limits, we use 307 MHz, 614 MHz, and 921 MHz as upper-bound operating GPU frequencies for the device. These are determined based on profiling training of different models and selecting the frequencies that assure a power limit is satisfied irrespective of what model or b is used for the training. Baseline 2: We select the value of b that minimizes Nsacc on the proxy dataset, but we use the same GPU frequencies as in Baseline 1, so no joint optimization is applied. Fastest configuration: This baseline serves as an upper bound where optimal f and b are jointly selected. Furthermore, this baseline validates the importance of jointly selecting f and b, as evidenced by its performance gap compared to other baselines under power constraints. Since the optimal configuration is not known prior to training, we determine it by exhaustively training the given model on the given dataset with all batch sizes to get Nsacc, substitute in Eq. (2) and finally select the best b and f configuration on the device. We repeat every experiment with different five seeds for image classification tasks and three seeds for next character prediction and record the mean and standard deviation. 5.2 Training Time Evaluation Table 2 and Table 3 report the training time comparison of our approach and the three baselines on Nvidia Jetson Nano. Table 4 provides additional comparison for image classification task on Nvidia Jetson TX2NX. Firstly, we start with the evaluation of image classification tasks reported in Table 2. For Res Net18 training, our method consistently outperforms baseline 1, achieving 1.5 2.3 reduction in training time. It also outperforms baseline 2 with 1.4 1.7 less training time. These two baselines are only able to explore a small subset of the design space, potentially missing the optimal configuration. Finally, compared to the fastest configuration, our method selects the same configurations in most of the evaluations, while in the case of a different selection, a minor performance decline (of 3%) is observed. Training of Mobile Net V2 has three distinct characteristics: lower peak power, better parallelization for larger batch sizes, and different r. Our method adapts to that by selecting larger batch sizes (and frequencies) than those selected for Res Net18. Compared with baseline 1, our method performs better with minor differences in training time since both select large batch sizes and maximum f utilizing the power. In low-power, the gain from our approach increases as the selected batch size enables higher frequencies, reducing training time by up to 1.4 . Baseline 2 misses the opportunity to use higher frequencies, especially as Mobile Net V2 consumes less power. Published in Transactions on Machine Learning Research (10/2025) Table 2: Training time comparison with baselines and upper bound over three different power limits (i.e., P 1 max, P 2 max, and N/A) on Jetson Nano, showing a peak gain of 2.3 . Recorded times are in seconds. CIFAR10 is used as proxy dataset in our proposed approach. (Evaluation time is excluded) Model Method SVHN CINIC P 1 max P 2 max N/A P 1 max P 2 max N/A Res Net18 Baseline 1 14347 3796 8204 2170 6187 1637 18154 4109 10381 2349 7829 1772 Res Net18 Baseline 2 11345 1616 7055 1005 5817 828 12436 1443 7733 897 6377 740 Res Net18 Ours 8477 1207 4607 725 4170 656 9292 1078 4454 788 4032 713 Res Net18 Fastest configuration 8477 1207 4607 725 4170 656 9030 1598 4454 788 4032 713 Mobile Net V2 Baseline 1 8708 2396 5107 1405 3990 1098 8909 1121 5226 657 4082 513 Mobile Net V2 Baseline 2 8082 912 4871 550 3912 441 8428 692 5079 417 4080 335 Mobile Net V2 Ours 5940 916 3912 441 3912 441 6194 509 4080 335 4080 335 Mobile Net V2 Fastest configuration 5940 916 3912 441 3912 441 6194 509 4080 335 4080 335 Efficient Vi T Baseline 1 13033 1233 7839 741 6369 602 8100 564 4800 339 3900 275 Efficient Vi T Baseline 2 14539 663 9269 422 7843 357 9223 584 5880 372 4975 315 Efficient Vi T Ours 11036 503 6925 655 6369 602 7001 444 4240 299 3900 275 Efficient Vi T Fastest configuration 11036 503 6925 655 6369 602 7001 444 4240 299 3900 275 Table 3: Training time comparison with baselines and upper bound for different power limits (i.e., P 1 max, P 2 max, and N/A). Recorded times are in seconds. Tiny Shakespeare is used as proxy dataset in our approach. Model Method Austen Dickens P 1 max P 2 max N/A P 1 max P 2 max N/A Transformer Baseline 1 8136 524 4607 296 3556 229 12347 439 6991 249 5397 192 Transformer Baseline 2 9425 131 6081 85 5193 72 14278 228 12941 532 7867 125 Transformer Ours 7232 101 3982 256 3556 229 10956 175 6043 215 5397 192 Transformer Fastest configuration 7232 101 3982 256 3556 229 10956 175 6043 215 5397 192 Table 4: Training time comparison with baselines and upper bound over three different power limits (i.e., P 1 max, P 2 max, and N/A) on Nvidia Jetson TX2NX. Recorded times are in seconds. CIFAR10 is used as proxy dataset in our proposed approach. (Evaluation time is excluded) Model Method SVHN CINIC P 1 max P 2 max N/A P 1 max P 2 max N/A Res Net18 Baseline 1 7343 1943 3944 1043 3021 799 9292 2103 4991 1129 3823 865 Res Net18 Baseline 2 7337 1943 4771 679 4086 582 8042 933 5230 897 4479 520 Res Net18 Ours 4283 708 2648 417 2431 382 5463 889 2560 453 2350 416 Res Net18 Fastest configuration 4283 708 2648 417 2431 382 5046 893 2510 625 2294 571 Ultimately, our method selects the fastest configurations when training Mobile Net V2 in all power-limits scenarios. Similarly, our method results in less training times compared to baselines 1 and 2 under P 1 max and P 2 max when finetuning Efficient Vi T, while selecting the fastest configuration in all scenarios. Next, we examine the next character prediction training. As shown in Table 3, our method improves the training time by 1.15 and 2.14 compared to baseline 1 and 2, while choosing the same configurations as the fastest configuration. In Table 4, we compare our solution with the baselines for Res Net18 training on Nvidia Jetson TX2NX, where the device receives r for Res Net18 training from the server. For training on SVHN, our method outperforms both the baselines 1 and 2, reducing the training time by peak of 1.7 and 1.8 , respectively. Furthermore, it selects the same configuration as the optimal one. Similarly, when training on the CINIC dataset, baselines 1 and 2 require up to 1.94 and 2.04 more training time compared to ours. Our method did not select the exact optimal configuration in this scenario; however, it selects a near-optimal configuration that results in only 2% 8.2% more training time compared to optimal. Published in Transactions on Machine Learning Research (10/2025) P 1 max P 2 max N/A 0 2 4 6 Energy [W s] Res Net18/SVHN Baseline 1 Baseline 2 Ours Fastest Configuration P 1 max P 2 max N/A 0 2 4 6 8 Energy [W s] Res Net18/CINIC P 1 max P 2 max N/A 0 Energy [W s] Mobile Net V2/SVHN P 1 max P 2 max N/A 0 1 2 3 4 104 Energy [W s] Mobile Net V2/CINIC P 1 max P 2 max N/A 0 Energy [W s] Efficient Vi T/SVHN P 1 max P 2 max N/A 0 Energy [W s] Efficient Vi T/CINIC P 1 max P 2 max N/A 0 1 2 3 4 Power Constraint Energy [W s] Transformers/Austen P 1 max P 2 max N/A 0 Power Constraint Energy [W s] Transformers/Dickens Figure 4: Energy consumption comparison between our approach and baselines during training under three power scenario, using Res Net18, Mobile Net V2, and Efficient VIT on the SVHN and CINIC datasets, and a transformers network on Austen and Dickens datasets, performed on Nvidia Jetson Nano. In summary, our method outperforms existing baselines, reducing training time to accuracy across various model architectures, tasks, and hardware, with an average speedup of 1.5 to baselines 1 and 2 under power constraints P 1 max and P 2 max. 5.3 Energy Evaluation We investigate the energy efficiency of our proposed approach for the image object classification and next character prediction tasks in Figure 4. For the image classification task, it is noticeable that our method always outperforms baseline 1 since it takes less training time given the power limits. In addition, selections made by baselines 1 and 2 can lead to up to 2 and 1.4 more energy usage compared to our method, respectively. It is important to state that minimizing the training time does not always mean minimizing energy consumption. An example is comparing the energy consumption for P 2 max on Res Net18 with the no power limit N/A. Despite having lower training time in N/A (as shown in Table 2), training under P 2 max is more energy efficient since the used frequencies for P 2 max provide a better tradeoff between performance and power. The same applies for Baseline 2 when training Mobile Net V2 and Baseline 1 on Efficient Vi T under P 2 max, that uses a lower Published in Transactions on Machine Learning Research (10/2025) Target Dataset Proxy Dataset Cifar10 SVHN CINIC FMNIST STL Cifar10 - 0 0 0 0 SVHN 0 - 0 0 0 CINIC 0 0 - 0 0 FMNIST 0 0 0 - 0 STL 0 0 0 0 - Target Dataset Cifar10 SVHN CINIC FMNIST STL Cifar10 - 0 0 0 0 SVHN 0 - 0 0 0 CINIC 0 0 - 0 0 FMNIST 0 0 0 - 0 STL 0 0 0 0 - Target Dataset Cifar10 SVHN CINIC FMNIST STL Cifar10 - 0 0 0 10.1 SVHN 0 - 0 0 10.1 CINIC 0 0 - 0 10.1 FMNIST 2.5 2 0.1 - 0 STL 2.5 2 0.1 0 - Target Dataset Proxy Dataset Shakes. Austen Dickens Shakes. - 0 0 Austen 0 - 0 Dickens 0 0 - Target Dataset Shakes. Austen Dickens Shakes. - 0 0 Austen 0 - 0 Dickens 0 0 - Target Dataset Shakes. Austen Dickens Shakes. - 0 0 Austen 0 - 0 Dickens 0 0 - Figure 5: Matrices showing the mean percentage increase in training time of our approach to the fastest configuration for image classification and next-character prediction datasets on Nvidia Jetson Nano across three power constraints. The top matrix corresponds to training Mobile Net V2 for image classification tasks, and the bottom to training a transformer network for next-character prediction tasks. Each entry reflect the mean percentage increase in training time when selecting configuration given the proxy-dataset s relation vector r (row) instead of the optimal configuration (column) for the target data. A value of 0% denote cases where the same configuration was selected, resulting in no time increase. The results indicate that the proposed method is not sensitive to the selection of proxy dataset, with minimal impact on the training time when the selected configuration differs from the fastest configuration for the target dataset. frequency with the same batch size of our approach. For the next character prediction task, our approach uses 9% less energy for P 1 max and nearly the same energy consumption for P 2 max while training in less time (the same configuration selection for P 3 max) compared to the baseline 1. In addition, our method records 1.25 less training energy for the baseline 2. 5.4 Proxy Dataset Analysis We conduct a sensitivity analysis for the selection of the proxy dataset for image classification and next character prediction tasks by testing all possible proxy and target dataset combinations. We report the mean percentage increase (averaged over runs; see Section 5.1) in training time relative to the fastest configuration. For the image classification task, we add two additional datasets namely, Fashion MNIST Xiao et al. (2017) (FMNIST) and STL Coates et al. (2011). The tasks include object, digit, and fashion classification, with dataset sizes ranging from 5K to 22.5K samples. We provide in Fig. 5 a confusion matrices showing the percentage of time increase for the choice of proxy dataset selection compared to the fastest configuration for different datasets for training Mobile Net V2 and the transformer network for next character prediction on Jetson Nano. For Mobile Net V2, the results show that our approach selects the correct configurations in most of the cases with only a few instances where the optimal configuration is not selected in no power Published in Transactions on Machine Learning Research (10/2025) constraint scenario. The same observation is also applied to Res Net18 as shown in Figure 7 (Appendix A.2). For the next character prediction, the optimal configuration is selected regardless of the selected proxy dataset. To sum up, this analysis shows that our approach is not sensitive to the selection of proxy dataset, demonstrating its practicality. 6 Conclusion In this work, we propose a power-aware training method aimed at accelerating training on power-constrained GPU devices. Our results show that great savings can be achieved in terms of training time and energy consumption, when carefully and jointly selecting the system and application parameters for training, without scarifying the training model quality. The proposed solution is applicable to a wide range of models (including, but not limited to, CNNs and transformers). Our solution can be employed to leverage nearby renewable energy sources by adapting GPU frequency and batch size to the available green energy at edge devices to reduce carbon footprint. A limitation of this work is that we assume that the power constraint on a device does not change through the training period, which may not hold for long-duration tasks that can take multiple days. For future work, we want to extend our study to the varying power constraints and distributed learning settings where devices with different constraints contribute to the training. Acknowledgments This research was supported in part by the EU Horizon project EXIGENCE under Grant Agreement No. 101139120. Luis Barba-Guaman, Jose Eugenio Naranjo, and Anthony Ortiz. Deep learning framework for vehicle and pedestrian detection in rural roads on an embedded gpu. Electronics, 9(4):589, 2020. Steven Bird, Ewan Klein, and Edward Loper. Natural language processing with Python: analyzing text with the natural language toolkit. " O Reilly Media, Inc.", 2009. Han Cai, Chuang Gan, Ligeng Zhu, and Song Han. Tinytl: Reduce memory, not parameters for efficient on-device learning. Advances in Neural Information Processing Systems, 33:11285 11297, 2020. Icaro Camelo and Ana-Maria Cretu. Evaluating compact convolutional neural networks for object recognition using sensor data on resource-constrained devices. Engineering Proceedings, 58(1), 2023. ISSN 2673-4591. URL https://www.mdpi.com/2673-4591/58/1/6. Jiasi Chen and Xukan Ran. Deep learning with edge computing: A review. Proceedings of the IEEE, 107 (8):1655 1674, 2019. doi: 10.1109/JPROC.2019.2921977. Hung-Yueh Chiang, Natalia Frumkin, Feng Liang, and Diana Marculescu. Mobiletl: on-device transfer learning with inverted residual blocks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pp. 7166 7174, 2023. Kyubaik Choi and Gerald E Sobelman. An efficient cnn accelerator for low-cost edge systems. ACM Transactions on Embedded Computing Systems (TECS), 21(4):1 20, 2022. Adam Coates, Andrew Ng, and Honglak Lee. An analysis of single-layer networks in unsupervised feature learning. In Proceedings of the fourteenth international conference on artificial intelligence and statistics, pp. 215 223. JMLR Workshop and Conference Proceedings, 2011. Ryan Cochran, Can Hankendi, Ayse K. Coskun, and Sherief Reda. Pack & cap: adaptive dvfs and thread packing under power caps. In Proceedings of the 44th Annual IEEE/ACM International Symposium on Microarchitecture, MICRO-44, pp. 175 185, New York, NY, USA, 2011. Association for Computing Machinery. ISBN 9781450310536. Published in Transactions on Machine Learning Research (10/2025) Luke N Darlow, Elliot J Crowley, Antreas Antoniou, and Amos J Storkey. Cinic-10 is not imagenet or cifar-10. ar Xiv preprint ar Xiv:1810.03505, 2018. Mauro Fernández-Sanjurjo, Manuel Mucientes, and Víctor Manuel Brea. Real-time multiple object visual tracking for embedded gpu systems. IEEE Internet of Things Journal, 8(11):9177 9188, 2021. Prakhar Ganesh, Yao Chen, Yin Yang, Deming Chen, and Marianne Winslett. Yolo-ret: Towards high accuracy real-time object detection on edge gpus. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pp. 3267 3277, 2022. Priya Goyal, Piotr Dollár, Ross Girshick, Pieter Noordhuis, Lukasz Wesolowski, Aapo Kyrola, Andrew Tulloch, Yangqing Jia, and Kaiming He. Accurate, large minibatch sgd: Training imagenet in 1 hour. ar Xiv preprint ar Xiv:1706.02677, 2017. João Guerreiro, Aleksandar Ilic, Nuno Roma, and Pedro Tomás. Dvfs-aware application classification to improve gpgpus energy efficiency. Parallel Computing, 83:93 117, 2019. Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770 778, 2016. Yihui He, Ji Lin, Zhijian Liu, Hanrui Wang, Li-Jia Li, and Song Han. Amc: Automl for model compression and acceleration on mobile devices. In Proceedings of the European conference on computer vision (ECCV), pp. 784 800, 2018. Chentao Jia, Ming Hu, Zekai Chen, Yanxin Yang, Xiaofei Xie, Yang Liu, and Mingsong Chen. Adaptivefl: Adaptive heterogeneous federated learning for resource-constrained aiot systems. In Proceedings of the 61st ACM/IEEE Design Automation Conference, pp. 1 6, 2024. Andrej Karpathy. char-rnn. https://github.com/karpathy/char-rnn, 2015. Diederik Kingma and Jimmy Ba. Adam: A method for stochastic optimization. In International Conference on Learning Representations (ICLR), San Diega, CA, USA, 2015. Alex Krizhevsky. One weird trick for parallelizing convolutional neural networks. ar Xiv preprint ar Xiv:1404.5997, 2014. Alex Krizhevsky, Geoffrey Hinton, et al. Learning multiple layers of features from tiny images. 2009. Etienne Le Sueur and Gernot Heiser. Dynamic voltage and frequency scaling: The laws of diminishing returns. In Proceedings of the 2010 international conference on Power aware computing and systems, pp. 1 8, 2010. Jungseob Lee, Vijay Sathisha, Michael Schulte, Katherine Compton, and Nam Sung Kim. Improving throughput of power-constrained gpus using dynamic voltage/frequency and core scaling. In 2011 International Conference on Parallel Architectures and Compilation Techniques, pp. 111 120, 2011. doi: 10.1109/PACT.2011.17. Ji Lin, Ligeng Zhu, Wei-Ming Chen, Wei-Chen Wang, Chuang Gan, and Song Han. On-device training under 256kb memory. Advances in Neural Information Processing Systems, 35:22941 22954, 2022. Siqin Liu and Avinash Karanth. Dynamic voltage and frequency scaling to improve energy-efficiency of hardware accelerators. In 2021 IEEE 28th International Conference on High Performance Computing, Data, and Analytics (Hi PC), pp. 232 241. IEEE, 2021. Xinyu Liu, Houwen Peng, Ningxin Zheng, Yuqing Yang, Han Hu, and Yixuan Yuan. Efficientvit: Memory efficient vision transformer with cascaded group attention. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 14420 14430, 2023. Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. ar Xiv preprint ar Xiv:1711.05101, 2017. Published in Transactions on Machine Learning Research (10/2025) Arian Maghazeh, Unmesh D. Bordoloi, Petru Eles, and Zebo Peng. General purpose computing on low-power embedded gpus: Has it come of age? In 2013 International Conference on Embedded Computer Systems: Architectures, Modeling, and Simulation (SAMOS), pp. 1 10, 2013. doi: 10.1109/SAMOS.2013.6621099. Sam Mc Candlish, Jared Kaplan, Dario Amodei, and Open AI Dota Team. An empirical model of large-batch training. ar Xiv preprint ar Xiv:1812.06162, 2018. Stephen Merity, Caiming Xiong, James Bradbury, and Richard Socher. Pointer sentinel mixture models. ar Xiv preprint ar Xiv:1609.07843, 2016. Seyed Morteza Nabavinejad, Hassan Hafez-Kolahi, and Sherief Reda. Coordinated dvfs and precision control for deep neural networks. IEEE Computer Architecture Letters, 18(2):136 140, 2019. Seyed Morteza Nabavinejad, Sherief Reda, and Masoumeh Ebrahimi. Batchsizer: Power-performance tradeoff for dnn inference. In 2021 26th Asia and South Pacific Design Automation Conference (ASP-DAC), pp. 819 824, 2021. Seyed Morteza Nabavinejad, Sherief Reda, and Masoumeh Ebrahimi. Coordinated batching and dvfs for dnn inference on gpu accelerators. IEEE Transactions on Parallel and Distributed Systems, 33(10):2496 2508, 2022. Yuval Netzer, Tao Wang, Adam Coates, Alessandro Bissacco, Baolin Wu, Andrew Y Ng, et al. Reading digits in natural images with unsupervised feature learning. In NIPS workshop on deep learning and unsupervised feature learning, volume 2011, pp. 7. Granada, Spain, 2011. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019. Kilian Pfeiffer, Ramin Khalili, and Jörg Henkel. Aggregating capacity in fl through successive layer training for computationally-constrained devices. Advances in Neural Information Processing Systems, 36:35386 35402, 2023. Flavio Ponzina, Marco Rios, Alexandre Levisse, Giovanni Ansaloni, and David Atienza. Overflow-free compute memories for edge ai acceleration. ACM Transactions on Embedded Computing Systems, 22(5s):1 23, 2023. Amine Saddik, Rachid Latif, Mohamed Elhoseny, and Abdelhafid El Ouardi. Real-time evaluation of different indexes in precision agriculture using a heterogeneous embedded system. Sustainable Computing: Informatics and Systems, 30:100506, 2021. Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, and Liang-Chieh Chen. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4510 4520, 2018. Samuel L. Smith, Pieter-Jan Kindermans, and Quoc V. Le. Don t decay the learning rate, increase the batch size. In International Conference on Learning Representations, 2018. Zhenheng Tang, Yuxin Wang, Qiang Wang, and Xiaowen Chu. The impact of gpu dvfs on the energy and performance of deep learning: An empirical study. In Proceedings of the Tenth ACM International Conference on Future Energy Systems, pp. 315 325, 2019. Robert J Wang, Xiang Li, and Charles X Ling. Pelee: A real-time object detection system on mobile devices. Advances in neural information processing systems, 31, 2018. Han Xiao, Kashif Rasul, and Roland Vollgraf. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. ar Xiv preprint ar Xiv:1708.07747, 2017. Published in Transactions on Machine Learning Research (10/2025) Yuedong Yang, Guihong Li, and Radu Marculescu. Efficient on-device training via gradient filtering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 3811 3820, 2023. Jie You, Jae-Won Chung, and Mosharaf Chowdhury. Zeus: Understanding and optimizing GPU energy consumption of DNN training. In 20th USENIX Symposium on Networked Systems Design and Implementation (NSDI 23), pp. 119 139, Boston, MA, April 2023. USENIX Association. ISBN 978-1-939133-33-5. URL https://www.usenix.org/conference/nsdi23/presentation/you. Guodong Zhang, Lala Li, Zachary Nado, James Martens, Sushant Sachdeva, George Dahl, Chris Shallue, and Roger B Grosse. Which algorithmic choices matter at which batch sizes? insights from a noisy quadratic model. Advances in neural information processing systems, 32, 2019. Xingyao Zhang, Chenhao Xie, Jing Wang, Weidong Zhang, and Xin Fu. Towards memory friendly long-short term memory networks (lstms) on mobile gpus. In 2018 51st Annual IEEE/ACM International Symposium on Microarchitecture (MICRO), pp. 162 174, 2018. Published in Transactions on Machine Learning Research (10/2025) A.1 Training on edge CPU In this section, we briefly discuss training on an edge device s CPUs (i.e., Jetson Nano) and the reasons of why to focus only on training on GPUs on such devices. In Fig. 6, we show the processing time and power for training over 50K samples with batch size 32, where both CPU and GPU are using the maximum frequency. We observe that the GPU is an order of magnitude faster compared to the CPU, while nearly consuming similar average power. Similarly, the GPU would be nearly an order of magnitude more energy efficient on these devices compared to the CPU. Thus, we focus in our work on edge GPUs as training is more time and energy efficient. Figure 6: Res Net18 time for training 50K samples (Ts) and average power (red triangles) of Cifar10 on Jetson Nano over CPU and GPU. A.2 Proxy dataset analysis extension We provide a continuation for the sensitivity analysis for the proxy dataset selection. In Fig. 7, provide the confusion matrices for selecting different proxy datasets to train Res Net18. The difference for Res Net18 compared to Mobile Net V2 (in Figure 5)is that the suboptimal configurations are selected for the lowest power constraint, while for other cases the selected configurations match the optimal. Target Dataset Proxy Dataset Cifar10 SVHN CINIC FMNIST STL Cifar10 0 0 2.9 27.3 0 SVHN 0 0 2.9 27.3 0 CINIC 1.3 10.2 0 0 3.7 FMNIST 1.3 10.2 0 0 3.7 STL 0 0 2.9 27.3 0 Target Dataset Cifar10 SVHN CINIC FMNIST STL Cifar10 0 0 0 0 0 SVHN 0 0 0 0 0 CINIC 0 0 0 0 0 FMNIST 0 0 0 0 0 STL 0 0 0 0 0 Target Dataset Cifar10 SVHN CINIC FMNIST STL Cifar10 0 0 0 0 0 SVHN 0 0 0 0 0 CINIC 0 0 0 0 0 FMNIST 0 0 0 0 0 STL 0 0 0 0 0 Figure 7: Matrices of mean time increase percentages to the fastest configuration for image classification datasets across three power constraints for Res Net18 training on Jetson Nano. Published in Transactions on Machine Learning Research (10/2025) A.3 Text datasets details For Jane Austen s dataset part, we extracted it from the nltk package (gutenberg corpus) where the text consists of 3 novel namely Emma, Persuasion, and Sense and Sensibility. For Dickens, we used eight novels namely The Pickwick Papers, Pictures from Italy, A Tale of Two Cities, A Story of the French Revolution, The Chimes, Mugby Junction, The Haunted Man and the Ghost s Bargain, and The Mystery of Edwin Drood. We filter the licensing text, author history, etc. from the training and testing text. A.4 Configuration selection visualization In this section, we provide additional visualization for our method. Figures 8 and 9 illustrate the configuration candidates from our method under power constraints, together with the cases of no joint optimization. Specifically, this includes the power constraint cases of P 1 max and P 2 max for Res Net18 in Figure 8 and transformers for the next character prediction in Figure 9. Our approach finds a wider range of operating points in the search space, effectively adapting to different training workloads. For instance, the next-character prediction task is less power-intensive, which allows exploring a broader set of operating points. Additionally, it selects the best configuration to minimize TTacc. 4 8 16 32 64 128 2 4 8 16 32 64 128 4 8 16 32 64 128 Figure 8: Visualization of the the power and resulting training times to accuracy for the configuration candidates of our method when finetuning Res Net18 on SVHN under power constraints of P 1 max (red) and P 2 max (green) respectively. We highlight baselines 1 and 2 (b1, b2), corresponding to the cases without joint optimization, together with b , denoting the point selected by our method. . Our method explores more operating points in the search space while selecting the best configuration. 8 16 32 64 128 2 8 16 32 64 128 8 16 32 64 128 Figure 9: Visualization of the the power and resulting training times to accuracy for the configuration candidates of our method when finetuning transformer on Austen dataset under power constraints of P 1 max (red) and P 2 max (green). We highlight baselines 1 and 2 (b1, b2), corresponding to the cases without joint optimization, together with b , denoting the point selected by our method. Our method explores more operating points in the search space while selecting the best configuration. Published in Transactions on Machine Learning Research (10/2025) 4 8 16 32 64 128 80 Res Net18 Mobile Net V2 Efficient VIT 4 8 16 32 64 128 60 Figure 10: Maximum test accuracy across different batch sizes for the image classification task, showing stable performance with only marginal variation. A.5 Effect of batch Size on model quality To complement our main results, we analyze the impact of batch size on training quality. Specifically, we train the image classification models under different batch sizes and show the mean of maximum accuracy achieved over three runs on test set in Fig. 10. The max accuracy remained consistent across different batch sizes, showing only marginal differences without signs of increased overfitting. Therefore, the proposed configuration selection approach can select the batch size and frequency combination that improves the training speed given a device s power constraint and capabilities without compromising model quality. This is consistent with prior work Goyal et al. (2017) showing that larger batch sizes can be used without loss in accuracy. A.6 Power measurement details We read the power sensor values during training with a sample rate of 1s. For the Nvidia Jetson Nano, the power measurements are available to read from using sysfs nodes. For example, on the jetson nano, the sysfs nodes are available in the path /sys/bus/i2c/drivers/ina3221x/6-0040/iio:device0/ .