# on_the_optimization_of_margin_distribution__4fe5e404.pdf On the Optimization of Margin Distribution Meng-Zhang Qian1 , Zheng Ai1 , Teng Zhang2 and Wei Gao1 1National Key Laboratory for Novel Software Technology, Nanjing University, China 2School of Computer Science and Technology, Huazhong University of Science and Technology, China {qianmz, aiz, gaow}@lamda.nju.edu.cn, tengzhang@hust.edu.cn Margin has played an important role on the design and analysis of learning algorithms during the past years, mostly working with the maximization of the minimum margin. Recent years have witnessed the increasing empirical studies on the optimization of margin distribution according to different statistics such as medium margin, average margin, margin variance, etc., whereas there is a relative paucity of theoretical understanding. In this work, we take one step on this direction by providing a new generalization error bound, which is heavily relevant to margin distribution by incorporating ingredients such as average margin and semi-variance, a new margin statistics for the characterization of margin distribution. Inspired by the theoretical findings, we propose the MSVMAv, an efficient approach to achieve better performance by optimizing margin distribution in terms of its empirical average margin and semi-variance. We finally conduct extensive experiments to show the superiority of the proposed MSVMAv approach. 1 Introduction Margin has played an important role on the design of learning algorithms from the pioneer work [Vapnik, 1982], which proposed the famous Support Vector Machines (SVMs) by maximizing the minimum margin, i.e. the smallest distance from the instances to the classification boundary. Boser et al. [1992] introduced the kernel technique for SVMs to relax the linear separation. Large margin has been one of the most important principles on the design of learning algorithms in the history of machine learning [Cortes and Vapnik, 1995; Schapire et al., 1998; Rosset et al., 2003; Shivaswamy and Jebara, 2010; Ji et al., 2021], even for recent deep learning [Sokoli c et al., 2017; Weinstein et al., 2020]. Various margin-based bounds have been presented to study the generalization performance of learning algorithms. Bartlett and Shawe-Taylor [1999] possibly presented the first generalization margin bounds based on VC dimension and fat-shattering dimension. Bartlett and Mendelson [2002] introduced the famous margin bounds based on Rademacher complexity, a data-dependent and finite-sample complexity measure. Kab an and Durrant [2020] took advantage of geometric structure to provide margin bounds for compressive learning. Grønlund et al. [2020] presented the near-tight margin generalization bound for SVMs. Margin has also been an ingredient to analyze the generalization performance for other algorithms such as boosting [Schapire et al., 1998; Breiman, 1999; Gao and Zhou, 2013], and deep learning [Bartlett et al., 2017; Wei and Ma, 2020]. Margin distribution has been considered as an important ingredient on the design and analysis of learning algorithms, and the basic idea is to optimize some margin statistics, relevant to the whole margin distribution rather than single margin. Garg and Roth [2003] introduced the model complexity measure to optimize margin distribution. Pelckmans et al. [2007] optimized margin distribution via average margin, while Aiolli et al. [2008] tried to maximize the minimum margin and average margin. Zhang and Zhou [2014] proposed the large margin distribution machine by considering average margin and margin variance simultaneously, which motivates the design of a series learning algorithms on the optimization of margin distribution [Cheng et al., 2016; Rastogi et al., 2020]. For deep learning, Jiang et al. [2019] introduced some margin distribution statistics, such as total variation, median quartile, etc., to analyze the generalization of neural networks. There is a relative paucity of theoretical understanding on how to correlate margin distribution with the generalization of learning algorithms. This work tries to fill the gap between theoretical and empirical studies on the optimization of margin distribution, and the main contributions can be summarized as follows: We present a new generalization error bound, which is heavily relevant to margin distribution by incorporating factors such as average margin and semi-variance. Here, semi-variance is a new statistics, counting the average of squared distances between average margin and the instances margin, that is smaller than average margin. Motivated from our theoretical result, we develop the MSVMAv approach, which tries to achieve better generalization performance by optimizing margin distribution in terms of empirical average margin and semi-variance. We find the closed-form solution in optimization, and improve its efficiency via Sherman-Morrison formula. Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) We conduct extensive empirical studies to validate the effectiveness of the MSVMAv approach in comparisons with the state-of-the-art algorithms on large-margin or margin distribution optimization. The rest of this paper is organized as follows. Section 2 introduces some preliminaries. Section 3 presents theoretical analysis. Section 4 proposes the MSVMAv approach. Section 5 conducts extensive empirical studies, and Section 6 concludes with future work. 2 Preliminaries Let X Rd and Y = {+1, 1} denote the instance and label space, respectively. Suppose that D is an underlying (unknown) distribution over the product space X Y. Let Sn = {(x1, y1), (x2, y2), , (xn, yn)} be a training sample with each element drawn independently and identically (i.i.d.) from distribution D. We use Pr D[ ] and ED[ ] to refer to the probability and expectation according to distribution D, respectively. Let H = {h: X [ 1, +1]} be a function space. We define the classification error (or generalization risk) with respect to function h H and distribution D, as E(h) = Pr D[sgn[h(x)] = y] = ED[I[yh(x) 0]] , where the sign function sgn[ ] returns +1, 0 and 1 if the argument is positive, zero and negative, respectively, and the indicator function I[ ] returns 1 when the argument is true, and 0 otherwise. Given an example (x, y), the margin of h H is defined as yh(x), which can be viewed as a measure of the confidence of the classification. We further define the average margin of h H over distribution D as θh = E(x,y) D[yh(x)] . (1) We also introduce the empirical Rademacher complexity [Bartlett and Mendelson, 2002] to measure the complexity of function space H as follows: b RSn(H) = Eσ1,σ2,...,σn i=1 σih(xi) where each σi is a Rademacher variable with Pr[σi = +1] = Pr[σi = 1] = 1/2 for i [n]. We finally introduce some notations used in this work. Write [d] = {1, 2, . . . , d} for integer d > 0, and w, x represents the inner product of w and x. Let Id be the identity matrix of size d d, and denote by the transpose of vectors or matrices. For positive f(n) and g(n), we write f(n) = O(g(n)) if g(n)/f(n) c for constant c < + . 3 Theoretical Analysis We begin with the squared margin loss as follows: Definition 1. For θ > 0, we define the squared margin loss ℓθ with respect to function h H as ℓθ h, (x, y) = (1 yh(x)/θ)+ 2 , where (a)+ = max(0, a). This is a simple extension from the traditional margin loss [Bartlett and Mendelson, 2002], while we consider the squared loss and unbounded constraint for the negative yh(x). The margin parameter θ is generally irrelevant to learned function h and data distribution in most previous theoretical and algorithmic studies. In this work, we select margin parameter θ as the average margin when θh > 0, to correlate generalization performance with margin distribution, that is, θ = θh = ED[yh(x)] , which is dependent on data distribution and learned function. Given training sample Sn, we try to learn a function h by minimizing the squared margin loss as follows: min h H: θh>0 " 1 yih(xi) For simplicity, we further introduce the notion of margin semi-variance [Markowitz, 1952] as follows: Definition 2. Given function h H and training sample Sn, we define the margin semi-variance as (θh yih(xi))+ 2 , where θh denotes the average margin defined by Eqn. (1). The margin semi-variance essentially counts the average of squared deviation between average margin and the margins yih(xi), which are smaller than average margin. This yields an equivalent expression for Eqn. (2) as follows: min h H:θh ν>0 SV(h)/θ2 h , that is, optimizing the squared margin loss with parameter θ = θh is equivalent to minimizing margin semi-variance and maximizing average margin simultaneously. For most real applications, we could learn some relativelygood functions from sufficient training data. Motivated from the notion of weak learner in boosting [Freund and Schapire, 1996], we formally define the set of relatively-good functions for function space H as follows: Hν = {h H, θh ν} for some small constant ν > 0 . Essentially, a relatively-good function is similar to a weak learner, which achieves slightly better performance than the randomly-guessed classifier. We now present the main theoretical result as follows: Theorem 1. For small constant ν 0, let H be a function space with relative-good set Hν. For any δ (0, 1) and for every h Hν, the following holds with probability at least 1 δ over the training sample Sn with empirical Rademacher complexity b Rn(Hν) b Rn(H). Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) This theorem presents a new generalization error bound, which is heavily relevant to margin distribution by incorporating factors such as average margin and semi-variance. This could shed some new insights on the design of algorithms on the optimization of margin distribution as shown in Section 4. The proof follows the empirical Rademacher complexity [Bartlett and Mendelson, 2002], while the challenge lies in the distribution-dependent average margin θh. We solve it by constructing a sequence of intervals for average margin θh, and the detailed proof is presented in [Qian et al., 2022]. It remains to study the empirical Rademacher complexity in Theorem 1, and we focus on linear and kernel functions. For instance space X = {x Rd : x r} and linear function space H = {h(x) = w x: w Λ}, let Hν denote the set of relatively-good classifiers. We upper bound the empirical Rademacher complexity as b RSn(Hν) b RSn(H) rΛ/ n from the work of [Shalev-Shwartz and Ben-David, 2014]. For kernel function κ( , ), we have the kernel function space H = n h(x) = i=1 aiκ(xi, x): i,j=1 aiajκ(xi, xj) Λ2o . We could upper bound the empirical Rademacher complexity for kernel functions, from [Bartlett and Mendelson, 2002], b RSn(Hν) b RSn(H) 2Λ i=1 κ(xi, xi) 1/2 . It is also noteworthy that the average margin θh is unknown on the design of new algorithms because of the unknown distribution D, and we resort to the empirical average margin from training sample Sn in practice. 4 The MSVMAv Approach Motivated from Theorem 1, this section develops the MSVMAv approach on the optimization of margin distribution, and we focus on linear and kernel functions. 4.1 Linear Functions For linear space H = {hw(x) = w, x : w = 1} and training sample Sn, we have the empirical average margin i=1 yi w, xi . For simplicity, we omit a bias term on the design of algorithm, and we will augment w and instance x with bias term b and 1 in experiments, respectively, as shown in Section 5. Our optimization problem can be formally written as where empirical average margin ˆθw > 0 and empirical margin semi-variance c SV(w) = Pn i=1[(ˆθw yi w, xi )+]2/n. Obviously, this is a non-convex optimization problem, and we would optimize the empirical margin semi-variance and average margin alternatively. Initialize the linear function w0 by optimizing empirical average margin, that is, w0 = arg max w 2 2=1 yixi Pn i=1 yixi 2 , (3) where we solve w0 from its dual problem using Lagrangian function, and the details are given by Qian et al. [2022]. Optimization of Empirical Margin Semi-variance In the k-th iteration (k 1) with previous classifier wk 1, we first calculate the empirical average margin ˆθwk 1 as i=1 yi wk 1, xi . (4) We then introduce the minimization of empirical margin semi-variance as follows: [(ˆθwk 1 yi w, xi )+]2 n + βk w wk 1 2 2 o , where βk is a proximal regularization parameter. We now introduce the following index set, to present a closed-form solution for the above minimization, Ak = i: yi wk 1, xi < ˆθwk 1 for i [n] , (5) i.e., the index set of instance with margins below the empirical average margin ˆθwk 1. We can rewrite the minimization of empirical margin semi-variance as (ˆθwk 1 yi w, xi )2 n + βk w wk 1 2 2 o . Denote by w k the minimizer of the above problem, and we obtain the closed-form solution for w k as follows i Ak yixi + wk 1 . (6) One problem is to calculate the inverse in Eqn. (6), which takes O(d3) computational costs (d is dimensionality). This remains one challenge to deal with high-dimensional tasks. We now present an efficient method to calculate of the inverse in Eqn. (6). For simplicity, we denote by Mk = Id + X i Ak xix i nβk 1 for k = 1, 2, , and it is easy to derive the following recursive relation: M 1 k = M 1 k 1 X xix i nβ + X with M0 = Id. We calculate Mk efficiently from Mk 1 and Sherman Morrison formula [Sherman and Morrison, 1950]. In other Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) Algorithm 1 The MSVMAv Approach Input: Training sample Sn, iteration number T, and proximal parameters αk and βk Output: w 1: Initialize M0 = Id, A0 = and w0 by Eqn. (3) 2: for k = 1, 2, , T do 3: Compute empirical average margin ˆθwk 1 by Eqn. (4) 4: Solve the index set Ak by Eqn. (5) 5: Compute Mk = Id + P i Ak xix i nβk 1 by Eqns. (7) and (8) 6: Compute the minimizer w k for empirical margin semivariance by Eqn. (9) 7: Solve the empirical average margin maximizer wk by Eqn. (10), and normalize wk = wk/ wk 2 8: end for 9: return w = w T words, we initialize M = Mk 1, and make the following updates iteratively, based on Sherman-Morrison formula, M = M M xix i M x i M xi nβk for i Ak 1 \ Ak , (7) M = M M xix i M x i M xi + nβk for i Ak \ Ak 1 . (8) We then obtain Mk = M , and the minimizer of empirical margin semi-variance is given by w k = M ˆθwk 1 i Ak yixi + wk 1 . (9) Optimization of Empirical Average Margin We now study the maximization of empirical average margin, which can be formalized as: wk = arg min w i=1 yi w, xi + αk w w k 2 2 o , where αk is a proximal regularization parameter. It is easy to obtain the closed-form solution as follows: wk = w k + 1 2αkn i=1 yixi . (10) We obtain wk = wk/ wk in the k-th iteration. Algorithm 1 presents a detailed description of our MSVMAv approach. 4.2 Kernelization This section focuses on kernel mapping φ : X H for Hilbert space H, we consider h(x) = w, φ(x) with w H and φ(x) H. The optimization problem is given by where average margin ˆθw = Pn i=1 yi w, φ(xi) /n > 0, and margin semi-variance c SV(w) = 1 ˆθw yi w, φ(xi) 1.0 0.5 0.0 0.5 1.0 1.5 2.0 Margin Cumulative Frequency MSVMAV SVM ODM FMM 0.2 0.0 0.2 0.4 0.6 Margin Cumulative Frequency MSVMAV SVM ODM FMM 0.05 0.00 0.05 0.10 0.15 Margin Cumulative Frequency MSVMAV SVM ODM FMM 0.4 0.2 0.0 0.2 0.4 0.6 0.8 Margin Cumulative Frequency MSVMAV SVM ODM FMM Figure 1: Cumulative frequency versus margin of our MSVMAv and other algorithms such as SVM, ODM and FMM. The more right the curve, the better the margin distribution. It is intractable to solve such optimization problem directly because of high or even infinity dimensionality. According to Representer theorem [Sch olkopf et al., 2002], we first have w = Pn i=1 aiφ(xi), spanned by {φ(xi), i [n]} with coefficients a1, , an. This follows the prediction h(x) = w, φ(x) = i=1 aiκ(x, xi) , where κ( , ) denotes the kernel function. For simplicity, denote by a = (a1, a2, , an) , and write the gram matrix of instances in Sn as K = (K1, K2, , Kn) = Kij n n = κ(xi, xj) where Ki denotes the k-th column of matrix K. Hence, our optimization problem can be further rewritten as 2 o = min a where the empirical average margin ˆθa = Pn i=1 yi Ki, a /n and semi-variance c SV(a) = Pn i=1[(ˆθa yi Ki, a )+]2/n. We first initialize the classifier a0 by maximizing the empirical average margin as follows: a0 = arg max a Ka=1 i=1 yi Ki, a o . In the k-th iteration with previous classifier ak 1, we minimize the margin semi-variance based on previous average margin ˆθak 1. We write a ak 1 In+K = (a ak 1) (In+K a ak 1) 1/2 , and the optimization problem can be given by [(ˆθak 1 yi Ki, a )+]2 n +βk a ak 1 2 In+K o , Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) Dataset MSVMAv SVM SVR LSSVM ODM MAMC FMM advertise .9837 .0015 .9823 .0021 .9825 .0024 .9819 .0019 .9701 .0021 .9132 .0007 .9799 .0025 australian .8314 .0079 .8167 .0077 .8171 .0048 .8220 .0036 .8104 .0065 .7700 .0201 .8295 .0100 bibtex .7417 .0040 .7378 .0069 .7469 .0060 .7478 .0046 .7469 .0053 .6689 .0123 .7371 .0062 biodeg .8712 .0087 .8741 .0068 .8490 .0107 .8741 .0068 .8681 .0075 .6872 .0000 .8613 .0089 breastw .9730 .0038 .9725 .0041 .9701 .0051 .9684 .0051 .8428 .0099 .4088 .0000 .9720 .0063 diabetes .7530 .0088 .7561 .0104 .7478 .0077 .7491 .0078 .6110 .0146 .5974 .0000 .7496 .0082 emotions .8216 .0074 .7983 .0130 .7675 .0181 .8036 .0127 .8084 .0111 .6975 .0000 .7697 .0218 german .7518 .0097 .7457 .0095 .7525 .0107 .7530 .0106 .7502 .0095 .6800 .0000 .7525 .0102 halloffame .9644 .0025 .9617 .0047 .9593 .0041 .9617 .0047 .9620 .0028 .9280 .0000 .9598 .0036 hill-valley .7771 .0631 .5972 .0195 .6999 .0572 .5972 .0195 .8898 .0079 .5000 .0000 .8526 .0296 kc1 .8713 .0025 .8711 .0039 .8642 .0026 .8711 .0039 .8690 .0035 .8649 .0000 .8701 .0028 parkinsons .9222 .0277 .8923 .0388 .9342 .0315 .9444 .0338 .8863 .0342 .7949 .0000 .8932 .0266 pbcseq .6626 .0074 .6439 .0147 .6595 .0104 .6555 .0104 .6562 .0125 .6700 .0187 .6549 .0117 sleepdata .6925 .0056 .6743 .0051 .6833 .0064 .6712 .0124 .5691 .0156 .5659 .0000 .6833 .0047 students .8957 .0062 .8913 .0088 .8870 .0064 .8867 .0069 .8893 .0046 .5133 .0129 .8898 .0040 titanic .7658 .0038 .7636 .0000 .7636 .0000 .7636 .0000 .7509 .0211 .6386 .0000 .7636 .0000 tokyo1 .9351 .0040 .9307 .0031 .9337 .0027 .9281 .0054 .9248 .0053 .7523 .0534 .9363 .0037 vehicle .9708 .0057 .9422 .0473 .9746 .0031 .9748 .0034 .9720 .0061 .7041 .0000 .9718 .0083 vertebra .8194 .0197 .7978 .0149 .7731 .0144 .7753 .0131 .7957 .0105 .7581 .0000 .7763 .0100 wdbc .9778 .0067 .9787 .0084 .9725 .0030 .9696 .0044 .9237 .0094 .5398 .1287 .9655 .0064 a9a .8433 .0009 .8417 .0007 .8403 .0007 .8358 .0006 .8430 .0012 .7577 .0000 .8390 .0012 acoustic .7494 .0011 .7321 .0037 .7394 .0004 N/A .7406 .0023 .7206 .0055 .7402 .0005 bank .9057 .0004 .8960 .0008 .9015 .0004 N/A .9021 .0006 .8854 .0000 .9021 .0005 eurgbp .5332 .0027 .5042 .0061 .5317 .0028 N/A .5111 .0084 .4985 .0000 .5288 .0028 jm1 .8125 .0015 .8132 .0012 .8119 .0016 .8123 .0015 .8065 .0000 .8065 .0000 .8076 .0010 magic .7998 .0005 .7976 .0012 .7928 .0008 .7912 .0009 .7943 .0007 .6514 .0000 .7987 .0014 nomao .9453 .0005 .9421 .0061 .9439 .0005 N/A .9452 .0004 .7062 .0000 .9442 .0008 phishing .9388 .0011 .9405 .0017 .9371 .0007 .9343 .0008 .9386 .0014 .5532 .0008 .9318 .0009 pol .9054 .0013 .8746 .0398 .9002 .0019 .9041 .0018 .6788 .0037 .6740 .0000 .8866 .0016 run-walk .7260 .0007 .7169 .0000 .7077 .0004 N/A .7104 .0061 .5431 .0757 .7261 .0054 Win/Tie/Loss 23/6/1 24/4/2 23/4/3 22/6/2 29/0/1 22/7/1 Table 1: Comparisons of the test accuracies (mean std.) on 30 datasets. / indicates that our MSVMAv approach is significantly better/worse than the corresponding algorithms (pairwise t-tests at 95% significance level). N/A indicates that LSSVM does not return results on the data set within 12 hours. where βk is a proximal regularization parameter. We introduce the index set Ak = {i: yi Ki, a < ˆθak 1 for i [n]}, and obtain the empirical margin semi-variance minimizer a k = Mk (K + In)ak 1 + ˆθa k X where we use the Sherman-Morrison formula to calculate Ki K i nβ + K + In We finally maximize the empirical average margin based on the following optimization problem: i=1 yi Ki, ak + αk(ak a k) K(ak a k) o , where αk is a proximal regularization parameter, and it is easy to get the closed-form solution as follows: ak = a k + [y1, y2, , yn] /2αkn . We get the final ak = ak/ ak K in the k-th iteration. 5 Empirical Study In this section, we present extensive empirical studies to verify the effectiveness of our proposed MSVMAv approach. We consider 30 datasets, including 20 regular and 10 largescale datasets. The number of instances varies from 208 to 88588 while the feature dimensionality ranges from 2 to 1836, covering a wide range of properties. The statistics for all datasets can be found in [Qian et al., 2022]. We compare our proposed MSVMAv approach with stateof-the-art algorithms on large-margin and margin distribution optimization: 1) SVM [Boser et al., 1992], 2) SVR [Drucker et al., 1997] with binary targets, 3) LSSVM [Suykens et al., 2002], 4) MAMC [Pelckmans et al., 2007], 5) ODM [Zhang and Zhou, 2019], 6) FMM [Ji et al., 2021]. The details of compared algorithms can be found in [Qian et al., 2022]. For each dataset, we scale all features into the interval [0, 1], and augment each instance x with constant 1 for the bias of linear model. The empirical average margin θw may be smaller than zero in experiments, when the proximal regularization parameter βk is set too small. In such case, we take the opposite model w so as to keep the positiveness of empirical average margin. For our MSVMAv approach, parameters αk and βk are set Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) Dataset MSVMAv SVM SVR LSSVM ODM MAMC FMM advertise .9837 .0014 .9820 .0023 .9835 .0019 .9848 .0016 .9838 .0016 .9547 .0026 .9810 .0028 australian .8580 .0078 .8225 .0087 .8210 .0080 .8357 .0111 .8329 .0088 .8302 .0122 .8237 .0060 bibtex .7554 .0036 .7506 .0050 .7508 .0053 .7505 .0056 .7570 .0045 .6714 .0397 .7509 .0050 biodeg .8834 .0081 .8687 .0115 .8580 .0089 .8712 .0093 .8845 .0095 .8559 .0098 .8915 .0096 breastw .9783 .0013 .9710 .0013 .9703 .0050 .9713 .0018 .9710 .0013 .9774 .0022 .9710 .0013 diabetes .7576 .0074 .7574 .0094 .7502 .0096 .7411 .0112 .7504 .0082 .6779 .0197 .7385 .0105 emotions .7986 .0122 .7899 .0156 .7756 .0164 .8115 .0124 .8101 .0147 .7952 .0120 .8078 .0126 german .7465 .0099 .7443 .0096 .7198 .0174 .7353 .0143 .7285 .0156 .7198 .0154 .7277 .0127 halloffame .9677 .0025 .9625 .0020 .9578 .0053 .9523 .0060 .9617 .0025 .9510 .0028 .9590 .0034 hill-valley .6826 .0184 .6668 .0177 .6482 .0135 .7116 .0678 .5950 .0360 .5402 .0320 .7886 .0299 kc1 .8739 .0042 .8701 .0057 .8689 .0045 .8703 .0044 .8716 .0055 .8764 .0027 .8706 .0038 parkinsons .9573 .0223 .9282 .0203 .9393 .0193 .9393 .0224 .9385 .0157 .9214 .0174 .9402 .0167 pbcseq .7350 .0143 .7214 .0152 .7317 .0151 .7312 .0165 .7269 .0221 .7076 .0149 .7238 .0184 sleepdata .7407 .0129 .7192 .0105 .7211 .0061 .7037 .0083 .7050 .0083 .7055 .0056 .7182 .0094 students .8977 .0119 .8920 .0079 .8665 .0098 .8898 .0072 .8805 .0142 .6543 .0185 .8993 .0062 titanic .7825 .0048 .7823 .0052 .7823 .0052 .7823 .0052 .7767 .0075 .7823 .0052 .7823 .0052 tokyo1 .9406 .0037 .9241 .0050 .9257 .0060 .9337 .0054 .9253 .0053 .9229 .0039 .9248 .0050 vehicle .9793 .0083 .9795 .0090 .9856 .0073 .9899 .0070 .9722 .0088 .9625 .0105 .9805 .0093 vertebra .8280 .0240 .7898 .0098 .7957 .0183 .8108 .0176 .8000 .0122 .7769 .0126 .7962 .0153 wdbc .9819 .0081 .9810 .0056 .9772 .0066 .9842 .0035 .9795 .0052 .9526 .0089 .9526 .0089 Win/Tie/Loss 15/5/0 16/3/1 13/3/4 14/5/1 17/2/1 14/3/3 Table 2: Comparisons of the test accuracies (mean std.) on 20 datasets. We use Gaussian kernel for all algorithms. / indicates that our MSVMAv approach is significantly better/worse than the corresponding algorithms (pairwise t-tests at 95% significance level). to be constant and selected by 5-fold cross validation from {2 10, 2 8, , 210}, and the width of Gaussian kernel is chosen from {2 10/d, 2 8/d, , 210/d}. We select the maximum iteration number T = 100 as a stopping criteria for MSVMAv . For SVM, SVR, LSSVM and ODM, we set regularization parameter C {2 10, 2 8, , 210} by 5fold cross validation again, and the others are set according to their respective references, also shown in [Qian et al., 2022]. We first compare the margin distributions of our proposed MSVMAv approach with other algorithms. Figure 1 illustrates the cumulative margin distributions of different algorithms on four datasets, and similar trends can be observed on other datasets. As can be seen, the curves of our MSVMAv approach generally lie on the rightmost side, which shows the margin distributions of MSVMAv are generally better than that of SVM, ODM and FMM. We further analyze the generalization performance of our proposed MSVMAv approach with other compared algorithms. All algorithms are evaluated by 30 times of random partitions of datasets with 80% and 20% of data for training and testing, respectively. The test accuracies are obtained by averaging over 30 times. Tables 1 and 2 show the empirical results of our MSVMAv and other algorithms with linear and Gaussian kernel functions, respectively. From Tables 1 and 2, our proposed MSVMAv approach takes significantly better performance than other algorithms for linear and kernel functions, since win/tie/loss counts show that our approach wins for most datasets, and rarely losses. One intuitive explanation is that our MSVMAv approach achieves better margin distribution by maximizing the empirical average margin and minimizing empirical margin semi-variance, as shown in Figure 1. SVM, SVR and FMM maximize the minimum margin, which ignores the margin distribution. LSSVM and MAMC essentially maximize average margin only, which fails to learn from other margin statistics. ODM takes the average margin and margin variance into consideration, but the process of margin variance minimization could constrain some large margins. This section omits partial empirical results due to the page limit, including the empirical curves of margin distributions, as well as running time comparisons for our MSVMAv and other compared algorithms. Relevant results can be found in our full work [Qian et al., 2022]. 6 Conclusion Large margin has been one of the most important principles on the design of algorithms in machine learning, and recent empirical studies show new insights on the optimization of margin distribution yet without theoretical supports. This work takes one step on this direction by providing a new generalization error bound, which is heavily relevant to margin distribution by incorporating factors such as average margin and semi-variance. Based on the theoretical results, we develop the MSVMAv approach for margin distribution optimization, and extensive experiments verify its superiority. An interesting future work is to exploit more effective statistics to characterize the whole margin distribution. Acknowledgements The authors want to thank the reviewers for helpful comments and suggestions. This work is partially supported by the NSFC (61921006, 61876078, 62006088). Wei Gao is the corresponding author. Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22) [Aiolli et al., 2008] F. Aiolli, G. Da San Martino, and A. Sperduti. A kernel method for the optimization of the margin distribution. In ICANN, pages 305 314, 2008. [Bartlett and Mendelson, 2002] P. Bartlett and S. Mendelson. Rademacher and gaussian complexities: Risk bounds and structural results. JMLR, 3:463 482, 2002. [Bartlett and Shawe-Taylor, 1999] P. Bartlett and J. Shawe Taylor. Generalization performance of support vector machines and other pattern classifiers. Advances in Kernel Methods: Support Vector Learning, pages 43 54, 1999. [Bartlett et al., 2017] P. Bartlett, D. J. Foster, and M. Telgarsky. Spectrally-normalized margin bounds for neural networks. In NIPS 30, pages 6240 6249. 2017. [Boser et al., 1992] B. E. Boser, I. M. Guyon, and V. Vapnik. A training algorithm for optimal margin classifiers. In COLT, pages 144 152, 1992. [Breiman, 1999] L. Breiman. Prediction games and arcing algorithms. Neural Comput., 11(7):1493 1517, 1999. [Cheng et al., 2016] F.-Y. Cheng, J. Zhang, and C.-H. Wen. Cost-sensitive large margin distribution machine for classification of imbalanced data. Pattern Recognit. Lett., 80:107 112, 2016. [Cortes and Vapnik, 1995] C. Cortes and V. Vapnik. Support-vector networks. Mach. Learn., 20(3):273 297, 1995. [Drucker et al., 1997] H. Drucker, J. C. Burges, L. Kaufman, A. Smola, and V. Vapnik. Support vector regression machines. In NIPS 9, pages 155 161. 1997. [Freund and Schapire, 1996] Y. Freund and R. E. Schapire. Experiments with a new boosting algorithm. In ICML, pages 148 156, 1996. [Gao and Zhou, 2013] W. Gao and Z.-H. Zhou. On the doubt about margin explanation of boosting. Artif. Intell., 203:1 18, 2013. [Garg and Roth, 2003] A. Garg and D. Roth. Margin distribution and learning. In ICML, pages 210 217, 2003. [Grønlund et al., 2020] A. Grønlund, L. Kamma, and K. G. Larsen. Near-tight margin-based generalization bounds for support vector machines. In ICML, pages 3779 3788, 2020. [Ji et al., 2021] Z.-W. Ji, N. Srebro, and M. Telgarsky. Fast margin maximization via dual acceleration. In ICML, pages 4860 4869, 2021. [Jiang et al., 2019] Y. Jiang, D. Krishnan, H. Mobahi, and S. Bengio. Predicting the generalization gap in deep networks with margin distributions. In ICLR, 2019. [Kab an and Durrant, 2020] Ata Kab an and Robert J Durrant. Structure from randomness in halfspace learning with the zero-one loss. JAIR, 69:733 764, 2020. [Markowitz, 1952] Harry Markowitz. Portfolio selection. Journal of Finance, 7(1):77 91, 1952. [Pelckmans et al., 2007] K. Pelckmans, J. A. K. Suykens, and B. De Moor. A risk minimization principle for a class of parzen estimators. In NIPS 20, pages 1137 1144. 2007. [Qian et al., 2022] M.-Z. Qian, Z. Ai, T. Zhang, and W. Gao. On the optimization of margin distribution. Co RR, abs/2204.14118, 2022. [Rastogi et al., 2020] R. Rastogi, P. Anand, and S. Chandra. Large-margin distribution machine-based regression. Neural Comput. Appl., 32(8):3633 3648, 2020. [Rosset et al., 2003] S. Rosset, J. Zhu, and T. Hastie. Margin maximizing loss functions. In NIPS 16, pages 1237 1244. 2003. [Schapire et al., 1998] R. E. Schapire, Y. Freund, P. Bartlett, and W. S. Lee. Boosting the margin: A new explanation for the effectiveness of voting methods. Ann. Stat., 26(5):1651 1686, 1998. [Sch olkopf et al., 2002] B. Sch olkopf, A. J. Smola, and F. Bach. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond. MIT Press, Cambridge, 2002. [Shalev-Shwartz and Ben-David, 2014] S. Shalev-Shwartz and S. Ben-David. Understanding Machine Learning: From Theory to Algorithms. Cambridge University Press, Cambridge, 2014. [Sherman and Morrison, 1950] J. Sherman and W. J. Morrison. Adjustment of an inverse matrix corresponding to a change in one element of a given matrix. Ann. Math. stat., 21(1):124 127, 1950. [Shivaswamy and Jebara, 2010] P. K. Shivaswamy and T. Jebara. Maximum relative margin and data-dependent regularization. JMLR, 11(2), 2010. [Sokoli c et al., 2017] J. Sokoli c, R. Giryes, G. Sapiro, and M. R. Rodrigues. Robust large margin deep neural networks. IEEE Trans. Signal Process., 65(16):4265 4280, 2017. [Suykens et al., 2002] J. A. K. Suykens, T. Van Gestel, J. De Brabanter, B. De Moor, and J. Vandewalle. Least Squares Support Vector Machines. World Scientific, Singapore, 2002. [Vapnik, 1982] V. Vapnik. Estimation of Dependences based on Empirical Data. Springer-Verlag, New York, 1982. [Wei and Ma, 2020] C. Wei and T.-Y. Ma. Improved sample complexities for deep neural networks and robust classification via an all-layer margin. In ICLR, 2020. [Weinstein et al., 2020] B. Weinstein, S. Fine, and Y. Hel Or. Margin-based regularization and selective sampling in deep neural networks. Co RR, abs/2009.06011, 2020. [Zhang and Zhou, 2014] T. Zhang and Z.-H. Zhou. Large margin distribution machine. In KDD, pages 313 322, 2014. [Zhang and Zhou, 2019] T. Zhang and Z.-H. Zhou. Optimal margin distribution machine. IEEE Trans. Knowl. Data Eng., 32(6):1143 1156, 2019. Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (IJCAI-22)