# stochastic_variancereduced_cubic_regularized_newton_methods__17733a3f.pdf Stochastic Variance-Reduced Cubic Regularized Newton Methods Dongruo Zhou 1 Pan Xu 1 Quanquan Gu 1 We propose a stochastic variance-reduced cubic regularized Newton method (SVRC) for nonconvex optimization. At the core of our algorithm is a novel semi-stochastic gradient along with a semi-stochastic Hessian, which are specifically designed for cubic regularization method. We show that our algorithm is guaranteed to converge to an ( , p )-approximate local minimum within e O(n4/5/ 3/2) second-order oracle calls, which outperforms the state-of-the-art cubic regularization algorithms including subsampled cubic regularization. Our work also sheds light on the application of variance reduction technique to high-order non-convex optimization methods. Thorough experiments on various non-convex optimization problems support our theory. 1 Introduction We study the following finite-sum optimization problem: min x2Rd F(x) = 1 fi(x), (1.1) where F(x) and each fi(x) can be non-convex. Such problems are common in machine learning, where each fi(x) is a loss function on a training example (Le Cun et al., 2015). Since F(x) is non-convex, finding its global minimum is generally NP-Hard (Hillar & Lim, 2013). As a result, one possible goal is to find an approximate first-order stationary point ( stationary point): kr F(x)k2 , for some given > 0. A lot of studies have been devoted to this problem including gradient descent (GD), stochastic gradient descent (SGD) (Robbins & Monro, 1951), and 1Department of Computer Science, University of California, Los Angeles, CA 90095, USA. Correspondence to: Quanquan Gu . Proceedings of the 35 th International Conference on Machine Learning, Stockholm, Sweden, PMLR 80, 2018. Copyright 2018 by the author(s). their extensions (Ghadimi & Lan, 2013; Reddi et al., 2016a; Allen-Zhu & Hazan, 2016; Ghadimi & Lan, 2016). Nevertheless, first-order stationary points can be non-degenerate saddle points or even local maximum in non-convex optimization, which are undesirable. Therefore, a more reasonable objective is to find an approximate second-order stationary point (Nesterov & Polyak, 2006), which is also known as an ( g, h)-approximate local minimum of F(x): kr F(x)k2 < g, λmin(r2F(x)) h, (1.2) for some given constant g, h > 0. In fact, in some machine learning problems like matrix completion (Ge et al., 2016), one finds that every local minimum is a global minimum, suggesting that finding an approximate local minimum is a better choice than a stationary point, and is good enough in many applications. One of the most popular method to achieve this goal is perhaps cubic regularized Newton method, which was introduced by Nesterov & Polyak (2006), and solves the following kind of subproblems in each iteration: h(x) = argmin h2Rd m(h, x), m(h, x) = hr F(x), hi + 1 where > 0 is a regularization parameter. Nesterov & Polyak (2006) proved that fixing a starting point x0, and performing the updating rule xt = xt 1 + h(xt 1), the algorithm can output a sequence xi that converges to a local minimum provided that the function is Hessian Lipschitz. However, it can be seen that to solve the subproblem (1.3), one needs to calculate the full gradient r F(x) and Hessian r2F(x), which is a big overhead in large scale machine learning problem because n is often very large. Some recent studies presented various algorithms to avoid the calculation of full gradient and Hessian in cubic regularization. Kohler & Lucchi (2017) used subsampling technique to get approximate gradient and Hessian instead of exact ones, and Xu et al. (2017c) also used subsampled Hessian. Both of them can reduce the computational complexity in some circumstance. However, just like other samplingbased algorithm such as subsampled Newton method Erdogdu & Montanari (2015); Xu et al. (2016); Roostakho- Stochastic Variance-Reduced Cubic Regularized Newton Methods rasani & Mahoney (2016a;b); Ye et al. (2017), their convergence rates are worse than that of the Newton method, especially when one needs a high-accuracy solution (i.e., the optimization error is small). This is because the subsampling size one needs to achieve certain accuracy may be even larger than the full sample size n. Therefore, a natural question arises as follows: When we need a high-accuracy local minimum, is there an algorithm that can output an approximate local minimum with better second-order oracle complexity than cubic regularized Newton method? In this paper, we give an affirmative answer to the above question. We propose a novel cubic regularization algorithm named Stochastic Variance-Reduced Cubic regularization (SVR Cubic, or SVRC for short), which incorporates the variance reduction techniques (Johnson & Zhang, 2013; Xiao & Zhang, 2014; Allen-Zhu & Hazan, 2016; Reddi et al., 2016a) into the cubic-regularized Newton method. The key component in our algorithm is a novel semi-stochastic gradient, together with a semi-stochastic Hessian, that are specifically designed for cubic regularization. Furthermore, we prove that, for -Hessian Lipschitz functions, to attain an ( , p )-approximate local minimum, our proposed algorithm requires O(n + n4/5/ 3/2) Second-order Oracle (SO) calls and O(1/ 3/2) Cubic Subproblem Oracle (CSO) calls. Here an SO oracle represents an evaluation of triple (fi(x), rfi(x), r2fi(x)), and a CSO oracle denotes an evaluation of the exact solution (or inexact solution) of the cubic subproblem (1.3). Compared with the original cubic regularization algorithm (Nesterov & Polyak, 2006), which requires O(n/ 3/2) SO calls and O(1/ 3/2) CSO calls, our proposed algorithm reduces the SO calls by a factor of (n1/5). We also carry out experiments on real data to demonstrate the superior performance of our algorithm. Our major contributions are summarized as follows: We present a novel cubic regularization method with im- proved oracle complexity. To the best of our knowledge, this is the first algorithm that outperforms cubic regularization without any loss in convergence rate. In sharp contrast, existing subsampled cubic regularization methods (Kohler & Lucchi, 2017; Xu et al., 2017b) suffer from worse convergence rates than cubic regularization. We also extend our algorithm to the case with inexact solution to the cubic regularization subproblem. Similar to previous work (Cartis et al., 2011a; Xu et al., 2017b), we also layout a set of sufficient conditions, under which the output of the inexact algorithm is still guaranteed to have the same convergence rate and oracle complexity as the exact algorithm. This further sheds light on the practical implementation of our algorithm. As far as we know, our work is the first to rigorously demonstrates the advantage of variance reduction for second-order optimization algorithms. Although there exist a few studies (Lucchi et al., 2015; Moritz et al., 2016; Rodomanov & Kropotov, 2016) using variance reduction to accelerate Newton method, none of them can deliver faster rates of convergence than standard Newton method. The remainder of this paper is organized as follows: We first review the most related work in Section 2. Then we present our algorithm and its main theoretical results in Sections 3 and 4 respectively. We discuss the proposed algorithm with inexact oracles in Section 5. In Section 6, we present our experimental results. Finally, we draw the conclusions in Section 7. Notation We use [n] to denote index set {1, 2, . . . , n}. We denote vector Euclidean norm by kvk2. For symmetric matrix H 2 Rd d, we denote its eigenvalues by λ1(H) . . . λd(H), its spectral norm by k Hk2 = max{|λ1(H)|, |λd(H)|}, and the Schatten r-norm by k Hk Sr = (Pd i=1 |λi(H)|r)1/r for r 1. We denote A B if λ1(A B) 0 for symmetric matrices A, B 2 Rd d. Note that k A Bk2 C ) k Ak2 k Bk2 C I, C > 0. We call a Rademacher random variable if P( = 1) = P( = 1) = 1/2. We use fn = O(gn) to denote that fn Cgn for some constant C > 0 and use fn = e O(gn) to hide the logarithmic terms of gn. 2 Related Work In this section, we briefly review the relevant work in the literature. The most related work to ours is the cubic regularized Newton method, which was originally proposed in Nesterov & Polyak (2006). Cartis et al. (2011a) presented an adaptive framework of cubic regularization, which uses an adaptive estimation of the local Lipschitz constant and approximate solution to the cubic subproblem. To connect cubic regularization with traditional trust region method (Conn et al., 2000; Cartis et al., 2009; 2012; 2013), Blanchet et al. (2016); Curtis et al. (2017); Mart ınez & Raydan (2017) showed that the trust-region Newton method can achieve the same iteration complexity as the cubic regularization method. To overcome the computational burden of gradient and Hessian matrix evaluations, Kohler & Lucchi (2017); Xu et al. (2017b;c) proposed to use subsampled gradient and Hessian in cubic regularization. On the other hand, in order to solve the cubic subproblem (1.3) more efficiently, Carmon & Duchi (2016) proposed to use gradient descent, while Agarwal et al. (2017) proposed a sophisticated algorithm based on approximate matrix inverse and approximate PCA. Tripuraneni et al. (2017) proposed a refined stochastic cubic regularization algorithm based on above subproblem solver. However, none of the aforementioned variants of cubic reg- Stochastic Variance-Reduced Cubic Regularized Newton Methods Table 1. Comparisons between different methods to find ( , p )-local minimum on the second-order oracle (SO) complexity and and the cubic sub-problem oracle (CSO) complexity. Algorithm SO calls CSO calls Gradient Lipschitz Hessian Lipschitz Cubic regularization O(n/ 3/2) O(1/ 3/2) no yes (Nesterov & Polyak, 2006) Subsampled cubic regularization e O(n/ 3/2 + 1/ 5/2)1 O(1/ 3/2) yes yes (Kohler & Lucchi, 2017; Xu et al., 2017c) SVRC e O(n + n4/5/ 3/2) O(1/ 3/2) no yes (this paper) ularization outperforms the original cubic regularization method in terms of oracle complexity. Another line of related research is the variance reduction method, which has been extensively studied for large-scale finite-sum optimization problems. Variance reduction was first proposed in convex finite-sum optimization (Roux et al., 2012; Johnson & Zhang, 2013; Xiao & Zhang, 2014; Defazio et al., 2014), which uses semi-stochastic gradient to reduce the variance of the stochastic gradient and improves the gradient complexity of both stochastic gradient descent (SGD) and gradient descent (GD). Representative algorithms include Stochastic Average Gradient (SAG) (Roux et al., 2012), Stochastic Variance Reduced Gradient (SVRG) (Johnson & Zhang, 2013) and SAGA (Defazio et al., 2014), to mention a few. Garber & Hazan (2015); Shalev-Shwartz (2016) studied non-convex finite-sum problems where each individual function may be non-convex, but their sum is still convex. Reddi et al. (2016a) and Allen-Zhu & Hazan (2016) extended the SVRG algorithm to the general non-convex finite-sum optimization, which outperforms SGD and GD in terms of gradient complexity as well. However, to the best of our knowledge, it is still an open problem whether variance reduction can improve the oracle complexity of second-order optimization algorithms. Last but not least is the vast literature of research which aims to escape from nondegenerated saddle points by finding the negative curvature direction. Ge et al. (2015); Jin et al. (2017a) showed that simple (stochastic) gradient descent with perturbation can escape from saddle points. Carmon et al. (2016); Royer & Wright (2017); Allen-Zhu (2017) showed that by calculating the negative curvature using Hessian information, one can find ( , p )-approximate local minima faster than first-order methods. Recent work (Xu & Yang, 2017; Allen-Zhu & Li, 2017; Jin et al., 2017b) proposed first-order algorithms that can escape from saddle points without using Hessian information. Yu et al. (2017b) proposed the GOSE algorithm to save negative curvature computation and Yu et al. (2017a) improved the gradient complexity by using third-order smoothness. Raginsky et al. (2017); Zhang et al. (2017); Xu et al. (2017a) proved that a family of algorithms based on discretizations of Langevin dynamics can find a neighborhood of the global minimum of nonconvex objective functions. For better comparison of our algorithm with the most related algorithms in terms of SO and CSO oracle complexities, we summarize the results in Table 1. It can be seen from Table 1 that our algorithm (SVRC) achieves the lowest (SO and CSO) oracle complexity compared with the original cubic regularization method (Nesterov & Polyak, 2006) which employs full gradient and Hessian evaluations and the subsampled cubic method (Kohler & Lucchi, 2017; Xu et al., 2017c). In particular, our algorithm reduces the SO oracle complexity of cubic regularization by a factor of n1/5 for finding an ( , p )-approximate local minimum. We will provide more detailed discussion in Section 4. 3 The Proposed Algorithm In this section, we present a novel algorithm, which utilizes stochastic variance reduction techniques to improve cubic regularization method. To reduce the computation burden of gradient and Hessian matrix evaluations in the cubic regularization updates in (1.3), subsampled gradient and Hessian matrix have been used in subsampled cubic regularization (Kohler & Lucchi, 2017; Xu et al., 2017c) and stochastic cubic regularization (Tripuraneni et al., 2017). Nevertheless, the stochastic gradient and Hessian matrix have large variances, which undermine the convergence performance. Inspired by SVRG (Johnson & Zhang, 2013), we propose to use a semi-stochastic version of gradient and Hessian matrix, which can control the variances automatically. Specifically, our algorithm has two loops. At the beginning of the s-th iteration of the outer loop, we denote bxs = xs 0. We first calculate the full gradient gs = r F(bxs) and Hessian matrix Hs = r2F(bxs), which are stored for further references in the inner loop. At the t-th iteration of the inner loop, we calculate the following semi-stochastic gradient and Hessian 1It is the refined rate proved by Xu et al. (2017c) for the subsampled cubic regularization algorithm proposed in Kohler & Lucchi (2017) Stochastic Variance-Reduced Cubic Regularized Newton Methods Algorithm 1 Stochastic Variance Reduction Cubic Regularization (SVRC) 1: Input: batch size bg, bh, cubic penalty parameter {Ms,t}, epoch number S, epoch length T and starting point x0. 2: Initialization bx1 = x0 3: for s = 1, . . . , S do 4: xs 5: gs = r F(bxs) = 1 i=1 rfi(bxs), Hs = 1 i=1 r2fi(bxs) 6: for t = 0, . . . , T 1 do 7: Sample index set Ig, Ih, |Ig| = bg, |Ih| = bh; 8: vs t) rfit(bxs) it2Ig r2fit(bxs) Hs* t) r2fjt(bxs) t = argminhvs th, hi + Ms,t t 12: end for 13: bxs+1 = xs T 14: end for 15: Output: xout = xs t, where s, t are uniformly random chosen from s 2 [S] and t 2 [T]. t) rfit(bxs) r2fit(bxs) Hs* t) r2fjt(bxs) + Hs, (3.2) where Ig and Ih are batch index sets, and the batch sizes will be decided later. In each inner iteration, we solve the following cubic regularization subproblem: t = argmin ms th, hi + Ms,t where {Ms,t} are cubic regularization parameter, which may depend on s and t. Then we perform the update xs t in the t-th iteration of the inner loop. The proposed algorithm is displayed in Algorithm 1. There are two notable features of our estimator of the full gradient and Hessian in each inner loop, compared with that used in SVRG (Johnson & Zhang, 2013). The first is that our gradient and Hessian estimators consist of minibatches of stochastic gradient and Hessian. The second one is that we use second order information when we construct the gradient estimator vs t, while classical SVRG only uses first order information to build it. Intuitively speaking, both features are used to make a more accurate estimation of the true gradient and Hessian with affordable oracle calls. Note that similar approximations of the gradient and Hessian matrix have been staged in recent work by Gower et al. (2017) and Wai et al. (2017), where they used this new kind of estimator for traditional SVRG in the convex setting, which radically differs from our setting. 4 Main Theory We first lay out the following Hessian Lipschitz assumption, which are necessary for our analysis and are widely used in the literature (Nesterov & Polyak, 2006; Xu et al., 2016; Kohler & Lucchi, 2017). Assumption 4.1 (Hessian Lipschitz). There exists a constant > 0, such that for all x, y and i 2 [n] ++r2fi(x) r2fi(y) In fact, this is the only assumption we need to prove our theoretical results. The Hessian Lipschitz assumption plays a central role in controlling the changing speed of second order information. It is obvious that Assumption 4.1 implies the Hessian Lipschitz assumption of F, which, according to Nesterov & Polyak (2006), is also equivalent to the following lemma. Lemma 4.2. Let function F : Rd ! R satisfy -Hessian Lipschitz assumption, then for any h 2 Rd, it holds that ++r2F(x) r2F(y) F(x + h) F(x) + hr F(x), hi + 1 2, ++r F(x + h) r F(x) r2F(x)h We then define the following optimal function gap between initial point x0 and the global minimum of F. Definition 4.3 (Optimal Gap). For function F( ) and the initial point x0, let F be F = inf{ 2 R : F(x0) F }, where F = infx2Rd F(x). Stochastic Variance-Reduced Cubic Regularized Newton Methods W.L.O.G., we assume F < +1 throughout this paper. Before we present nonasympotic convergence results of Algorithm 1, we define kr F(x)k3/2 By definition in (4.1), µ(x) 3/2 holds if and only if kr F(x)k2 , λmin > p . (4.2) Therefore, in order to find an ( , p )-approximate local minimum of the non-convex function F, it suffices to find a point x which satisfies µ(x) < 3/2. Next we define our oracles formally: Definition 4.4 (Second-order Oracle). Given an index i and a point x, one second-order oracle (SO) call returns such a triple: [fi(x), rfi(x), r2fi(x)]. (4.3) Definition 4.5 (Cubic Subproblem Oracle). Given a vector g 2 Rd, a Hessian matrix H and a positive constant , one Cubic Subproblem Oracle (CSO) call returns hsol, where hsol can be solved exactly as follows hsol = argmin h2Rd hg, hi + 1 Remark 4.6. The second-order oracle is a special form of Information Oracle introduced by Nesterov, which returns gradient, Hessian and all high order derivatives of the objective function F(x). Here, our second-order oracle will only returns first and second order information at some point of single objective fi instead of F. We argue that it is a reasonable adaption because in this paper we focus on finitesum objective function. The Cubic Subproblem Oracle will return an exact or inexact solution of (3.3), which plays an important role in both theory and practice. Now we are ready to give a general convergence result of Algorithm 1: Theorem 4.7. Under Assumption 4.1, suppose that the cubic regularization parameter Ms,t of Algorithm 1 satisfies that Ms,t = CM , where is the Hessian Lipschitz parameter and CM 100 is a constant. The batch sizes bg and bh satisfy that bg 5T 4, bh 100T 2 log d, (4.4) where T 2 is the length of the inner loop of Algorithm 1 and d is the dimension of the problem. Then the output of Algorithm 1 satisfies E[µ(xout)] 240C2 Remark 4.8. According to (4.1), to ensure that xout is an ( , p )-approximate local minimum, we can set the right hand side of (4.5) to be less then 3/2. This immediately implies that the total iteration complexity of Algorithm 1 is ST = O( F 1/2 3/2), which matches the iteration complexity of cubic regularization (Nesterov & Polyak, 2006). Remark 4.9. Note that there is a log d term in the expression of parameter, and it is only related to Hessian batch size bh. The log d term comes from matrix concentration inequalities, which is believed to be unavoidable (Tropp et al., 2015). In other words, the batch size of Hessian matrix bh has a inevitable relation to dimension d, unlike the batch size of gradient bg. The complexity result in Theorem 4.7 depends on a series of parameter. In the following corollary, we will show how to choose these parameters in practice to achieve a better oracle complexity. Corollary 4.10. Under Assumption 4.1, let the cubic regularization parameter Ms,t = M = CM , where CM 100 is a constant. Let the epoch length T = n1/5, batch sizes bg = 5n4/5 and bh = 100n2/5 log d, and the number of epochs S = max{1, 240C2 M 1/2 F n 1/5 3/2}. Then Algorithm 1 will find an ( , p )-approximate local minimum xout within n + F p n4/5 SO calls (4.6) CSO calls. (4.7) Remark 4.11. Corollary 4.10 states that we can reduce the SO calls by setting the batch size bg, bh related to n. In contrast, in order to achieve an ( , p ) local minimum, original cubic regularization method in Nesterov & Polyak (2006) needs O(n/ 3/2) second-order oracle calls, which is by a factor of n1/5 worse than ours. And subsampled cubic regularization (Kohler & Lucchi, 2017; Xu et al., 2017c) requires e O(n/ 3/2 + 1/ 5/2) SO calls, which is also worse than our algorithm. 5 SVRC with Inexact Oracles In practice, the exact solution to the cubic subproblem (3.3) cannot be obtained. Instead, one can only get an approximate solution by some inexact solver. Thus we replace the CSO oracle in (4.5) with the following inexact CSO oracle ehsol argmin h2Rd hg, hi + 1 To analyze the performance of Algorithm 1 with inexact cubic subproblem solver, we relax the exact solver in Line Stochastic Variance-Reduced Cubic Regularized Newton Methods (a) a9a (b) covtype (c) ijcnn1 (d) a9a (e) covtype (f) ijcnn1 Figure 1. Logarithmic function value gap for nonconvex regularized logistic regression on different datasets. (a), (b) and (c) present the oracle complexity comparison; (d), (e) and (f) present the runtime comparison. 10 of Algorithm 1 with t argmin ms t(h). (5.1) The ultimate goal of this section is to prove that the theoretical results of our SVRC algorithm still hold with inexact subproblem solvers. To this end, we present the following sufficient condition, under which inexact solution can ensure the same oracle complexity as the exact solution: Condition 5.1 (Inexact condition). For each s, t and given δ > 0, ehs t satisfies δ-inexact condition if ehs s,t δ2/3, ..kehs Remark 5.2. Similar inexact conditions have been studied in the literature of cubic regularization. For instance, Nesterov & Polyak (2006) presented a practical way to solve the cubic subproblem without termination condition. Cartis et al. (2011a); Kohler & Lucchi (2017) presented termination criteria for approximate solution to cubic subproblem, which is slightly different from Condtion 5.1. Now we present the convergence result of SVRC with inexact CSO oracles: Theorem 5.3. Suppose that for each s, t, ehs t is an inexact solver of cubic subproblem ms t(h), which satisfies Condition 5.1. Under the same conditions of Theorem 4.7, the output of Algorithm 1 satisfies E[µ(xout)] 240C2 M 1/2δ. (5.2) Remark 5.4. By the definition of µ(x), in order to attain an ( , p )-approximate local minimum, we require E[µ(xout)] 3/2 and thus 480C2 M 1/2δ < 3/2, which implies that δ in Condition 5.1 should satisfy δ < (480C2 M 1/2) 1 3/2. Thus the total iteration complexity of Algorithm 1 with inexact oracle is still O( F 1/2 3/2). By the same choice of parameters, Algorithm 1 with inexact oracle can achieve a reduction in SO calls. Corollary 5.5. Suppose that for each s, t, ehs t is an inexact solver of cubic subproblem ms t(h), which satisfies Condition 5.1 with δ = (960C2 M) 1 1/2 3/2. Under Assumption 4.1, let the cubic regularization parameter Ms,t = M = CM , where CM 100 is a constant. Let the epoch length T = n1/5, batch sizes bg = 5n4/5 and bh = 100n2/5 log d, and the number of epochs S = max{1, 480C2 M 1/2 F n 1/5 3/2}. Then Algorithm 1 will find an ( , p )-approximate local minimum within n + F p n4/5 SO calls (5.3) CSO calls. (5.4) Stochastic Variance-Reduced Cubic Regularized Newton Methods (a) a9a (b) covtype (c) ijcnn1 (d) a9a (e) covtype (f) ijcnn1 Figure 2. Logarithmic function value gap for nonlinear least square on different datasets. (a), (b) and (c) present the oracle complexity comparison; (d), (e) and (f) present the runtime comparison. Remark 5.6. It is worth noting that even with the inexact CSO oracle satisfying Condition 5.1, the SO and CSO complexities of SVRC remain the same as that of SVRC with exact CSO oracle. Furthermore, this result always holds with any inexact cubic sub-problem solver. 6 Experiments In this section, we present numerical experiments on different non-convex Empirical Risk Minimization (ERM) problems and on different datasets to validate the advantage of our SVRC algorithm in finding approximate local minima. Baselines: We compare our algorithm with adaptive cubic regularization (Adaptive Cubic) (Cartis et al., 2011a), subsampled cubic regularization (Subsampled Cubic) (Kohler & Lucchi, 2017), stochastic cubic regularization (Stochastic Cubic) (Tripuraneni et al., 2017), gradient cubic regularization (Gradient Cubic) (Carmon & Duchi, 2016) and trust region Newton method (TR) (Conn et al., 2000). All algorithms are carefully tuned for a fair comparison. Calculation for SO calls: For Subsampled Cubic, each loop takes (Bg + Bh) SO calls, where Bg and Bh are the subsampling sizes of gradient and Hessian. For Stochastic Cubic, each loop costs (ng+nh) SO calls, where ng and nh denote the subsampling sizes of gradient and Hessian-vector operator. Gradient Cubic, Adaptive Cubic and TR cost n SO calls in each loop. We define the amount of epochs to be the amount of SO calls divided by n. Parameters and subproblem solver: For each algorithm and each dataset, we choose different bg, bh, T for the best performance. Meanwhile, we choose Ms,t = /(1 + β)(s+t/T ), , β > 0 for each iteration. When β = 0, it has been proved to enjoy good convergence performance. This choice of parameter is similar to the choice of penalty parameter in Subsampled Cubic and Adaptive Cubic, which sometimes makes some algorithms behave better in our experiment. For the subproblem solover of (3.3) in each loop, we choose the Lanczos-type method (Cartis et al., 2011a). Datasets: The datasets we use are a9a, covtype, ijcnn1, which are common datasets used in ERM problems. The detailed information about these datasets are in Table 2. Table 2. Overview of the datasets used in our experiments Dataset sample size n dimension d a9a 32,561 123 covtype 581,012 54 ijcnn1 35,000 22 Non-convex regularized logistic regression: The first nonconvex problem we study is a binary logistic regression problem with a non-convex regularizer Pd (i)/(1 + w2 (i)) (Reddi et al., 2016b). Specifically, suppose we are given training data {xi, yi}n i=1, where xi 2 Rd and yi 2 {0, 1} are feature vectors and labels corresponding to the i-th data Stochastic Variance-Reduced Cubic Regularized Newton Methods (a) a9a (b) covtype (c) ijcnn1 (d) a9a (e) covtype (f) ijcnn1 Figure 3. Logarithmic function value gap for robust linear regression on different datasets. (a), (b) and (c) present the oracle complexity comparison; (d), (e) and (f) present the runtime comparison. points. The minimization problem is as follows yi log φ(x> i w) + (1 yi) log[1 φ(x> (i)/(1 + w2 where φ(x) = 1/(1 + exp( x)) is the sigmoid function. We fix λ = 10 in our experiments. Recall the definition Ms,t = /(1+β)(s+t/T ), , β > 0. We set = 0.05, β = 0 for a9a and ijcnn1 datasets and = 5e3, β = 0.15 for covtype. The experiment results are shown in Figure 1. Nonlinear linear squares: The second problem is a nonlinear least squares problem which focuses on the task of binary linear classification (Xu et al., 2017b). Given training data {xi, yi}n i=1, where xi 2 Rd and yi 2 {0, 1} are feature vectors and labels corresponding to the i-th data points. The minimization problem is Here φ is the sigmoid function. We set = 0.05, 1e8, 0.003 and β = 0, 1, 0.5 for a9a, covtype and ijcnn1 datasets respectively. The experiment results are shown in Figure 2. Robust linear regression: The third problem is a robust linear regression problem where we use a non-convex robust loss function log(x2/2+1) (Barron, 2017) instead of square loss in least square regression. Given a training sample i=1, where xi 2 Rd and yi 2 {0, 1} are feature vectors and labels corresponding to the i-th data point. The minimization problem is where (x) = log(x2/2 + 1). We set = 0.1, 1e9, 2 and β = 0.1, 1, 0 for a9a, covtype and ijcnn1 datasets respectively. The experimental results are shown in Figure 3. From Figures 1, 2 and 3, we can see that our algorithm SVRC outperforms all the other baseline algorithms on all the datasets. The only exception happens in the non-linear least square problem and the robust linear regression problem on the covtype dataset, where our algorithm behaves a little worse than Adaptive Cubic at the high accuracy regime in terms of epoch counts. However, under this setting, our algorithm still outperforms the other baselines in terms of the cpu time. 7 Conclusions In this paper, we propose a novel second-order algorithm for non-convex optimization called SVRC. Our algorithm is the first algorithm which improves the oracle complexity of cubic regularization and its subsampled variants under certain regime using variance reduction techniques. We also show that similar oracle complexity also holds with inexact oracles. Under both settings our algorithm outperforms the state-of-the-art methods. Stochastic Variance-Reduced Cubic Regularized Newton Methods Acknowledgements We would like to thank the anonymous reviewers for their helpful comments. This research was sponsored in part by the National Science Foundation IIS-1618948 and IIS1652539. The views and conclusions contained in this paper are those of the authors and should not be interpreted as representing any funding agencies. Agarwal, N., Allenzhu, Z., Bullins, B., Hazan, E., and Ma, T. Finding approximate local minima for nonconvex optimization in linear time. 2017. Allen-Zhu, Z. Natasha 2: Faster non-convex optimization than sgd. ar Xiv preprint ar Xiv:1708.08694, 2017. Allen-Zhu, Z. and Hazan, E. Variance reduction for faster non- convex optimization. In International Conference on Machine Learning, pp. 699 707, 2016. Allen-Zhu, Z. and Li, Y. Neon2: Finding Local Minima via First- Order Oracles. Ar Xiv e-prints, abs/1711.06673, November 2017. Full version available at http://arxiv.org/abs/1711. 06673. Barron, J. T. A more general robust loss function. ar Xiv preprint ar Xiv:1701.03077, 2017. Blanchet, J., Cartis, C., Menickelly, M., and Scheinberg, K. Con- vergence rate analysis of a stochastic trust region method for nonconvex optimization. ar Xiv preprint ar Xiv:1609.07428, 2016. Carmon, Y. and Duchi, J. C. Gradient descent efficiently finds the cubic-regularized non-convex newton step. 2016. Carmon, Y., Duchi, J. C., Hinder, O., and Sidford, A. Accelerated methods for non-convex optimization. 2016. Cartis, C., Gould, N. I., and Toint, P. L. Trust-region and other regularisations of linear least-squares problems. BIT Numerical Mathematics, 49(1):21 53, 2009. Cartis, C., Gould, N. I., and Toint, P. L. Adaptive cubic regularisa- tion methods for unconstrained optimization. part i: motivation, convergence and numerical results. Mathematical Programming, 127(2):245 295, 2011a. Cartis, C., Gould, N. I. M., and Toint, P. L. Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case functionand derivative-evaluation complexity. Springer-Verlag New York, Inc., 2011b. Cartis, C., Gould, N. I., and Toint, P. L. Complexity bounds for second-order optimality in unconstrained optimization. Journal of Complexity, 28(1):93 108, 2012. Cartis, C., Gould, N. I., and Toint, P. L. On the evaluation com- plexity of cubic regularization methods for potentially rankdeficient nonlinear least-squares problems and its relevance to constrained nonlinear optimization. SIAM Journal on Optimization, 23(3):1553 1574, 2013. Chen, R. Y., Gittens, A., and Tropp, J. A. The masked sample covariance estimator: an analysis using matrix concentration inequalities. Information and Inference: A Journal of the IMA, 1(1):2 20, 2012. Conn, A. R., Gould, N. I., and Toint, P. L. Trust region methods. SIAM, 2000. Curtis, F. E., Robinson, D. P., and Samadi, M. A trust region algorithm with a worst-case iteration complexity of o( 3/2) for nonconvex optimization. Mathematical Programming, 162 (1-2):1 32, 2017. Defazio, A., Bach, F., and Lacoste-Julien, S. Saga: A fast incre- mental gradient method with support for non-strongly convex composite objectives. In Advances in Neural Information Processing Systems, pp. 1646 1654, 2014. Durrett, R. Probability: theory and examples. Cambridge univer- sity press, 2010. Erdogdu, M. A. and Montanari, A. Convergence rates of sub- sampled newton methods. In Proceedings of the 28th International Conference on Neural Information Processing Systems Volume 2, pp. 3052 3060. MIT Press, 2015. Garber, D. and Hazan, E. Fast and simple pca via convex optimiza- tion. ar Xiv preprint ar Xiv:1509.05647, 2015. Ge, R., Huang, F., Jin, C., and Yuan, Y. Escaping from saddle pointsonline stochastic gradient for tensor decomposition. In Conference on Learning Theory, pp. 797 842, 2015. Ge, R., Lee, J. D., and Ma, T. Matrix completion has no spurious local minimum. In Advances in Neural Information Processing Systems, pp. 2973 2981, 2016. Ghadimi, S. and Lan, G. Stochastic first-and zeroth-order meth- ods for nonconvex stochastic programming. SIAM Journal on Optimization, 23(4):2341 2368, 2013. Ghadimi, S. and Lan, G. Accelerated gradient methods for non- convex nonlinear and stochastic programming. Mathematical Programming, 156(1-2):59 99, 2016. Gower, R. M., Roux, N. L., and Bach, F. Tracking the gradients using the hessian: A new look at variance reducing stochastic methods. ar Xiv preprint ar Xiv:1710.07462, 2017. Hillar, C. J. and Lim, L.-H. Most tensor problems are np-hard. Journal of the ACM (JACM), 60(6):45, 2013. Jin, C., Ge, R., Netrapalli, P., Kakade, S. M., and Jordan, M. I. How to escape saddle points efficiently. 2017a. Jin, C., Netrapalli, P., and Jordan, M. I. Accelerated gradient descent escapes saddle points faster than gradient descent. ar Xiv preprint ar Xiv:1711.10456, 2017b. Johnson, R. and Zhang, T. Accelerating stochastic gradient descent using predictive variance reduction. In Advances in neural information processing systems, pp. 315 323, 2013. Kohler, J. M. and Lucchi, A. Sub-sampled cubic regularization for non-convex optimization. ar Xiv preprint ar Xiv:1705.05933, 2017. Le Cun, Y., Bengio, Y., and Hinton, G. Deep learning. Nature, 521 (7553):436 444, 2015. Stochastic Variance-Reduced Cubic Regularized Newton Methods Lucchi, A., Mc Williams, B., and Hofmann, T. A variance reduced stochastic newton method. ar Xiv preprint ar Xiv:1503.08316, 2015. Mackey, L., Jordan, M. I., Chen, R. Y., Farrell, B., Tropp, J. A., et al. Matrix concentration inequalities via the method of exchangeable pairs. The Annals of Probability, 42(3):906 945, 2014. Mart ınez, J. M. and Raydan, M. Cubic-regularization counter- part of a variable-norm trust-region method for unconstrained minimization. Journal of Global Optimization, 68(2):367 385, 2017. Moritz, P., Nishihara, R., and Jordan, M. A linearly-convergent stochastic l-bfgs algorithm. In Artificial Intelligence and Statistics, pp. 249 258, 2016. Nesterov, Y. and Polyak, B. T. Cubic regularization of newton method and its global performance. Mathematical Programming, 108(1):177 205, 2006. Raginsky, M., Rakhlin, A., and Telgarsky, M. Non-convex learning via stochastic gradient langevin dynamics: a nonasymptotic analysis. ar Xiv preprint ar Xiv:1702.03849, 2017. Reddi, S. J., Hefny, A., Sra, S., Poczos, B., and Smola, A. Stochas- tic variance reduction for nonconvex optimization. pp. 314 323, 2016a. Reddi, S. J., Sra, S., P oczos, B., and Smola, A. Fast incremental method for smooth nonconvex optimization. In Decision and Control (CDC), 2016 IEEE 55th Conference on, pp. 1971 1977. IEEE, 2016b. Robbins, H. and Monro, S. A stochastic approximation method. The annals of mathematical statistics, pp. 400 407, 1951. Rodomanov, A. and Kropotov, D. A superlinearly-convergent proximal newton-type method for the optimization of finite sums. In International Conference on Machine Learning, pp. 2597 2605, 2016. Roostakhorasani, F. and Mahoney, M. W. Sub-sampled newton methods i: Globally convergent algorithms. 2016a. Roostakhorasani, F. and Mahoney, M. W. Sub-sampled newton methods ii: Local convergence rates. 2016b. Roux, N. L., Schmidt, M., and Bach, F. R. A stochastic gradient method with an exponential convergence rate for finite training sets. In Advances in Neural Information Processing Systems, pp. 2663 2671, 2012. Royer, C. W. and Wright, S. J. Complexity analysis of second- order line-search algorithms for smooth nonconvex optimization. ar Xiv preprint ar Xiv:1706.03131, 2017. Shalev-Shwartz, S. Sdca without duality, regularization, and in- dividual convexity. In International Conference on Machine Learning, pp. 747 754, 2016. Tripuraneni, N., Stern, M., Jin, C., Regier, J., and Jordan, M. I. Stochastic cubic regularization for fast nonconvex optimization. ar Xiv preprint ar Xiv:1711.02838, 2017. Tropp, J. A. The expected norm of a sum of independent ran- dom matrices: An elementary approach. In High Dimensional Probability VII, pp. 173 202. Springer, 2016. Tropp, J. A. et al. An introduction to matrix concentration inequal- ities. Foundations and Trends R in Machine Learning, 8(1-2): 1 230, 2015. Wai, H.-T., Shi, W., Nedic, A., and Scaglione, A. Curvature- aided incremental aggregated gradient method. ar Xiv preprint ar Xiv:1710.08936, 2017. Xiao, L. and Zhang, T. A proximal stochastic gradient method with progressive variance reduction. SIAM Journal on Optimization, 24(4):2057 2075, 2014. Xu, P., Yang, J., Roosta-Khorasani, F., R, C., and Mahoney, M. W. Sub-sampled newton methods with non-uniform sampling. 2016. Xu, P., Chen, J., Zou, D., and Gu, Q. Global convergence of langevin dynamics based algorithms for nonconvex optimization. ar Xiv preprint ar Xiv:1707.06618, 2017a. Xu, P., Roosta-Khorasan, F., and Mahoney, M. W. Second-order optimization for non-convex machine learning: An empirical study. ar Xiv preprint ar Xiv:1708.07827, 2017b. Xu, P., Roosta-Khorasani, F., and Mahoney, M. W. Newton-type methods for non-convex optimization under inexact hessian information. ar Xiv preprint ar Xiv:1708.07164, 2017c. Xu, Y. and Yang, T. First-order stochastic algorithms for escap- ing from saddle points in almost linear time. ar Xiv preprint ar Xiv:1711.01944, 2017. Ye, H., Luo, L., and Zhang, Z. Approximate newton methods and their local convergence. In International Conference on Machine Learning, pp. 3931 3939, 2017. Yu, Y., Xu, P., and Gu, Q. Third-order smoothness helps: Even faster stochastic optimization algorithms for finding local minima. ar Xiv preprint ar Xiv:1712.06585, 2017a. Yu, Y., Zou, D., and Gu, Q. Saving gradient and negative curvature computations: Finding local minima more efficiently. ar Xiv preprint ar Xiv:1712.03950, 2017b. Zhang, Y., Liang, P., and Charikar, M. A hitting time analysis of stochastic gradient langevin dynamics. ar Xiv preprint ar Xiv:1702.05575, 2017.