# moment_distributionally_robust_tree_structured_prediction__9199fcd7.pdf Moment Distributionally Robust Tree Structured Prediction Yeshu Li Danyal Saeed Xinhua Zhang Brian D. Ziebart Department of Computer Science University of Illinois at Chicago {yli299, dsaeed3, zhangx, bziebart}@uic.edu Kevin Gimpel Toyota Technological Institute at Chicago kgimpel@ttic.edu Structured prediction of tree-shaped objects is heavily studied under the name of syntactic dependency parsing. Current practice based on maximum likelihood or margin is either agnostic to or inconsistent with the evaluation loss. Risk minimization alleviates the discrepancy between training and test objectives but typically induces a non-convex problem. These approaches adopt explicit regularization to combat overfitting without probabilistic interpretation. We propose a momentbased distributionally robust optimization approach for tree structured prediction, where the worst-case expected loss over a set of distributions within bounded moment divergence from the empirical distribution is minimized. We develop efficient algorithms for arborescences and other variants of trees. We derive Fisher consistency, convergence rates and generalization bounds for our proposed method. We evaluate its empirical effectiveness on dependency parsing benchmarks. 1 Introduction Structured prediction is an important learning setting for joint prediction of interdependent variables. The output space typically consists of an exponential number of structured objects whose inherent relations can be exploited to develop efficient learning algorithms and capture key properties of data [Ciliberto et al., 2019]. Trees are widely used structures that offer expressiveness and simplicity. We distinguish between two different tree structured prediction tasks in the literature. The first task is a structure learning problem in graphical models [Bradley and Guestrin, 2010], aimed at constructing trees underlying a predictive model from training data. The optimal tree is found easily with greedy algorithms for generative models [Chow and Liu, 1968], while it is NP-hard for the discriminative max-margin setting [Meshi et al., 2013]. The second task requires prediction itself to be a tree-shaped object (e.g., an incidence vector). Dependency parsing is a crucial application of this problem that has inspired a flurry of work in natural language processing. The first-order spanning tree prediction assuming factorization over arcs can be done in O(n2) [Stanojevi c and Cohen, 2021], whereas exact inference is NP-hard for certain (non-projective) higher-order trees (e.g., considering siblings) [Mc Donald and Satta, 2007]. We study the latter in this work. A common evaluation criterion in dependency parsing is the attachment score, namely, the score we would like to maximize on test data. It is cost-sensitive to allow partially correct prediction. Ideally, the training objective should be aligned with the test objective. An early attempt to directly mimic test conditions leads to a non-convex piece-wise constant objective [Och, 2003]. Risk minimization in appropriate parametric form has a non-convex smooth objective, solvable with gradient descent, 36th Conference on Neural Information Processing Systems (Neur IPS 2022). but still losing global convergence and generalization guarantees. Maximum likelihood approaches formulate a convex smooth problem minimizing a logistic loss, consistent with conditional probability estimates but oblivious to test losses. Maximum margin methods have convex objectives able to implicitly incorporate custom losses by scaling margins, but are known to be inconsistent with test losses generally [Nowak-Vila et al., 2021]. Unfortunately, none of these approaches yield a Bayes optimal estimator for test losses with global convergence and finite-sample generalization guarantees. Consistent structured prediction methods include Ciliberto et al. [2016], Blondel [2019], Nowak-Vila et al. [2020], the latter two of which are based on Fenchel-Young losses [Blondel et al., 2020]. However, none of them have addressed the tree structured prediction problem explicitly. For instance, Blondel [2019] calls for Euclidean or Kullback-Leibler projection oracles, which do not exist in an efficient sense from what we know for arborescence (directed tree) polytopes. In addition, the Frank Wolfe type algorithm adopted by Nowak-Vila et al. [2020] requires a max-min oracle and converges in a rate of O( 1 ϵ ). Furthermore, all of the above methods belong to empirical risk minimization (ERM) that requires explicit regularization to combat overfitting, which can be quite vulnerable in high-dimensional settings (e.g., scarce data). To address the above issues, we propose an estimator from first principles in distributionally robust optimization (DRO). It minimizes the worst-case risk over an ambiguity set of distributions within bounded moment divergence from the empirical distribution. We seek probabilistic prediction by assuming non-deterministic groundtruth labels, which, together with the ambiguity set, models uncertainty about the unknown true distribution. We interpret the primal problem as a dual-normregularized surrogate loss minimization problem. Note that prior art applying moment-based DRO to tree-structured graphical models [Fathony et al., 2018b] and bipartite matching [Fathony et al., 2018a] adopts a special case of our ambiguity set in which the empirical feature moments are matched exactly and regularization has to be imposed manually. This moment-based DRO also allows us to derive generalization bounds regarding true worst-case risks. When the ambiguity radius is zero, the DRO estimator is shown to be consistent. We develop two practical algorithms, one based on game theory and the other based on marginal probabilities of tree parts. We further propose efficient Euclidean projection oracles onto the arborescence polytope with linearly convergent guarantees. We conduct experiments on three common dependency parsing datasets, suggesting that our method is particularly effective with little training data. Contributions. Our contributions are summarized as follows. (1) We propose a distributionally robust tree structured prediction method and show its equivalence to regularized surrogate minimization. (2) We derive its generalization bounds and consistency. (3) We propose efficient algorithms based on projection oracles for arborescence polytopes. (4) We perform empirical study on real-world datasets. Paper structure. We begin with problem setup and existing work in Section 2. We present our method with theoretical analysis in Section 3. Section 4 proposes efficient projection oracles. Section 5 discusses extensions beyond first-order directed trees. Experimental results of comparing our method with a competitive baseline are given in Section 6. We conclude the paper in Section 7. 2 Background and Related Works 2.1 Tree Structured Prediction Consider a weighted directed multi-graph G = (V, E) where each arc (i, j, l) E from node i to j has a label l. By designating a root node r V, we say that A E is an r-arborescence of G if (V, A) is a directed spanning tree rooted at r. For any v V, denote by δ (v) := {(i, j, l) E : j = v} the set of its incoming arcs, and δ+(v) := {(i, j, l) E : i = v} the set of its outgoing arcs. Let X be the input space and Y S x X Y(x) be the output space where Y(x) represents the set of rarborescences of a graph G(x) formed by x. Dependence on x is suppressed when context is clear. Let R 2E be a set of parts with E R. Each part s R is a subset of arcs. It is convenient to represent y Y as a binary vector with ys = 1 iff part s appears in y. Let wθ(x, y) P s R wθ(x, ys) be a score function decomposing over parts, parameterized by θ. Let {(x(i), y(i))}m i=1 be a set of m training examples drawn i.i.d. from a distribution P P(X Y), where each y(i) is an r-arborescence. The goal of tree structured prediction is to learn a function h : X Y from training data. Assume that the evaluation criterion is a loss function ℓ: Y Y R 0. We introduce existing methods in the setting of (graph-based, non-projective, syntactic) dependency parsing where x is a sequence of tokens and G(x) encodes dependencies among tokens. 2.2 Maximum Likelihood A probabilistic modeling approach based on exponential family distributions maximizes the conditional log-likelihood of the training data: i=1 log pθ(y(i)|x(i)) := i=1 log [exp (wθ(x(i), y(i)))/Z(x(i))], where Z(x) P y Y(x) exp (wθ(x, y)). This problem is convex for log-linear models, but intractable for general R [Koller and Friedman, 2009]. The first-order arc-factored model (R = E) is equivalent to a loop-free factor graph, rendering it tractable via the matrix-tree theorem [Kirchhoff, 1847, William, 1984, Koo et al., 2007, Mc Donald and Satta, 2007, Smith and Smith, 2007]. Neural parsers either leverage the same theorem to compute the partition function [Ma and Hovy, 2017] or consider the parent node distribution independently for each node by local normalization [Dozat and Manning, 2017, Zhang et al., 2017]. Higher-order models require approximate algorithms such as loopy belief propagation [Murphy et al., 1999] and Markov chain Monte Carlo [Brooks, 1998]. This approach does not incorporate task-specific losses. In fact, with maximum a posteriori (MAP) decoding, it is not consistent with any specific loss in general [Nowak-Vila et al., 2019]. 2.3 Maximum Margin An alternative approach based on maximum margin Markov networks [Taskar et al., 2003] or structured support vector machines [Tsochantaridis et al., 2005] optimizes a hinge-type surrogate: i=1 wθ(x(i), y(i)) + max y ℓ(y(i), y) + wθ(x(i), y), which inspires a rich line of work based on MAP inference with manual features [Taskar et al., 2004, Mc Donald et al., 2005, Mc Donald and Pereira, 2006, Martins et al., 2009, 2010, 2015, Zhang et al., 2014] or deep learning [Kiperwasser and Goldberg, 2016, Wang and Chang, 2016]. Approximate MAP inference is required for models beyond first-order. A smooth variant called softmax-margin [Gimpel and Smith, 2010] incorporates the task-specific loss ℓbut still implicitly minimizes it. Margin-based objectives are known to be consistent only under very restrictive conditions [Liu, 2007, Nowak-Vila et al., 2021] (i.e., data with majority label, loss being a distance). 2.4 Minimum Risk Empirical risk minimization suggests directly optimizing the expected target loss on training data: y pθ(y|x(i))ℓ(y(i), y), which is commonly non-convex due to normalization of pθ. There are a few parsers optimizing this objective via back-propagation [Stoyanov and Eisner, 2012], k-best lists [Smith and Eisner, 2006], semirings [Li and Eisner, 2009, Zmigrod et al., 2021] and other differentiable approximations [Gormley et al., 2015, Mensch and Blondel, 2018]. Local optima found by these algorithms do not satisfy the premise of Fisher consistency and make it difficult to quantify generalization errors. 2.5 Distributionally Robust Optimization Distributionally robust optimization has attracted emerging interests in improving machine learning models due to its connections to robustness, regularization and generalization. It proposes to minimize a risk with respect to the worst-case distribution chosen by an adversary in some uncertainty set: min θ max Q B EQ[ℓ(Y , hθ(X))], where B is an ambiguity set that can be defined by discrepancies [Shafieezadeh-Abadeh et al., 2019, Duchi and Namkoong, 2019], moments [Delage and Ye, 2010, Farnia and Tse, 2016], shapes [Popescu, 2005, Hanasusanto et al., 2015] and kernels [Shang et al., 2017, Staib and Jegelka, 2019]. A thorough review can be found in Rahimian and Mehrotra [2019]. We focus on moment-matching discriminative approaches while a similar generative method is proposed in Ganapathi et al. [2008]. We introduce the formulation, followed by practical algorithms for learning and inference. Afterwards, we present the theoretical guarantees. We defer all proofs to Appendix A. 3.1 Formulation We assume that the evaluation criterion is the Hamming loss ℓ(y, y ) := P i 1(yi = y i) with 1( ) being the 0-1 indicator function, but the results in this paper generalize to losses with affine decomposition [Ramaswamy et al., 2013] easily. Let Ptrue be the true distribution and Pemp be the empirical distribution. Our approach builds upon a probabilistic predictor that non-parametrically minimizes the expected loss with regard to the most adverse distribution in an uncertainty set where the distributions are ε away from the empirical distribution in terms of feature moment difference: min P max Q B(Pemp) EQX, ˇ Y ,P ˆ Y |Xℓ( ˆY , ˇY ), (1) where B(Pemp) := {Q : QX = Pemp X EPempϕ( ) EQϕ( ) ε} with ε 0 and ϕ : X Y Rd is a joint feature mapping decomposable over parts: ϕ(x, y) P s ϕ(x, ys). In Farnia and Tse [2016], cross-moments are adopted: ϕ(x, y) := ϕX(x) ϕY (y) where is the tensor product. By Fenchel duality [Altun and Smola, 2006] and strong duality [Von Neumann and Morgenstern, 1947], we show that Eq. (1) is analogous to dual-norm-regularized surrogate loss minimization: Proposition 1. The distributionally robust tree structured prediction problem based on moment divergence in Eq. (1) can be rewritten as min θ EPemp X,Y min P max Q EP ˆ Y |X,Q ˇ Y |Xℓ( ˆY , ˇY ) + θ (ϕ(X, ˇY ) ϕ(X, Y )) + ε θ | {z } ℓadv(θ,(X,Y )) where θ Rd is the vector of Lagrangian multipliers and is the dual norm of . 3.2 Constraint Generation Solution From a game-theoretic rationale [Topsøe, 1979, Grünwald and Dawid, 2004], Eq. (1) is considered as an adversary-constrained zero-sum game. A prediction player chooses a set of stochastic strategies (conditional distributions over arborescences) in order to minimize the expected payoff whereas an adversarial player chooses constrained strategies to maximize it. The payoff for a pair of pure strategies is the incurred loss, ℓ(ˆy, ˇy). The constrained game is transformed to a set of unconstrained ones in Eq. (2) whose payoffs are parameterized by θ: payoff(ˆy, ˇy) ℓ(ˆy, ˇy) + θ ϕ(x, ˇy). Note that the games in Eq. (1) are jointly constrained for all x s in the support of Pemp X while the ones in Eq. (2) are conditionally independent given x. The unconstrained game can be solved by a linear program [Von Neumann and Morgenstern, 1947]. However, there are O(nn) spanning trees in a complete graph, thus making explicit construction of the full payoff matrix impractical. We adopt a constraint generation algorithm named double oracle [Mc Mahan et al., 2003], shown in Appendix B. It builds a payoff sub-matrix starting from small initial sets of strategies. In each iteration, each player takes their turn based on the game payoff sub-matrix by finding the best response among all possible strategies to the opponent s optimal mixture strategies. The response is added to a player s strategy set if it improves the value of the game, with the sub-matrix updated. The algorithm terminates and converges to a Nash equilibrium of the original game when the strategy sets no longer grow. The size of the final sub-matrix is usually small in practice but there are no known theoretical guarantees, thus no way to analyze the convergence behavior. Finding the best response requires an oracle, equivalent to finding the minimum weight arborescence. The objective in Eq. (2) is a convex function of θ, so we can optimize it with sub-gradients based on solutions of the inner zero-sum games. Although lacking convergence guarantees, this algorithm is flexible with custom losses and provides a game-theoretic perspective to a typical DRO problem. 3.3 Marginal Distribution Formulation The r-arborescence polytope is defined as the convex hull of all vectors representing r-arborescences: Aarb(x) := Conv({y R|R| : y Y(x)}). Note that each p Aarb is a convex combination of all r-arborescences: p P y Prob(y)y, where ps denotes the marginal probability of part s. Here we adopt the squared ℓ2 norm as the dual norm and an ambiguity radius of ε = λ/2. By substituting the marginal probability vectors and switching min-max optimization orders, we simplify Eq. (2) into max q(i) Aarb min θ 1 m i=1 min p Aarb(q(i) p(i) emp) Φ(i)θ p, q(i) + µ 2 q(i) 2 2 + λ 2 θ 2 2, (3) where Φ(i) R|R| d denotes the feature matrix of the i-th training data, µ R 0 is a smoothing parameter to induce strong convexity. We push the maximization over q to the outermost level because of its large computational cost. If µ = 0, the solution to Eq. (3) is also optimal to Eq. (2) by strong duality but the problem becomes non-smooth. Therefore we expect θ obtained with a very small positive µ to be a good approximation of θ obtained with µ = 0. To optimize it, with fixed q, due to strong convexity, the unconstrained minimization over θ yields θ = 1 mλ Pm i=1(Φ(i)) (q(i) p(i) emp). In contrast, the constrained minimization over p admits no closed-form solution but can be cast as Euclidean projection onto Aarb instead, independently for each i [m]: p = minp Aarb p 1 µq(i) 2 2 Proj Aarb( 1 µq(i)). Given θ and p , the outermost maximization can be solved by a projected quasi-Newton algorithm [Schmidt et al., 2009] that also requires the projection oracle Proj Aarb( ), elaborated in Section 4. 3.4 Inference We propose two algorithms to make inference with given θ . Weight construction. Construct the part weights as Φθ R|R| and find the maximum weight arborescence: y arg maxy y Φθ by the Gabow-Tarjan (GT) algorithm [Gabow et al., 1986, Zmigrod et al., 2020] or approximate methods for higher-order trees. Minimum Bayes risk decoding. The optimal probabilistic prediction P or p can be obtained from Eq. (2) or Eq. (3). The marginal probabilities enable minimum Bayes risk decoding: y arg miny EP ˆ Y |xℓ(y, ˆY ) arg maxy P s:ys=1 p s, a maximum weight arborescence problem. 3.5 Statistical Properties Basic generalization bounds of DRO methods derived from measure concentration are not appropriate for an ambiguity set defined by low-order moments in this paper since it fails to converge [Shafieezadeh-Abadeh et al., 2019]. We take an alternate approach following Farnia and Tse [2016] to obtain excess out-of-sample risk bounds by assuming boundedness on features and losses. Theorem 2. Given m samples, a non-negative loss ℓ( , ) such that |ℓ( , )| K, a feature function ϕ( , ) such that ϕ( , ) B, a positive ambiguity level ε > 0, then, for any ρ (0, 1], with a probability at least 1 ρ, the following excess true worst-case risk bound holds: max Q B(Ptrue) RL Q(θ emp) max Q B(Ptrue) RL Q(θ true) 4KB where θ emp and θ true are the optimal parameters learned in Eq. (2) under Pemp and Ptrue respectively. The original risk of θ under Q is RL Q(θ) := EQX,Y ,Pθ ˆ Y |Xℓ( ˆY , Y ) with Bayes prediction Pθ Y |x arg min P max Q EQ ˇ Y |x P ˆ Y |xℓ( ˆY , ˇY ) + θ ϕ(x, ˇY ). Theorem 2 presents a bound based on uniform convergence and Rademacher complexities [Bartlett and Mendelson, 2002], which improves the results in Asif et al. [2015], who merely show that the worst-case risk upper bounds the risk under any distribution in the ambiguity set. The dual problem in Eq. (2) suggests an adversarial surrogate loss ℓadv(θ, (x, y)) in a ERM form. The special case of ε = 0 in our DRO estimator has a similar form to the max-min surrogate loss in Nowak-Vila et al. [2020] except that we assume probabilistic prediction. A conclusion of its Fisher consistency can thus be drawn based on Fathony et al. [2018a], Nowak-Vila et al. [2020]. Corollary 3. When ε = 0, ℓadv is Fisher consistent with respect to ℓ. Namely, Pθ true ˆ Y |X is the probabilistic prediction made by the Bayes optimal decision rule, where θ true is defined in Theorem 2. If ε > 0, the decoded prediction for each x will not belong to the convex hull of true conditional distributions, thus not a minimizer of ℓ. On the other hand, if ε is chosen as m α for 0 < α < 1/2, ℓadv will be universally consistent according to the comparison inequality in Nowak-Vila et al. [2020]. 4 Projection onto Arborescence Polytopes The Euclidean projection onto an r-arborescence polytope is a quadratic programming problem1: min x Aarb f(x) := x w 2 2. We focus on first-order models and discuss the extensions to other classes of trees in Section 5. 4.1 Frank-Wolfe Algorithm The Frank-Wolfe (FW) method [Frank et al., 1956] is an iterative first-order algorithm that enforces constraints by optimizing a linear objective over the feasible set at each iteration t: st arg min s Aarb s f(xt), (4) which is a minimum weight arborescence problem with weights f(xt) in our case. The solution is updated and stays feasible: xt+1 xt + γt(st xt), where γt is a step size typically set to 2 t+2. FW style algorithms are known to have a convergence rate of O( 1 ϵ ) [Jaggi, 2013]. 4.2 Martin s Polytope A compact representation of Aarb with a polynomial number of linear constraints is attractive to lead to efficient algorithms. To the best of our knowledge, there is no existing projection method exploiting special structures of this polytope. An extended formulation of the arborescence polytope [Friesen, 2019, Martin, 1991] follows a lift-and-project approach. It relates each element to existence of k-arboresences of the underlying undirected graph for all k V. We extend it to multi-graphs: Amarb := {zr : zk 0 X a δ (j) zk a = 1(j = k) k, j V X a Eij zr a k = r, i, j V zr 0}, where zr R|E| is associated with the original arcs E, zk R|E | for k = r is associated with a simple directed graph (V, E ) formed by removing directions and splitting each edge {i, j} into two directed ones, Eij := {a E : a = {i, j}} is the set of arcs connecting i and j with a (i, j, l) := {i, j} denoting the underlying undirected edge. We show exact correspondence between Amarb and Aarb based on a similar argument for simple graphs [Friesen, 2019]: Proposition 4. Let G be a multi-graph. Amarb Aarb. 1This is a well-defined convex optimization problem, different from that in differentiable structured prediction methods [Peng et al., 2018, Mihaylova et al., 2020] which elicit gradients with respect to inputs. To solve minx Amarb x w 2 2, we propose to adopt the alternating direction method of multipliers (ADMM) and rewrite it into the following separable form: min u g(u) := X 1 |V| uk w 2 2 + IUk(uk) s.t. Uk := {x R|E| : z R|E | 0 X a δ (j) za = 1(j = k) X a E ij za = X a Eij xa i, j V} ur = uk k V \ r, Ur := {x R|E| 0 : X a δ (j) xa = 1(j = r) j V}, where IU( ) is the characteristic function with IU(x) = 0 if x U and otherwise. Let λ k be the dual variables and λk := 1 ρk λ k. The scaled augmented Lagrangian function is Lρ(u, λ)=g(u) + P 2 ur uk + λk 2 2 ρk The ADMM algorithm updates the parameters as follows: ut+1 k := arg min uk Uk Lρ((ut r, ut k), λt) Proj Uk(2w + ρk|V|(ut r + λt k) 2 + ρk|V| ) k = r ut+1 r := arg min ur Ur Lρ((ut r, ut+1 k ), λt) Proj Ur( 2w + |V| P k =r ρk(ut+1 k λt k) λt+1 k := λt k + (ut+1 r ut+1 k ) k = r. This decomposes the original projection problem into simpler projection problems. Projection onto Uk for k = r decomposes over j V into |V| projections onto simplex, solvable as fast as O(n) in the worst case [Condat, 2016]. For k = r, computation of ut+1 k can be done in parallel. The Lagrange dual problem of Proj Uk( ) can be written as max α R|V| X {i,j} E hij(α) X j =k αj s.t. hij(α) = ( w2 ij/nij if αij > 2wij/nij, nijα2 ij/4 + αijwij if αij 2wij/nij, where wij := P a Eij wa, nij := |Eij|, αij := min(αi, αj) and αk := + . Strong duality holds by linear constraint qualification. Primal solutions are recovered by x a = wa min(α a/2, w a/n a). Convergence. The dual objective of Proj Uk( ) is strongly concave on {α R|V| : i j {i, j} E αi αj αi 2wij/nij}, with a unique global maximizer. This implies fast convergence in practice given good initialization. The negative Lagrange dual function has restricted strong convexity with ν = minij(nij/2), near the optimum, suggesting linear convergence [Zhang and Cheng, 2015]. Alternatively, exact solutions can be found by enumerating rankings (with duplicates) of α in O(|V||V|). In this manner, the ADMM algorithm with a strongly convex objective has a linear convergence rate O(log 1 ϵ ) with either exact [Deng and Yin, 2016] or linearly convergent approximate solution [Hager and Zhang, 2020] of Proj Uk( ). Using Nesterov s accelerated gradient algorithm [Nesterov, 2003] to optimize Eq. (3) leads to iteration complexity O(C log 1 ϵ ) with constant C dependent on Lipschitz constants of gradients and µ. 5 Extensions 5.1 Undirected Spanning Trees An straight-forward way of extending to undirected spanning trees is to split {i, j} into two arcs (i, j), (j, i) and make the feature mapping direction-invariant, i.e., ϕ(x, ys) = ϕ(x, ys ) for s and s having the same underlying undirected graph. We post-process the prediction by removing directions. Alternatively, we seek projection oracles for undirected graphs. Projection via FW is done by using any minimum spanning tree algorithm in Eq. (4). For ADMM, the formulation in Martin [1991] is originally for undirected trees: Amund := {x : z 0 P a δ (j) zk a = 1(j = k) zk ij + zk ji = x{i,j} k, i, j V}. ADMM is easily adapted to this case with P a Eij xa replaced by x{i,j}. 5.2 Dependency Trees The spanning tree structure in dependency parsing is a special one where the outdegree of root is restricted to be one. We can use the GT algorithm for inference with either the same training objective or an aligned objective where a dependency tree polytope is considered: Adep(x) := Conv({y Y(x) : |δ+(r)| = 1}). A straightforward extension of Amarb to characterizing dependency trees is Amdep := {zr : zr Amarb P a δ+(r) zr a = 1}, equivalent to Adep by the following proposition: Proposition 5. Let G be a multi-graph. Amdep Adep. FW methods leverage the GT algorithm in Eq. (4). As for ADMM, the dual problem of projection onto U r := {x : x Ur P a δ+(r) xa = 1} becomes a E ha(α, β) X j =r αj β s.t. ha(α, β) = ( w2 a γa > 2wa, waγa γ2 a/4 γa 2wa, where γ(i,j,l) := αj + 1(i = r)β. This can be solved in O(|E| log |E|) [Zhang et al., 2010]. 5.3 Higher-order Polytope Compact higher-order polytope descriptions exist for undirected spanning trees but are still unknown for arborescences with even one monomial [Friesen, 2019]. FW requires a linear oracle that is NP-hard to solve exactly in higher-order settings [Mc Donald and Pereira, 2006]. Instead, we can approximate it with a local polytope where the marginal probabilities of each part s is required to be locally consistent with that of each arc a. For simplicity, we consider only features for the all-true assignments, i.e., all arcs exist in part s. The resulting polytope can be written as Amloc := {x : x E Amarb s R, a s ps pa}, which suggests an ADMM algorithm with additional constraint sets for each part: Us := {x R|R| 0 : xs xa a s}, the projection onto which can be done in O(|s| log |s|). See Appendix D for details. 6 Experiments We evaluate our proposed method on dependency parsing tasks and compare its ability to Bi AF [Dozat and Manning, 2017], arguably the state-of-the-art neural dependency parser. We implement our methods in Python and C2. We leverage the implementations in Su Par3 [Zhang et al., 2020] for the baseline. All experiments are conducted on a computer with an Intel Core i7 CPU (2.7 GHz) and an NVIDIA Tesla P100 GPU (16 GB). We adopt three public datasets, the English Penn Treebank (PTB v3.0) [Marcus et al., 1993], the Penn Chinese Treebank (CTB v5.1) [Xue et al., 2002] and the Universal Dependencies (UD v2.3) [Nivre et al., 2016]. See Appendix C for data-processing details. Representation learning is not the focus of this paper. We follow Levy et al. [2020] and compare our method with the last biaffine classification layer in Bi AF on top of pretrained features preceding this layer (backbone s output). The pretrained embeddings produced by complicated non-linear models make Fisher consistency s assumption of optimizing over all measurable functions less violated. To featurize the data, for each dataset, we train a Bi AF network with the whole training set to obtain a pretrained model. Note that this may create unfair advantages for the baseline because the last layer was optimized together with the backbone network in an end-to-end manner during pretraining. Moreover, pretraining uses a standard ERM objective with the cross-entropy loss and local normalization over head nodes. The pretrained features are thus more adequate for the ERM objective than for our DRO objective. To make use of the features as inputs in our method, we take the outer product of the embedding vectors for two nodes as the arc feature vector. Our method and the biaffine layer therefore share the same number of parameters (501 501, including bias terms). We focus on predicting the unlabeled dependency tree while relying on pretrained models for relation label prediction. The evaluation criteria are the labeled/unlabeled attachment scores (LAS/UAS) and 2Our code is publicly available at https://github.com/Daniel Leee/drtreesp. 3https://github.com/yzhangcs/parser Table 1: Comparison of mean UAS and execution time under different training set sizes. Time refers to the CPU time taken to finish one gradient descent step. Statistically significant differences compared to Bi AF are marked with (paired t-test, p < 0.05). The best UAS are highlighted in bold. PTB CTB UD Dutch UD Turkish (low resource) Method Time (s) m = 10 50 100 1000 m = 10 50 100 1000 m = 10 50 100 1000 m = 10 50 100 1000 Bi AF 0.34 93.48 96.87 96.95 97.16 88.45 90.89 91.15 91.70 90.86 93.80 94.15 94.98 17.64 26.59 30.75 42.82 Marginal 0.28 94.51 96.81 96.92 97.12 89.19 91.03 91.27 91.67 92.41 94.22 94.50 95.15 24.85 32.83 33.75 43.18 Stochastic 2.72 94.62 96.81 96.93 97.14 89.27 91.03 91.27 91.66 92.40 94.23 94.47 95.14 25.06 31.35 33.62 41.20 Game 7.25 94.51 96.86 96.92 97.08 89.22 91.06 91.22 91.57 92.32 94.34 94.59 95.01 19.85 23.18 27.12 36.30 100 101 102 103 Iteration Figure 1: Convergence of ADMM and FW for random points with 95% confidence intervals. 10 7 10 6 10 5 10 4 10 3 10 2 10 1 100 Lambda 1e-8 1e-6 1e-4 1e-2 1e0 Figure 2: The best UAS with the Marginal algorithm as µ and λ vary in logarithmic scales. labeled/unlabeled complete matches (LCM/UCM). The attachment score can be transformed to the Hamming loss with linear mapping: AS(y, y ) |V| 1 ℓ(y, y )/2. Full batch learning is adopted for Marginal (Eq. (3)). Mini-batch training is adopted for Game, the game-theoretic algorithm, and Stochastic, which solves the inner min-max problem in Eq. (2) using Eq. (3) with fixed θ. All models are trained with the training set only. The optimal hyperparameters and parameters are chosen based on the validation set. See Appendix C for detailed parameter values. To showcase the ability of DRO methods tackling scarce data, in each run, we randomly draw m {10, 50, 100, 1000} samples without replacement from the training set and keep the original validation and test sets. All the models are trained on the same set of sampled data. The process is repeated 5 times for each m. The main UAS results on the PTB, CTB and UD Dutch Lassy Small datasets are reported in Table 1 with complete results provided in Appendix C. Our methods consistently deliver higher UAS than Bi AF especially with a small amount of data4. With little training data, DRO approaches minimize the worst-case risk to avoid overfitting. With more training data available, our method is still comparable to Bi AF which is not significantly better than our methods by statistical tests. This illustrates the advantages of replacing conditional log-likelihood with our Fisher consistent surrogate loss without changing the number of model parameters. Moreover, we study a low-resource setting with the UD Turkish dataset in which only the sampled data is used for pretraining without BERT embeddings. The binary cross-entropy loss (single normalization) is adopted during pretraining in this setting to avoid pretrained features biased towards the multi-class cross-entropy loss (local normalization) adopted by Bi AF. We observe consistently competitive performance of our methods in the low-resource setting in Table 1 as well. We report computational time of one gradient descent step in the second column of Table 1, averaged across 10 runs. For fair comparisons, all the models are run with CPU only, with a batch size of 200. All the methods achieve their optimal validation set performance in 150-300 steps. Bi AF and Marginal are the fastest because the most time-consuming step of computing dot products of features and parameters is only performed once whereas the other two methods perform it multiple times. However, since Marginal is unable to leverage stochastic gradients, its execution time grows linearly in the full batch size. Henceforth, there is a trade-off between Marginal and Stochastic/Game for computational efficiency. The extra cost compared to Bi AF with cross entropy is expected because distributional robustness against a set of adversarial distributions is guaranteed. 4The UAS is high with 10 training samples possibly because (1) the backbone sub-network and linear layer were trained together with the whole training set; (2) BERT embeddings yield data representation that is easily linearly separable; (3) 10 samples result in as many as 10 20 20 balanced head-selection instances for Bi AF. We compare ADMM and FW by performing for 100 times projection of random points in [ 5, 5]75 on a graph with 5 nodes and 3 parallel arcs between each (i, j). We subtract the integral part of the observed minimum values in each run for better illustration. As shown in Figure 1, ADMM usually finds a better solution in the arborescence polytope than FW does within 1000 iterations5. That being said, the per-iteration cost of ADMM is about 8n times higher than that of FW due to consensus optimization of n subproblems. In practice, the solution computed with FW usually leads to an approximately good sub-derivative to optimize the DRO objective. We have verified that the solutions suggested by ADMM satisfy the polytope constraints for graphs of up to 10 nodes. We conduct sensitivity analysis by varying µ and λ on UD Dutch with 100 training samples. Figure 2 implies that moderate smoothing is beneficial to generalization. The ambiguity radius should be judiciously chosen because a small λ causes overfitting while a large λ leads to conservative models. 7 Discussion and Conclusion We proposed a distributionally robust and consistent tree structured prediction method. We showed its equivalence to regularized surrogate loss minimization. We put forward a provably convergent algorithm based on efficient projection oracles for arborescence polytopes. Our proposed method enjoys Fisher consistency and robustness against noise in conditional distributions in terms of feature moments. Theoretical and empirical results validate its effectiveness. We assume that an expressive feature mapping is given such that a sufficiently good linear discriminant rule can be learned. The class-sensitive form ϕ(x, y) is general but consumes more memory than the decomposable form ϕX(x) ϕY (y). The ADMM projection algorithm is efficient theoretically with high per-iteration costs in practice. We expect this work to be a principled way of learning to predict tree-structured objects. Future directions include a more efficient implementation and general structured prediction with DRO. Potential negative societal impacts of our work include using its prediction without verification to guide human-centered design in policy-making. Representation learning. Our method can be easily adapted to a representation learning framework with automatic differentiation. Although this may lead to a non-convex problem without the theoretical guarantees derived in this paper, it is highly desired in practice if feature mappings are optimized as well. We discuss a possible approach as follows. Modern neural networks for supervised learning typically have a linear layer in the end without activation. Assume the penultimate layer outputs Φ(x) for input x, the last layer with parameters θ will typically output ψ(x) := Φ(x)θ Rk, sometimes called logits, with k = n2 labels for all arcs when parsing a sentence of n tokens. Note that θ in our formulation naturally serves as the parameters of this linear layer. Moreover, knowing ψ(x) is sufficient for us to solve the inner minimax problem in Eq. (2) to get P ˆ Y |x and Q ˇ Y |x. In this way, our DRO method can be considered a loss layer without learnable parameters, which backpropagates the sub-derivative of the objective with respect to ψ(x): i=1 (q(i) p(i) emp ), where B is the batch size. The sub-derivative of the regularization term with respect to θ should be added to the linear layer. Now we are able to take advantage of automatic differentiation and focus on solving the inner adversarial problem given ψ(x) and y. Since the computational bottleneck lies in computing ψ(x) and backward passes, the overhead of computing the adversarial loss may be dominated and not significant compared to the cross-entropy loss. We leave investigations on its effective applications to future work. Acknowledgments and Disclosure of Funding This material is based upon work supported by the National Science Foundation under Grant Nos. 1652530, 1910146, and 1934915. 5One explanation is that FW relies on first-order approximations while there are exponential number of facets in the arborescence polytope. Yasemin Altun and Alex Smola. Unifying divergence minimization and statistical inference via convex duality. In International Conference on Computational Learning Theory, pages 139 153. Springer, 2006. Kaiser Asif, Wei Xing, Sima Behpour, and Brian D Ziebart. Adversarial cost-sensitive classification. In Proceedings of the Conference on Uncertainty in Artificial Intelligence, pages 92 101, 2015. Peter L Bartlett and Shahar Mendelson. Rademacher and gaussian complexities: Risk bounds and structural results. Journal of Machine Learning Research, 3(Nov):463 482, 2002. Mathieu Blondel. Structured prediction with projection oracles. Advances in Neural Information Processing Systems, 32:12145 12156, 2019. Mathieu Blondel, André FT Martins, and Vlad Niculae. Learning with Fenchel-Young losses. J. Mach. Learn. Res., 21(35):1 69, 2020. Stephen Boyd, Neal Parikh, and Eric Chu. Distributed optimization and statistical learning via the alternating direction method of multipliers. Now Publishers Inc, 2011. Joseph K Bradley and Carlos Guestrin. Learning tree conditional random fields. In Proceedings of the 27th International Conference on International Conference on Machine Learning, pages 127 134, 2010. Stephen Brooks. Markov chain Monte Carlo method and its application. Journal of the Royal Statistical Society: Series D (the Statistician), 47(1):69 100, 1998. Danqi Chen and Christopher D Manning. A fast and accurate dependency parser using neural networks. In Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), pages 740 750, 2014. C.K. Chow and Cong Liu. Approximating discrete probability distributions with dependence trees. IEEE transactions on Information Theory, 14(3):462 467, 1968. Carlo Ciliberto, Lorenzo Rosasco, and Alessandro Rudi. A consistent regularization approach for structured prediction. Advances in neural information processing systems, 29:4412 4420, 2016. Carlo Ciliberto, Francis Bach, and Alessandro Rudi. Localized structured prediction. Advances in Neural Information Processing Systems, 32, 2019. Laurent Condat. Fast projection onto the simplex and the l_1 ball. Mathematical Programming, 158 (1-2):575, 2016. Alexis Conneau, Kartikay Khandelwal, Naman Goyal, Vishrav Chaudhary, Guillaume Wenzek, Francisco Guzmán, Edouard Grave, Myle Ott, Luke Zettlemoyer, and Veselin Stoyanov. Unsupervised cross-lingual representation learning at scale. Co RR, abs/1911.02116, 2019. URL http://arxiv.org/abs/1911.02116. Yiming Cui, Wanxiang Che, Ting Liu, Bing Qin, Shijin Wang, and Guoping Hu. Revisiting pre-trained models for Chinese natural language processing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, pages 657 668, Online, November 2020. Association for Computational Linguistics. URL https://www.aclweb.org/anthology/2020. findings-emnlp.58. Marie-Catherine De Marneffe and Christopher D Manning. The Stanford typed dependencies representation. In Coling 2008: proceedings of the workshop on cross-framework and crossdomain parser evaluation, pages 1 8, 2008. Erick Delage and Yinyu Ye. Distributionally robust optimization under moment uncertainty with application to data-driven problems. Operations research, 58(3):595 612, 2010. Wei Deng and Wotao Yin. On the global and linear convergence of the generalized alternating direction method of multipliers. Journal of Scientific Computing, 66(3):889 916, 2016. Timothy Dozat and Christopher D. Manning. Deep biaffine attention for neural dependency parsing. In 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, April 24-26, 2017, Conference Track Proceedings. Open Review.net, 2017. URL https://openreview. net/forum?id=Hk95PK9le. John Duchi and Hongseok Namkoong. Variance-based regularization with convex objectives. The Journal of Machine Learning Research, 20(1):2450 2504, 2019. Chris Dyer, Miguel Ballesteros, Wang Ling, Austin Matthews, and Noah A Smith. Transition-based dependency parsing with stack long short-term memory. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 334 343, 2015. Farzan Farnia and David Tse. A minimax approach to supervised learning. Advances in Neural Information Processing Systems, 29:4240 4248, 2016. Rizal Fathony, Sima Behpour, Xinhua Zhang, and Brian Ziebart. Efficient and consistent adversarial bipartite matching. In International Conference on Machine Learning, pages 1457 1466. PMLR, 2018a. Rizal Fathony, Ashkan Rezaei, Mohammad Ali Bashiri, Xinhua Zhang, and Brian D Ziebart. Distributionally robust graphical models. In Advances in Neural Information Processing Systems, pages 8354 8365, 2018b. Marguerite Frank, Philip Wolfe, et al. An algorithm for quadratic programming. Naval research logistics quarterly, 3(1-2):95 110, 1956. Mirjam Friesen. Extended formulations for higher order polytopes in combinatorial optimization. Ph D thesis, Otto von Guericke University Magdeburg, 2019. Harold N Gabow, Zvi Galil, Thomas Spencer, and Robert E Tarjan. Efficient algorithms for finding minimum spanning trees in undirected and directed graphs. Combinatorica, 6(2):109 122, 1986. Varun Ganapathi, David Vickrey, John Duchi, and Daphne Koller. Constrained approximate maximum entropy learning of Markov random fields. In Proceedings of the Twenty-Fourth Conference on Uncertainty in Artificial Intelligence, pages 196 203, 2008. Kevin Gimpel and Noah A Smith. Softmax-margin crfs: Training log-linear models with cost functions. In Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, pages 733 736, 2010. Matthew R Gormley, Mark Dredze, and Jason Eisner. Approximation-aware dependency parsing by belief propagation. Transactions of the Association for Computational Linguistics, 3:489 501, 2015. Peter D Grünwald and A Philip Dawid. Game theory, maximum entropy, minimum discrepancy and robust Bayesian decision theory. the Annals of Statistics, 32(4):1367 1433, 2004. William W Hager and Hongchao Zhang. Convergence rates for an inexact admm applied to separable convex optimization. Computational Optimization and Applications, 77(3):729 754, 2020. Grani A Hanasusanto, Vladimir Roitch, Daniel Kuhn, and Wolfram Wiesemann. A distributionally robust perspective on uncertainty quantification and chance constrained programming. Mathematical Programming, 151(1):35 62, 2015. Martin Jaggi. Revisiting Frank-Wolfe: Projection-free sparse convex optimization. In International Conference on Machine Learning, pages 427 435. PMLR, 2013. Eliyahu Kiperwasser and Yoav Goldberg. Simple and accurate dependency parsing using bidirectional lstm feature representations. Transactions of the Association for Computational Linguistics, 4: 313 327, 2016. Gustav Kirchhoff. Ueber die auflösung der gleichungen, auf welche man bei der untersuchung der linearen vertheilung galvanischer ströme geführt wird. Annalen der Physik, 148(12):497 508, 1847. Daphne Koller and Nir Friedman. Probabilistic graphical models: principles and techniques. MIT Press, 2009. Terry Koo, Amir Globerson, Xavier Carreras Pérez, and Michael Collins. Structured prediction models via the matrix-tree theorem. In Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-Co NLL), pages 141 150, 2007. Daniel Levy, Yair Carmon, John C Duchi, and Aaron Sidford. Large-scale methods for distributionally robust optimization. Advances in Neural Information Processing Systems, 33:8847 8860, 2020. Zhifei Li and Jason Eisner. First-and second-order expectation semirings with applications to minimum-risk training on translation forests. In Proceedings of the 2009 Conference on Empirical Methods in Natural Language Processing, pages 40 51, 2009. Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized BERT pretraining approach. Co RR, abs/1907.11692, 2019. URL http://arxiv.org/abs/1907.11692. Yufeng Liu. Fisher consistency of multicategory support vector machines. In Artificial Intelligence and Statistics, pages 291 298. PMLR, 2007. Xuezhe Ma and Eduard Hovy. Neural probabilistic model for non-projective MST parsing. In Proceedings of the Eighth International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 59 69, 2017. Mitchell Marcus, Beatrice Santorini, and Mary Ann Marcinkiewicz. Building a large annotated corpus of English: The Penn Treebank. 1993. R Kipp Martin. Using separation algorithms to generate mixed integer model reformulations. Operations Research Letters, 10(3):119 128, 1991. André Filipe Torres Martins. The geometry of constrained structured prediction: applications to inference and learning of natural language syntax. Ph D thesis, Carnegie Mellon University, 2012. André FT Martins, Noah A Smith, and Eric Xing. Concise integer linear programming formulations for dependency parsing. In Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, pages 342 350, 2009. André FT Martins, Noah A Smith, Eric Xing, Pedro Aguiar, and Mario Figueiredo. Turbo parsers: Dependency parsing by approximate variational inference. In Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing, pages 34 44, 2010. André FT Martins, Mário AT Figueiredo, Pedro MQ Aguiar, Noah A Smith, and Eric P Xing. Ad3: Alternating directions dual decomposition for map inference in graphical models. The Journal of Machine Learning Research, 16(1):495 545, 2015. Ryan Mc Donald and Fernando Pereira. Online learning of approximate dependency parsing algorithms. In 11th Conference of the European Chapter of the Association for Computational Linguistics, 2006. Ryan Mc Donald and Giorgio Satta. On the complexity of non-projective data-driven dependency parsing. In Proceedings of the Tenth International Conference on Parsing Technologies, pages 121 132, 2007. Ryan Mc Donald, Fernando Pereira, Kiril Ribarov, and Jan Hajic. Non-projective dependency parsing using spanning tree algorithms. In Proceedings of Human Language Technology Conference and Conference on Empirical Methods in Natural Language Processing, pages 523 530, 2005. H Brendan Mc Mahan, Geoffrey J Gordon, and Avrim Blum. Planning in the presence of cost functions controlled by an adversary. In Proceedings of the 20th International Conference on Machine Learning (ICML-03), pages 536 543, 2003. Arthur Mensch and Mathieu Blondel. Differentiable dynamic programming for structured prediction and attention. In International Conference on Machine Learning, pages 3462 3471. PMLR, 2018. Ofer Meshi, Elad Eban, Gal Elidan, and Amir Globerson. Learning max-margin tree predictors. In Proceedings of the Twenty-Ninth Conference on Uncertainty in Artificial Intelligence, pages 411 420, 2013. Tsvetomila Mihaylova, Vlad Niculae, and André FT Martins. Understanding the mechanics of spigot: Surrogate gradients for latent structure learning. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 2186 2202, 2020. Kevin P Murphy, Yair Weiss, and Michael I Jordan. Loopy belief propagation for approximate inference: an empirical study. In Proceedings of the Fifteenth Conference on Uncertainty in Artificial Intelligence, pages 467 475, 1999. Yurii Nesterov. Introductory lectures on convex optimization: A basic course, volume 87. Springer Science & Business Media, 2003. Joakim Nivre, Marie-Catherine De Marneffe, Filip Ginter, Yoav Goldberg, Jan Hajic, Christopher D Manning, Ryan Mc Donald, Slav Petrov, Sampo Pyysalo, Natalia Silveira, et al. Universal dependencies v1: A multilingual treebank collection. In Proceedings of the Tenth International Conference on Language Resources and Evaluation (LREC 16), pages 1659 1666, 2016. Alex Nowak-Vila, Francis Bach, and Alessandro Rudi. A general theory for structured prediction with smooth convex surrogates. ar Xiv preprint ar Xiv:1902.01958, 2019. Alex Nowak-Vila, Francis Bach, and Alessandro Rudi. Consistent structured prediction with max-min margin Markov networks. In Proceedings of the International Conference on Machine Learning (ICML), 2020. Alex Nowak-Vila, Alessandro Rudi, and Francis Bach. Max-margin is dead, long live max-margin! ar Xiv preprint ar Xiv:2105.15069, 2021. Franz Josef Och. Minimum error rate training in statistical machine translation. In Proceedings of the 41st annual meeting of the Association for Computational Linguistics, pages 160 167, 2003. Hao Peng, Sam Thomson, and Noah A Smith. Backpropagating through structured argmax using a spigot. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1863 1873, 2018. Ioana Popescu. A semidefinite programming approach to optimal-moment bounds for convex classes of distributions. Mathematics of Operations Research, 30(3):632 657, 2005. Hamed Rahimian and Sanjay Mehrotra. Distributionally robust optimization: A review. ar Xiv preprint ar Xiv:1908.05659, 2019. Harish G Ramaswamy, Shivani Agarwal, and Ambuj Tewari. Convex calibrated surrogates for lowrank loss matrices with applications to subset ranking losses. In Advances in Neural Information Processing Systems, pages 1475 1483, 2013. Mark Schmidt, Ewout Berg, Michael Friedlander, and Kevin Murphy. Optimizing costly functions with simple constraints: A limited-memory projected quasi-newton algorithm. In Artificial Intelligence and Statistics, pages 456 463. PMLR, 2009. Soroosh Shafieezadeh-Abadeh, Daniel Kuhn, and Peyman Mohajerin Esfahani. Regularization via mass transportation. Journal of Machine Learning Research, 20(103):1 68, 2019. Shai Shalev-Shwartz and Shai Ben-David. Understanding machine learning: From theory to algorithms. Cambridge University Press, 2014. Chao Shang, Xiaolin Huang, and Fengqi You. Data-driven robust optimization based on kernel learning. Computers & Chemical Engineering, 106:464 479, 2017. David A Smith and Jason Eisner. Minimum risk annealing for training log-linear models. In Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions, pages 787 794, 2006. David A Smith and Noah A Smith. Probabilistic models of nonprojective dependency trees. In Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-Co NLL), pages 132 140, 2007. Matthew Staib and Stefanie Jegelka. Distributionally robust optimization and generalization in kernel methods. Advances in Neural Information Processing Systems, 32:9134 9144, 2019. Miloš Stanojevi c and Shay B Cohen. A root of a problem: Optimizing single-root dependency parsing. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 10540 10557, 2021. Veselin Stoyanov and Jason Eisner. Minimum-risk training of approximate CRF-based NLP systems. In Proceedings of the 2012 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 120 130, 2012. Ben Taskar, Carlos Guestrin, and Daphne Koller. Max-margin Markov networks. Advances in Neural Information Processing Systems, 16, 2003. Ben Taskar, Dan Klein, Mike Collins, Daphne Koller, and Christopher D Manning. Max-margin parsing. In Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, pages 1 8, 2004. Flemming Topsøe. Information-theoretical optimization techniques. Kybernetika, 15(1):8 27, 1979. Kristina Toutanova, Dan Klein, Christopher D Manning, and Yoram Singer. Feature-rich part-ofspeech tagging with a cyclic dependency network. In Proceedings of the 2003 Human Language Technology Conference of the North American Chapter of the Association for Computational Linguistics, pages 252 259, 2003. Ioannis Tsochantaridis, Thorsten Joachims, Thomas Hofmann, Yasemin Altun, and Yoram Singer. Large margin methods for structured and interdependent output variables. Journal of Machine Learning Research, 6(9), 2005. John Von Neumann and Oskar Morgenstern. Theory of games and economic behavior, 2nd rev. 1947. Wenhui Wang and Baobao Chang. Graph-based dependency parsing with bidirectional LSTM. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 2306 2315, 2016. T William. Tutte. graph theory. Encyclopedia of Mathematics and its Applications, 21, 1984. Richard T Wong. Integer programming formulations of the traveling salesman problem. In Proceedings of the IEEE international Conference of Circuits and Computers, pages 149 152. IEEE Press Piscataway NJ, 1980. Zheng Xu, Gavin Taylor, Hao Li, Mário AT Figueiredo, Xiaoming Yuan, and Tom Goldstein. Adaptive consensus ADMM for distributed optimization. In International Conference on Machine Learning, pages 3841 3850. PMLR, 2017. Nianwen Xue, Fu-Dong Chiou, and Martha Palmer. Building a large-scale annotated Chinese corpus. In COLING 2002: The 19th International Conference on Computational Linguistics, 2002. Hui Zhang and Lizhi Cheng. Restricted strong convexity and its applications to convergence analysis of gradient-type methods in convex optimization. Optimization Letters, 9(5):961 979, 2015. Xingxing Zhang, Jianpeng Cheng, and Mirella Lapata. Dependency parsing as head selection. In Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume 1, Long Papers, pages 665 676, 2017. Xinhua Zhang, Ankan Saha, and SVN Vishwanathan. Regularized risk minimization by Nesterov s accelerated gradient methods: Algorithmic extensions and empirical studies. ar Xiv preprint ar Xiv:1011.0472, 2010. Yu Zhang, Zhenghua Li, and Min Zhang. Efficient second-order Tree CRF for neural dependency parsing. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 3295 3305, 2020. Yuan Zhang, Tao Lei, Regina Barzilay, and Tommi Jaakkola. Greed is good if randomized: New inference for dependency parsing. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1013 1024, 2014. Ran Zmigrod, Tim Vieira, and Ryan Cotterell. Please mind the root: Decoding arborescences for dependency parsing. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 4809 4819, 2020. Ran Zmigrod, Tim Vieira, and Ryan Cotterell. Efficient computation of expectations under spanning tree distributions. Transactions of the Association for Computational Linguistics, 9:675 690, 2021. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] (c) Did you discuss any potential negative societal impacts of your work? [Yes] (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes] (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [Yes] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [Yes] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [Yes] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [Yes] (b) Did you mention the license of the assets? [No] (c) Did you include any new assets either in the supplemental material or as a URL? [No] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]