# optimal_dynamic_regret_in_lqr_control__3f2c30a1.pdf Optimal Dynamic Regret in LQR Control Dheeraj Baby Department of Computer Science UC Santa Barbara dheeraj@ucsb.edu Yu-Xiang Wang Department of Computer Science UC Santa Barbara yuxiangw@cs.ucsb.edu We consider the problem of nonstochastic control with a sequence of quadratic losses, i.e., LQR control. We provide an efficient online algorithm that achieves an optimal dynamic (policy) regret of O(max{n1/3T V(M1:n)2/3, 1}), where T V(M1:n) is the total variation of any oracle sequence of Disturbance Action policies parameterized by M1, ..., Mn chosen in hindsight to cater to unknown nonstationarity. The rate improves the best known rate of O( p n(T V(M1:n) + 1)) for general convex losses [Zhao et al., 2022] and we prove that it is informationtheoretically optimal for LQR. Main technical components include the reduction of LQR to online linear regression with delayed feedback due to Foster and Simchowitz [2020], as well as a new proper learning algorithm with an optimal O(n1/3) dynamic regret on a family of minibatched quadratic losses, which could be of independent interest. 1 Introduction This paper studies the linear quadratic regulator (LQR) control problem which is a specific instantiation of the more general RL framework where the evolution of states follows a predefined linear dynamics. At each round t [n] := {1, . . . , n}, the agent is at state xt Rdx. Based on the state, the agent select a control input ut Rdu. The next state evolves according to the law: xt+1 = Axt + But + wt, where A and B are system matrices known to the agent. wt Rdx is a disturbance term that can be selected by a potentially adaptive adversary. We assume that wt 2 1. This disturbance term reflects the perturbation from the ideal linear state transition arising due to environmental factors that could be difficult to model. The loss suffered by playing the control u at state x is given by ℓ(x, u) := x T Rxx + u T Ruu, where Rx, Ru 0, that are apriori fixed and known. Recently there has been a surge of interest in viewing this classical LQR problem under the lens of online learning [Hazan, 2016]. The work of Agarwal et al. [2019] places regret of the agent against a set of benchmark policies as the central notion to evaluate learner s performance. Following Agarwal et al. [2019], Foster and Simchowitz [2020] we adopt the class of disturbance action policies (DAP) as our benchmark class: Definition 1. (Disturbance action policies, [Foster and Simchowitz, 2020]). Let M = (M [i])m i=1 denote a sequence of matrices M [i] Rdu dx. We define the corresponding disturbance action policies (DAP) πM as: πM t (xt) = K xt q M(w1:t 1), (1) where q M(w1:t 1) = Pm i=1 M [i]wt i and K as in Eq.(4). We are interested in DAPs for which the sequence M belongs to the set: M(m, R, γ) := {M = (M [i])m i=1 : M [i] op Rγi 1}, (2) 36th Conference on Neural Information Processing Systems (Neur IPS 2022). where m, R and γ are algorithm parameters. This class is known to be sufficiently rich to approximate many linear controllers. A policy takes in the past history and current state as input and produces a control signal as output. Let s denote M1:n := (M1, . . . , Mn) to be a sequence of DAP policies such that at time t, the control signal is selected using the policy parameterized by Mt (see Eq.(1)). We denote x M1:n t to be the state reached at round t by playing the sequence of policies defined by parameters M1:t 1 in the past. Similarly u M1:n t is used to denote the control signal produced by the policy Mt. The universal dynamic regret of the learner against the policy sequence M1:n is defined as: t=1 ℓ(xalg t , ualg t ) ℓ(x M1:n t , u M1:n t ), (3) where (xalg t , ualg t ) denotes the state and control signal of the learner at round t. Note that the policy sequence M1:n can be any valid sequence of DAP polices. The main focus of this paper is to design algorithms that can control the dynamic regret against a sequence of reference policies as a function of the time horizon n and the a path variation of the DAP parameters of the comparator M1:n. We remark that the comparator polices M1:n can be chosen in hindsight and potentially unknown to the learner. Whenever M1:n = (M, . . . , M) for a fixed parameter M, we recover the notion of static regret. However the notion of static regret is not befitting for non-stationary environments. For best performance under a non-stationary environment, a controller may have to choose different policies at different time steps that can counter-act the disturbances and guide the dynamics properly. Hence, we aim to control the dynamic regret which allows us to be competent against a sequence of potentially time-varying polices chosen in hindsight. We remark that our algorithm automatically adapts to the level of non-stationarity in the hindsight sequence of policies. Next, we take a digression and discuss a desirable property for the design of algorithms for LQR control. Proper learning in LQR control. Proper learning is an online learning paradigm where the decisions of the learner are required to obey some user specified safety constraints. On the other hand, improper learning framework allows the learner to disregard such safety constraints. The paradigm of improper learning may not be attractive in certain applications where safety is a paramount concern. Improper algorithms can possibly take the system through trajectories that are deemed to be risky. It is desirable to avoid such behaviours in physical systems such as self driving cars, control of medical ventilators, robotic control [Levine et al., 2016] and cooling data centers [Cohen et al., 2018]. When translated into the LQR control problem, we regard the benchmark pool M(m, R, γ) defined in Eq.(2) as a space of safe policies. So to avoid risky behaviours, at any round, the learner plays a control signal that is recommended by a DAP policy in the class M(m, R, γ). Below are our contributions: We develop an optimal universal dynamic regret minimization algorithm for the general mini-batch linear regression problem (see Theorem 5). Applying the reduction of Foster and Simchowitz [2020] from LQR problem to online linear regression, the above result lends itself to an algorithm for controlling the dynamic regret of the LQR problem (Eq.(3)) to be O (n1/3[T V(M1:n)]2/3), where T V denotes the total variation incurred by the sequence of DAP policy parameters in hindsight (see Corollary 10). O hides the dependencies in dimensions and system parameters. We show that the aforementioned dynamic regret guarantee is minimax optimal modulo dimensions and factors of log n (see Theorem 11). The resulting algorithm is also strongly adaptive, in the sense that the static regret against a DAP policy in any local time window is O (log n). Notes on novelty and impact. As discussed before, the reduction of Foster and Simchowitz [2020] casts LQR problem to an instance of proper online linear regression. In the context of regression, proper learning means that the decisions of the learner belongs to a user specified convex domain. The main challenge in developing aforementioned contributions rests on the design of an optimal universal dynamic regret minimization algorithm for online linear regression under the setting of proper learning. We are not aware of any such algorithms in the literature to-date and the problem remains open. However, there exists an improper algorithm from Baby and Wang [2021] for controlling the desired dynamic regret. Given this fact, the design of our algorithm is facilitated by coming up with new black-box reductions (see Section 4) that can convert an improper algorithm for non-stationary online linear regression to a proper one. There are improper to proper black-box reduction schemes given in the influential work of Cutkosky and Orabona [2018]. However, they are developed to support general convex or strongly convex (see Definition 4) losses. The linear regression losses arising in our setting are exp-concave (see Definition 3) which enjoy strong curvature only in the direction of the gradients as opposed to uniformly curved strongly convex losses. Hence the reduction scheme of Cutkosky and Orabona [2018] is inadequate to provide fast regret rates in our setting. In contrast, we develop novel reduction schemes that carefully take the non-uniform curvature of the linear regression losses into account so as to facilitate fast dynamic regret rates (see Section 4.2). The construction of this new reduction scheme requires non-trivial adaptation of the ideas in Cutkosky and Orabona [2018]. We remark that the algorithm Pro DR.control developed in Section 4 can be impactful in general online learning literature. That the non-stationary LQR problem can be optimally solved using Pro DR.control is a testament to this fact. Further our algorithm is out-of-the-box applicable to more general settings such as non-stationary multi-task linear regression, which is beyond the current scope. The lower bound we provide in Theorem 11 is also applicable to the more general problem of online non-parametric regression against a Besov space / class of Total Variation bounded functions [Rakhlin and Sridharan, 2014] (see Section 5 for more details). The main contribution here is that we provide a new lower bounding strategy that characterizes the correct rate wrt both n and the radius (or path-variation) of the non-parametric function class. This is in contrast with Rakhlin and Sridharan [2014] who establish the correct dependency only wrt n. Attaining the correct dependencies wrt both n and the radius / path-variation is imperative in implying a dynamic regret lower bound for the LQR problem. The rest of the paper is organized as follows. In Section 2, we cover the necessary preliminaries on LQR control. Section 3 discusses relevant literature. In Section 4, we develop a proper algorithm for non-stationary online linear regression. In Section 5, we apply the results of Section 4 to provide an algorithm for non-sationary LQR control and prove its minimax optimality. This is followed by conclusion and open problems in Section 6. We provide a concise overview of the results from Baby and Wang [2022] in Appendix A which we build upon. All proofs are given in Appendix B. 2 Preliminaries We start with a brief overview of the LQR problem for the sake of completeness. The material of this section closely follows Foster and Simchowitz [2020]. The definitions and notations introduced in this section will be used throughout the paper. A linear control law is given by ut = Kxt for a controller K Rdu dx. A linear controller K is said to be stabilizing if ρ(A BK) < 1 where ρ(A BK) is the maximum of the absolute values of the eigenvalues of A BK. We assume that there exists a stabilizing controller for the system (A, B). For such systems, there exists a unique matrix P which is the solution to the equation: P = AT PA + Rx AT PB(Ru + BT PB) 1BT PA. The solution P is called the infinite horizon Lyapunov matrix. It is an intrinsic property of the system (A, B) and characterizes the optimal infinite horizon cost for control in the absence of noise [Bertsekas, 2005]. We also define the optimal state feedback controller K := (Ru + BT P B) 1BT P A, (4) the steady state covariance matrix: Σ := Ru + BT P B, and the closed loop dynamics matrix: Acl, := A BK . Foster and Simchowitz [2020] shows that the problem of controlling the regret in the LQR problem can be reduced to online linear regression problem with delays. Specifically we have the following fundamental result due to Foster and Simchowitz [2020]. Proposition 2. Suppose the learner plays policy of the form πalg t (x) = K x + q M alg t (w1:t 1). Let the comparator policies take the form πt(x) = K x + q Mt(w1:t 1) for a sequence of matrices M1:n chosen in hindsight. Then the dynamic regret against the policies π := (π1, . . . , πn) satisfies: Rn(π) O(1) + t=1 ˆAt(M alg t , wt:t+h) ˆAt(Mt, wt:t+h), where the parameters involved in the inequality are defined as below: ˆAt(M, wt:t+h) := q M(w1:t 1) q ;h(wt:t+h) 2 Σ . q ;h(wt:h+t) := Pt+h i=t+1 Σ 1 BT (Acl, )i 1 t P wi. h := 2(1 γ ) 1 log(κ2 β2 Ψ Γ2 n2). γ := I P+ 1/2Rx P 1/2 1/2 op . κ := P 1/2 op P 1/2 op. β := max{1, λ 1 min(Ru), λ 1 min(rx)}. Ψ = max{1, A op, B op, Rx op, Ru op}. Γ := max{1, P op} Observe that the losses ˆAt(M, wt:t+h) := q M(w1:t 1) q ;h(wt:t+h) 2 Σ = ˆAt(M, wt:t+h) := Σ1/2 q M(w1:t 1) Σ1/2 q ;h(wt:t+h) 2 2 are essentially linear regression losses. The quantity Σ1/2 q M(w1:t 1) is a linear map from the matrix sequence M to Rdu. However, there is one caveat in that the bias vector at round t given by Σ1/2 q ;h(wt:t+h) is only available at round t + h = t + O(log n). This issue of delayed feedback can be directly handled using the delayed to non-delayed online learning reduction from Joulani et al. [2013]. 3 Related work In this section, we review recent progress at the intersection of control and online convex optimization (OCO) that are most relevant to our work. Online control. The idea of using tools from OCO for general control problem was proposed in Agarwal et al. [2019]. They place the notion of regret against the class of DAP policies as the central performance measure. The DAP class is also shown to be sufficiently rich to approximate a wide class of linear state-feedback controllers. Under general convex losses, they propose a reduction to OCO with memory [Merhav et al., 2000, Anava et al., 2015] and derives O( n) regret when the system matrices (A, B) are known. For the case of unknown system, Hazan et al. [2020] provides O(n2/3) regret via system identification techniques. When the losses are strongly convex and sub-quadratic, Simchowitz [2020] strengthens these results to attain O(n) regret for known systems and O( n) when the system is unknown. For partially observable systems strong regret guarantees are provided in Simchowitz et al. [2020]. Luo et al. [2022] provides an O(n3/5) dynamic regret bound for the case when the system matrices (At, Bt) can change over time. Their results are incompatible to ours in that they consider unknown dynamics, stochastic disturbances and the dynamic regret compete with controllers that are pointwise optimal (restricted dynamic regret), while we assume known dynamics, adversarial disturbances and compete with an arbitrary sequence of controllers (i.e., universal dynamic regret). Next, we clarify the differences from other existing work on the nonstochastic control problems [Gradu et al., 2020a,b, Cassel and Koren, 2020, Zhang et al., 2021b, Shi et al., 2020, Goel and Hassibi, 2021, Zhao et al., 2022]. Gradu et al. [2020a], Cassel and Koren [2020] studied online control in the partially observed cases with bandit feedback, but did not consider the problem of non-stationarity with dynamic regret. Gradu et al. [2020b], Zhang et al. [2021b] studied the adaptive regret in nonstochastic control problems, which is an alternative metric to capture the performance of the learning controller in non-stationary environments. Our algorithm uses a reduction to adaptive regret too, but it is highly nontrivial to show that one can tweak adaptive regret minimizing algorithms into ones that achieve optimal dynamic regret. Moreover, our algorithm is the first that achieves logarithmic adaptive regret for nonstochastic LQR control problems too. In contrast, Gradu et al. [2020b], Zhang et al. [2021b] focused on the slower adaptive regret in the general convex loss cases. To the best of our knowledge, Goel and Hassibi [2021] and Zhao et al. [2022] are the only existing work that considered dynamic regret in non-stochastic control. Goel and Hassibi [2021] used tools from H control and derived a controller with exact minimax optimal dynamic regret against an oracle controller that sees the whole sequence of disturbances and chooses an optimal sequence of control actions. But the optimal dynamic regret against the sequence of control actions given by the unrealizable oracle controller is linear in n in general (see an explicit lower bound from Goel and Hassibi [2020]). It is unclear whether this oracle controller can be realized by a sequence of time-varying DAP controllers. If so, then our results would imply regret bound against the optimal sequence of control actions too. Comparing to the exact minimax regret of the H style controller, our regret bound would adapt to each problem instance, and is sublinear whenever the approximating sequence of DAP controllers has sublinear total variation. Zhao et al. [2022] studied the universal dynamic (policy) regret problem similar to ours, but works for a broader family convex loss functions. Their regret bound O( p n(1 + Cn)) is optimal for the convex loss family. Our results show that the optimal regret improves to O(n1/3C2/3 n 1) when specializing to the LQR problem where the losses are quadratic. On the technical level, Zhao et al. [2022] used a reduction to the dynamic regret of OCO with memory, while we reduced to the dynamic regret of OCO with delayed feedback. Dynamic regret minimization in online learning. There is a rich body of literature on dynamic regret (Eq.(5)) minimization. As discussed in Section 2, the non-staionary LQR problem can be reduced to an instance of linear regression losses which are exp-concave on compact domains. There is a recent line of research [Baby and Wang, 2021, 2022] that provides optimal universal dynamic regret rates under exp-concave losses. However, the algorithm of Baby and Wang [2021] is improper, in the sense that the iterates of the learner can lie outside the feasibility set. The work of Baby and Wang [2022] ameliorates this issue to some extend by providing proper algorithms for the particular case of L constrained (box) decisions sets. The DAP policy space in Definition 1 is indeed not an L ball. We note that if improper learning is allowed in the LQR problem, one can run the algorithms of Baby and Wang [2021, 2022] to attain optimal dynamic regret rates. The proper learning algorithms such as Zinkevich [2003], Zhang et al. [2018a], Cutkosky [2020], Jacobsen and Cutkosky [2022] control dynamic regret for general convex losses. However, they are not adequate to optimally minimize dynamic regret under curved losses that are strongly convex or exp-concave. The notion of restrictive dynamic regret introduced in Besbes et al. [2015] competes with a sequence of minimizers of the losses. This notion of regret can sometimes be overly pessimistic as noted in Zhang et al. [2018a]. There is a series of work in the direction of dynamic regret minimization in OCO such as Jadbabaie et al. [2015], Yang et al. [2016], Mokhtari et al. [2016], Chen et al. [2018], Zhang et al. [2018b], Goel and Wierman [2019], Baby and Wang [2019], Zhao et al. [2020], Zhao and Zhang [2021], Zhao et al. [2022], Chang and Shahrampour [2021], Baby and Wang [2020], Baby et al. [2021b], Chatterjee and Goswami [2022], Baby et al. [2021a], Raj et al. [2020]. However, to the best of our knowledge none of these works are known to attain the optimal universal dynamic regret rate for the setting of online linear regression. Dynamic regret for OCO vs Dynamic (Policy) regret for Control. We emphasize that the regret in Eq.(3) is dynamic policy regret [Anava et al., 2015, Zhao et al., 2022]. The states visited by the reference policy is counterfactual and is different from that of the learner s trajectory which we observe. This is very different from the standard OCO framework where the state of both the learner and adversary are same. So bounding the policy regret seems qualitatively harder than bounding the regret in an OCO setting. Nevertheless, for the LQR problem, the fact that there exists a reduction [Foster and Simchowitz, 2020] from the problem of controlling policy regret to the problem of controlling the standard OCO regret is remarkable. Strongly adaptive regret minimization. There is also a complementary body of literature on strongly adaptive algorithms that focus on controlling the static regret in any local time window. For example, the algorithm of Daniely et al. [2015], Jun et al. [2017] can lead to O( p |I|) static regret in any interval of I [n] under convex losses. When the losses are exp-concave the algorithm of Hazan and Seshadhri [2007], Adamskiy et al. [2016], Zhang et al. [2021a] can lead to O(log n) static regret in any interval. 4 Non-stationary mini-batch Linear Regression In view of Proposition 2, the losses of interest are linear regression type losses. So we take a digression in this section and study the problem of controlling dynamic regret in a general linear regression setting. 4.1 Linear regression framework Consider the following linear regression protocol. At round t, nature reveals a co-variate matrix At Rp d. Learner plays zt D Rd. Nature reveals the loss ft(z) = Atz bt 2 2. Under the above regression framework, we are interested in controlling the universal dynamic regret against an arbitrary sequence of predictors u1, . . . , un D (abbreviated as u1:n) : t=1 ft(zt) ft(ut). (5) Dynamic regret is usually expressed as a function of n and a path variational that captures the smoothness of the comparator sequence. We will focus on the path variational defined by: T V(u1:n) = t=2 ut ut 1 1. Below are the list of assumptions made: Assumption 1. Let at,i Rd be the ith row vector of At. We assume that at,i 1 α for all t [n] and i [p]. Further bt 1 σ for all t. Assumption 2. For any x D, x 1 χ and x R. We refer this setting as mini-batch linear regression since the loss at round t can be written as a sum of a batch of quadratic losses: ft(z) = Pp i=1 z T at,i bt[i] 2. Terminology. For a convex loss function f, we abuse the notation and take f(x) to be a sub-gradient of f at x. We denote D ( R) := {x Rd : x R}. Linear regression losses belong to a broad family of convex loss functions called exp-concave losses: Definition 3. A convex function f is α exp-concave in a domain D if for all x, y D we have f(y) f(x) + f(x)T (x y) + α 2 ( f(x)T (x y))2. We refer the reaeder to Cesa-Bianchi and Lugosi [2006] and Hazan et al. [2007] for more detailed expositions on exp-concavity. The losses ft(z) = Atz bt 2 2 are (2R) 1 exp-concave if f(z) R for all z D (see Lemma 2.3 in Foster and Simchowitz [2020]). Definition 4. A convex function f is σ strongly convex wrt 2 norm in a domain D if for all x, y D we have f(y) f(x) + f(x)T (x y) + σ We note that if the matrix At is rank deficient, then the losses ft(z) cannot be strongly convex. Moving forward we do not impose any restrictive assumptions on the rank of At. As mentioned in Remark 12, the covariate matrix that arise in the reduction of the LQR problem to linear regression is not in general full rank. So we target a solution that can handle general covariate matrices irrespective of their rank. 4.2 The Algorithm Starting point of our algorithm design is the work of Baby and Wang [2022]. They provide an algorithm that attains optimal dynamic regret when the losses are exp-concave. However, their setting works only in a very restrictive setup where the decision set is an L constrained box. Consequently, we cannot directly apply their results to the linear regression problem of Section 4 whenever the decision set D is a general convex set. An online learner is termed proper if the decisions of the learner are guaranteed to lie within the feasibility set D. Otherwise it is called improper. A recent seminal work of Cutkosky and Orabona [2018] proposes neat reductions that can convert an improper online learner to a proper one, whenever Pro DR.control: Inputs - Decision set D, G > 0, a surrogate algorithm A which ensures low dynamic regret under general exp-concave losses against any comparator sequence in some D D. Here D is a compact and convex set. Note that such an algorithm A may produce iterates outside D. (See Theorem 5 for a specific choice of A.) 1. At round t, receive wt from A. 2. Receive co-variate matrix At := [at,1, . . . , at,p]T . 3. Play ˆwt argminx D maxi=1,...,p|a T t,i(x wt)|. 4. Let ℓt(w) = ft(w) + G St(w), where ft(w) = Atw bt 2 2 and St(w) = minx D maxi=1,...,p|a T t,i(x w)|. 5. Send ℓt(w) to A. Figure 1: Pro DR.control: An algorithm for non-stationary and proper linear regression. the losses are convex. Following this line of research, we can aim to convert the algorithm of Baby and Wang [2022] that works exclusively on box decision set to one that can support arbitrary convex decision sets by coming up with suitable reduction schemes. However, the specific reduction scheme proposed in Cutkosky and Orabona [2018] is inadequate to yield fast dynamic rates for exp-concave losses. Our algorithm Pro DR.control (Fig.1, Proper Dynamic Regret.control) is a by-product of constructing new reduction schemes to circumvent the aforementioned problem for the case of linear regression losses. We expand upon these details below. In Pro DR.control, we maintain a surrogate algorithm A, which is chosen to be the algorithm of Baby and Wang [2022] that produces iterates wt in an L norm ball (box), D , that encloses the actual decision set D. We take D = D as per the notations in Fig.1. Since wt can be infeasible, we play ˆwt obtained via a special type of projection of wt onto D which is formulated as a min-max problem in Line 3 of Fig.1. In Line 4, we construct surrogate losses ℓt to be passed to the algorithm A. The surrogate loss penalises A for making predictions outside D. We will show (see Lemma 15 in Appendix) that the instantaneous regret satisfies ft( ˆwt) ft(ut) ℓt(wt) ℓt(ut), where ut D is the comparator at round t. Thus the dynamic regret of the proper iterates ˆwt wrt linear regression losses is upper bounded by the dynamic regret of the surrogate algorithm A on the losses ℓt and box decision set. The design of the min-max barrier St(w) is driven to ensure exp-concavity of the surrogate losses ℓt(w) = ft(w) + G St(w). We capture its intuition as follows. We start by observing that since 2ft(w) = 2AT t At, the linear regression losses ft exhibits strong curvature along the row-space of At, denoted by row(At). Further we have ft(w) = 2AT t (Atw bt) row(At). So the loss ft exhibits strong curvature along the direction of its gradient too. This is the fundamental reason behind the exp-concavity of ft. The min-max barrier St(w) is designed such that its gradient is guaranteed to lie in the row(At) (see Lemma 16 in Appendix for a formal statement). So the overall gradient ℓt(w) also lies in the row(At). Since the function ft already exhibits strong curvature along row(At), we conclude that the sum ℓt(w) = ft(w) + G St(w) exhibits strong curvature along its gradient ℓt(w). This maintains the exp-concavity of the losses ℓt over D (see Lemma 17 in Appendix). Such curvature considerations along with the fact that St(w) has to be sufficiently large to facilitate the instantaneous regret bound ft( ˆwt) ft(ut) ℓt(wt) ℓt(ut) results in functional form for St(w) displayed in Fig.1. Consequently the fast dynamic regret rates derived in Baby and Wang [2022] becomes directly applicable. The reduction scheme used by Cutkosky and Orabona [2018] for producing proper iterates ˆwt and their accompanying surrogate loss design ℓt also allows one to upper bound the regret wrt linear regression losses ft by the regret of the algorithm A wrt surrogate losses ℓt. However, the surrogate loss ℓt they construct is not guaranteed to be exp-concave and consequently not amenable to fast dynamic regret rates. 4.3 Main Results We have the following guarantee for Pro DR.control: Theorem 5. Let u1:n D be any comparator sequence. In Fig.1, choose G such that supw1,w2 D ( R),t [n] At(w1 + w2) 2bt 1 G. Let α be as in Assumption 2. Let L be such that supw D ( R),j [p] 2 Atw bt 2 2+2G2 L for all t [n]. Choose A as the algorithm from Baby and Wang [2022] (see Appendix A) with parameters γ = 2Gα R p 2L and ζ = min{ 1 d, 1/(4γ2)} and decision set D ( R). Under Assumptions 1 and 2, a valid of assignment of G and L are 2pχ + 2σ and 6(pχ + σ)2 respectively. Then the algorithm Pro DR.control yields a dynamic regret rate of t=1 ft( ˆwt) ft(ut) = O(d3n1/3[T V(u1:n)]2/3 1), where (a b) := max{a, b}. Remark 6. In view of Proposition 10 in Baby and Wang [2021], the dynamic regret guarantee in Theorem 5 is optimal modulo dependencies in d and log n. Further the algorithm does not require apriori knowledge of the path length T V(u1:n). Proof sketch for Theorem 5. First step is to show that ft( ˆwt) ℓt(wt). This is accomplished by Lipschitzness type arguments. For any u D, one observes that ℓt(u) = ft(u). So the instantaneous regret of Pro DR.control, ft( ˆwt) ft(ut), is upper bounded by the instantaneous regret, ℓt(wt) ℓt(ut) of the surrogate algorithm A. The crucial step is to show the exp-concavity of the losses ℓt across D ( R). For this, we prove that there is a sub-gradient St(w) that is aligned with at,j for some j [p]. This observation followed by few algebraic manipulations (see proof of Lemma 17 in Appendix) allows us to show the exp-concavity of ℓt over D ( R). Now the overall regret can be controlled if the surrogate algorithm A provides optimal dynamic regret under exp-concave losses and box decision sets, D ( R). This is accomplished by choosing A as the algorithm in Baby and Wang [2022] which is also strongly adaptive. Since the surrogate algorithm A we used in Theorem 5 is strongly adaptive (see for eg. Appendix A), we also have the following performance guarantee in terms of adaptive regret: Proposition 7. Consider the instantiation of Pro DR.control in Theorem 5. Then for any time window [a, b] [n] we have that: Pb t=a ft( ˆwt) infu D Pb t=a ft(u) = O(d1.5 log n). Remark 8. Theorem 5 and Proposition 7 together makes the algorithm Pro DR.control a good candidate for performing proper online linear regression in non-stationary environments. 4.4 Linear regression with delayed feedback In this section, we consider a linear regression protocol with feedback delayed by τ time steps. At round t, nature reveals a co-variate matrix At Rp d. Learner plays zt D Rd. Nature reveals the loss ft τ+1(z) = At τ+1z bt τ+1 2 2. This delayed setting can be handled by the framework developed in Joulani et al. [2013]. Although these authors focus on bounding the regret as a function of time horizon n, the extension to dynamic regret bounds expressed in terms of both n and T V(u1:n) can be handled straight-forwardly in the analysis. We include the analysis in Appendix B for the sake of completeness. The entire algorithm is as shown in Fig.2. We have the following regret guarantee for Algorithm Pro DR.control.delayed. Theorem 9. Let xt be the prediction of the algorithm in Fig. 2 at time t. Instantiating each Pro DR.control instance by the parameter setting described in Theorem 5. Let τ be the feedback delay. We have that t=1 ft(xt) ft(ut) = O(d3τ 2/3n1/3[T V(u1:n)]2/3 τ). Pro DR.control.delayed: Inputsdelay τ > 0 Maintain τ separate instances of Pro DR.control (Fig.1). Enumerate them by 0, 1, . . . , τ 1. At time t: 1. Update instance (t 1) mod τ with loss ft τ. 2. Predict using instance (t 1) mod τ. Figure 2: Pro DR.control.delayed: An instance of delayed to non-delayed reduction from Joulani et al. [2013] Further for any interval [a, b] [n]: t=a ft(xt) ft(u) = O(d1.5τ log n). 5 Instantiation for the LQR Problem In view of Proposition 2, the LQR problem is reduced to a mini-batch linear regression problem with delayed feedback, where the delay is given by h = O(log n) in Proposition 2. In this section, we provide explicit form of the linear regression losses arising in the LQR problem and instantiate Algorithm Pro DR.control.delayed (Fig.2). First we need to define certain quantities: For a sequence of matrices (M [i])m i=1 define flatten((M [i])m i=1) as follows: Let M [i] k be the kth column of M [i]. Let s define M k 1... M k dx flatten((M [i])m i=1) := For a sequence of DAP parameters M1:n, let T V(M1:n) := Pn t=2 Pm i=1 M [i] t M [i] t 1 1. We define deflatten as the natural inverse operation of flatten. We have the following Corollary of Theorem 9 and Proposition 2. Corollary 10. Assume the notations in Fig.1 and Section 2. Let Σ = U T Λ U be the spectral decomposition of the positive semi definite (PSD) matrix Σ Rdu du. . Let the covariate matrix At := [w T t 1 . . . w T t m] Λ1/2 U Rdu mdudx, where denotes the Kronecker product. Let the bias vector bt := Λ1/2 U q ;h(wt:t+h). Let the delay factor of Pro DR.control.delayed (Fig.2) be τ = h as defined in Proposition 2 and let the decision set given to the Pro DR.control instances in Fig.2 be the DAP space defined in Eq.(2). Let zt be the prediction at round t made by the Pro DR.control.delayed algorithm and let M alg t := deflatten(zt). At round t, we play the control signal ualg t (xt) = πM alg t t (xt) according to Eq.(1). There exists a choice of input parameters (see Corollary 19 in Appendix B) for the Pro DR.control instances in Fig.2 such that t=1 ℓ(xalg t , ualg t ) ℓ(x M1:n t , u M1:n t ) = O m3d4d5 x(du dx)(n1/3[T V(M1:n)]2/3 1) , where M1:n is a sequence of DAP policies where each Mt M (eq.(2)). Further the algorithm Pro DR.control.delayed also enjoys a strongly adaptive regret guarantee for any interval [a, b] [n]: t=a ℓ(xalg t , ualg t ) ℓ(x M t , u M t ) = O((mdudx)1.5 log n), for any fixed DAP policy M M. The following theorem provides a nearly matching lower bound. Theorem 11. There exists an LQR system, a choice of the perturbations wt and a DAP policy class such that: sup M1:n with T V(M1:n) Cn E[R(M1:n)] = Ω(n1/3C2/3 n 1), where the expectation is taken wrt randomness in the strategies of the agent and adversary. We refer the reader to Appendix B for a proof of the above theorem along with its connections to online non-parametric regression framework of Rakhlin and Sridharan [2014]. Remark 12. The covariate matrix At Rdu mdudx that arises in Corollary 10 is rank deficient whenever mdx > 1. In such cases, the linear regression losses ft(w) as in Fig.1 cannot be strongly convex. So the proper universal dynamic regret minimizing algorithm for strongly convex losses from Baby and Wang [2022] is inapplicable in general except potentially for the particular setting of m = dx = 1. Moreover, in the setting of m = dx = 1 a non-zero strong convexity parameter can exist only if the magnitude of the perturbations |wt| are bounded away from zero which is restrictive in its scope. 6 Conclusion and Future Work In this paper, we proposed a new algorithm for minimizing dynamic regret of the non-stationary linear regression problem. We applied this algorithm to obtain a non-stationary LQR controller. The techniques developed in this work can be of independent interest in the broader literature of online learning. We defer the task of deriving similar dynamic regret rates for general strongly convex losses in the LQR problem as a future work. As mentioned in Section 1, there has been a recent surge of interest in applying tools from online learning to develop non-stochastic controllers. The present work also falls under this umbrella. However, existing literature lacks experimental studies in this vein. It would be a good direction to do a thorough survey of the strengths and limitations of various online learning based controllers when deployed in practice. Acknowledgements The construction of the lower bound in Theorem 11 is due to an early discussion with Daniel Lokshtanov on a related problem. We also thank the anonymous reviewers for their detailed suggestions in improving the paper. Dmitry Adamskiy, Wouter M. Koolen, Alexey Chernov, and Vladimir Vovk. A closer look at adaptive regret. Journal of Machine Learning Research, 2016. Naman Agarwal, Brian Bullins, Elad Hazan, Sham Kakade, and Karan Singh. Online control with adversarial disturbances. In Proceedings of the 36th International Conference on Machine Learning, 2019. Oren Anava, Elad Hazan, and Shie Mannor. Online learning for adversaries with memory: Price of past mistakes. In Advances in Neural Information Processing Systems, 2015. Dheeraj Baby and Yu-Xiang Wang. Online forecasting of total-variation-bounded sequences. In Neural Information Processing Systems (Neur IPS), 2019. Dheeraj Baby and Yu-Xiang Wang. Adaptive online estimation of piecewise polynomial trends. Neural Information Processing Systems (Neur IPS), 2020. Dheeraj Baby and Yu-Xiang Wang. Optimal dynamic regret in exp-concave online learning. In COLT, 2021. Dheeraj Baby and Yu-Xiang Wang. Optimal dynamic regret in proper online learning with strongly convex losses and beyond. AISTATS, 2022. Dheeraj Baby, Hilaf Hasson, and Yuyang Wang. Dynamic regret for strongly adaptive methods and optimality of online krr. Ar Xiv, abs/2111.11550, 2021a. Dheeraj Baby, Xuandong Zhao, and Yu-Xiang Wang. An optimal reduction of tv-denoising to adaptive online learning. AISTATS, 2021b. Dimitri P. Bertsekas. Dynamic Programming and Optimal Control. Athena Scientific, 3rd edition, 2005. Omar Besbes, Yonatan Gur, and Assaf Zeevi. Non-stationary stochastic optimization. Operations research, 63(5):1227 1244, 2015. Asaf Cassel and Tomer Koren. Bandit linear control. In Proceedings of the 34th International Conference on Neural Information Processing Systems, 2020. Nicolo Cesa-Bianchi and Gabor Lugosi. Prediction, Learning, and Games. Cambridge University Press, New York, NY, USA, 2006. ISBN 0521841089. Ting-Jui Chang and Shahin Shahrampour. On online optimization: Dynamic regret analysis of strongly convex and smooth problems. Proceedings of the AAAI Conference on Artificial Intelligence, 2021. Sabyasachi Chatterjee and Subhajit Goswami. Spatially adaptive online prediction of piecewise regular functions. Ar Xiv, abs/2203.16587, 2022. Xi Chen, Yining Wang, and Yu-Xiang Wang. Non-stationary stochastic optimization under lp, q-variation measures. 2018. Alon Cohen, Avinatan Hassidim, Tomer Koren, Nevena Lazic, Y. Mansour, and Kunal Talwar. Online linear quadratic control. In ICML, 2018. Ashok Cutkosky. Parameter-free, dynamic, and strongly-adaptive online learning. In ICML, 2020. Ashok Cutkosky and Francesco Orabona. Black-box reductions for parameter-free online learning in banach spaces. In COLT, 2018. Amit Daniely, Alon Gonen, and Shai Shalev-Shwartz. Strongly adaptive online learning. In International Conference on Machine Learning, pages 1405 1411, 2015. Ronald A. De Vore and George G. Lorentz. Constructive approximation. In Grundlehren der mathematischen Wissenschaften, 1993. Dylan J. Foster and Max Simchowitz. Logarithmic regret for adversarial online control. In ICML, 2020. Gautam Goel and Babak Hassibi. The power of linear controllers in lqr control. ar Xiv preprint ar Xiv:2002.02574, 2020. Gautam Goel and Babak Hassibi. Regret-optimal estimation and control. Transactions of Automatic Control (TAC), 2021. Gautam Goel and Adam Wierman. An online algorithm for smoothed regression and lqr control. In Proceedings of the Twenty-Second International Conference on Artificial Intelligence and Statistics, 2019. Paula Gradu, John Hallman, and Elad Hazan. Non-stochastic control with bandit feedback. In Advances in Neural Information Processing Systems, 2020a. Paula Gradu, Elad Hazan, and Edgar Minasyan. Adaptive regret for control of time-varying dynamics. Ar Xiv, abs/2007.04393, 2020b. Elad Hazan. Introduction to online convex optimization. Foundations and Trends in Optimization, 2(3-4):157 325, 2016. Elad Hazan and Comandur Seshadhri. Adaptive algorithms for online decision problems. In Electronic colloquium on computational complexity (ECCC), volume 14, 2007. Elad Hazan, Amit Agarwal, and Satyen Kale. Logarithmic regret algorithms for online convex optimization. Machine Learning, 69(2):169 192, 2007. Elad Hazan, Sham Kakade, and Karan Singh. The nonstochastic control problem. In Proceedings of the 31st International Conference on Algorithmic Learning Theory, 2020. Andrew Jacobsen and Ashok Cutkosky. Parameter-free mirror descent. COLT, 2022. Ali Jadbabaie, Alexander Rakhlin, Shahin Shahrampour, and Karthik Sridharan. Online optimization: Competing with dynamic comparators. In Artificial Intelligence and Statistics, pages 398 406, 2015. Pooria Joulani, András György, and Csaba Szepesvari. Online learning under delayed feedback. In ICML, 2013. Kwang-Sung Jun, Francesco Orabona, Stephen Wright, and Rebecca Willett. Improved Strongly Adaptive Online Learning using Coin Betting. In Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, 2017. Sergey Levine, Chelsea Finn, Trevor Darrell, and Pieter Abbeel. End-to-end training of deep visuomotor policies. 2016. Yuwei Luo, Varun Gupta, and Mladen Kolar. Dynamic regret minimization for control of nonstationary linear dynamical systems. Proc. ACM Meas. Anal. Comput. Syst., 2022. N. Merhav, E. Ordentlich, G. Seroussi, and M.J. Weinberger. On sequential strategies for loss functions with memory. In 2000 IEEE International Symposium on Information Theory, 2000. Aryan Mokhtari, Shahin Shahrampour, A. Jadbabaie, and Alejandro Ribeiro. Online optimization in dynamic environments: Improved regret rates for strongly convex problems. 2016 IEEE 55th Conference on Decision and Control (CDC), pages 7195 7201, 2016. Anant Raj, Pierre Gaillard, and Christophe Saad. Non-stationary online regression. Ar Xiv, abs/2011.06957, 2020. Alexander Rakhlin and Karthik Sridharan. Online non-parametric regression. In Conference on Learning Theory, pages 1232 1264, 2014. Guanya Shi, Yiheng Lin, Soon-Jo Chung, Yisong Yue, and Adam Wierman. Online optimization with memory and competitive control. In Proceedings of the 34th International Conference on Neural Information Processing Systems, 2020. Max Simchowitz. Making non-stochastic control (almost) as easy as stochastic. In Advances in Neural Information Processing Systems, 2020. Max Simchowitz, Karan Singh, and Elad Hazan. Improper learning for non-stochastic control. In COLT, 2020. Tianbao Yang, Lijun Zhang, Rong Jin, and Jinfeng Yi. Tracking slowly moving clairvoyant: optimal dynamic regret of online learning with true and noisy gradient. In International Conference on Machine Learning (ICML-16), pages 449 457, 2016. Lijun Zhang, Shiyin Lu, and Zhi-Hua Zhou. Adaptive online learning in dynamic environments. In Advances in Neural Information Processing Systems (Neur IPS-18), pages 1323 1333, 2018a. Lijun Zhang, Tianbao Yang, Zhi-Hua Zhou, et al. Dynamic regret of strongly adaptive methods. In International Conference on Machine Learning (ICML-18), pages 5877 5886, 2018b. Lijun Zhang, G. Wang, Wei-Wei Tu, and Zhi-Hua Zhou. Dual adaptivity: A universal algorithm for minimizing the adaptive regret of convex functions. Neur IPS, 2021a. Zhiyu Zhang, Ashok Cutkosky, and Ioannis Ch. Paschalidis. Strongly adaptive oco with memory. Ar Xiv, abs/2102.01623, 2021b. Peng Zhao and Lijun Zhang. Improved analysis for dynamic regret of strongly convex and smooth functions. L4DC, 2021. Peng Zhao, Y. Zhang, L. Zhang, and Zhi-Hua Zhou. Dynamic regret of convex and smooth functions. Neur IPS, 2020. Peng Zhao, Yu-Xiang Wang, and Zhi-Hua Zhou. Non-stationary online learning with memory and non-stochastic control. AISTATS, 2022. Martin Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In International Conference on Machine Learning (ICML-03), pages 928 936, 2003. 1. For all authors... (a) Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? [Yes] (b) Did you describe the limitations of your work? [Yes] We do not support general strongly convex losses as mentioned in Section 6. (c) Did you discuss any potential negative societal impacts of your work? [N/A] Paper of theoretical nature. (d) Have you read the ethics review guidelines and ensured that your paper conforms to them? [Yes] 2. If you are including theoretical results... (a) Did you state the full set of assumptions of all theoretical results? [Yes] (b) Did you include complete proofs of all theoretical results? [Yes] 3. If you ran experiments... (a) Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [N/A] Paper of theoretical nature. (b) Did you specify all the training details (e.g., data splits, hyperparameters, how they were chosen)? [N/A] (c) Did you report error bars (e.g., with respect to the random seed after running experiments multiple times)? [N/A] (d) Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [N/A] 4. If you are using existing assets (e.g., code, data, models) or curating/releasing new assets... (a) If your work uses existing assets, did you cite the creators? [N/A] (b) Did you mention the license of the assets? [N/A] (c) Did you include any new assets either in the supplemental material or as a URL? [N/A] (d) Did you discuss whether and how consent was obtained from people whose data you re using/curating? [N/A] (e) Did you discuss whether the data you are using/curating contains personally identifiable information or offensive content? [N/A] 5. If you used crowdsourcing or conducted research with human subjects... (a) Did you include the full text of instructions given to participants and screenshots, if applicable? [N/A] (b) Did you describe any potential participant risks, with links to Institutional Review Board (IRB) approvals, if applicable? [N/A] (c) Did you include the estimated hourly wage paid to participants and the total amount spent on participant compensation? [N/A]