# the_power_of_predictions_in_online_control__13b3925e.pdf The Power of Predictions in Online Control Chenkai Yu IIIS, Tsinghua University yck17@mails.tsinghua.edu.cn Guanya Shi CMS, Caltech gshi@caltech.edu Soon-Jo Chung CMS, GALCIT, JPL, Caltech sjchung@caltech.edu Yisong Yue CMS, Caltech yyue@caltech.edu Adam Wierman CMS, Caltech adamw@caltech.edu We study the impact of predictions in online Linear Quadratic Regulator control with both stochastic and adversarial disturbances in the dynamics. In both settings, we characterize the optimal policy and derive tight bounds on the minimum cost and dynamic regret. Perhaps surprisingly, our analysis shows that the conventional greedy MPC approach is a near-optimal policy in both stochastic and adversarial settings. Specifically, for length-T problems, MPC requires only O(log T) predictions to reach O(1) dynamic regret, which matches (up to lower-order terms) our lower bound on the required prediction horizon for constant regret. 1 Introduction This paper studies the effect of using predictions for online control in a linear dynamical system governed by xt+1 = Axt + But + wt, where xt, ut, and wt are the state, control, and disturbance (or exogenous input) respectively. At each time step t, the controller incurs a quadratic cost c(xt, ut). Recently, considerable effort has been made to leverage and integrate ideas from learning, optimization and control theory to study the design of optimal controllers under various performance criteria, such as static regret [2, 3, 12, 13, 15, 20, 29], dynamic regret [16, 23] and competitive ratio [17, 28]. However, the study of online convergence when incorporating predictions has been largely absent. Indeed, a key aspect of online control is considering the amount of available information when making decisions. Most recent studies focus on the basic setting where only past information, x0, w0, , wt 1, is available for ut at every time step [2, 13, 15, 28]. However, this basic setting does not effectively characterize situations where we have accurate predictions, e.g., when x0, w0, , wt 1+k are available at step t. These types of accurate predictions are often available in many applications, including robotics [8, 27], energy systems [30], and data center management [22]. Moreover, there are many practical algorithms that leverage predictions, such as the popular Model Predictive Control (MPC) [6 9, 18, 19]. While there has been increased interest in studying online guarantees for control with predictions, to our knowledge, there has been no such study for the case of a finite-time horizon with disturbances. Several previous works studied the economic MPC problem by analyzing the asymptotic performance without disturbances [6, 7, 18, 19]. Rosolia and Borrelli [25, 26] studied learning for MPC but focused on the episodic setting with asymptotic convergence guarantees. Li et al. [23] considered a linear system where finite predictions of costs are available, and analyzed the dynamic regret of their new algorithm; however, they neither consider disturbances nor study the more practically relevant MPC approach. Goel and Hassibi [16] characterized the offline optimal policy (i.e., with infinite predictions) and cost in LQR control with i.i.d. zero-mean stochastic disturbances, but those results do not apply to limited predictions or non-i.i.d. disturbances. Other prior works analyze the power of 34th Conference on Neural Information Processing Systems (Neur IPS 2020), Vancouver, Canada. predictions in online optimization [11, 24], but the connection to online control in dynamical systems is unclear. From this literature, fundamental questions about online control with predictions have emerged: 1. What are the cost-optimal and regret-minimizing policies when given k predictions? What are the corresponding cost and regret of these policies? 2. What is the marginal benefit from each additional prediction used by the policy, and how many predictions are needed to achieve (near-)optimal performance? 3. How well does MPC with k predictions perform compared to cost-optimal and regretminimizing policies? Main contributions. We systematically address each of the questions above in the context of LQR systems with general stochastic and adversarial disturbances in the dynamics. In the stochastic case, we explicitly derive the cost-optimal and dynamic regret minimizing policies with k predictions. In both the stochastic and adversarial cases, we derive (mostly tight) upper bounds for the optimal cost and minimum dynamic regret given access to k predictions. We also show that the marginal benefit of an extra prediction exponentially decays as k increases. Additionally, for MPC specifically, we show that it has a bounded performance ratio against the cost-optimal policy in both stochastic and adversarial settings. We further show that MPC is near-optimal in terms of dynamic regret, and needs only O(log T) predictions to achieve O(1) dynamic regret (the same order as is needed by the dynamic regret minimizing policy) in both settings. We would like to emphasize the generality of the results. The model we consider is the general LQR setting with disturbance in the dynamics, where only the stabilizability of [A, B] and [A , Q] is assumed [4]. Further, in the stochastic setting we consider general distributions, which are not necessarily i.i.d. or zero-mean. Additionally, our results compare to the globally optimal policies for cost and regret rather than compare to the optimal linear or static policy. Finally, our upper bounds are (almost) tight, i.e., there exist some systems such that the bounds are (nearly) reached, up to lower-order terms. It is perhaps surprising that classic MPC, which is a simple greedy policy (up to the prediction horizon), is near-optimal even with adversarial disturbances in the dynamics. Our results thus highlight the power of predictions to reduce the need for algorithmic sophistication. In that sense, our results somewhat mirror recent developments in the study of exploration strategies in online LQR control with unknown dynamics {A, B}: after a decade s research beginning with the work of Abbasi-Yadkori and Szepesvári [1], Simchowitz and Foster [29] recently show that naive exploration is optimal. Taken together with the result from [29], our paper provides additional evidence for the idea that the structure of LQR allows simple algorithmic ideas to be effective, which sheds light on key algorithmic principles and fundamental limits in continuous control. 2 Background and model We consider the Linear Quadratic Regulator (LQR) optimal control problem with disturbances in the dynamics. In particular, we consider a linear system initialized with x0 Rn and controlled by ut Rd, with dynamics xt+1 = Axt + But + wt and cost J = t=0 (x t Qxt + u t Rut) + x T Qf x T , where T 1 is the total length of the control period. The goal of the controller is to minimize the cost given A, B, Q, R, Qf , x0, and the characterization of the disturbance wt. Throughout this paper, we use ρ( ) to denote the spectral radius of a matrix and to denote the 2-norm of a vector or the spectral norm of a matrix. We assume Q, Qf 0, R 0 and the pair (A, B) is stabilizable, i.e., there exists a matrix K0 Rd n such that ρ(A BK0) < 1. Further, we assume the pair (A, Q) is detectable, i.e., (A , Q) is stabilizable, to guarantee stability of the closed-loop. Note that detectability of (A, Q) is more general than Q 0, i.e., Q 0 implies (A, Q) is detectable. For wt, in the stochastic case, we assume {wt}t=0,1, are sampled from a joint distribution with bounded cross-correlation, i.e., E w t wt m for any t, t ; in the adversarial case, we assume wt is picked from a bounded set Ω. These are standard assumptions in the literature, e.g., [13, 15, 29] and it is worth noting that our notion of stochasticity is much more general than typically considered [10, 12, 13]. We also note that many important problems can be straightforwardly converted to our model for example, input-disturbed systems and the Linear Quadratic (LQ) tracking problem [4]. Example: linear quadratic tracking. The standard quadratic tracking problem is defined with dynamics xt+1 = Axt + But + wt and cost function J = PT 1 t=0 (xt+1 dt+1) Q(xt+1 dt+1) + u t Rut, where {dt}T t=1 is the desired trajectory to track. To map this to our model, let xt = xt dt. Then, we get J = PT 1 t=0 x t+1Q xt+1 + u t Rut and xt+1 = A xt + But + wt, which is an LQR control problem with disturbance wt = wt + Adt dt+1 in the dynamics. 2.1 Predictions In the classic model, at each step t, the controller decides ut after observing wt 1 and xt. In other words, ut is a function of all the previous information: x0, x1, . . . , xt 1 and w0, w1, . . . , wt 1, or equivalently, of x0, w0, w1, , wt 1. We describe this scenario via the following event sequence: x0 u0 w0 u1 w1 u T 1 w T 1, where each ut denotes the decision of a control policy, each wt denote the observation of a disturbance, and each decision may depend on previous events. However, in many real-world applications the controller may have some knowledge about future. In particular, at time step t, the controller may have predictions of immediate k future disturbances and make decision ut based on x0, w0, . . . , wt+k 1. In this case, the event sequence is given by: x0 w0 w1 wk 1 u0 wk u1 wk+1 u T k 1 w T 1 u T k u T 1. The existence of predictions is common in many applications such as disturbance estimation in robotics [27] and model predictive control (MPC) [9], which is a common approach for the LQ tracking problem. When given k predictions of dt, the LQ tracking problem can be formulated as a LQR problem with k 1 predictions of future disturbances. In this paper we assume all the predictions are exact, and leave inexact predictions [11, 28] as future work. This is common in the literature on online algorithms with predictions, e.g., [23, 24]. 2.2 Disturbances The characteristics of the disturbances have a fundamental impact on the optimal control policy and cost. We consider two types of disturbance: stochastic disturbances, which are drawn from a joint distribution (not necessarily i.i.d.), and adversarial disturbances, which are chosen by an adversary to maximize the overall control cost of the policy. In the stochastic setting, we model the disturbance sequence {wt}T 1 t=0 as a discrete-time stochastic process with joint distribution W which is known to the controller. Let Wt = Wt(w0, . . . , wt 1) be the conditional distribution of wt given w0, . . . , wt 1. Then the cost of the optimal online policy with k predictions is given by: STOT k = E w0 W0,...,wk 1 Wk 1 min u T k 1 E w T 1 WT 1 min u T k,...,u T 1 J . Note that the cost J = J(x0, u0, , u T 1, w0, , w T 1). Two extreme cases are noteworthy: k = 0 reduces to the classic case without prediction and k = T reduces to the offline optimal. In the adversarial setting, each disturbance wt is selected by an adversary from a bounded set Ω Rn in order to maximize the cost. The controller has no information about the disturbance except that it is in Ω. Similar to the stochastic setting, we define: ADV T k = sup w0,...,wk 1 Ω min u T k 1 sup w T 1 Ω min u T k,...,u T 1 J . This can be viewed as online H control [31] with predictions. Algorithm 1: Model predictive control with k predictions Parameter: {A, B, Q, R} and Qf Rn n Input: x0, w0, . . . , wk 1 1 for t = 0 to T 1 do Input: xt, wt+k 1 // The controller now knows x0, . . . , xt, w0, . . . , wt+k 1 2 (ut, . . . , ut+k 1) = arg minu Pt+k 1 i=t (x i Qxi + u i Rui) + x t+k Qf xt+k subject to xi+1 = Axi + Bui + wi for i = t, . . . , t + k 1 Output: ut The average cost in an infinite horizon is particularly important in both control and learning communities to understand asymptotic behaviors. We use separate notation for it: STOk = lim T 1 T STOT k , ADVk = lim T 1 T ADV T k . We emphasize that we do not have any constraints (like linearity) on the policy space, and both STOT k and ADV T k are globally optimal with the corresponding type of disturbance. This point is important in light of recent results that show that linear policies cannot make use of predictions at all [16, 28], i.e., the cost of the best linear policy with infinite predictions (k = ) is asymptotically equal to that with no predictions (k = 0) in the setting with i.i.d. zero-mean stochastic disturbances. In this paper, we explicitly derive the optimal policy for every k > 0, which is nonlinear in general. 2.3 Model predictive control Model predictive control (MPC) is perhaps the most common control policy for situations where predictions are available. MPC is a greedy algorithm with a receding horizon based on all available current predictions. Algorithm 1 provides a formal definition, and we additionally refer the reader to the book [9] for a literature review on MPC. We adopt a conventional definition of MPC as an online optimal control problem with a finite-time horizon with dynamics constraints. Note that other prior work on MPC sometimes considers other input and state constraints [9]. MPC is a practical algorithm in many scenarios like robotics [8], energy system [30] and data center cooling [22]. The existing theoretical studies of MPC focus on asymptotic stability and performance [6, 7, 18, 19, 25]. To our knowledge, we provide the first general, dynamic regret guarantee for MPC in this paper. In this paper, we study the performance of MPC in three different cases, where disturbances are i.i.d. zero-mean stochastic, generally stochastic, and adversarial, corresponding to Sections 3 to 5 respectively. We define the performance of MPC in the stochastic and adversarial settings as follows: MPCS T k = E w0,...,w T 1 JMPCk, MPCSk = lim T 1 T MPCS T k , MPCAT k = sup w0,...,w T 1 JMPCk, MPCAk = lim T 1 T MPCAT k , where JMPCk is the cost of MPC given a specific disturbance sequence, i.e., JMPCk(w) = J(u, w) where for each t, ut = φ(xt, wt, . . . , wt+k 1) and φ( ) is the function that maps xt, wt, . . . , wt+k 1 to the policy ut, as defined in Algorithm 1. By definition, MPCSk STOk and MPCAk ADVk for every k 1 since they use the same information but the latter ones are defined to be optimal. 2.4 Dynamic regret and the performance ratio In this paper, we focus on two performance metrics, the dynamic regret and the performance ratio. Dynamic regret. Regret is a standard metric in online learning and provides a bound on the cost difference between an online algorithm and the optimal static policy given complete information. We focus on the dynamic regret, which compares to the optimal dynamic offline policy, rather than the optimal static offline policy. Note that the optimal offline policy may be nonlinear. It is important to consider nonlinear policies because recent results highlight that the optimal offline policy can have cost that is arbitrarily lower than the optimal linear policy in hindsight [16, 28]. More specifically, we compare the cost of an online algorithm with k predictions to that of the offline optimal (nonlinear) algorithm, i.e., one that has predictions of all disturbances. For MPC with k predictions, we define its dynamic regret in the stochastic and adversarial settings, respectively, as: Reg S(MPCk) = E (w0, ,w T 1) W JMPCk(w) min u 0,...,u T 1 J(u , w) , Reg A(MPCk) = sup w0, ,w T 1 Ω JMPCk(w) min u 0,...,u T 1 J(u , w) . As compared to (static) regret, dynamic regret does not have any restriction on the policies u 0, . . . , u T 1 used for comparison and thus differs from other notions of regret where u 0, . . . , u T 1 are limited in special cases. For example, in the classic form of regret, u 0 = = u T 1; and in the regret compared to the best offline linear controller [2, 12], u t = K xt. In this work, we obtain both upper bounds and lower bounds on dynamic regret. For lower bounds, we define the minimum possible regret that an algorithm with k predictions can achieve (i.e., the regret of the algorithm that minimizes the regret): Reg S k = E w0, ,wk 1 min u0 E wk min u T k 1 E w T 1 min u T k, ,u T 1 J(u, w) min u 0,...,u T 1 J(u , w) , Reg A k = sup w0, ,wk 1 min u0 sup wk min u T k 1 sup w T 1 min u T k, ,u T 1 J(u, w) min u 0,...,u T 1 J(u , w) . Finally, we end our discussion of dynamic regret with a note highlighting an important contrast between stochastic and adversarial settings. In the stochastic setting, Reg S k = E w0, ,wk 1 min u0 E wk min u T k 1 E w T 1 min u T k, ,u T 1 J(u, w) min u 0,...,u T 1 J(u , w) = E w0, ,wk 1 min u0 E wk min u T k 1 E w T 1 min u T k, ,u T 1 J(u, w) E w0,...,w T 1 min u 0,...,u T 1 J(u , w) = STOT k STOT T . This equality still holds if we take arg min instead of min and thus the regret-optimal policy is the same as the cost-optimal policy. However, in the adversarial case, a similar reasoning gives an inequality: Reg A k ADV T k ADV T T , and correspondingly, the regret-optimal and cost-optimal policies can be different. Similarly, for MPC, we have Reg S(MPCk) = MPCS T k STOT T while Reg A(MPCk) MPCAT k ADV T T . Performance ratio. The second metric we study is a new metric that we term the performance ratio. It characterizes the ratio of the cost of an online algorithm with k predictions to the cost of the optimal online algorithm using k predictions. Thus, it gives a way of comparing to a weaker benchmark than regret one that has the same amount of information as the algorithm. Note that it is related to, but different than, the competitive ratio in this context. Formally, the performance ratio of the MPC algorithm in stochastic and adversarial settings, respectively, is defined as: PRS(MPCk) = MPCSk STOk , PRA(MPCk) = MPCAk While the dynamic regret indicates whether the algorithm can match the optimal offline policy (which has complete information), the performance ratio measures whether the algorithm is using the information available to it in as efficient a manner as possible. Thus, the contrast between the two separates the efficiency of the algorithm from the inefficiency created by the lack of information about future disturbances. Finally, one may wonder if there are connections between dynamic regret and performance ratio. As might be expected, in both the stochastic and adversarial settings, the performance ratio of an online policy with k predictions provides a lower bound of its dynamic regret: PRS(MPCk) 1 MPCSk STO 1 = MPCSk STO STO = 1 STO lim T 1 T Reg S(MPCk), PRA(MPCk) 1 MPCAk ADV 1 = MPCAk ADV ADV 1 ADV lim T 1 T Reg A(MPCk). 3 Zero-mean i.i.d. disturbances We begin our analysis with the simplest of the three settings we consider: the disturbances wt are independent and identically distributed with zero mean. Though i.i.d. zero-mean is a limited setting, it is still complex enough to study predictions and the first results characterizing the optimal policy with predictions appeared only recently [15, 16], focusing only on the optimal policy when k . Before delving into our results, we first recap the classic Infinite Horizon Linear Quadratic Stochastic Regulator [4, 5], i.e., the case when k = 0: Proposition 3.1 (Anderson and Moore [5]). Let wt be i.i.d. with zero mean and covariance matrix W. Then, the optimal control policy corresponding to STO0 is given by: ut = (R + B PB) 1B PAxt =: Kxt, where P is the solution of discrete-time algebraic Riccati equation (DARE) P = Q + A PA A PB(R + B PB) 1B PA. (1) The corresponding closed-loop dynamics A BK is exponentially stable, i.e., ρ(A BK) < 1. Further, the optimal cost is given by STO0 = Tr{PW}. This result has been extensively studied in optimal control theory [4, 21] as well as in reinforcement learning [13, 14, 29]. We want to emphasize two important properties of the optimal policy ut = Kxt. First, the policy is linear in the state xt. In contrast, we show later that the optimal policy when k = 0 is, in general, nonlinear. Second, under the assumptions of our model, this policy is exponentially stable, i.e., ρ(A BK) < 1. We leverage this to show the power of predictions later in the paper. Optimal policy. Let F = A BK and λ = 1+ρ(F ) 2 < 1. From Gelfand s formula, there exists a constant c(n) such that F k c(n)λk for all k 1. Theorem 3.2. Let wt be i.i.d. with zero mean and covariance matrix W. Suppose the controller has k 1 predictions. Then, the optimal control policy at each step t is given by: ut = (R + B PB) 1B i=0 (A A PH)i Pwt+i where P is the solution of DARE in Equation (1). The cost under this policy is: i=0 P(A HPA)i H(A A PH)i P where H = B(R + B PB) 1B . Following the approach developed in [15, 16], the proof is based on an analysis of quadratic costto-go functions in the form Vt(xt) = x t Ptxt + v t xt + qt. Note that A HPA = A B(R + B PB) 1B PA = A BK = F. Thus, the online optimal cost STOk with k predictions approaches the offline optimal cost STO by an exponential rate. In other words, STOk/STO = 1 + O( F k 2) = 1 + O(λ2k). Two extreme cases of our result are noteworthy. When k = 0, it reduces to the classic Proposition 3.1. When k , it reduces to the offline optimal case derived by Goel and Hassibi [16]. Model predictive control. As might be expected, since the disturbances are i.i.d., future disturbances have no dependence on the current. As a result, MPC gives the optimal policy. Theorem 3.3. In Algorithm 1, let Qf = P. Then, the MPC policy with k predictions is also given by Equation (2). Assuming i.i.d. disturbance with zero mean, the MPC policy is optimal. Due to the greedy nature, MPC does not utilize any properties of the disturbance, so the first part in Theorem 3.3 holds not only for i.i.d. disturbance, but also other types of disturbance considered in the later sections, i.e., MPC policy with k predictions is always given by Equation (2). 4 General stochastic disturbances In this section, we consider a general form of stochastic disturbance, more general than typically considered in this context [10, 12, 13]. Suppose the disturbance sequence {wt}t=0,1,2,... is sampled from a joint distribution W such that the trace of the cross-correlation of each pair is uniformly bounded, i.e., there exist m > 0 such that for all t, t 1, E w t wt m. Optimal policy. In the case of general stochastic disturbances, we cannot obtain as clean a form for STOk as in the i.i.d. case in Section 3. However, the marginal benefit of having an extra prediction decays with the same (exponential) rate and the optimal policy is similar to that in Section 3, but with some additional terms that characterize the expected future disturbances given the current information. Theorem 4.1. The optimal control policy with general stochastic disturbance is given by: ut = (R + B PB) 1B i=0 F i Pwt+i + i=k F i Pµt+i|t+k 1 where µt |t = E[wt | w0, . . . , wt]. Under this policy, the marginal benefit of obtaining an extra prediction decays exponentially fast in the existing number k of predictions. Formally, for k 1, STOk STOk+1 = O( F k 2) = O(λ2k).1 This proof leverages a novel difference analysis of cost-to-go functions. Note that for some distributions, STOk may approach STO much faster than exponential rate. It is even possible that STOk = STO for finite k, as we show in Example 4.2 below. On the other hand, there are scenarios where STOk approaches STO in an exactly exponential manner, as we show in Example 4.3 below. Example 4.2. Define the joint distribution W such that with probability 1/2, all wt = w, and otherwise all wt = w. In this case, one prediction is equivalent to infinite predictions since it is enough to distinguish these two scenarios with only w0. As a result, STO1 = STO . Example 4.3. Suppose the system is 1-d (n = d = 1) and the disturbance is i.i.d. with zero mean, i.e., the setting of Section 3. Then, according to Equation (3), as long as F, P, H, W are non-zero, i=k F 2i P 2HW = Θ(F 2k). Model predictive control. The comparison between the MPC policy in Equation (2) and the optimal policy in Equation (4) reveals that MPC is a truncation of the optimal policy and is no longer optimal because MPC is a greedy policy without considering future dependence on current information. Nevertheless, it is still a near-optimal policy, as characterized by the following results. Theorem 4.4. MPCSk MPCSk+1 = O( F k 2) = O(λ2k). Moreover, in Example 4.3, MPCSk MPCSk+1 = Θ( F k 2). In other words, the marginal benefit for the MPC algorithm of an extra prediction decays exponentially fast, paralleling the result for optimal policy in Equation (4). Theorem 4.4 implies that MPC has a bounded performance ratio, which converges to 1 with an exponential rate in the number of available predictions. Formally: Corollary 4.5. PRS(MPCk) = MPCSk STO = MPCSk MPCS = 1 + O( F k 2) = 1 + O(λ2k). Moreover, in Example 4.2, we have PRS(MPCk) = 1 + Θ( F k 2). Besides, the dynamic regret of MPC (nearly) matches the order of the optimal dynamic regret. Theorem 4.6 (Main result). Reg S(MPCk) = MPCS T k STOT T = O( F k 2T +1) = O(λ2k T +1), where the second term results from the difference between finite/infinite horizons. 1We say that f(k) = O(g(k)) if C > 0, k 1, |f(k)| C g(k); Ω() is similar except that the last is replaced by ; Θ() means both O() and Ω(). This is stronger than the standard definition where f(k) = O(g(k)) if C > 0, k > 0, k k , |f(k)| C g(k). Theorem 4.7. The optimal dynamic regret Reg S k = STOT k STOT T = O( F k 2T + 1) = O(λ2k T + 1) and there exist A, B, Q, R, Qf , x0, and W such that Reg S k = Θ( F k 2(T k)). Note that, in the stochastic case, the regret-optimal policy is the same as the cost-optimal policy, i.e., the policy for STOT k is the same as Reg S k . 5 Adversarial disturbances We now move from stochastic to adversarial disturbances. In this section, the disturbances are chosen from a bounded set Ω Rn by an adversary in order to maximize the controller s cost. Maintaining small regret is more challenging in adversarial models than in stochastic ones, so one may expect weaker bounds. Perhaps surprisingly, we obtain bounds with the same order. Optimal policy. In the adversarial setting, the cost of the optimal policy, defined with a sequence of min s and sup s, is the equilibrium value of a two-player zero-sum game. In general, it is impossible to give an analytical expression of either ADVk or the corresponding optimal policy. However, we prove a result that is structurally similar to the results from the stochastic setting, highlighting the exponential improvement from predictions. Theorem 5.1. For k 1, ADVk ADVk+1 = O( F k 2) = O(λ2k). Similarly to Example 4.2 for the stochastic case, in the adversarial setting, the optimal cost with k predictions may approach the offline optimal cost (under infinite predictions) much faster than exponential rate, and it is possible that ADVk = ADV for finite k, as shown in Example 5.2. Example 5.2. Let A = B = Q = R = 1 and Ω= [ 1, 1]. In this case, one prediction is enough to leverage the full power of prediction. Formally, we have ADV1 = ADV = 1. In other words, for all k 1, ADVk = 1. The optimal control policy (as T ) is a piecewise function: (x + w) , 1 x + w 1 (x + w) + 3 5 2 (x + w 1) , x + w > 1 (x + w) + 3 5 2 (x + w + 1) , x + w < 1 . The proof leverages two different cost-to-go functions for the min player and the sup player. Note that the optimal policy could be much more complex. Unlike Example 5.2, where the optimal policy is piecewise linear with only 3 pieces, for other values of A, B, Q, R, this function may have many more pieces. Model predictive control. Under adversarial disturbances, MPC is suboptimal, e.g., in Example 5.2. However, its performance ratio and dynamic regret bounds turn out to be the same as those in the stochastic setting. Theorem 5.3. MPCAk MPCAk+1 = O( F k 2) = O(λ2k). Corollary 5.4. For k 1, PRA(MPCk) = MPCAk ADV = MPCAk MPCA = 1 + O( F k 2) = 1 + O(λ2k). This highlights that MPC has a bounded performance ratio, which converges to 1 with exponential rate. Additionally, MPC has the same order of dynamic regret as the stochastic case: Theorem 5.5 (Main result). Reg A(MPCk) = O( F k 2T + 1) = O(λ2k T + 1). This dynamic regret is linear in the horizon T if we fix the number of predictions. However, if k is a super-constant function of T an increasing function of T that is not upper-bounded by a constant then the regret is sub-linear. Furthermore, if we let k = log T 2 log(1/λ), then Reg A(MPCk) = O(1). In other words, we can get constant regret with O(log T) predictions, even with adversarial disturbances. Finally, as implied by the following result, the O(log T) horizon cannot be improved since even the regret minimizing algorithm needs the same order of predictions to reach constant regret. Theorem 5.6. Reg A k = O( F k 2T + 1) = O(λ2k T + 1). Moreover, there exist A, B, Q, R, Qf , x0, and Ωsuch that Reg A k = Ω( F k 2(T k)).2 2Ω( ) is the growth order notation and has nothing to do with the bounded set Ω. 5 10 15 k 10-7 cost difference Figure 1: The power of predictions in online tracking. The left four figures show the desired trajectory (blue) and the actual trajectories (orange). The rightmost figure shows the cost difference (regret) between MPC using k predictions and the offline optimal policy. Note that the y-axis of the rightmost figure is in log-scale. 6 Numerical experiments To illustrate our theoretical results, we test MPC with different numbers of predictions in a Linear Quadratic (LQ) tracking problem, where the desired trajectory is given by: dt = 8 sin(t/3) cos(t/3) 8 sin(t/3) We consider following double integrator dynamics: pt+1 = pt + vt + ht, vt+1 = vt + ut + ηt, where pt R2 is the position, vt is the velocity, ut is the control, and ht, ηt U[ 1, 1]2 are i.i.d. noises. The objective is to minimize t=0 pt dt 2 + ut 2, where we let T = 200. This problem can be converted to the standard LQR with disturbance wt by letting xt = [ pt vt ] and wt = ht ηt and then using the reduction in the LQ tracking example in Section 2. Note that after the reduction, the disturbances are the combination of a deterministic trajectory and i.i.d. noises, which corresponds to the case discussed in Section 4. Figure 1 shows the tracking results with MPC using different numbers of predictions. We see that the regret exponentially decreases as the number of predictions increases, which is consistent with our theoretical results. 7 Concluding remarks We conclude with several open problems and potential future research directions. Our results highlight the power of predictions and show that, given predictions, a simple greedy policy (MPC) is nearoptimal for LQR control with disturbances in the dynamics, in terms of dynamic regret. Building on our results, it will be interesting to understand if MPC has a constant competitive ratio in this setting. In a different but related setting, Chen et al. [11] show for negative results on the competitive ratio so the answer is unclear at this point. Additionally, in this paper predictions are assumed to be perfect. Of course, in real applications predictions are noisy and are derived based on historical data. An important extension will be to understand how the analysis and results in this paper can extend to models with imperfect predictions learned from history, such as done in related models [11, 28]. Finally, real-world MPC problems often require non-linear dynamics and/or constraints, and the learning-theoretic study of such settings remains largely unexplored. Broader Impact Linear quadratic control is a common and powerful model with a variety of commercial and industrial applications, e.g., in robotics, chemical process control, and energy systems. This paper provides new fundamental insights about the role of predictions in online linear quadratic control with disturbances and provides the first finite time performance guarantees for the most commonly used policy in the linear quadratic setting, model predictive control (MPC). The guarantees provided by the theoretical analysis in this paper offer the potential for ensuring safety and robustness in industry applications where predictions are common and MPC is used. However, like many other theoretical contributions, this paper s results are limited to its assumptions, e.g., linear system and fixed system parameters {A, B, Q, R}. The performance of MPC and the fundamental limits in other scenarios, e.g., nonlinear dynamics or time-variant {A, B, Q, R}, are still open research problems. We see no ethical concerns related to the results in this paper. Acknowledgments and Disclosure of Funding This project was supported in part by funding from Raytheon, DARPA PAI, Ait F-1637598 and CNS-1518941, with additional support for Guanya Shi provided by the Simoudis Discovery Prize. [1] Yasin Abbasi-Yadkori and Csaba Szepesvári. Regret bounds for the adaptive control of linear quadratic systems. In Proceedings of the 24th Annual Conference on Learning Theory, pages 1 26, 2011. [2] Naman Agarwal, Brian Bullins, Elad Hazan, Sham M Kakade, and Karan Singh. Online control with adversarial disturbances. In International Conference on Machine Learning (ICML), 2019. [3] Naman Agarwal, Elad Hazan, and Karan Singh. Logarithmic regret for online control. In Advances in Neural Information Processing Systems, pages 10175 10184, 2019. [4] Brian DO Anderson and John B Moore. Optimal control: linear quadratic methods. Courier Corporation, 2007. [5] Brian DO Anderson and John B Moore. Optimal filtering. Courier Corporation, 2012. [6] David Angeli, Rishi Amrit, and James B Rawlings. On average performance and stability of economic model predictive control. IEEE transactions on automatic control, 57(7):1615 1626, 2011. [7] David Angeli, Alessandro Casavola, and Francesco Tedesco. Theoretical advances on economic model predictive control with time-varying costs. Annual Reviews in Control, 41:218 224, 2016. [8] Tomas Baca, Daniel Hert, Giuseppe Loianno, Martin Saska, and Vijay Kumar. Model predictive trajectory tracking and collision avoidance for reliable outdoor deployment of unmanned aerial vehicles. In 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 6753 6760. IEEE, 2018. [9] Eduardo F Camacho and Carlos Bordons Alba. Model predictive control. Springer Science & Business Media, 2013. [10] Asaf Cassel, Alon Cohen, and Tomer Koren. Logarithmic regret for learning linear quadratic regulators efficiently. ar Xiv preprint ar Xiv:2002.08095, 2020. [11] Niangjun Chen, Anish Agarwal, Adam Wierman, Siddharth Barman, and Lachlan LH Andrew. Online convex optimization using predictions. In Proceedings of the 2015 ACM SIGMETRICS International Conference on Measurement and Modeling of Computer Systems, pages 191 204, 2015. [12] Alon Cohen, Tomer Koren, and Yishay Mansour. Learning linear-quadratic regulators efficiently with only T regret. In International Conference on Machine Learning (ICML), 2019. [13] Sarah Dean, Horia Mania, Nikolai Matni, Benjamin Recht, and Stephen Tu. Regret bounds for robust adaptive control of the linear quadratic regulator. In Neural Information Processing Systems (Neur IPS), 2018. [14] Maryam Fazel, Rong Ge, Sham M Kakade, and Mehran Mesbahi. Global convergence of policy gradient methods for the linear quadratic regulator. ar Xiv preprint ar Xiv:1801.05039, 2018. [15] Dylan J Foster and Max Simchowitz. Logarithmic regret for adversarial online control. ar Xiv preprint ar Xiv:2003.00189, 2020. [16] Gautam Goel and Babak Hassibi. The power of linear controllers in LQR control. ar Xiv preprint ar Xiv:2002.02574, 2020. [17] Gautam Goel and Adam Wierman. An online algorithm for smoothed regression and LQR control. Proceedings of Machine Learning Research, 89:2504 2513, 2019. [18] Lars Grüne and Simon Pirkelmann. Economic model predictive control for time-varying system: Performance and stability results. Optimal Control Applications and Methods, 2018. [19] Lars Grüne and Marleen Stieler. Asymptotic stability and transient optimality of economic MPC without terminal conditions. Journal of Process Control, 24(8):1187 1196, 2014. [20] Elad Hazan, Sham M Kakade, and Karan Singh. The nonstochastic control problem. In Conference on Algorithmic Learning Theory (ALT), 2020. [21] Donald E Kirk. Optimal control theory: an introduction. Courier Corporation, 2004. [22] Nevena Lazic, Craig Boutilier, Tyler Lu, Eehern Wong, Binz Roy, MK Ryu, and Greg Imwalle. Data center cooling using model-predictive control. In Advances in Neural Information Processing Systems, pages 3814 3823, 2018. [23] Yingying Li, Xin Chen, and Na Li. Online optimal control with linear dynamics and predictions: Algorithms and regret analysis. In Advances in Neural Information Processing Systems, pages 14858 14870, 2019. [24] Yiheng Lin, Gautam Goel, and Adam Wierman. Online optimization with predictions and non-convex losses. ar Xiv preprint ar Xiv:1911.03827, 2019. [25] Ugo Rosolia and Francesco Borrelli. Learning model predictive control for iterative tasks. a data-driven control framework. IEEE Transactions on Automatic Control, 63(7):1883 1896, 2017. [26] Ugo Rosolia and Francesco Borrelli. Sample-based learning model predictive control for linear uncertain systems. ar Xiv preprint ar Xiv:1904.06432, 2019. [27] Guanya Shi, Xichen Shi, Michael O Connell, Rose Yu, Kamyar Azizzadenesheli, Animashree Anandkumar, Yisong Yue, and Soon-Jo Chung. Neural lander: Stable drone landing control using learned dynamics. In International Conference on Robotics and Automation (ICRA), 2019. [28] Guanya Shi, Yiheng Lin, Soon-Jo Chung, Yisong Yue, and Adam Wierman. Beyond no-regret: Competitive control via online optimization with memory. ar Xiv preprint ar Xiv:2002.05318, 2020. [29] Max Simchowitz and Dylan J Foster. Naive exploration is optimal for online LQR. ar Xiv preprint ar Xiv:2001.09576, 2020. [30] Sergio Vazquez, Jose Rodriguez, Marco Rivera, Leopoldo G Franquelo, and Margarita Norambuena. Model predictive control for power converters and drives: Advances and trends. IEEE Transactions on Industrial Electronics, 64(2):935 947, 2016. [31] Kemin Zhou and John Comstock Doyle. Essentials of robust control, volume 104. Prentice hall Upper Saddle River, NJ, 1998.