# neural_eventtriggered_control_with_optimal_scheduling__4671e050.pdf Neural Event-Triggered Control with Optimal Scheduling Luan Yang * 1 Jingdong Zhang * 1 2 Qunxi Zhu 1 3 4 Wei Lin 1 2 3 4 Learning-enabled controllers with stability certificate functions have demonstrated impressive empirical performance in addressing control problems in recent years. Nevertheless, directly deploying the neural controllers onto actual digital platforms requires impractically excessive communication resources due to a continuously updating demand from the closed-loop feedback controller. We introduce a framework aimed at learning the event-triggered controller (ETC) with optimal scheduling, i.e., minimal triggering times, to address this challenge in resource-constrained scenarios. Our proposed framework, denoted by Neural ETC, includes two practical algorithms: the path integral algorithm based on directly simulating the event-triggered dynamics, and the Monte Carlo algorithm derived from new theoretical results regarding lower bound of inter-event time. Furthermore, we propose a projection operation with an analytical expression that ensures theoretical stability and schedule optimality for Neural ETC. Compared to the conventional neural controllers, our empirical results show that the Neural ETC significantly reduces the required communication resources while enhancing the control performance in constrained communication resources scenarios. 1. Introduction Stabilizing the complex nonlinear systems represents a formidable focal task within the realms of mathematics and *Equal contribution 1Research Institute of Intelligent Complex Systems, Fudan University, China. 2School of Mathematical Sciences, LMNS, and SCMS, Fudan University, China. 3State Key Laboratory of Medical Neurobiology and MOE Frontiers Center for Brain Science, Institutes of Brain Science, Fudan University, China. 4Shanghai Artificial Intelligence Laboratory, China. Correspondence to: Qunxi Zhu , Wei Lin . Proceedings of the 42 nd International Conference on Machine Learning, Vancouver, Canada. PMLR 267, 2025. Copyright 2025 by the author(s). Neural ETC NLC Inter-event Time Triggering Times Neural ETC: 6 NLC: 1999 0 1 Time -3 Trajectories Figure 1. Comparison of Neural ETC (yellow) and neural Lyapunov control (NLC, purple) in stabilizing the Lorenz system under the event-triggered control setting. (a): The inter-event time of consecutive triggering events. The dashed lines represent the minimal inter-event time of each control. (b): The control values acting on variable x in the control process. (c): The controlled trajectories of the variable x. engineering. Previous research in the field of cybernetics has applied the Lyapunov stability theory to formulate stabilizing policies for linear or polynomial dynamical systems, including the linear quadratic regulator (LQR) (Khalil, 2002) and the sum-of-squares (SOS) polynomials, using the semidefinite planning (SDP) (Parrilo, 2000). Stabilizing more intricate dynamical systems with high dimension and nonlinearity, as encountered in real applications, has prompted the integration of machine learning techniques into the cybernetics community(Tsukamoto et al., 2021). Recent advancements in learning neural networks based controllers with certificate functions, such as Lyapunov function (Chang et al., 2019; Zhang et al., 2022a), La Salle function (Zhang et al., 2022b), barrier functions (Qin et al., 2020) and contraction metrics (Sun et al., 2021), have demonstrated outstanding performance in controlling diverse dynamics (Dawson et al., 2022). Nevertheless, it is noteworthy that all these controllers require updating the control signal continuously over time, leading to a considerable communication cost between controller and platform. The periodic control is mostly advocated for implementing feedback control laws on digital platforms (Franklin et al., 2002). However, such implementations often incur significant over-provisioning of the communication network, especially in the recently developed large-scale resourceconstrained wireless embedded control systems (Lemmon, 2010). To mitigate this issue, event-triggering mechanism is introduced to generate sporadic transmissions across the feedback channels of the system. Compared to periodic Neural Event-Triggered Control with Optimal Scheduling control which updates the control signal at a series of predefined explicit times, event-triggered control updates the control signal at the instants when the current measurement violates a predefined triggering condition, thereby triggering a state-dependent event (Heemels et al., 2012). Given that these instants are implicitly determined by the state trajectories, the scheduling of computation and communication resources for event-triggered control becomes a very challenging problem, involving the minimization of the number of events and the increase of inter-event time. While significant strides have been made in stabilizing specific dynamics with event-triggered control in recent years, the task of designing event-triggered control for general nonlinear and large-scale dynamics with optimal scheduling remains an open problem ( Aarz en, 1999; Tabuada, 2007; Heemels et al., 2008; Henningsson et al., 2008). Our goal is to design event-triggered control for general complex dynamics, ensuring both stability guarantee and optimal scheduling, i.e., to implement event-triggered control with the minimal triggering times and the maximal inter-event time. Fig. 1 depicts the comparison of control performance of the Neural ETC and the NLC in the event-triggered realization to stabilize a Lorenz dynamic. In Fig. 1(a)-1(b), it is evident that the triggering times of Neural ETC are significantly fewer than those of NLC, and the minimal inter-event time of consecutive triggering times of Neural ETC considerably exceeds that of NLC. These disparities lead to the different behaviors of the controlled trajectories, as depicted in Fig. 1(c).Under Neural ETC, the trajectory rapidly converges to the target state, while the NLC exhibits violent oscillation around the target. Contribution. The principal contributions of this paper can be summarized as follows: We propose Neural ETC, a framework for learning event-triggered controllers ensuring both stability guarantee and optimal scheduling, where the exponential stability comes from the devised event function. Specifically, we firstly propose a path integral approach to realize the implementation of the machine learning framework based on the root solver and neural event ODE solver that calculate the trainable event triggering times. Secondly, we theoretically address the estimation of the minimal inter-event time of the event triggered controlled system, which leading to the Monte Carlo approach of our framework that circumvents the expensive computation cost of back-propagation through ODE solvers. The two approaches trade off in terms of stabilization performance and training efficiency, which is convenient for users to flexibly choose the specific approach according to the task in hand. To theoretically guarantee the stability of the controlled system under our Neural ETC, we propose the projection operation that rigorously endows our Neural ETC with stability and schedule optimality. Instead of solving an optimization problem to obtain the projection, we provide analytical expression for our projection operation, leading to a fast implementation for our framework. Finally, we evaluate Neural ETCs on a variety of representative physical and engineering systems. Compared to existing stabilizing controllers, we find that Neural ETCs exhibit significant superiority in decreasing the triggering times and maximizing the minimal inter-event time. The code for reproducing all the numerical experiments is released at https://github.com/jingddong-zhang/Neural Event-triggered-Control (hyperlink of Neural ETC). 2. Background Notations. Denote by the L2-norm for any given vector in Rd. Denote by C(D) the maximum norm on continuous function space C(D). For A = (aij), a matrix of dimension d r, denote by A 2 F = Pd i=1 Pr j=1 a2 ij the Frobenius norm. Denote max(a, 0) by (a)+. Denote x y as the inner product of two vectors. 2.1. Neural Lyapunov Control To begin with, we consider the feedback controlled dynamical system of the following general form: x = f(x, u(x)) fu(x), x Rd, u Rm, (1) where fu(x) : D Rd is the Lipschitz-continuous vector field acting on some prescribed open set D Rd. The solution initiated at time t0 from x0 under controller u is denoted by xu(t; t0, x0). For brevity, we let the unstable equilibrium x D be origin, i.e., f(0, 0) = 0. One major problem in cybernetics field is to design stabilizing controller u(x) (Wiener, 2019) such that lim t xu(t; t0, x0) = 0, for any initial value x0 D. Theorem 2.1. (Mao, 2007) Suppose there exists a continuously differentiable function V : D R that satisfies the following conditions: (i) V (0) = 0, (ii) V (x) c x p for some constants c, p > 0, (iii) and Lfu V < δV , for some δ > 0. 1 Then, the system is exponentially stable at the origin, that is, lim supt 1 t log xu(t; t0, x0) δ p. Here V is called a Lyapunov function. Previous works parameterize the controller and the Lyapunov function as uϕ, Vθ, and integrate the sufficient con- 1Lfu V represent the Lie derivative of V along the direction fu, i.e., Lfu V = V fu. Neural Event-Triggered Control with Optimal Scheduling ditions (i)-(iii) in Theorem 2.1 for Lyapunov stability into the loss function (Chang et al., 2019; Zhang et al., 2022a; Dawson et al., 2023). The learned Lyapunov Vθ plays a role of stability certificate function. Remark 2.2. Unlike model-free reinforcement learning (RL) approaches that search for an online control policy guided by a reward function along the trajectories of the dynamical systems. (Kaelbling et al., 1996), the neural Lyapunov control searches for an offline policy and a certificate function V that proves the soundness of the learned policy (Dawson et al., 2022). Nevertheless, updating the feedback policy continuously in the implementation process incurs prohibitive high communication cost. 2.2. Event-triggered Control Although the feedback controller works well in numerical simulations, updating and implementing the controller continuously is impractical in most real-world digital platforms under communication constraints ( Astr om & Bernhardsson, 1999). To conquer this weakness, event-triggered stabilizing control is introduced as follows (Heemels et al., 2012), Definition 2.3. (Event-triggered Control) Consider the controlled system (1), the event-triggered controller is defined as u(t) = u(x(tk)), tk t < tk+1, where the triggering time is decided by tk+1 = inf{t > tk : h(x(t)) = 0} for some predefined event function h. The largest lower bound τ of {tk+1 tk} is called as minimal inter-event time. For example, if there exists a Lyapunov function V for the feedback controlled system (1), then the event function is set to guarantee the Lyapunov condition on each event triggering time interval, i.e., V f(x(t), u(xtk)) < 0, t [tk, tk+1). Problem Statement. We assume that the zero solution of the uncontrolled system in Eq. (1) is unstable, i.e. limt xu=0(t; t0, x0) = 0. We aim at stabilizing the zero solution using event-triggered control based on neural networks (NNs) with optimal scheduling, i.e., the least triggering times, which is urgently required by the digital platforms wherein the communication resources of updating the control value are limited. Notice that in an average sense, the triggering times are inversely proportional to the inter-event time, our goal is equivalently to leverage the NNs to design an appropriate controller u with u(0) = 0 such that the controlled system under event-triggered implementation is steered to the zero solution with the maximal inter-event time. We summarize the problem formulation as the following optimization problem, where the triggering time {tk : tk T} depends on the controller u and the triggering mechanism, and T is the prefixed time limit according to the specific tasks. We aim at devising controller u and triggering mechanism to solve this problem based on the known model f and time limit T. 1 min{tk T }(tk+1 tk) + λ1 u(x) C(D) s.t. x(t) = f(x(t), u(x(tk)), t [tk, tk+1), x(0) = x0 D, lim t T x(t) = 0, The major difficulty of this problem comes from that the implicitly defined triggering times are not equidistant, and are only known when the events are triggered (Miskowicz, 2018). The majority of existing works focus on the stabilization performance of event-triggered control and often omit the communication cost of updating the control value at triggering moments. In what follows, we propose neural event-triggered control (Neural ETC) framework to address both the stabilization and the communication cost issues of event-triggered control. Closed-loop controlled dynamics. The dynamics under event-triggered control is generally an open-loop system with controller varying from different triggering time intervals. In order to simplify the theoretical analysis and to utilize the existing numerical tools for ODE solvers, we transform the event-triggered controlled system to the closed-loop version via augmenting the dynamics with an error state e(t) = x(tk) x(t), t [tk, tk+1) and an update operation e(tk+1) = 0. Then we obtain the closed-loop controlled dynamics as x = f(x, u(x + e)), e = f(x, u(x + e)), t [tk, tk+1). In the next sections, we construct the event function with exponential stability guarantee and deduce the theoretical estimation of minimal inter-event time based on the augmented dynamics of (x, e). Event function for exponential stability. We consider the exponential Lyapunov stability for controlled system (1) such that the corresponding Lyapunov function defined in Theorem 2.1 satisfies the stability condition Lfu V δV and V (x) α( x ), where α is a class-K function2. For brevity, we fix δ = 1 in this paper such that the decay exponent of the Lyapunov function is 1. Since the event-triggered controller is a discrete time realization of the original feedback controller u, the corresponding exponential decay rate of the Lyapunov function is less than 1. Therefore, we design the event function h = h(x, e) as h = V (x) (f(x, u(x + e)) f(x, u(x))) σV (x) (3) 2A continuous function α : (0, ) (0, ) is said to belong to class-K if it is strictly increasing and α(0) = 0. Neural Event-Triggered Control with Optimal Scheduling untriggered trigger Communication Cost Neural ETC(ours) 𝑉 𝑓𝑥, 𝑢𝑥+ 𝑒 𝑓𝑥, 𝑢𝑥 𝜎𝑉(𝒙) < = Figure 2. Illustration of the Neural ETC with optimal scheduling. with 0 < σ < 1, such that the event-triggered controlled system satisfies V (x) f(x, u(x + e)) V (x) f(x, u(x)) + σV (x) (1 σ)V (x). (4) Hence, the exponential stability of the event-triggered controlled system is assured with exponential decay rate 1 σ. As illustrated in Fig. 2, the NLC is used as an example method to be compared with the Neural ETC. The control value is updated when an event is triggered, i.e., the event function h equals to zero. Our method achieves the exponential stability and has the least triggered events, leading to the lowest communication cost of updating the control value. 3.1. Path Integral Approach Parameterization. In order to design the feedback controller such that its event-triggered implementation stabilize the unstable equilibrium efficiently and has the largest minimal inter-event time, we consider the following parameterized optimization problem. min {tk T } 1 tk+1 tk + λ1 uϕ(x) C(D) s.t. x = f(x, uϕ(x + e)), t [tk, tk+1), e = f(x, uϕ(x + e)), t [tk, tk+1), x(0) = x0, e(tk) = 0, Vθ(0) = 0, uϕ(0) = 0, tk+1 = inf t>tk{t : hθ,ϕ(x(t), e(t)) = 0} α( x ) Vθ(x) 0, Lfuϕ Vθ(x) + Vθ(x) 0. Here, λ1 is a predefined weight factor, T is the temporal length of the controlled trajectory, α is a class-K function, and hθ,ϕ(e, x) = Lfuϕ Vθ (f(x, uϕ(x + e)) f(x, uϕ(x))) σVθ(x(t)) is the parameterized event function. To ensure the neural functions Vθ, uϕ satisfy some constraints naturally, we adopt the parametrization in (Zhang et al., 2022a) as follows, Vθ = ICNNθ(x) ICNNθ(0) + ε x 2, uϕ = diag(x)NNϕ(x) or NNϕ(x) NNϕ(0), (5) where diag(x) transforms a vector to a diagonal matrix with (diag(x))ij = δijxi, ICNNθ and NNϕ represent the input convex neural network and the feedforward neural networks, respectively, the detailed formulation is provided in Appendix A.3.1. We minimize the continuous function norm uϕ C(D) by regularizing the Lipschitz constant of the neural network, we apply the spectral norm regularization method in (Yoshida & Miyato, 2017) to minimize the spectral norm of the weight matrices {Wϕ,i}l i=1 in uϕ with the regularization term Llip = Pl i=1 σ(Wϕ,i)2. To solve the substantially non-convex optimization problem, we relax the original hard constraint Lfuϕ Vθ(x) + Vθ(x) 0 to a soft constraint in the loss function as N PN i=1 Lfuϕ Vθ(xi) + Vθ(xi) + . Calculate gradients of tk. To proceed, we handle the objective function related to the triggering times. Instead of directly training the parameters ϕ, θ based on the direct samples of Vθ, uϕ and f(x, uϕ(x)) as done in neural certificate-based controllers, we have to numerically solve the controlled ODEs to identify the triggering times. To proceed, we need to calculate the gradients of tk for optimizing 1 tk+1 tk term in loss function during gradient-based optimization. We employ the neural event ODE method as: tk+1, x(tk+1) = ODESolve Event(x(tk), f, uϕ, tk), where ODESolve Event is proposed by (Chen et al., 2020), which introduces root solver and adjoint method (Pontryagin, 2018) to the numerical solver and deduce the gradient tk ϕ from the implicit function theorem (Krantz & Parks, 2002). Reduce computation complexity. We denote by tk(x) the kth triggering time from initial value t0 = 0, x(0) = x. The computation cost of ODESolve Event is O(M KLd2), Neural Event-Triggered Control with Optimal Scheduling where M is the batch size of the initial value {xi(0)}M i=1, K = 1 M PM i=1 K(xi(0)), K(xi(0)) = #{tk(xi(0)) : tk T} is the number of triggering times before T, and L is the iteration times in the root solver. In this case, the computation cost pivots on the sampled batch and its variance is hard to decrease. In addition, the numerical error in ODE solver accumulates over the triggering time sequence {tk}. To mitigate these issues, according to the time invariance property of ODEs, i.e., tk+1(x(0)) tk(x(0)) = t1(x(tk)), we recast the problem of solving M batch trajectories {xi(tk), tk T|xi(0) q0(x)}M i=1 of controlled ODE as solving MK trajectories {xi(t1), t1 T|xi(0) q0(x)}MK k=1 up to t1. Here, K represent the expectation of K. In practice, we directly treat MK together as a single hyperparameter M. Then the triggering times contribute into the loss function as Levent = 1 M PM i=1 1 t1(xi(0)). Finally, we train the overall parameterized model with the total loss function as follows, L(ϕ, θ) = Lstab + λ1Llip + λ2Levent (6) The whole training procedure is summarized in Algorithm 1. Remark 3.1. A more reasonable augmented distribution should takes the form as q0(x) = 1 K 1 PK 1 k=0 qk(x), where xtk qk(x) is deduced from the initial distribution q0 and the ODE integration from 0 to tk. Since tk varies for different initial value and cannot be determined in advance, we fix q0 = q0 for simplicity. 3.2. Monte Carlo Approach Although the proposed algorithm works efficiently in low dimensional ODEs, the high computation cost and accumulate error caused by the ODE solver affect its performance in higher dimensional tasks. To circumvent this drawback, we propose a Monte Carlo approach for training the Neural ETC. Inspired by the event-triggered scheduling theory in (Tabuada, 2007), we provide the following estimation of minimal inter-event time. Theorem 3.2. Consider the event-triggered controlled dynamics in Eq. (1), if the following assumptions are satisfied: (i) f(x , u ) f(x, u) lf ( x x + u u ); (ii) u(x ) u(x) lu x x ; (iii) Lfu V (x, u(x + e)) α( x ) + γ( e ) for some class-K functions α, γ with α 1(γ( e )) P e . Then, the minimal inter-event time implicitly defined by event function h = α( x ) γ( e ) is lower bounded by τh = 1 lf log P +1 P + lu 1+lu . The detailed proof is provided in Appendix A.1.2. According to the theorem, the lower bound of minimal inter-event time increases as Lipschitz constants of α 1 γ and u decrease. Therefore, we can maximize the minimal inter-event time by regularizing these Lipschitz constants. Nonetheless, directly integrating the conditions and results of The- orem 3.2 into the training process is unrealistic and cumbersome, because the error state e in condition (iii) should depend on x and we cannot determine the sampling distribution of e before training. To solve this problem, we split the inequality in condition (iii) into a sufficient inequality group as V (f(x, u(x + e)) f(x, u(x))) γ( e ) (7) Lfu V (x) α( x ) (8) Lfu V (x, u(x + e)) α( x ) + γ( e ) The dependence on x of right term γ in Eq. (7) can be omitted when the state space D is bounded, which occurs in most real-world scenarios. The Eq. (7) implies that the Lipschitz constant of γ is related to the Lipschitz constant of u. Furthermore, we notice that if we replace the event function in Theorem 3.2 by the following event function with stability guarantee, h = V (f(x, u(x + e)) f(x, u(x))) α( x ), (9) then the inter-event time of these two event functions hold the relation tk+1,h tk,h tk+1, h tk, h due to Eq. (7). Therefore, the inter-event time of event function h is also lower bounded by τh in Theorem 3.2. We summarize the results in the following theorem. Theorem 3.3. For the event-triggered controlled dynamics in Eq. (1) with event function h defined in Eq. (9), if the state space D is bounded, the Eqs. (7),(8) and the conditions (i), (ii) in Theorem 3.2 hold, then the minimal inter-event time is lower bounded by τ h = 1 lf log clα 1lu+1 clα 1lu+ lu 1+lu , here lα 1 is the Lipschitz constant of α 1, c is a constant depending on V, f, D. The proof is provided in Appendix A.1.3. With this theorem, we come to a Monte Carlo approach for training the Neural ETC framework by directly learning the parameterized functions Vθ, αθα, and control function uϕ simultaneously, as well as regularizing the Lipschitz constants of uϕ and α 1 θα . For constructing neural class-K functions, we adopt the monotonic NNs to construct the candidate class-K function as αθα(x) = Z x 0 qθα(s)ds, (10) where qθα( ) 0 is the output of the NNs (Wehenkel & Louppe, 2019). We regularize the inverse of integrand to minimize the Lipschitz constant of α 1 θα . We apply the spectral norm regularization Llip defined above to minimize the Lipschitz constant of controller uϕ. Finally, we train Neural Event-Triggered Control with Optimal Scheduling the overall model with the loss function as follows, Lfuϕ Vθ(xi) + αθα(xi) + , Lα 1 = 1 Mα L(ϕ, θ, θα) = Lstab + λ1Llip + λ2Lα 1. The specific training procedure of this algorithm, dubbed Neural ETC-MC, is shown in Algorithm 2. Remark 3.4. To obtain a stronger exponential decay rate of V , we multiply the term α( x ) in Eq. (9) by σ (0, 1) in realization. Similarly to Eq. (4), the controlled vector under event h satisfied V (f(x, u(x + e)) (1 σ)α( x ). (12) Then the exponential decay rate of V is 1 σ. The lower bound of inter-event time can be obtained by replacing lα 1 with σ 1lα 1 in Theorem 3.3. 4. Theoretical Guarantee for Stability and Optimality In this section, we provide several theoretical results for rigorously guaranteeing the stability and optimality of our neural controllers. Firstly, we note that the NNs trained on finite samples cannot guarantee the Lyapunov stability condition in the loss function is satisfied in the whole state space with infinite data points. To circumvent this weakness, we introduce the projection operation in the following theorem. Theorem 4.1. (Stability guarantee) For a candidate controller u and the stable controller space U(V ) = {u : Lfu V + V 0}, we define the projection operator as, π(u, U(V )) u max(0, Lfu V V ) If the controller has affine actuator, then we have π(u, U(V )) U(V ), the projected controller is Lipschitz continuous over the state space D if and only if D is bounded. Furthermore, under the triggering mechanism V (x) [f(x, u(x + e)) f(x, u(x))] σV (x) = 0, σ (0, 1), e = x(tk) x(t), t [tk, tk+1) the controlled system under π(u, U) is assured exponential stable with decay rate 1 σ, and the inter-event time has positive lower bound. We provide the proof in Appendix A.1.4. By applying the projection operation to the learned controller uϕ and potential function Vθ, we obtain the theoretical stability guarantee for our approach. Based on the Theorem 4.1 and Theorem 3.2, we could provide necessary condition for the optimal event-triggered control with the largest minimal inter-event time by utilizing the lower bound of the interevent time and the projection operation. Theorem 4.2. (Optimality guarantee) Denote the Lipschitz constant of the controller u on state space as lu, then the optimal control with the largest minimal inter-event time satisfies, u arg min U(V ) lu. (13) Furthermore, for any candidate controllers u, the optimal condition can be simplified as π(u, U(V )) arg min lπ(u,U(V )). (14) This theorem is a direct result from the Theorem 4.1 and Theorem 3.2, and the projection operation simplifies the constrained necessary condition in Eq. (13) to the unconstrained condition Eq. (14). We can easily provide optimality guarantee for the neural network controller uϕ and the Lyapunov function Vθ by regularizing the Lipschitz constant of π(uϕ, U(Vθ)) . 5. Experiments and Analysis In this section, we demonstrate the superiority of the Neural ETCs over existing methods using series of experiments from low dimensional tasks to high dimensional tasks, then we unravel the key factors of Neural ETCs. More details of the experiments can be found in Appendix A.3. 5.1. Benchmark Experiments Benchmark dynamical systems. (1) Gene Regulatory Network (GRN) plays a central role in describing the gene expression levels of m RNA and proteins in cell (Davidson & Levin, 2005), here we consider a two-node GRN, x1 = a1 xn 1 sn+xn 1 + b1 sn sn+xn 2 kx1, x2 = a2 xn 2 sn+xn 2 + b2 sn sn+xn 1 kx2, where the tunable parameters a1, a2, b1 and b2 represent the strengths of auto or mutual regulations. We aim at stabilizing the system from one attractor to another attractor via only tuning a1 in time interval [0, 20]. (2) Lorenz system is a fundamental model in atmospheric science (Lorenz, 1963): x = σ(y x), y = ρx y xz, z = xy βz. For this chaotic system, we stabilize its unstable zero solution by an fully actuated controller u = (ux, uy, uz) in time interval [0, 2]. (3) Michaelis Menten model for subcellular dynamics (Cell) captures the collective behavior of the coupled cells (San- Neural Event-Triggered Control with Optimal Scheduling Table 1. Comparison studies of benchmark models and dynamical systems. Best results bolded. Averaged over 5 runs. The dimension of tasks are: GRN (2-D), Lorenz (3-D), Cell (100-D). Method Number of triggers Minimal inter-event time MSE under finite triggers GRN Lorenz Cell GRN Lorenz Cell GRN Lorenz Cell BALSA (Fan et al., 2020) 12( 4) 273( 24) 18( 4) 0.29( 0.07) 6e-4( 6e-4) 3e-3( 1e-3) 0.05( 0.07) 7.20( 2.25) 29.75( 12.72) LQR (Heemels et al., 2012) 1816( 14) 2000( 0) 449( 1) 6e-3( 3e-3) 2e-5( 1e-6) 0.02( 0.02) 2.19( 0.54) 53.02( 7.03) 2e-3( 2e-3) Quad-NLC (Jin et al., 2020) 1914( 107) 433( 379) 551( 220) 6e-6( 1e-6) 3e-5( 4e-5) 4e-6( 10e-7) 2.29( 0.54) 7.00( 1.28) 62.50( 18.44) NLC (Chang et al., 2019) 23( 2) 1602( 795) 15( 13) 5e-8( 8e-9) 4e-8( 1e-7) 6e-6( 1e-5) 0.20( 0.05) 40.56( 21.14) 27.76( 5.52) IRL ETC (Xue et al., 2022) 131( 28) 2000( 0) 370( 14) 8e-3( 6e-4) 0.00( 0.00) 3e-3( 8e-5) 4.94( 0.77) 9.76( 2.13) 38.12( 0.11) Cirtic-Actor ETC (Cheng et al., 2023) 605( 293) 69( 8) 330( 5) 5e-4( 3e-5) 2e-3( 8e-4) 1e-3( 8e-5) 4.09( 0.81) 7.22( 2.05) 38.13( 0.04) ETS (Li & Liu, 2019) 81( 87) 1120( 5.62) 11( 1) 0.05( 2e-16) 9e-4( 5e-4) 0.02( 1e-3) 0.19( 2.27) 2.42( 1.02) 4.96( 7.53) PGDNLC (Yang et al., 2024) 23( 2) 340( 246) 60( 15) 0.37( 0.15) 7e-4( 8e-4) 6e-5( 6e-5) 0.39( 8e-3) 11.35( 7.69) 41.09( 3.36) Neural ETC-PI (ours) 20( 3) 20( 4) 11( 0.00) 0.26( 0.17) 0.02( 0.02) 0.95( 0.03) 0.05( 4e-3) 0.11( 0.12) 5e-8( 1e-9) Neural ETC-MC (ours) 4( 4) 13( 1) 2( 0.00) 15.52( 8.54) 0.01( 8e-3) 27.18( 0.60) 0.07( 0.03) 0.29( 0.31) 1.66( 0.12) hedrai et al., 2022): xi = Bxi + Pn i=1 Aij x2 j 1+x2 j . This model has two important equilibrium phase, inactive phase indicating the malignant state and active state indicating the benign state. We consider n = 100 and regulate the high dimensional model from the inactive phase to the active phase by only tuning the topology structure {Aii}n i=1 in time interval [0, 30]. All these systems have application scenarios and urgently call for event-triggered control with minimal communication burden, we summarize the motivation for selecting the them in Appendix A.3.5. Benchmark methods. We benchmark against the extensively used neural Lyapunov control (NLC) method (Chang et al., 2019), an improvement version of neural Lyapunov control via constructing quadratic Lyapunov function proposed in (Jin et al., 2020), dubbed as Quad-NLC here, a integral reinforcement learning (IRL) based ETC (Xue et al., 2022), a critic-actor neural network based ETC method (Cheng et al., 2023), and two latest SOTA methods ETS (Li & Liu, 2019) and PGDNLC (Yang et al., 2024). We also compare with the classic linear quadratic regulator (LQR) method, BALSA (Fan et al., 2020), an online control policy based on the quadratic programming (QP) solver, and our Neural ETC variants: Neural ETC-PI and Neural ETC-MC. We implement all the control methods with the similar kinds of event functions proposed in Eqs. (3),(9). For a fair comparison, we set the number of hidden units per layer such that all learning models have nearly the same number of total parameters. We provide further details of model selection, hyperparameter selection and experimental configuration in Appendix A.3. Results. Table 1 summarizes the control performance results in terms of the triggering times in the same temporal length, the minimal inter-event time and the mean square error (MSE) between the target state and the controlled trajectories with no larger than 10 triggering events, representing the control performance in limited communication resources. We see that our Neural ETC variants achieve superior performance compared to the other online and offline For the communication cost, our Neural ETCs need the least number of triggers in the same time interval while have the largest minimal inter-event time compared to other methods, leading to the most optimal scheduling in actual implementation. In addition, the MSE results illustrate our Neural ETCs have the ability in stabilizing the systems at various scales with limited communication resources. We also find the Neural ETC-PI and Neural ETC-MC form the trade off in scheduling and the stabilization performance, we further compare them in the next section. The results underpin the practicability of the Neural ETCs. Take GRN model for an example, the auto regulation strength a1 can be adjusted externally through the application of repressive or inductive drugs in a typical experimental setting (Wang et al., 2016). In reality, the drugs can only be administered a few times and it takes time for the drug to take effect, requiring the controller should only be updated at several times with large interval. Therefore, while all the benchmark methods successfully regulate the GRN to the target gene expression level in simulation, only the Neural ETC-MC is acceptable. Combining online and offline policy. In the context of event-triggered control, Table 1 demonstrates that the online control method outperforms other offline policies. However, the online policy s computational cost is high due to solving the quadratic programming (QP) problem at each realization time. In contrast, our Neural ETCs achieve superior performance compared to online methods while maintaining the same computation cost as the offline policy during the control process. The event-triggered control employs an event function that continuously assesses whether an event is triggered, effectively acting as an online solver to determine real-time control values. Consequently, we can view event-triggered control as an online realization of the offline policy, inheriting the advantages of both online and offline approaches Neural Event-Triggered Control with Optimal Scheduling 5.2. Comparison Between Neural ETCs We further evaluate the strengths and weaknesses of the Neural ETC-PI and Neural ETC-MC. As shown in Table 2, the Neural ETC-MC is more efficient in training process, especially in the high dimensional tasks. Nevertheless, the temporal variance of the controlled trajectories of Neural ETC-PI is far below that of Neural ETC-MC, implying Neural ETC-PI is more robust in the control process. These two algorithms thus are complementary in applications. In addition, the training time of Neural ETC-PI in 2-D GRN and 3-D Lorenz has significant difference, the reason is that the minimal inter-event time of the former is larger than the latter (see Table 1), requiring more time to solve t1. Table 2. Comparison of Neural ETCs (denoted by NETC) in terms of training time and variance of stabilized trajectories. Model Training time Temporal variance NETC-PI NETC-MC NETC-PI NETC-MC GRN 1230 32 5e-4 7e-3 Lorenz 503 29 4e-3 0.09 FHN 4634 62 3e-15 2.78 5.3. Ablation Study 0.1 0.9 σ 10 Triggering Times Neural ETC-PI Neural ETC-MC 0.1 0.9 σ 0 Inter-event Time Figure 3. The solid lines are obtained through averaging the 5 sampled trajectories, while the shaded areas stand for the variance regions. The parameter σ corresponds to the exponential decay rate of Lyapunov function along controlled trajectory in Eqs. (3),(12). We investigate the influence of σ in applying Neural ETC variants to Lorenz dynamic. The results in Fig. 3 suggests the best choice is σ = 0.8. Then we investigate the influence of weight factor λ2 of event loss in Eqs. (6),(11) and summarize the results in Table 3. We find the small λ2 leads to poor triggering scheduling because the event loss does not play a leading role in training, the large λ2 will break the stabilization performance because the optimization function of event loss is not guaranteed to satisfy the stabilization loss. This phenomenon inspires us to extend the framework to the setting where the parame- terized controllers are already stabilization controllers in the future work. For reference, in Table 1 Neural ETC-PI is using σ = 0.5, λ2 = 0.05 and Neural ETC-MC is using σ = 0.5, λ2 = 0.1. Table 3. Performance under various event loss weight λ2. Method Neural ETC-PI Neural ETC-MC λ2 0.005 0.05 0.5 0.01 0.1 1.0 Triggering times 114 29 34 37 10 10 Min Inter-event time 0.010 0.008 0.025 0.02 0.07 0.06 MSE [1.8,2] 8e-8 7e-4 0.32 3.92 0.25 0.53 5.4. Essential Factor of Neural ETC We investigate the essential factor in the Neural ETC framework that determine the optimization of scheduling. We plot the convexity of V function (Tr( 2V )) and the strength of the variation of controller ( u ) in the training process, and compare their evolution with triggering times of the corresponding trained controller. Fig. 4 shows that the u plays a leading role in minimizing the triggering times as it has strong negative correlation to the triggering times while the convexity of V function does not. We also observe an early convergence phenomenon of the triggering times and u simultaneously in Neural ETC-PI from Fig. 4(b),(c). 0 10 Epochs 0 Convexity of V Neural ETC-PI Neural ETC-MC 0 10 Epochs 10 Triggering Times 0 10 Epochs Figure 4. (a) Convexity of V is calculated as the trace of the 2V on 1000 points in [ 2.5, 2.5]3. (b) Triggering times and (c) norm of u in the training process. 6. Related Work Neural control with certificate functions. Previous works in neural control establish the performance guarantee via using the certificate functions, including Lyapunov function for stability (Giesl & Hafstein, 2015; Chang et al., 2019), barrier function for safety (Zhang et al., 2022b; Ames et al., 2016; Taylor et al., 2020; 2019; Taylor & Ames, 2020; Peruffo et al., 2021), and contraction metrics for stability in trajectory tracking (Singh et al., 2021; Tsukamoto et al., 2021). However, all these feedback controllers require impractically high communication cost for updating the controller continuously when deployed on the digital platforms. We solve this challenge in limited communication resources and improve the performance guarantee at the same time. Neural Event-Triggered Control with Optimal Scheduling Event-triggered control. The pioneering works ( Astr om & Bernhardsson, 1999; Aarz en, 1999) highlighted the advantages of event-based control against the periodic implementation in reducing the communication cost. Since then, (Tabuada, 2007) investigates the sufficient conditions for avoiding the Zeno behavior in event-triggered implementations of stabilizing feedback control laws, (Henningsson et al., 2008) extends the event-triggered control to the linear stochastic system and (Heemels et al., 2008) gives the system theory of event-triggered control scheme for perturbed linear systems. Machine learning methods have also been introduced to the ETC settings, (Xue et al., 2022; Cheng et al., 2023) employ the critic-actor RL structure to solve the dynamic Hamilton-Jacobi-Bellman equation under the ETC, (Funk et al., 2021) cultivates a model-free hierarchical RL method to optimize both the control and communication policies for discrete dynamics, and (Baumann et al., 2018) applies deep RL to ETC in the nonlinear systems. All the previous works focus on the stabilization analysis of the controlled systems, the existence of the minimal inter-event time (and hence avoids the Zeno behavior), and directly introducing machine learning methods to ETC. To our knowledge, we are the first to study the optimization scheduling problem of ETC in the continuous dynamics. 7. Scope and Limitations ODE solver. The use of the fixed step ODE solvers in finding the triggering times in the training process is less optimal than the adaptive ODE solver. One can still improve the performance of the framework by applying the adaptive solvers with higher accuracy tolerance with a stronger computing platform. However, in practice the performance of the Neural ETC did not decrease substantially when using adaptive solvers. In addition, the employ of ODE solvers in the Neural ETC-PI may not always work, especially for systems described by stiff equations, stiff-based ODE solvers can be introduced to mitigate this issue (Kim et al., 2021). Neural ETC for SDEs. Although the current Neural ETC framework works efficiently in ODEs, many real-world scenarios affected by the noise are described by stochastic differential equations (SDEs). The major challenge for establishing the Neural ETC framework for SDEs ensues from the stochasticity of the triggering time. Specifically, the triggering time in SDEs, t1 = inft 0{t : h(x(t)) = 0} initiated from any fixed x(0) with h(x(0)) < 0, is a stopping time. Therefore, t1 is a random variable and can take different values in different sample paths. In this case, none of the existing methods can find t1 for SDEs as a counterpart of ODESolve Event for ODEs. 8. Conclusion This work focuses on a new connection of machine learning and control field in the context of learning event-triggered stabilization control with optimal scheduling. In contrast to the existing learning control methods, the learned eventtriggered control, named Neural ETC, only updates the control value in very few times when an event is triggered. As a consequence, our Neural ETC can be deployed on the actual platform where the communication cost for updating the control value is limited (e.g. tuning the protein regulation strength in cell via drugs). The superiority of the Neural ETC over the existing methods is demonstrated through a series of representative dynamical systems. Acknowledgements L. Yang is supported by the China Scholarship Council (No. 202406100251). Q. Zhu is supported by the China Postdoctoral Science Foundation (Grant No. 2022M720817), by the Shanghai Postdoctoral Excellence Program (Grant No. 2021091), and by the STCSM (Grants No. 21511100200, No. 22ZR1407300, and No. 23YF1402500). W. Lin is supported NSFC (Grant No. 11925103), the IPSMEC (Grant No. 2023ZKZD04), and the STCSM (Grants No. 22JC1401402, No. 22JC1402500, and No. 2021SHZDZX0103). Neural Event-Triggered Control with Optimal Scheduling Impact Statement This paper presents work whose goal is to advance the field of Machine Learning. There are many potential societal consequences of our work, none which we feel must be specifically highlighted here. Aarz en, K.-E. A simple event-based pid controller. IFAC Proceedings Volumes, 32(2):8687 8692, 1999. Abdelrahim, M., Postoyan, R., Daafouz, J., and Neˇsi c, D. Stabilization of nonlinear systems using event-triggered output feedback controllers. IEEE transactions on automatic control, 61(9):2682 2687, 2015. Ames, A. D., Xu, X., Grizzle, J. W., and Tabuada, P. Control barrier function based quadratic programs for safety critical systems. IEEE Transactions on Automatic Control, 62(8):3861 3876, 2016. Amos, B., Xu, L., and Kolter, J. Z. Input convex neural networks. In International Conference on Machine Learning, pp. 146 155. PMLR, 2017. Astr om, K. J. and Bernhardsson, B. Comparison of periodic and event based sampling for first-order stochastic systems. IFAC Proceedings Volumes, 32(2):5006 5011, 1999. Baumann, D., Zhu, J.-J., Martius, G., and Trimpe, S. Deep reinforcement learning for event-triggered control. In 2018 IEEE Conference on Decision and Control (CDC), pp. 943 950. IEEE, 2018. Boccaletti, S., Grebogi, C., Lai, Y.-C., Mancini, H., and Maza, D. The control of chaos: theory and applications. Physics reports, 329(3):103 197, 2000. Chang, Y.-C., Roohi, N., and Gao, S. Neural lyapunov control. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, pp. 3245 3254, 2019. Chen, R. T., Amos, B., and Nickel, M. Learning neural event functions for ordinary differential equations. In International Conference on Learning Representations, 2020. Cheng, S., Li, H., Guo, Y., Pan, T., and Fan, Y. Eventtriggered optimal nonlinear systems control based on state observer and neural network. Journal of Systems Science and Complexity, 36(1):222 238, 2023. Davidson, E. and Levin, M. Gene regulatory networks. Proceedings of the National Academy of Sciences, 102 (14):4935 4935, 2005. Dawson, C., Qin, Z., Gao, S., and Fan, C. Safe nonlinear control using robust neural lyapunov-barrier functions. In Conference on Robot Learning, pp. 1724 1735. PMLR, 2022. Dawson, C., Gao, S., and Fan, C. Safe control with learned certificates: A survey of neural lyapunov, barrier, and contraction methods for robotics and control. IEEE Transactions on Robotics, 2023. Fan, D. D., Nguyen, J., Thakker, R., Alatur, N., Aghamohammadi, A.-a., and Theodorou, E. A. Bayesian learning-based adaptive control for safety critical systems. In 2020 IEEE international conference on robotics and automation (ICRA), pp. 4093 4099. IEEE, 2020. Franklin, G. F., Powell, J. D., Emami-Naeini, A., and Powell, J. D. Feedback control of dynamic systems, volume 4. Prentice hall Upper Saddle River, 2002. Funk, N., Baumann, D., Berenz, V., and Trimpe, S. Learning event-triggered control from data through joint optimization. IFAC Journal of Systems and Control, 16:100144, 2021. Giesl, P. and Hafstein, S. Review on computational methods for lyapunov functions. Discrete and Continuous Dynamical Systems-B, 20(8):2291 2331, 2015. Heemels, W., Sandee, J., and Van Den Bosch, P. Analysis of event-driven controllers for linear systems. International journal of control, 81(4):571 590, 2008. Heemels, W. P., Johansson, K. H., and Tabuada, P. An introduction to event-triggered and self-triggered control. In 2012 ieee 51st ieee conference on decision and control (cdc), pp. 3270 3285. IEEE, 2012. Henningsson, T., Johannesson, E., and Cervin, A. Sporadic event-based control of first-order linear stochastic systems. Automatica, 44(11):2890 2895, 2008. Jin, W., Wang, Z., Yang, Z., and Mou, S. Neural certificates for safe control policies. ar Xiv preprint ar Xiv:2006.08465, 2020. Kaelbling, L. P., Littman, M. L., and Moore, A. W. Reinforcement learning: A survey. Journal of Artificial Intelligence Research, 4:237 285, 1996. Khalil, H. K. Nonlinear systems third edition. Patience Hall, 115, 2002. Kim, S., Ji, W., Deng, S., Ma, Y., and Rackauckas, C. Stiff neural ordinary differential equations. Chaos: An Interdisciplinary Journal of Nonlinear Science, 31(9), 2021. Neural Event-Triggered Control with Optimal Scheduling Krantz, S. G. and Parks, H. R. The implicit function theorem: history, theory, and applications. Springer Science & Business Media, 2002. Lemmon, M. Event-triggered feedback in control, estimation, and optimization. Networked control systems, pp. 293 358, 2010. Li, F. and Liu, Y. Event-triggered stabilization for continuous-time stochastic systems. IEEE Transactions on Automatic Control, 65(10):4031 4046, 2019. Lorenz, E. N. Deterministic nonperiodic flow. Journal of atmospheric sciences, 20(2):130 141, 1963. Mao, X. Stochastic differential equations and applications. Elsevier, 2007. Miskowicz, M. Event-based control and signal processing. CRC press, 2018. Ott, E., Grebogi, C., and Yorke, J. A. Controlling chaos. Physical review letters, 64(11):1196, 1990. Parrilo, P. A. Structured Semidefinite Programs and Semialgebraic Geometry Methods in Robustness and Optimization. California Institute of Technology, 2000. Peruffo, A., Ahmed, D., and Abate, A. Automated and formal synthesis of neural barrier certificates for dynamical models. In International conference on tools and algorithms for the construction and analysis of systems, pp. 370 388. Springer, 2021. Pontryagin, L. S. Mathematical theory of optimal processes. Routledge, 2018. Qin, Z., Zhang, K., Chen, Y., Chen, J., and Fan, C. Learning safe multi-agent control with decentralized neural barrier certificates. In International Conference on Learning Representations, 2020. Sanhedrai, H., Gao, J., Bashan, A., Schwartz, M., Havlin, S., and Barzel, B. Reviving a failed network through microscopic interventions. Nature Physics, 18(3):338 349, 2022. Singh, S., Richards, S. M., Sindhwani, V., Slotine, J.-J. E., and Pavone, M. Learning stabilizable nonlinear dynamics with contraction-based regularization. The International Journal of Robotics Research, 40(10-11):1123 1150, 2021. Sun, D., Jha, S., and Fan, C. Learning certified control using contraction metric. In Conference on Robot Learning, pp. 1519 1539. PMLR, 2021. Tabuada, P. Event-triggered real-time scheduling of stabilizing control tasks. IEEE Transactions on Automatic control, 52(9):1680 1685, 2007. Taylor, A., Singletary, A., Yue, Y., and Ames, A. Learning for safety-critical control with control barrier functions. In Learning for Dynamics and Control, pp. 708 717. PMLR, 2020. Taylor, A. J. and Ames, A. D. Adaptive safety with control barrier functions. In 2020 American Control Conference (ACC), pp. 1399 1405. IEEE, 2020. Taylor, A. J., Dorobantu, V. D., Le, H. M., Yue, Y., and Ames, A. D. Episodic learning with control lyapunov functions for uncertain robotic systems. In 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 6878 6884. IEEE, 2019. Tsukamoto, H., Chung, S.-J., and Slotine, J.-J. E. Contraction theory for nonlinear stability analysis and learningbased control: A tutorial overview. Annual Reviews in Control, 52:135 169, 2021. Wang, L.-Z., Su, R.-Q., Huang, Z.-G., Wang, X., Wang, W.- X., Grebogi, C., and Lai, Y.-C. A geometrical approach to control and controllability of nonlinear dynamical networks. Nature communications, 7(1):11323, 2016. Wehenkel, A. and Louppe, G. Unconstrained monotonic neural networks. Advances in neural information processing systems, 32, 2019. Wiener, N. Cybernetics or Control and Communication in the Animal and the Machine. MIT press, 2019. Xue, S., Luo, B., Liu, D., and Gao, Y. Neural networkbased event-triggered integral reinforcement learning for constrained h tracking control with experience replay. Neurocomputing, 513:25 35, 2022. Yang, L., Dai, H., Shi, Z., Hsieh, C.-J., Tedrake, R., and Zhang, H. Lyapunov-stable neural control for state and output feedback: A novel formulation. In Salakhutdinov, R., Kolter, Z., Heller, K., Weller, A., Oliver, N., Scarlett, J., and Berkenkamp, F. (eds.), Proceedings of the 41st International Conference on Machine Learning, volume 235 of Proceedings of Machine Learning Research, pp. 56033 56046. PMLR, 21 27 Jul 2024. URL https://proceedings.mlr.press/ v235/yang24f.html. Yoshida, Y. and Miyato, T. Spectral norm regularization for improving the generalizability of deep learning. ar Xiv preprint ar Xiv:1705.10941, 2017. Zhang, J., Zhu, Q., and Lin, W. Neural stochastic control. In Oh, A. H., Agarwal, A., Belgrave, D., and Cho, K. (eds.), Advances in Neural Information Processing Systems, 2022a. URL https://openreview.net/ forum?id=5w I7g Nop MHW. Neural Event-Triggered Control with Optimal Scheduling Zhang, J., Zhu, Q., Yang, W., and Lin, W. Sync: Safetyaware neural control for stabilizing stochastic delaydifferential equations. In The Eleventh International Conference on Learning Representations, 2022b. Neural Event-Triggered Control with Optimal Scheduling A. Appendix A.1. Proofs and Derivations In this section, we introduce some basic notations and then provide the proofs of the theoretical results. A.1.1. NOTATIONS Notations. Throughout the paper, we employ the following notation. Let x, y be the inner product of vectors x, y Rd. For a second continuous function f(x) : Rd R, let f denote the gradient of f(x), that is, 2f denote the Hessian matrix of f. For the two sets A, B, let A B denote that A is covered in B. Denote by log the base e logarithmic function. Denote by the L2-norm for any given vector in Rd. Denote by | | the absolute value of a scalar number or the modulus length of a complex number. For A = (aij), a matrix of dimension d r, denote by A 2 F = Pd i=1 Pr j=1 a2 ij the Frobenius norm. A.1.2. PROOF OF THEOREM 3.2 Theorem A.1. Consider the event-triggered controlled dynamics in Eq. (1), if the following assumptions are satisfied: (i) f(x , u ) f(x, u) lf ( x x + u u ); (ii) u(x ) u(x) lu x x ; (iii) Lfu V (x, u(x + e)) α( x ) + γ( e ) for some class-K functions α, γ with α 1(γ( e )) P e . Then, the minimal inter-event time implicitly defined by event function h = α( x ) γ( e ) is lower bounded by τh = 1 lf log P +1 P + lf lu lf (1+lu) . From the condition (iii) and the definition of the event function, we have the triggering time happens after P e = x . Therefore, the inter-event time is lower bounded by the minimal inter-event time defined by the event function h = P( e ) x . Now we come to deduce the estimation of the inter-event time of h, i.e., the time from e = 0 to P x . Consider the dynamic of e x , we have d dt e x = d dt (e e)1/2 1 2(e e) 1/22e e(x x)1/2 1 2(x x) 1/22x x(e e)1/2 = e e e x x x x x e x e x x x x x e x = f(x, u(x + e) lf x + lflu( x + e ) = lf(1 + lu) + lflu e x By denoting z = e x , we have the triggering time of h happens after the variable z increases from 0 to 1 P . The dynamic of z is z = (lf(1 + lu) + lfluz) (1 + z) Neural Event-Triggered Control with Optimal Scheduling We have dz (1 + az)(1 + y) = bdt, where a = lf lu lf (1+lu), b = lf(1 + lu). Then we have dz (1 + az)(1 + z) = a a 1 1 1 + az 1 a(1 + z) = 1 a 1 (d log(1 + az) d log(1 + z)) By integrating the above equation, we have P ) log(1 + 1 T = 1 b(a 1) log 1 + a = 1 b(1 a) log 1 + 1 lf log P + 1 P + lf lu lf (1+lu) , which completes the proof. A.1.3. PROOF OF THEOREM 3.3 Theorem A.2. For the event-triggered controlled dynamics in Eq. (1) with event function h defined in Eq. (9), if the state space D is bounded, the Eqs. (7),(8) and the conditions (i), (ii) in Theorem 3.2 hold, then the minimal inter-event time is lower bounded by τ h = 1 lf log clα 1lu+1 clα 1lu+ lf lu lf (1+lu) , here lα 1 is the Lipschitz constant of α 1. From the Eqs. (7), we know that the triggering time defined by h in Eq. 9 is larger than that defined by h in Theorem 3.2. Notice in Theorem 3.2 P is a tight upper bound Lipschitz constant of α 1 γ. Since the state space D is bounded, from Eq. 7, if we set γ as the tight estimation of V (f(x, u(x + e)) f(x, u(x))), the Lipschitz constant of γ can be bounded by max x D V (x) lflu. Then we get Lip(α 1 γ) max x D V (x) lflulα 1. By denoting c = maxx D V (x) lf and replace P with clα 1lu in Theorem 3.2, we obtain the final estimation of τ h. Neural Event-Triggered Control with Optimal Scheduling A.1.4. PROOF OF THEOREM 4.1 Theorem A.3. (Stability guarantee) For a candidate controller u and the stable controller space U(V ) = {u : Lfu V +V 0}, we define the projection operator as, π(u, U(V )) u max(0, Lfu V V ) If the controller has affine actuator, then we have π(u, U(V )) U(V ), the projected controller is Lipschitz continuous over the state space D if and only if D is bounded. Furthermore, under the triggering mechanism V (x) [f(x, u(x + e)) f(x, u(x))] σV (x) = 0, σ (0, 1), e = x(tk) x(t), t [tk, tk+1) the controlled system under π(u, U) is assured exponential stable with decay rate 1 σ, and the inter-event time has positive lower bound. Proof. To begin with, we check the inequality constraint in U(V ) is satisfied by the projection element, that is Lfu V u=π(u,U(V )) V. Since the controller has affine actuator, from the definition of the Lie derivative operator, we have Lfu V u=π(u,U(V )) = V (f + u max(0, Lu V + V ) = V (f + u) V max(0, Lu V + V ) = Lu V max(0, Lu V + V ) V. Next, we show the equivalent condition of the Lipschitz continuity of projection element. Notice that u Lip(D), then we have ˆπ(u, U(V )) Lip(D) max(0, Lu V + V ) V 2 V Lip(D). V is a continuous unit vector, and naturally is Lipschitz continuous, we only need to consider the remaining term max(0, Lu V + V ) V . According to the definition, all the functions occured in this term are continuous, so we only need to bound this term to obtain the global Lipschitz continuity, that is max(0, Lu V + V ) V Lip(D) sup x D max(0, Lu V + V ) When Lu V V , obviously we have max(0, Lu V + V ) = 0 < + , otherwise, since V ε x p, we have Lu V + V Lu V + ε x p O( x p) ( x ). Thus, we have max(0, Lu V + V ) V < + sup x D x < + . The positive lower bound of the inter-event time comes from the Theorem 3.2. We complete the proof. Neural Event-Triggered Control with Optimal Scheduling A.2. Algorithms In this section, we provide the algorithms of Neural ETC-PI (1) and Neural ETC-MC (2). Firstly, we supplement the warm up stage for path integral algorithm to accelerate the convergence of training process. Warm up. At the beginning of the training process, the stability constraint is not satisfied, which leads to the solution t1 of the event function hθ,ϕ does not exist. To ensure the training process can proceed smoothly, we pre-train the parameterized model with L(ϕ, θ, {ci}) = Lstab + λ1Llip. (15) Algorithm 1 Neural ETC-PI: Path Integral Algorithm 1: hyperparameters: N, M Sample size and batch size β, m Learning rate and max iterations µ(D), λ1, λ2 Data distrituion, weight factors 2: initialize w = (ϕ, θ) From Eq. (5) 3: generate dataset DN = {xi}N i=1 µ(D) 4: for r = 1 : m do 5: w w β w L(w) Warm up in Eq. (15) 6: end for 7: for r = 1 : m do 8: {xi(0)}M i=1 DN Sample batch data 9: ti,1, xi(ti,1) = ODESolve Event(xi(0), f, uϕ, 0) 10: w w β w L(w) From Eq. (6) 11: end for 12: return uϕ, Vθ Algorithm 2 Neural ETC-MC: Monte Carlo Algorithm 1: hyperparameters: N, Mα, λ1, λ2 Sample sizes and weight factors β, m Learning rate and max iterations µ(D), µ(X) Distributions of state and error 2: initialize w = (ϕ, θ, θα) Eqs. (5),(10) 3: generate dataset {xi}N i=1 {xi}Mα i=1 µ(D) µ(X) 4: for r = 1 : m do 5: w w β w L(w) From Eq. (11) 6: end for 7: return uϕ, Vθ, αθα Neural Event-Triggered Control with Optimal Scheduling A.3. Experimental Configurations In this section, we provide the detailed descriptions for the experimental configurations of the benchmark dynamical systems and control methods in the main text. We implement the code on a single i7-10870 CPU with 16GB memory, and we train all the parameters with Adam optimizer. A.3.1. NEURAL NETWORK STRUCTURES For constructing the potential function V , we utilize the ICNN as (Amos et al., 2017): z1 = σ(W0x + b0), zi+1 = σ(Uizi + Wix + bi), i = 1, , k 1, V (x) = σ(p(x) p(0)) + ε x 2, where σ is the smoothed Re LU function as defined in the main text, Wi Rhi d, Ui (R+ {0})hi hi 1, x Rd, and, for simplicity, this ICNN function is denoted by ICNN(h0, h1, , hk 1). We set ε = 1e-3 as default value for all the experiments; The class-K function α is constructed as: q1 = Re LU(W0s + b0), qi+1 = Re LU(Wiqi + bi), i = 1, , k 2, qk = ELU(Wk 1qk 1 + bk 1), where Wi Rhi+1 hi, and this class-K function is denoted by K(h0, h1, , hk); The neural control function (nonlinear version) is constructed as: z1 = F(Spectral Norm(W0x + b0)), zi+1 = F(Spectral Norm((Wizi + bi)), i = 1, , k 1, NN(x) Spectral Norm(Wkzk), u(x) = diag(x x )NN(x) or NN(x) NN(x ), where F( ) is the activation function, Spectral Norm is the spectral norm function from (Yoshida & Miyato, 2017), Wi Rhi+1 hi, and this control function is denoted by Control(h0, h1, , hk+1). Since we deploy the Spectral Norm package in our algorithm, the weight factor λ1 for Lipschitz constant of u is automatically set as the default value in this package and we do not tune it in our experiments due to its good performance. The standard neural network is constructed as: z1 = F(W0x + b0), zi+1 = F(Wizi + bi), i = 1, , k 1, NN(x) Wkzk, where F( ) is the activation function, and this standard function is denoted by MLP(h0, h1, , hk+1) A.3.2. GENE REGULATORY NETWORK Here we model the controlled gene regulatory network (GRN) as x1 = a1 xn 1 sn + xn 1 + b1 sn sn + xn 2 kx1 + u xn 1 sn + xn 1 , x2 = a2 xn 2 sn + xn 2 + b2 sn sn + xn 1 kx2, Neural Event-Triggered Control with Optimal Scheduling where the under-actuated control u only acts on the protein regulation strength a1. We specify a1 = a2 = 1, b1 = b2 = 0.2, n = 2, k = 1.1, s = 0.5. The two attractors of the original model is P1 : (x 1, x 2) = (0.62562059, 0.62562059), P2 : (x0 1, x0 2) = (0.0582738, 0.85801853). We aims at stabilize the attractor P2 with low protein concentration to P1 with high protein expression level. We slightly modify the neural networks s.t. V (P1) = 0, u(P1) = 0, e.g. V = V (x) V (P1), u = u(x) u(P1). Since these two attractors are close in the Euclidean space, it hard for algorithms to identify them from states with numerical error. To address this issue, we rescale the original system as x1 = 10x1, x2 = 10x2 to enlarge the attractors. For training controller u, we uniformly sample 1000 data from the state region [ 10, 10]. We test the performance under different learning rate lr {0.01, 0.03, 0.05} and pick the best one, the considered control methods are set as following, Neural ETC-PI. We parameterize V (x) as ICNN(2, 10, 10, 1), u(x) as Control(2, 20, 20, 1) with F = Re LU. We set the iterations for warm up as 500, the iterations and batch size for calculating the triggering times as 50 and 10, the learning rate as lr = 0.01, the weight factor for event loss as λ2 = 10 1000. Neural ETC-MC. We parameterize V (x) as ICNN(2, 20, 1), u(x) as Control(2, 20, 20, 1). We set the iterations as 500 + 50, the learning rate as lr = 0.05, the weight factor for event loss as λ2 = 0.1. NLC. We parameterize V (x) as MLP(2, 20, 20, 1), u(x) as MLP(2, 20, 20, 1). We set the iterations as 500 + 50, the learning rate as lr = 0.05, the loss function is Lfuϕ Vθ(xi) + + (Vθ(xi))+ + Vθ(P1)2 Quad-NLC. We parameterize V (x) as (x P1) MLP(2, 20, 2) MLP(2, 20, 2)(x P1), u(x) as MLP(2, 20, 20, 1). We set the iterations as 500 + 50, the learning rate as lr = 0.05, the loss function is Lfuϕ Vθ(xi) + Vθ(xi) + + Vθ(P1)2. BALSA. For this QP based method, we set the object function as min u,d1,d2 1 2 u 2 + p1d2 1, s.t.Lfu V V d1, where d1 is the relaxation number. We choose V = 1 2 x P1 2, p1 = 50. We solve this problem with the QP solver in cvxopt in Python package. LQR. We linearize the controlled dynamic near the target P1 as x = A(x P1) + Bu, a1 n(x 1)n 1 (sn + (x 1)n)2 k b1 n(x 2)n 1 (sn + (x 2)n)2 a2 n(x 1)n 1 (sn + (x 1)n)2 b2 n(x 2)n 1 (sn + (x 2)n)2 k (sn + (x 1)n)2 k Neural Event-Triggered Control with Optimal Scheduling We set the cost matrix in LQR as Q = 10 0 0 10 and solve the problem via lqr method in Matlab. The obtained Riccati solution S forms the Lyapunov function V = 1 2(x P1) S(x P1), the controller is u = K(x P1) where K R1 2 is returned by the lqr solver. The Lie derivative of the Lyapunov function is (x P1) Q1(x P1) with Q1 = Q + K RK. Critic-Actor ETC. According to the implementation setting in (Cheng et al., 2023), we consider the following eventtriggered controller parametrized by the critic neural network Wc and the actor neural network Wa, x = f(x) + g(x)u(x), V (x) = min u x Qx + u Ru dt, V (x) = x Wcx, u (x) = Wax, ea = Wax + 1 ec = V (x) [f(x) + g(x)u(x)] + x Qx + u Ru, here f is the original dynamics described above, g is the actuator taking the form, xn 1 (sn + xn 1)2 0 The cost matrix Q, R are the same as that in LQR. In the event-triggered mode, the weights of critic and actor NN, Wc and Wa, obeying the evolution dynamics as follows, Wa = 0, t [tk, tk+1), W + a = Wa αa Ka Wa , t = tk+1, Wc = 0, t [tk, tk+1), W + c = Wc αc Kc Wc , t = tk+1, where αa and αc are the learning rates of the critic and actor NNs, respectively. For the initial value of Wc and Wa, we employ the solutions from the above LQR solver as Wc = S, Wa = K. We set the learning rate as αc = αa = 1e 2, the event function is set as h = |e| ethres, ethres = 0.2 according to (Cheng et al., 2023), here e = (ex1, ex2) are the variables of error dynamics. The event function here is different to our proposed stability guaranteed function because of the lack of Lyapunov function in this method, we note that V is only an auxiliary function used to find the dynamics of Wc, Wa and cannot be verified as a Lyapunov function. We have tuned the hyperparameters αc, αa {5e 4, 1e 3, 1e 2, , 1e 1}, ethres {0.1, 0.2, 0.3, 0.4, 0.5} and fix the parameters with the best performance. Neural Event-Triggered Control with Optimal Scheduling IRL ETC. Similarly to the Critic-Actor ETC, (Xue et al., 2022) transformed the optimization control problem to a RL problem via abstracting the Hamilton-Jacobi-Bellman equation as the value function and approximating the optimal value function based on a preset basis activation function. Specifically, we consider the control problem parametrized by the critic neural network W as follows, x = f(x) + g(x)u(x), V (x) = min u x Qx + u Ru dt, V (x) = x W x, u (x) = ησ 1 2η R 1g V (x) , t e α(τ t) " 0 2ησ 1(η 1s)rids dτ, diag(R) = (r1, , rm), here f is the original dynamics described above, g is the actuator taking the form, xn 1 (sn + xn 1)2 0 The cost matrix Q, R are the same as that in LQR. In the event-triggered mode, the weight W of critic NN is updated as, W = 0, t [tk, tk+1), W + = W β K W , t = tk+1, with β being the learning rates of the weight. We initialize the weight as W = (Wij = 4)2 2. We set the learning rate as β = 1e 2, the event function is set as h = e 2 (1 λ2 y)λ(Q) η2λ2x x 2, λ2 y = 0.6 according to (Cheng et al., 2023). In the original work (Xue et al., 2022), historical data is considered as multiple integral on time interval [tj, tj + l] [tk, tk+1) like E = E[t,t+l]. To simplify the calculation, here we merge the multiple historical data to a single integral on time interval [tk, tk +l] with l = min(l , tk+1 tk), here l = 1.2 is a predefined length of historical data. We tuned the hyperparameters in the same way with Critic-Actor ETC, and the final results are α = 0.1, η = 1.0, λx = 0.1, σ( ) = Id( ). Test configurations. For implementing the controller in the event-triggered mode, we set the event function for Neural ETC-PI, Quad-NLC, BALSA as V (f(x, u(x + e)) f(x, u(x))) σV (x), the event function for Neural ETC-MC as V (f(x, u(x + e)) f(x, u(x))) σα( x ), the event function for NLC as V (f(x, u(x + e)) σf(x, u(x))) , the event function for LQR as (Heemels et al., 2012) (σ 1)(x P1) Q1(x P1) + 2(x P1) SBKe, where the σ is set as 0.5 for all models. For the initial value, we set x0 = P2 + ξi, ξi U[ 1, 1], i = 1, , 5, the random seed is (2, 4, 5, 6, 7). Neural Event-Triggered Control with Optimal Scheduling A.3.3. LORENZ SYSTEM Here we model the state of the Lorenz system under fully actuated control u = (u1, u2, u3) as x = (x, y, z) , x = σ(y x) + u1, y = ρx y xz + u2, z = xy βz + u3. We aim to stabilize the zero solution of this chaotic system. We consider σ = 10, ρ = 28, β = 8/3. For training controller u, we uniformly sample 5000 data from the state region [ 10, 10]. We construct the controllers as follows. Neural ETC-PI. We parameterize V (x) as ICNN(3, 64, 1), u(x) as Control(3, 64, 64, 3) with F = Re LU. Since the Ode solver in the training process require high computational resources, we down-sample 2000 data from the original dataset for training. We set the iterations for warm up as 500, the iterations and batch size for calculating the triggering times as 100 and 10, the learning rate as lr = 0.05, the weight factor for event loss as λ2 = 100 2000. Neural ETC-MC. We parameterize V (x) as ICNN(3, 64, 1), u(x) as Control(3, 64, 64, 3). We set the iterations as 500 + 100, the learning rate as lr = 0.05, the weight factor for event loss as λ2 = 0.1. NLC. We parameterize V (x) as MLP(3, 64, 64, 1), u(x) as MLP(3, 64, 64, 3). We set the iterations as 500 + 100, the learning rate as lr = 0.05, the loss function is Lfuϕ Vθ(xi) + + (Vθ(xi))+ + (Vθ(0))+, notice that we select the last term in the right hand side as (Vθ(0))+ instead of Vθ(0)2 since the former performs better than the latter. We also resample 5000 data from [ 5, 5] since the NLC performs poorly in the original dataset, the similar case holds for Quad-NLC. Quad-NLC. We parameterize V (x) as x MLP(3, 64, 3) MLP(3, 64, 3)x, u(x) as MLP(3, 64, 64, 3). We set the iterations as 500 + 100, the learning rate as lr = 0.05, the loss function is Lfuϕ Vθ(xi) + Vθ(xi) + + Vθ(0)2. BALSA. For this QP based method, we set the object function as min u,d1,d2 1 2 u 2 + p1d2 1, s.t.Lfu V V d1. We choose V = 1 2 x 2, p1 = 20. LQR. We linearize the controlled dynamic near the zero solution as x = Ax + Bu, σ σ 0 ρ 1 0 0 0 β 1 0 0 0 1 0 0 0 1 Neural Event-Triggered Control with Optimal Scheduling We set the cost matrix in LQR as 5 0 0 0 10 0 0 0 5 0.1 0 0 0 0.1 0 0 0 0.1 The obtained Riccati solution S forms the Lyapunov function V = 1 2x Sx, the controller is u = Kx where K R3 3 is returned by the lqr solver. The Lie derivative of the Lyapunov function is x Q1x with Q1 = Q + K RK. Critic-Actor ETC. The updating procedure is the same as that in Appendix A.3.2, we set the hyperparameters as αc = αa = 5e 4, ethres = 0.3, we note that for the chaotic system, the event-triggered dynamics is easy to explode when αc,a are slightly larger than 1e 3. The actuator in this example is the identity matrix as g = I3 3. IRL ETC. The updating procedure is the same as that in Appendix A.3.2, we set the hyperparameters as β = 1e 2, α = 0.1, σ( ) = tanh( ), η = 10, λx = 1.0, λ2 y = 0.6. The actuator in this example is the identity matrix as g = I3 3. Test configurations. We select the same event functions as those for GRN to implement the event-triggered control, except for setting σ = 0.99 for LQR since it fails in the case σ = 0.5. For the initial value, we randomly select 5 points in the original dataset using numpy.random.choice method in Python, and the random seeds are set as {3, 5, 7, 8, 9}. A.3.4. MICHAELIS MENTEN MODEL Consider the coupled subcellular model under topology control as i=1 Aij x2 j 1 + x2 j + δAii x2 i 1 + x2 i . This dynamic has two attractor, inactive state P1 = 0 represents the cell apoptosis and the active P2 represents the reviving cell state. We aim at regulating the cell state to the reviving state through only tuning the diagonal topology structure, which can be achieved experimentally via drugs or electrical stimulation. Therefore, an ideal control should be updated as little as possible since the frequent stimulation may do harm to the cells. For training controller u = (δA11, , δA100,100), we uniformly sample 1000 data from the state region [ 10, 10]. Similarly to that in GRN, we modify the parameterized V and u functions s.t. V (P2) = 0, u(P2) = 0. We construct the controllers as follows. Neural ETC-PI. We parameterize V (x) as ICNN(100, 64, 1), u(x) as Control(100, 64, 64, 100) with F = Re LU. Since the dimension of the task is very high, the ODE solver has very high computational cost in solving the triggering times. We set the the iterations and batch size for calculating the triggering times as 10 and 5. If the readers have more powerful computing device, larger iterations and batch size are recommended. We set the iterations for warm up as 500, the learning rate as lr = 0.01, the weight factor for event loss as λ2 = 100 1000. In the case, we try a combination of Neural ETC-PI and Neural ETC-MC by setting the stabilzaition loss as Lfuϕ Vθ(xi) + αθα( xi ) + and we also penalize the Lipschitz constant of α 1 by adding term Lα 1 to the loss function with weight 0.1. The dataset {xi} for Lα 1 is generated by equidistant sampling on [0, 10]. Neural ETC-MC. We parameterize V (x) as ICNN(100, 200, 1), u(x) as Control(100, 200, 200, 100). We set the iterations as 500, the learning rate as lr = 0.05, the weight factor for event loss as λ2 = 0.1. The dataset {xi} for Lα 1 is generated by equidistant sampling on [0, 5]. Neural Event-Triggered Control with Optimal Scheduling NLC. We parameterize V (x) as MLP(100, 200, 200, 1), u(x) as MLP(100, 200, 200, 100). We set the iterations as 500, the learning rate as lr = 0.01, the loss function is Lfuϕ Vθ(xi) + + (Vθ(xi))+ + Vθ(0)2. Quad-NLC. We parameterize V (x) as (x P2) MLP(100, 200, 100) MLP(100, 200, 100)(x P2), u(x) as MLP(100, 200, 200, 100). We set the iterations as 500, the learning rate as lr = 0.01, the loss function is Lfuϕ Vθ(xi) + Vθ(xi) + + Vθ(0)2. BALSA. For this QP based method, we set the object function as min u,d1,d2 1 2 u 2 + p1d2 1, s.t.Lfu V V d1. We choose V = 1 2 x 2, p1 = 50. LQR. We linearize the controlled dynamic near the P2 solution as x = A(x P2) + Bu, i=1 Aij 2x j (1 + (x j)2)2 + δAii (x i )2 (1 + (x i )2)2 We set the cost matrix in LQR as Q = 10I100 100, R = 0.01I100 100. The obtained Riccati solution S forms the Lyapunov function V = 1 2(x P2) S(x P2), the controller is u = K(x P2) where K R100 100 is returned by the lqr solver. The Lie derivative of the Lyapunov function is (x P2) Q1(x P2) with Q1 = Q + K RK. Critic-Actor ETC. Similarly, we set the hyperparameters as αc = αa = 1e 2, ethres = 0.2. The actuator in this example is g = diag( x2 1 (1 + x2 1)2 , , x2 100 (1 + x2 100)2 ). (16) Since the dynamics of Wc and Wa are both 1002-D, leading to a significantly high dimensional system, we reduce the dynamics as V (x) = W c (x2 1, , x2 100) , Wc R100, u (x) = diag(Wa)(x1, , x100) , Wa R100, to the 100-D systems, for the sake of limited computational resources. IRL ETC. We set the hyperparameters as β = 1e 2, α = 0.1, σ( ) = Id( ), η = 1, λx = 0.1, λ2 y = 0.6. The actuator in this example is g = diag( x2 1 (1 + x2 1)2 , , x2 100 (1 + x2 100)2 ). (17) We reduce the dimension of the dynamics as the same with that in Critic-Actor ETC above. Neural Event-Triggered Control with Optimal Scheduling Test configurations. We select the same event functions as those for GRN to implement the event-triggered control. For the initial value, we set x0 = P1 + ξi, ξi U[ 0.5, 0.5], i = 1, c . . . , 5, and the random seeds are set as {0, 3, 4, 5, 6}. A.3.5. MOTIVATION OF SELECTING THE BENCHMARK SYSTEMS In (Wang et al., 2016), a geometrical approach for switching the system from ROA of one equilibrium to another, through finite changes of the experimentally feasible parameters, wherein GRN system is investigated in their paper. Since our Neural ETC has similarity to the geometrical approach in terms of adding finite non-invasive control to the system, we also study GRN in our work. The Lorenz system is a classic chaotic systems possessing plentiful shapes of dynamical trajectories, hence, the control of Lorenz (or control of chaos in a more common sense) is of important position in control literature (Ott et al., 1990; Boccaletti et al., 2000), and the control of Lorenz system under event-triggered implementation is also investigated in (Abdelrahim et al., 2015). In (Sanhedrai et al., 2022), a topological reconstruction method to the structure of complex dynamics is proposed to revive the degenerate complex system via minimal interventions, i.e., reconstructing links or nodes as small as possible, and the Michaelis Menten model describing the evolution dynamics of sub-cellular behavior is considered as an illustration. Since the event-triggered control aims at adding feasible control to the complex system intermittently, e.g., changing the network structure slowly in time, we think it s meaningful to consider the Michaelis Menten model in our work to see if there are essentially same parts between our method with the topological reconstruction method.