Next Article in Journal
New Approach to Cross-Correlation Reflectometry Diagnostics of Nonlocality of Plasma Turbulence
Previous Article in Journal
Numerical Model Validation for Detection of Surface Displacements over Twin Tunnels from Metro Line 1 in the Historical Area of Seville (Spain)
Previous Article in Special Issue
A Continuous Terminal Sliding-Mode Observer-Based Anomaly Detection Approach for Industrial Communication Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Event-Triggered Integral Sliding Mode Adaptive Optimal Tracking Control for Uncertain Nonlinear Systems

School of Mathematics, Southeast University, Nanjing 211189, China
*
Author to whom correspondence should be addressed.
Submission received: 13 May 2022 / Revised: 9 June 2022 / Accepted: 13 June 2022 / Published: 18 June 2022
(This article belongs to the Special Issue Recent Advances in Sliding Mode Control/Observer and Its Applications)

Abstract

:
In this paper, we study the event-triggered integral sliding mode optimal tracking problem of nonlinear systems with matched and unmatched disturbances. The goal is to design an adaptive dynamic programming-based sliding-mode controller, which stabilizes the closed-loop system and guarantees the optimal performance of the sliding-mode dynamics. First, in order to remove the effects of the matched uncertainties, an event-triggered sliding mode controller is designed to force the state of the systems on the sliding mode surface without Zeno behavior. Second, another event-triggered controller is designed to suppress unmatched disturbances with a nearly optimal performance while also guaranteeing Zeno-free behavior. Finally, the benefits of the proposed algorithm are shown in comparison to several traditional triggering and learning-based mechanisms.

1. Introduction

Traditional control systems are implemented with time-triggered (TT) sampling (i.e., periodic sampling). Data are sent from the controller to the actuator (or from the sensor to the controller) using a fixed sample length. However, in modern control systems, especially networked control systems, control signals are often implemented aperiodically [1]. The advantages of the aperiodic sampling in terms of update times have been elaborated in [2] in detail. In fact, the limited communication bandwidth in networked control systems has stimulated a huge attention in event-triggered control (ETC) [3,4,5,6,7,8,9,10,11], as an alternative to TT. Despite the progress in the filed, some problems remain open, such as how to simultaneously counteract the effect of matched and unmatched disturbances while guaranteeing the optimal performance of the ETC systems, which is one of our concerns and is an issue to be studied.
Owing to the distinguished features such as fast dynamic response, robustness, order reduction, and implementation simplicity, sliding mode control (SMC) is widely studied in the fields of transportation, power grid, industrial communication networks, and uncertain serial industrial robots [12,13,14,15,16,17,18,19,20]. Especially, SMC has been recognized as one of the popular and powerful control tools in power converters based on deriving from the variable structure systems theorem. In addition, its popularity comes from the robustness feature, which eliminates the burden of the necessity of system parameters required for accurate modeling, yet it may lack robustness against unmatched disturbances. Although many achievements in codesign of sliding mode control and event-triggered have emerged [21,22,23], they all neglect how to guarantee the optimal performance of the controlled system in the presence of matched and unmatched disturbances.
Optimal control theory is now quite mature [24,25,26,27]. Optimal control for nonlinear systems requires one to solve the Hamilton–Jacobi–Bellman (HJB) or Hamilton–Jacobi–Isaac (HJI) equations. However, the nonlinearity of the above equation makes it impossible to get an analytical solution [28]. Although numerical solutions via dynamic programming [29] can be obtained, issues of offline optimization and curse of dimensionality make these numerical methods not feasible in practical applications. To address these shortcomings, adaptive dynamic programming (ADP), oriented from reinforcement learning (RL) technique, has been widely researched in recent decades [30,31,32,33,34,35].
Event-triggered ADP [36,37,38] and event-triggered ADP-based ISMC [39,40] have been investigated in the past years. However, some open issues still remain, which are explained as follows. Although event-triggered implementations can reduce the update times of standard TT implementations, it needs to be pointed out that the event-triggering techniques developed in [39,40] depend only on the current state (static event-triggered control). It is well known that static event-triggered techniques may effectively reduce the communication cost when the sampling error is large. However, as the sampling error becomes smaller, this technique will become conservative, i.e., it can cause several unnecessary triggering. So, developing more flexible event-triggering conditions to further save resources is urgently needed. Moreover, it is still an open problem to design a learning-based event-triggered strategy while at the same time relaxing the excitation condition typically required for ADP methods to further reduce triggered times. Enlightened by the aforementioned discussions, in this article, we will provide a new learning-based framework for ETC-based ISM optimal tracking problem. The novelties are presented below.
  • Different from the combined SMC and ADP frameworks of [39,40], this paper proposes a new dynamic event-triggered (DET) mechanism. By introducing an auxiliary variable, which is non-negative. This can increase the length of the time intervals between triggering events, further reducing the communication burden compared with [39,40].
  • A novel Integral Sliding Mode Control (ISMC) scheme based on DET for uncertain nonlinear systems is proposed, consisting of two control laws. A first event-triggered controller is designed to tackle the matched uncertainties and force the trajectory of the system on the sliding mode surface. A second event-triggered controller is designed to tackle the unmatched uncertainties and guarantee optimal performance.
  • To solve the resulting optimal control problem, a critic-only neural network (NN) based on ADP is proposed via the experience replay technique, which helps relaxing the excitation condition typically required for ADP methods to work. Stability of the closed-loop system is proven in the sense of uniformly ultimately boundedness, while guaranteeing Zeno-free behavior of the triggering mechanism.
The paper is arranged as follows. In Section 2, we introduce the model formulation and preliminaries. Section 3 covers the event-triggered-based ISMC design. Section 4 presents the framework of a dynamic triggered ADP strategy, along with stability analysis. Section 5 illustrates the effectiveness of the novel algorithm via comparative simulations. Section 6 gives the conclusions and possible future works.
Notations: R + represents the sets of the positive real numbers. R n and R n × m denote the space of all real n-vectors and the space of all n × m real matrices, respectively. ≜ means “equal by definition”, and I n is the identity matrix of dimension n × n . T is the transposition symbol. C 1 represents the class of functions with continuous derivative. λ m i n ( X ) is defined as the minimum eigenvalue of matrix X, · represents the 2-norm of a vector or matrix. For any full column rank matrix F ( · ) , its left pseudoinverse is F ( · ) ( F T ( · ) F ( · ) ) 1 F T ( · ) .

2. System Formulation and Preliminaries

In this section, an ISMC design oriented toward optimal tracking is discussed. To target optimality, it is useful to introduce an augmented system associated with the tracking error system and the desired reference system.
Consider the following nonlinear system with matched and unmatched uncertainties [39]:
x ˙ = p ( x ) + q ( x ) ( u + d ) + h ¯ ( x ) w
where x R n , u R m , d R m and w R q are the system state, control input, unknown bounded matched disturbance, and unknown bounded unmatched disturbance, respectively. The system dynamics p ( x ) , q ( x ) and h ¯ ( x ) are known Lipschitz functions with p ( 0 ) = 0 , and q ( x ) and h ¯ ( x ) having a positive upper bound. Meanwhile, let the system (1) be controllable, and the matrix function q ( x ) be full column rank with q ( · ) h ¯ ( · ) .
Remark 1.
Herein, it should be noted that system (1) is favored by scholars in theoretical studies [18,33,35] as well as has been widely studied and explored in practical applications, such as a single link robot arm [39,40]; power system [34]; a spacecraft [41] et al.
The desired signal x d is subject to
x ˙ d = r ( x d )
where the bounded x d R n is Lipschitz continuous with r ( 0 ) = 0 . Define the tracking error
e d = x x d .
Combining (1)–(3), e d satisfies the following dynamics
e ˙ d = p ( x ) + q ( x ) ( u + d ) + h ¯ ( x ) w x ˙ d .
Define the augmented state ξ = [ e d T , x d T ] T R 2 n . Combining (2) with (4) generates the following augmented system
ξ ˙ = f ( ξ ) + g ( ξ ) ( u + d ) + h ( ξ ) w
with
f ( ξ ) = p ( e d + x d ) r ( x d ) r ( x d ) g ( ξ ) = q ( e d + x d ) 0 h ( ξ ) = h ¯ ( e d + x d ) 0
where f ( ξ ) R 2 n , g ( ξ ) R 2 n × m and h ( ξ ) R 2 n × q .
For simplicity and convenience of analysis, we will omit some arguments of the functions ( f ( ξ ) , g ( ξ ) , and h ( ξ ) will be written as f , g , and h , respectively) in the later part of the paper. To this end, the following two standard assumptions are required.
Assumption 1
([39]). System dynamics (5) with f ( 0 ) = 0 is Lipschitz continuous, and g and h satisfy g b q ( b q R + ) and h b h ( b h R + ) , respectively. The disturbances d and w satisfy d b d ( b d R + ) and w b w ( b w R + ) , respectively.
Assumption 2
([39]). The matrix function g is full column rank and its left pseudoinverse is given by g = ( g T g ) 1 g T and there exists some positive values b q ( b q R + ) such that g b q . Then, the following equality holds
h w = g g h w + ( I g g ) h w
where g g h b g ( b g R + ) .
Control objective: This article aims to achieve an optimal tracking control of system state x for the desired trajectory x d , so that the tracking error is uniformly ultimately bounded UUB. The control input should eliminate the effect of the matched disturbance and reduce the effects of the unmatched disturbance.
To achieve this goal, first, a new composite control law in the form u = u 0 + u 1 will be considered. Second, according to the robustness to nonlinearities and uncertainties of SMC technique, the Section 3 designs an integral sliding surface and event-triggered controller u 1 aiming to suppress the matched effects of the systems while forcing the state of the systems on the sliding mode surface without Zeno-free behavior. Then, event-triggered optimal controller u 0 is designed aiming to reduce the unmatched effect and guarantee the optimal performance of the sliding-mode dynamics in Section 4. Finally, a nonlinear single link robot arm is considered to verify the effectiveness of the proposed algorithm.

3. DET-Based ISMC Design

To tackle the uncertainty affecting the nonlinear system (5), an integral type sliding surface is designed as
S ( ξ ) = M [ ξ ξ 0 0 t f + g u 0 d τ ]
where M R m × 2 n ( m < 2 n ) is a projection matrix satisfying that M g is invertible. Here, u 0 is an optimal control input that will be designed in the next section. In this section, we aim to design the control u 1 to force the augmented system (5) onto the manifold { ξ | S ( ξ , t ) = 0 } in finite time, to remove the effect of matched disturbances. This can be achieved via
u 1 ( t ) = X sgn ( g T M T S ( ξ ) )
where X > 0 . The sign function sgn ( · ) is
sgn ( α ) = 1 , α > 0 , 0 , α = 0 , 1 , α < 0 .
Next, we introduce the dynamics of the sliding variable and the notion of practical sliding mode, which will be required later. The dynamics of the sliding variable can be obtained by differentiating (7) with respect to time
S ˙ ( ξ ) = M [ ξ ˙ ( f + g u 0 ) ] .
According to the SMC theory, when the system trajectories reach the manifold, one has S ˙ = 0 . Then, by combining (5), (6) and (9), one has an equivalent control u 1 e q satisfying
u 1 e q = d g h w ( M g ) 1 M ( I g g ) h w .
Substituting the equivalent control (10) into the augmented system (5), one gets thew dynamics on the sliding manifold
ξ ˙ = f + g u 0 + [ I g ( M g ) 1 M ] w u
where w u = [ I g g ] h w . The following result is recalled.
Lemma 1
([18]). The optimal solution to the following optimization problem
arg min M R m × 2 n [ I g ( M g ) 1 M ] w u 2
is M = g .
Remark 2.
Since M is the left pseudoinverse of g, one knows
(1) 
M = g can minimize w u e q , which leads to w u e q = w u , refer to [18] for the detailed proof;
(2) 
The modulation gain associated with u 1 ( t ) is minimized, which means that the amplitude of chattering can be reduced;
(3) 
M = g avoids amplifying the effect of the unmatched disturbance.
To extend the continuous-time control u 1 in the event-triggered paradigm, we define a virtual control input μ ( t ) satisfying μ ( t ) u ( t k ) , t [ t k , t k + 1 ) , where
μ ( t ) = X sgn ( g T M T S ( ξ ( t k ) ) )
and { t k } k = 0 is a sampling sequence with t k < t k + 1 , k N with N = { 0 , 1 , 2 , } . Define the following measurement error
e u 1 = μ ( t k ) μ ( t ) = X sgn ( g T M T S ( ξ ) ) X sgn ( g T ( ξ k ) M T ( ξ k ) S ( ξ k ) )
satisfying the following triggering condition
t k + 1 = inf { t > t k | X b d b q e u 1 ( t ) 0 } , t 0 = 0 .
The following theorem establishes a sufficient condition to the reachability of the specified sliding surface (7).
Theorem 1.
Suppose Assumptions 1 and 2 hold. Consider the augmented system (5) with the sliding surface (7) under event-triggered controller (12). Then, practical sliding mode is achieved if the triggering condition satisfies (14).
Proof. 
Consider the following Lyapunov function V 1 = 1 2 S T S . For t [ t k , t k + 1 ) , the time derivative V 1 concerning system (5) with (12) is
V ˙ 1 = S T S ˙ = S T M [ ξ ˙ f g u 0 ] = S T M [ f + g ( u 0 ( t k ) + u 1 ( t k ) + d ) + h w ( f + g u 0 ( t k ) ) ] = S T M [ g ( u 1 ( t k ) + d + g h w ) + w u ] = S T M g [ ( e u 1 ( t ) X sgn ( g T M T S ) + d + g h w ] + S T M g ( M g ) 1 M w u S T M g [ X + b d + b q + e u 1 ( t ) ] + S T M g ( M g ) 1 M w u .
Substituting M = g into (15), we have V ˙ 1 S T [ X b d b q e u 1 ( t ) ] . Noticing the triggering condition (14), we know that e u 1 X b d b q holds all the time, then V ˙ 1 < 0 for S 0 , which implies that the system state ξ starting from the initial state ξ ( 0 ) slide robustly on the switching surface S from initial time t = 0 and can gradually reach the sliding mode surface S = 0 . The proof is completed.    □
Theorem 2.
The event-triggered rule (14) avoids the Zeno behavior, since a minimal triggering interval is given by
T u 1 ( k ) ν ( X b d b q ) n X ( X + b d + b q ) .
Proof. 
Define T u 1 ( k ) = t k + 1 t k as the interexecution time for u 1 . Recall that the control u 1 is updated at t = t k and e u 1 ( t k ) = 0 . During the event intervals
d e u 1 ( t ) d t d e u 1 ( t ) d t = d d t [ μ ( t ) u 1 ( t ) ] = d d t [ X sgn ( g T ( ξ k ) M T ( ξ k ) S ( ξ k ) ) + X sgn ( g T ( ξ ) M T S ( ξ ) ) ] d d t [ X sgn ( g T ( ξ ) M T S ( ξ ) ) ] .
The technique [39] is utilized to approximate the sign function sgn ( S ) . That is, sgn ( S ) tanh ( S / ν ) where ν 1 . Substituting M = g into (16), we have
d | | e u 1 ( t ) | | d t d d t [ X tanh ( g T ( ξ ) M T ( ξ ) S ) ] = X I n tanh 2 ( g T ( ξ ) M T ( ξ ) S ) 1 ν S ˙ .
As I n tanh 2 ( g T ( ξ ) M T ( ξ ) S ) I n = n , using (7), we have
d e u 1 ( t ) d t ( n / ν ) X M ξ ˙ [ f + g u 0 ] ( n / ν ) X M g u 1 ( ξ k ) + d ( t ) + g k w ( n / ν ) X ( X + b d + b q ) .
After integrating both sides with respect to (17) for t [ t k , t k + 1 ) , we get
e u 1 ( t t k ) ( n / ν ) X ( X + b d + b q ) .
Using the event-triggered rule (14), when t [ 0 , τ ) , one has
| | e u 1 | | X b d b q .
For t [ t k , t k + 1 ) , there is t t k t k + 1 t k = T u 1 ( k ) . Then, combined with (18) and (19), one has
T u 1 ( k ) ν ( X b d b q ) n X ( X + b d + b q ) .
The proof is completed.    □
Next, a dynamically triggered controller u 0 will be designed to suppress unmatched disturbances and guarantee the event-triggered stability of sliding manifold dynamics system (20) with a nearly optimal performance. We shall find the dynamic event-triggered condition and solve the optimal control problem using critic-only NN approximation strategy.

4. DET-Based Optimal Controller Design

The design u 0 comprises two steps. First, find the event-triggered rule to guarantee stability and optimal performance; second, solve the resulting optimal control problem approximately by using ADP and NN approximation strategy. With M = g , the sliding mode dynamics (11) can be revised as
ξ ˙ = f + g u 0 + k ¯ w
where k ¯ = ( I g g ) h .
The discounted cost function V ( ξ ) , which is subject to the above dynamics (20), is defined as
V ( ξ ) = t e α ( τ t ) ( ξ T Q ξ + u 0 T R u 0 γ 2 w T w ) d τ
where α > 0 is a discount factor and Q = Q 0 n × n 0 n × n 0 n × n . Moreover, Q R n × n and R R m × m are symmetric positive definite matrices to weight the system state and input, respectively. Meanwhile, if u 0 is admissible and V ( V = V ( ξ ) ) is C 1 , the corresponding nonlinear Bellman equation is
ξ T Q ξ + u 0 T R u 0 γ 2 w T w α V + V T ( f + g u 0 + k ¯ w ) = 0 .
Define the Hamiltonian
H ( ξ , u 0 , w , V ) ξ T Q ξ + u 0 T R u 0 γ 2 w T w α V + V T ( f + g u 0 + k ¯ w )
where V = V / ξ .
According to the zero-sum game [27], we get the optimal cost function V * ( ξ ) by
V * ( ξ ) = min u 0 max w V ( ξ , u 0 , w ) = max w min u 0 V ( ξ , u 0 , w ) ,
which satisfies the HJI equation
min u 0 max w H ( ξ , u 0 , w , V * ) = 0 .
Consider the stationarity conditions [25]:
H ( ξ , u 0 , w , V * ) / u 0 = 0 , H ( ξ , u 0 , w , V * ) / w = 0 .
Solving the stationary conditions (25), we obtain the optimal control input and worst-case disturbance
u 0 * = 1 2 R 1 g T V * , w * = 1 2 γ 2 k ¯ T V * .
Substituting the control input (26) into (22), the HJI equation is written as
ξ T Q ξ + V * T f α V * 1 4 V * T g R 1 g T V * + 1 4 γ 2 V * T k ¯ T k ¯ V * = 0 .

4.1. Dynamically Triggering Rule for Optimal Input

To propose a dynamically triggering rule, we define again a new sampling sequence { t i } i = 0 with t i < t i + 1 , i N and N = { 0 , 1 , 2 , } . Define the error between the sampled state ξ i and the current state ξ as
e i = ξ i ξ ( t ) , t [ t i , t i + 1 ) , i N
where ξ i ξ ( t ) | t = t i . Thus, for t [ t i , t i + 1 ) , the sampled-data version of the system (20) can be rewritten as
ξ ˙ ( t ) = f + g u 0 ( ξ + e i ) + k ¯ w .
Considering the event-based sampling rule, (26) is written as
u 0 * ( ξ i ) = 1 2 R 1 g T ( ξ i ) V * ( ξ i ) w * ( ξ ) = 1 2 γ 2 k ¯ T V * ( ξ )
where V * ( ξ i ) = V * ( ξ ) / ξ | ξ = ξ i .
The HJI equation with event-triggered law can be written as
ξ T Q ξ + V * T f α V * + 1 4 V * T ( ξ i ) g ( ξ i ) R 1 g T ( ξ i ) V * ( ξ i ) 1 2 V * T g R 1 g T ( ξ i ) V * ( ξ i ) 1 4 γ 2 V * T k ¯ T k ¯ V * = 0 .
Now, a necessary assumption is introduced for stability analysis.
Assumption 3
([36]). The optimal controller u 0 * is Lipschitz continuous on Ω,viz. there exists L > 0 , ∀ ξ , ξ ¯ Ω that satisfy
u 0 * ( ξ ) u 0 * ( ξ ¯ ) L ξ ¯ ξ .
We define an internal dynamic variable η evolving according to the following differential equation:
η ˙ = λ η + e T 2 2 L 2 R e i 2 , η ( 0 ) 0
where 0 < λ < 1 , and e T = ( 1 ϵ 2 ) λ m i n ( Q ) ξ 2 . Here, η is designed as a filtered value associated with e T 2 2 L 2 R e i 2 . This new DET technique can avoid e T 2 2 L 2 R e i 2 to be always nonnegative if the following condition is used
t 0 = 0 , t i + 1 = inf { t > t i | η + θ ( e T 2 2 L 2 R e i 2 ) 0 }
where θ R + . This is stated in the following result.
Lemma 2.
The dynamic variable η defined by (32), with the event triggered rule (33), is always non-negative.
Proof. 
Using the dynamic rule (33), for ∀ t [ 0 , + ) , one has
η + θ ( e T 2 2 L 2 R e i 2 ) 0 .
First, if θ = 0 , then η 0 is true.
Second, if θ 0 , by combining (32) and (34), one can deduce the following relation:
η ˙ + λ η = e T 2 2 L 2 R e i 2 η θ , η ( 0 ) 0 .
By the comparison lemma, for ∀ t [ 0 , + ) , one gets
η ˙ ( t ) η ( 0 ) e ( λ + 1 θ ) t .
Since the initial condition η ( 0 ) 0 and e ( λ + 1 θ ) t > 0 for ∀ t [ 0 , + ) , so one can obtain η 0 . The proof is completed.    □
Two main stability results for u 0 * follow.
Theorem 3.
Considering the sliding mode dynamics (20) with the optimal cost V * and (30). Let Assumptions 1–3 hold. The tracking error e d and the closed-loop system (29) achieve UUB via DET (33).
Proof. 
Choose W 1 = V * + η as the Lyapunov function for t [ t i , t i + 1 ) . With (30), taking the derivative of the Lyapunov function along the trajectory of (29) gives
W ˙ 1 = V ˙ * + η ˙ = ( V * ) T ( f + g u 0 * ( ξ i ) + k ¯ w * ) + η ˙ .
Note that (26) implies that
V * T g = 2 R u 0 * T , V * T = 2 γ 2 k ¯ w * T .
According to the time-triggered HJI (27), one has
V * T f = ξ T Q ξ u * 0 T R u 0 * + γ 2 w * T w * + α V * V * T g u 0 * V * T k ¯ w * ,
Thus, based on (38) and (39), we have
V ˙ * = V * T ( f + g u 0 * ( ξ i ) + k ¯ w * ) = ξ T Q ξ u 0 * T R u 0 * + γ 2 w * T w * + α V * V * T g u 0 * V * T k ¯ w * + V * T g u 0 * ( ξ i ) + V * T k ¯ w * = R u 0 * ( ξ ( t i ) ) u 0 * 2 + γ 2 w * T w * u 0 * T ( ξ i ) R u 0 * ( ξ i ) ξ T Q ξ + α V * ξ T Q ξ + 2 R L 2 e i 2 + α V * R | u 0 * ( ξ i ) 2 + γ 2 w * 2 ϵ 2 λ m i n ( Q ) ξ 2 + γ 2 w * 2 + 2 L 2 R e i 2 + α V * ( 1 ϵ 2 ) λ m i n ( Q ) ξ 2 .
Since V * is continuously differentiable on Ω , one can conclude that both V * and its derivative V ξ * are bounded on Ω . Here, we have max { V * , V * } b v * , where b v * R + is a constant. Recalling the triggering condition (33), one obtains
W ˙ 1 = V ˙ * + η ˙ = ϵ 2 λ m i n ( Q ) ξ 2 + 2 L 2 R e i 2 ( 1 ϵ 2 ) λ m i n ( Q ) ξ 2 + γ 2 w * 2 λ η + ( 1 ϵ 2 ) λ m i n ( Q ) ξ 2 2 L 2 R e i 2 + α V * ϵ 2 λ m i n ( Q ) ξ 2 + Λ
where Λ = α b v * + 1 4 γ 2 b k 2 b v * 2 and k ¯ b k ( b k R + ) .
Thus, we have W ˙ 1 < 0 once ξ > Λ ϵ 2 λ m i n ( Q ) is established.
Accordingly, the UUB of the closed-loop system is ensured. The proof is completed.    □
Next, we prove that the dynamic triggering rule (33) avoids the Zeno behavior.
Assumption 4
([36]). f + g u * ( ξ i ) is Lipschitz continuous with Lipschitz constants L f and L g , for all ξ and ξ i , one has
f + g u * ( ξ i ) = f + g u * ( ξ + e i ) L f ξ + L g e i .
Theorem 4.
Let Assumptions 1 and 4 hold. The dynamic triggering rule (33) avoids the Zeno behavior, and a minimal triggering interval is given by
T u 0 ( i ) 1 L g + L f ln ( η / θ + e T 2 ( L g + L f ) 2 L 2 R ( L f ξ ( t i ) + b k b w ) + 1 ) .
Proof. 
Define T u 0 ( i ) = t i + 1 t i as the interexecution time for u 0 . The control u 0 is updated at t = t i and e ˙ i = 0 . During the event intervals
ξ ˙ = f + g u 0 * ( ξ i ) + k ¯ w * = f + g u 0 * ( ξ i ) + k ¯ w * L f ξ + L g e i + b k b w .
For t [ t i , t i + 1 ) , we have ξ ( t ) = ξ ( t i ) e i and ξ ˙ = e ˙ i . From (43), we have
e ˙ i L f ξ ( t i ) e i + L g e i + b k b w .
Using the comparison lemma, we have
e i L f ξ ( t i ) + b k b w L g + L f ( e ( L g + L f ) ( t t i ) 1 ) .
Recall the triggering condition in (33) associated with
η + θ ( e T 2 2 L 2 R e i 2 ) 0 ,
and resulting in
| | e i | | 2 η / θ + e T 2 2 L 2 R .
For t [ t i , t i + 1 ) , one has t t i T u 0 , so
T u 0 ( i ) 1 L g + L f ln ( η / θ + e T 2 ( L g + L f ) 2 L 2 R ( L f ξ ( t i ) + b k b w ) + 1 ) ,
where η , θ are positive for all t > 0 . The proof is completed.    □

4.2. Dynamically Triggered ADP with Single Critic NN

The dynamic event-triggered optimal control u 0 * has been analyzed before. In the sequence, to guarantee UUB of sliding-mode dynamics (20), approximate optimal solution V * is to obtain by a critic-only NN approximation structure in the following part using reinforcement learning method.
Next, the solution V * of HJI (31) is approximated using NN using Weierstrass high-order approximation theorem:
V * = ω c T ϕ + ε
where ω c R l is the unknown ideal weight vector, ϕ R l is the activation function vector, l is the number of hidden neurons, and ε R is the critic NN approximation error.
The gradient of V * is
V * = ϕ T ω c + ε
where ϕ T = ϕ / ξ and ε = ε / ξ .
According to (26), (30), and (46), it is easy to obtain
u 0 * ( ξ i ) = 1 2 R 1 g T ( ξ i ) ( ϕ T ( ξ i ) ω c + ε ( ξ i ) ) , w * = 1 2 γ 2 k ¯ T ( ϕ T ω c + ε ) .
Because the ideal weights ω c are unknown, we first define the approximate expression of the optimal cost function V * as
V ^ * = ω ^ c T ϕ
where ω ^ c R l is the estimation of ω c .
According to (26) and (30), the approximate expression of (26) are
u ^ 0 ( ξ i ) = 1 2 R 1 g T ( ξ i ) ϕ T ( ξ i ) ω ^ c , w ^ = 1 2 γ 2 k ¯ T ϕ T ω ^ c .
Hence, the event-triggered ISMC becomes
u ( t ) = 1 2 R 1 g T ( ξ i ) ϕ T ( ξ i ) ω ^ c X sgn ( g T ( ξ k ) M T ( ξ k ) s ( t k ) ) .
Then, the approximation of the Hamiltonian is
H ^ ( ξ , u ^ 0 ( ξ i ) , w ^ , V ^ ) ξ T Q ξ + u ^ 0 T ( ξ i ) R u ^ 0 ( ξ i ) γ 2 w ^ T w ^ α V ^ + ω ^ c T ϕ ( ξ ) ( f + g u ^ 0 ( ξ i ) + k ¯ w ^ ) .
Owing to H ( ξ , u 0 * ( ξ i ) , w * , V * ) = 0 , we have
H ^ ( ξ , u ^ 0 ( ξ i ) , w ^ , V ^ ) H ( ξ , u 0 * ( ξ i ) , w * , V * ) = + ω ^ c T χ e H
where = ξ T Q ξ + u ^ 0 T ( ξ i ) R u ^ 0 ( ξ i ) γ 2 w ^ T w ^ α V ^ and χ = ϕ ( ξ ) ( f + g u ^ 0 ( ξ i ) + k ¯ w ^ ) .
Subsequently, a feasible training method is proposed by minimizing the residual error E = ( 1 / 2 ) e H 2 , combining codesign of the gradient descent technique and the experience replay (ER) technique, that is
ω ^ ˙ c = ζ χ ( 1 + χ T χ ) 2 ( χ T ω ^ c + ) ζ j = 1 l χ j ( 1 + χ j T χ j ) 2 ( χ j T ω ^ c + j ) ,
where ζ > 0 is the adaptive learning rate, κ = ( κ 1 , . . . , κ l ) T ,
χ j = ϕ ( ξ j ) ( f ( ξ j ) + g ( ξ j ) u ^ 0 ( ξ j ) + k ¯ ( ξ j ) ω ^ ( ξ j ) ) ,
for j { 1 , 2 , , l } . The term 1 ( 1 + χ T χ ) 2 is used for the purpose of normalization.
Based on the weight estimation error ω ˜ c = ω c ω ^ c , together with ω ˜ ˙ c = ω ^ ˙ c , one has
ω ˜ ˙ c = ζ χ ( 1 + χ T χ ) 2 ε H + ζ j = 1 l ζ χ i ( 1 + χ i T χ i ) 2 ε H i ζ χ χ T ( 1 + χ T χ ) 2 + ζ j = 1 l χ j χ j T ( 1 + χ j T χ j ) 2 ω ˜ c
where ε H = ε T ( f + g u ^ 0 ( ξ i ) + k ¯ ω ^ ) , and ε H j = ε ( ξ j ) ( f ( ξ j ) + g ( ξ j ) u ^ 0 ( ξ j ) + k ¯ ( ξ j ) ω ^ ) .

4.3. Stability Analysis

According to the critic-only NN strategy presented above, the ISMC can be obtained by solving the HJI Equation (31) approximately by DET way. Thus, the weight estimation error and the tracking error are proven to be UUB by the Lyapunov function in this section. To this end, the assumption 5 is necessary as follows for the following analysis.
Assumption 5
([36]). The NN approximation error and activation functions corresponding to their gradients are bounded, i.e., ε ( ξ ) b ε and ε ( ξ ) b ε ξ and ϕ ( ξ ) b ϕ and ϕ ( ξ ) b ϕ ξ .
In the sequence, an important theorem will be emerged to guarantee that the weight estimation error and the tracking error are UUB under dynamic event-triggered condition (33) and controller (49).
Theorem 5.
Considering the sliding-mode dynamics (20) with (49), let Assumptions 1–5 hold. The tracking error and the critic weight estimation error areUUBvia DET (33).
Proof. 
We will discuss two cases, i.e., the continuous dynamics and the jump dynamics. For ∀ t [ t i , t i + 1 ) , define the Lyapunov function L as
L = L ξ + L ξ i + L 0 + η = V * ( ξ ) + V * ( ξ i ) + ζ 1 2 t r w ˜ c T w ˜ c + η .
(1)
Event is not triggered:
Obviously, for ∀ t [ t i , t i + 1 ) , we have V ˙ * ( ξ i ) = 0 . Then,
L ˙ = L ˙ ξ + L ˙ 0 + η ˙ .
Taking the derivative of L ξ with (38) and (39) along the trajectory ξ ˙ = f + g u ^ 0 ( ξ i ) + k ¯ w ^ , we have
L ˙ ξ = V ˙ * ( ξ ) = V * T [ f + g u ^ 0 ( ξ i ) + k ¯ w ^ ] = ξ T Q ξ u * 0 T R u 0 * + γ 2 w * T w * + α V * + 2 R u * 0 T u 0 * 2 γ 2 w * T w * 2 R u * 0 T u ^ 0 ( ξ i ) + 2 γ 2 w * T w ^ = ξ T Q ξ + u * 0 T R u 0 * 2 R u * 0 T u ^ 0 * ( ξ i ) + u ^ 0 T ( ξ i ) R u ^ 0 ( ξ i ) + α V * γ 2 w * T w * + 2 γ 2 w * T w ^ u ^ 0 T ( ξ i ) R u ^ 0 * ( ξ i ) = ξ T Q ξ + R | | u 0 * u ^ 0 ( ξ i ) | | 2 + α V * γ 2 w * T w * + 2 γ 2 w * T w ^ u ^ 0 T ( ξ i ) R u ^ 0 ( ξ i ) λ m i n ( Q ) ξ 2 + α V * + R u 0 * u ^ 0 ( ξ i ) 2 R u ^ 0 ( ξ i ) 2 + 2 γ 2 w * T ( w ^ w * ) + γ 2 w * 2 λ m i n ( Q ) ξ 2 + α V * + R u 0 * u ^ 0 ( ξ i ) 2 R u ^ 0 ( ξ i ) 2 + 2 γ 2 w * T ( w ^ w * ) + γ 2 w * 2 λ m i n ( Q ) ξ 2 + α V * + R u 0 * u ^ 0 ( ξ i ) 2 R u ^ 0 ( ξ i ) 2 + γ 2 ( w ^ w * ) + 2 γ 2 w * 2 .
From (47) and (49), we have
u 0 * u ^ 0 ( ξ i ) 2 = u 0 * u 0 * ( ξ i ) + u 0 * ( ξ i ) u ^ 0 ( ξ i ) 2 2 u 0 * u 0 * ( ξ i ) 2 + 2 u 0 * ( ξ i ) u ^ 0 ( ξ i ) 2 = 2 u 0 * u 0 * ( ξ i ) 2 + R 1 g T ( ξ i ) [ ϕ T ( ξ i ) ( w ^ c w c ) + ε ( ξ i ) ] 2 = 2 | | u 0 * u 0 * ( ξ i ) | | 2 + R 1 g T ( ξ i ) [ ϕ T ( ξ i ) w ˜ c + ε ( ξ i ) ] 2 2 L 2 e i 2 + 2 ( R ) 2 b q 2 ( ϕ w ˜ c 2 + ε 2 ) ,
and
( w ^ w * ) = 1 2 γ 2 k T ( ξ ) [ ϕ T ( ξ ) ( ω ^ c ω c ) ε ( ξ ) ] = 1 2 γ 2 k T ( ξ ) [ ϕ T ( ξ ) w ˜ c ε ( ξ ) ] b k 2 γ 2 ( ϕ w ˜ c 2 + ε 2 ) .
Then, we further obtain
L ˙ ξ λ m i n ( Q ) ξ 2 R u ^ 0 ( ξ i ) 2 + 2 γ 2 w * 2 + 2 L 2 R e i 2 + 2 R 1 b q 2 ( ϕ w ˜ c 2 + ε 2 ) + b k 2 γ 2 ( ϕ w ˜ c 2 + ε 2 ) + α V *
and
L ˙ 0 = ζ w ˜ c T [ χ χ T ( 1 + χ T χ ) 2 + j = 1 l χ j χ j T ( 1 + χ T χ j ) 2 ] w ˜ c + ζ w ˜ c T [ χ ( 1 + χ T χ ) 2 ε H + j = 1 l χ j ( 1 + χ j T χ j ) 2 ε H j ] .
Because of
w ˜ c T χ ( 1 + χ T χ ) 2 ε H = 1 ( 1 + χ T χ ) 2 w ˜ c T χ ε H 1 2 ( 1 + χ T χ ) 2 ( w ˜ c T χ χ T w ˜ c + ε H ε H T ) 1 2 ( w ˜ c T χ χ T w ˜ c + ε H ε H T ) ,
and
w ˜ c T j = 1 l 1 ( 1 + χ j T χ j ) 2 ε H i = j = 1 l 1 ( 1 + χ j T χ j ) 2 w ˜ c T χ j ε H j j = 1 l 1 2 ( 1 + χ j T χ j ) 2 ( w ˜ c T χ j χ j T w ˜ c + ε H j ε H j T ) 1 2 j = 1 l w ˜ c T χ j χ j T w ˜ c + 1 2 j = 1 l ε H j ε H j T .
Using (57)–(59), it is easy to obtain
L ˙ 0 1 2 λ m i n ( Ψ ) w ˜ c 2 + l + 1 2 b ε H 2
where Ψ = χ χ T ( 1 + χ T χ ) 2 + j = 1 l χ j χ j T ( 1 + χ T χ j ) 2 .
With the aid of (56) and (60), we obtain
L ˙ ϵ 2 λ m i n ( Q ) ξ 2 ( 1 ϵ 2 ) λ m i n ( Q ) ξ 2 R u ^ 0 ( ξ i ) 2 + 2 γ 2 w * 2 + 2 L 2 R e i 2 + 2 R 1 g 2 ( ϕ w ˜ c 2 + ε 2 ) + α V * + b k 2 γ 2 ( ϕ w ˜ c 2 + ε 2 ) 1 2 λ m i n ( Ψ ) w ˜ c 2 + l + 1 2 b ε H 2 + η ˙ 2 ϵ 2 λ m i n ( Q ) ξ 2 + 2 L 2 R e i 2 R u ^ 0 ( ξ i ) 2 + 2 γ 2 w * 2 + α V * ( 1 ϵ 2 ) λ m i n ( Q ) ξ 2 ( 1 2 λ m i n ( Ψ ) 2 R 1 b q 2 ϕ 2 b k 2 γ 2 ) w ˜ c 2 + ( 2 R 1 b q 2 + b k 2 γ 2 ) b ε ξ 2 + l + 1 2 b ε H 2 + η ˙ 2 .
Recalling the triggering condition (33), one obtains
L ˙ ϵ 2 λ m i n ( Q ) ξ 2 + 2 L 2 R e i 2 ( 1 ϵ 2 ) λ m i n ( Q ) ξ 2 + α b v * λ η + ( 1 ϵ 2 ) λ m i n ( Q ) ξ 2 2 L 2 R e i 2 ( 1 2 λ m i n ( Ψ ) 2 R 1 b q 2 ϕ 2 b k 2 γ 2 ) w ˜ c 2 + l + 1 2 b ε H 2 + ( 2 R 1 b q 2 + b k 2 γ 2 ) b ε ξ 2 + 1 2 γ 2 b k 2 b v * 2 ϵ 2 λ m i n ( Q ) ξ 2 + l + 1 2 b ε H 2 + Π + Γ w ˜ c 2
where
Π = α b v * + ( 2 R 1 b q 2 + b k 2 γ 2 ) b ε ξ 2 + 1 2 γ 2 b k 2 b v * 2 , Γ = ( 1 2 λ m i n ( Ψ ) 2 R 1 b q 2 ϕ 2 b k 2 γ 2 ) .
Hence, (61) implies L ˙ < 0 only if one of the following inequalities holds
ξ > Π ϵ 2 λ m i n ( Q ) and w ˜ c > l + 1 2 b ε H 2 Γ .
(2)
On the triggering instant:
At the trigger moments t = t i , L ˙ is
L ˙ + ( t i ) = V * ( ξ + ) V * ( ξ i ) + V * ( ξ ^ i + 1 ) V * ( ξ ^ i ) + V * ( w ˜ c + ) V * ( w ˜ c ( ξ ^ i ) )
where V * ( w ˜ c + ) = lim h 0 + V * ( w ˜ c ) ( t i + h ) , V * ( w ˜ c ) = lim h 0 V * ( w ˜ c ) ( t i + h ) . From (56) and (57), we know V * ( ξ + ) V * ( ξ i ) < 0 and V * ( w ˜ c + ) V * ( w ˜ c ( ξ ^ i ) ) < 0 . So, we can get V * ( ξ ^ i + 1 ) V * ( ξ ^ i ) < 0 , so L ˙ < 0 .
Finally, UUB of the tracking error and of the weight approximation error are concluded. The proof is completed.    □

4.4. Algorithm Design of the Event-Triggered ISM Optimal Tracking Control

In the framework of all assumptions holding in this article, the following Algorithm 1 is to show the procedures of the event-triggered ISM optimal tracking control.
Algorithm 1 Event-Triggered ISM Optimal Tracking Control.
Input: initial states of the sliding mode dynamics (11)
1: Select an initial admissible policy u 0 0 ( X ) , ω 0 ( X ) and a proper small scalar ϵ > 0 .
ω ^ ˙ c = ζ χ ( 1 + χ T χ ) 2 ( χ T ω ^ c + ) ζ j = 1 l χ j ( 1 + χ j T χ j ) 2 ( χ j T ω ^ c + j ) .
while the optimal continuous control law with input constraints is approximated as
u ^ 0 ( ξ i ) = 1 2 R 1 g T ( ξ i ) ϕ T ( ξ i ) ω ^ c .
2: To tackle the uncertainty affecting the nonlinear system (5), an integral type sliding surface is designed as
S ( ξ ) = M [ ξ ξ 0 0 t f + g u 0 d τ ] .
3: To extend the continuous-time control u 1 in the event-triggered paradigm, we define a virtual control input μ ( t ) satisfying μ ( t ) u ( t k ) , t [ t k , t k + 1 ) , where
μ ( t ) = X sgn ( g T M T S ( ξ ( t k ) ) )
4: Hence, the event-triggered ISMC becomes
u ( t ) = 1 2 R 1 g T ( ξ i ) ϕ T ( ξ i ) ω ^ c X sgn ( g T ( ξ k ) M T ( ξ k ) s ( t k ) ) .
Remark 3.
An optimal tracking composite controller u = u 0 + u 1 , subject to two different dynamic event-triggered conditions is presented in this paper. Subsequent numerical experiments (cf. Table 1 and Table 2 and Figure 1 and Figure 2, respectively) show that the proposed new algorithm not only can reduce the communication burden but also improve the speed of convergence.

5. Simulation

In this section, a nonlinear single link robot arm is considered to verify the effectiveness of the proposed algorithm.
Consider the following system dynamics [39]
θ ¨ = M g l G ˜ sin ( θ ( t ) ) D G ˜ θ ˙ + 1 G ˜ u ( t ) + h ¯ w ( t ) .
where θ ( t ) is the angle position of robot arm, and u ( t ) is the control input. Moreover, M is the mass of the payload, G ˜ is the moment of inertia, g is the acceleration of gravity, l is the length of the arm, and D is the viscous friction, where g , l , D are the system parameters and M , G ˜ are the design parameters. Set the values of the system parameters as g = 9.81 , D = 1 , and l = 1 , and the design parameters M and G ˜ are alterable. Assuming x 1 ( t ) = θ ( t ) and x 2 ( t ) = θ ˙ ( t ) . Herein, consider the effect of interference on the actuator, so the dynamics by selecting the system parameters can be written as
x 1 ˙ x 2 ˙ = x 2 4 x 1 0.5 x 2 + 0 0.04 ( u ( t ) + d ( t ) ) + 0.2 sin ( t ) + 2 x 2 0.2 sin ( t ) + 2 x 1 ) . w ( t )
Let us take the initial state x ( 0 ) = [ 0.02 ; 0.5 ] . The matched disturbance is taken as
d ( t ) = 0.5 cos ( 0.35 t ) , t < 60 , 0.5 sin ( 0.5 t ) , t 60 .
The desired trajectory is
x ˙ d = x d 1 + sin ( x d 2 ) 4.9 sin ( x d 1 ) 0.5 x d 2
with the initial state x d ( 0 ) = [ 0.1 ; 0.65 ] .
Therefore, based on (64)–(66), the augmented system is reformulated as
ξ ˙ = f ( ξ ) + g ( ξ ) ( u ( t ) + d ( t ) ) + h ( ξ ) ω ( t )
with
f ( ξ ) = x 2 + x d 1 sin ( x d 2 ) 4 x 1 0.5 x 2 + 4.9 sin ( x d 1 ) + 0.5 x d 2 x d 1 + sin ( x d 2 ) 4.9 sin ( x d 1 ) 3 0.5 x d 2 , g ( ξ ) = 0 0.04 0 0 , h ( ξ ) = 0.2 ( sin ( t ) + 2 x 2 ) 0.2 ( sin ( t ) + 2 x 1 ) 0 0 .
Based on Lemma 1, one has g = [ 0 ; 25 ; 0 ; 0 ] and k = ( I g g ) h = [ 0.1013 ; 0 ; 0 ; 0 ] . The integral sliding mode surface is as in (7) with the sliding mode gain X = 1 .
According to the ISMC u 1 , the sliding-mode dynamics with unmatched disturbances is obtained as
ξ ˙ = f ( ξ ) + g ( ξ ) u 0 ( t ) + k ¯ ( ξ ) ω ( t )
where k ¯ ( ξ ) = ( I g ( ξ ) g ( ξ ) ) h ( ξ ) .
For simulation, the parameters of the algorithm are chosen as Q = diag ( 100 , 100 , 0 , 0 ) , γ = 5 , ζ = 0.9 , ϵ = 0.5 . The parameters of the triggering condition are selected as η 1 ( 0 ) = 0.1 , θ 1 = 1 , η ( 0 ) = 0.1 , L = 10 , λ = 0.1 , θ = 1 and α = 0.5 . The initial NN weight is selected as ω c = [ 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 , 0 ] and the NN activation function is designed as ϕ ( ξ ) = [ ξ 1 2 , ξ 2 2 , ξ 3 2 , ξ 4 2 , ξ 1 ξ 2 , ξ 1 ξ 3 , ξ 1 ξ 4 , ξ 2 ξ 3 , ξ 2 ξ 4 , ξ 3 ξ 4 ] T .
To make a comparison with DET via PE, during the neural network implementation process, a small probing noise is injected in u for the first 90 s. Figure 1 presents the evolution of the tracking system states and the augmented system states with DET via PE technique. Figure 2 presents the evolution of the tracking system states and the augmented system states with DET-based ER technique. From Figure 1 and Figure 2, it is obvious that the convergence of states and tracking trajectories via ER method is faster than the PE method. The optimal control u 0 , the ISMC law u 1 , the composite law u, and the sliding mode function S are shown in Figure 3 and Figure 4, for DET via PE technique and DET via ER technique, respectively. From Figure 3 and Figure 4, we can see that the convergence via ER method is faster than the PE method. The initial weights are selected randomly in the interval [ 0 , 1 ] . After a learning process, Figure 5 presents the convergence process of the critic NN weights ω c . The evolution of the event error | | e i | | 2 and triggering threshold | | e T | | 2 are shown in Figure 6. The interevent time of the control u 0 under under the four strategies is shown in Figure 7. Next, the characteristics of the different strategies are analyzed by comparison through Table 1, Table 2 and Table 3.
First, Table 1 shows the data comparison for the four control strategies. As is well known, the number of triggering events is one of most important factors in evaluating the triggering mechanism. Furthermore, a smaller number of triggering events can reduce the communication burden and save resources. To implement this goal, it is needed to reduce the unnecessary update of the controller guaranteeing system performance. Based on this, a novel adaptive adjustment technique consisting of DET via ER technique is used. In addition, with the help of the simulation experiment platform and MATLAB data package, using respectively the technology in [39,40] and our technique in this paper, some experimental results are shown in Table 1. From Table 1, it is obvious that the DET via ER technique is the best, and which can better reduce the communication burden and save resources, since there are only 418 samples that occurred because of the larger average triggering interval time 0.2392 s. In particular, one may notice that the ER technique is also beneficial in reducing the number of event-triggering comparing with the PE technique in the framework of the same event-triggered conditions. Similarly, the DET is also beneficial in reducing the number of event-triggering comparing with ET in the framework of the same PE technique (ER technique). In addition, the minimal interval indicates Zeno-free behaviors. The minimal interval values of the four strategies are consistent with Figure 7. Moreover, according to Figure 7, we know that DET via ER technique can generate a biggest minimal interval comparing with other technique, so the effectiveness of the current technique designed in this paper is verified.
Next, we define another important factor to evaluate the control strategy, (called a triggering rate based on TT). The triggering rate is calculated by DET number TT number . When θ = 1 and λ = 0.01 , Table 2 and Table 3 present that the rate of DET via PE and the rate of DET via ER are 23.75 % and 20.08 % , respectively. Generally speaking, a small triggering rate is more favorable than a large triggering rate. To investigate the characteristics of the changes utilized to execute the triggering rule, some data of the triggering rates are shown in Table 2 and Table 3.The tables also show the effect of the parameters on the triggering rate generated by θ and λ with DET via PE and ER respectively. To summary, a bigger θ value means a reduction of the number of events; in contrast, a bigger λ value means an increase in the number of events.
Remark 4.
ET PE is ET via PE technique, ET ER is ET via ER technique, DET PE is DET via PE technique, and DET ER is DET via ER technique.

6. Conclusions

In this article, a learning-based event-triggered optimal tracking control technique for nonlinear systems was developed via ADP. Matched uncertainties have been eliminated by the ISMC proposed in this paper. Unmatched uncertainties have been attenuated, utilizing projection matrix and an optimal controller. A critic NN via a novel dynamic event-triggered rule has been constructed to ensure the existence of the solution of the HJI equation and all errors have been proved UUB using the Lyapunov analysis method. Moreover, the simulation results revealed that our control algorithm is more favorable than traditional event-triggered control algorithms. Future work is to further explore this framework, e.g., in the presence of various delays or system constraints.
Remark 5.
(1) The pros: This study first provides codesign of dynamic event mechanism and experience replay technique-based weighted adaptive adjustment technique to guarantee the optimal approximation of the cost function V * . In addition, this codesign technique not only speeds up the approximation speed of the cost function but also enlarges the average interval of internal events aiming to reduce the updating times of the controller and save the calculation burden and communication resources.
(2) The cons: In the proof Theorem 2, using the approximate the sign function sgn ( S ) tanh ( S / ν ) where ν 1 that makes us to only obtain an approximation condition rather than a sufficient condition, namely, which weakens the effect of sliding-mode control. This issue should be investigated further.

Author Contributions

Conceptualization, W.T. and H.W.; methodology, W.T. and H.W.; validation, W.T. and H.W.; formal analysis, W.Y.; investigation, W.T. and H.W.; writing original draft preparation, W.T.; writing review and editing, W.Y. and H.W.; visualization, W.T. and H.W.; supervision, W.Y.; project administration, W.Y.; funding acquisition, W.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Science and Technology Project of State Grid Zhejiang Electric Power CO., LTD. (5211JY19000X).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Isermann, R. Digital Control Systems; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2013. [Google Scholar]
  2. Astrom, K.J.; Bernhardsson, B.M. Comparison of Riemann and Lebesque sampling for first order stochastic systems. In Proceedings of the 41st IEEE Conference on Decision and Control, Las Vegas, NV, USA, 10–13 December 2002; pp. 2011–2016. [Google Scholar]
  3. Tabuada, P. Event-triggered real-time scheduling of stabilizing control tasks. IEEE Trans. Autom. Control 2007, 52, 1680–1685. [Google Scholar] [CrossRef] [Green Version]
  4. Girard, A. Dynamic Triggering Mechanisms for Event-Triggered Control. IEEE Trans. Autom. Control 2015, 60, 1992–1997. [Google Scholar] [CrossRef] [Green Version]
  5. Ding, L.; Han, Q.L.; Ge, X.; Zhang, X.M. An overview of recent advances in event-triggered consensus of multi-agent systems. IEEE Trans. Cybern. 2018, 48, 1110–1123. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Peng, C.; Li, F. A survey on recent advances in event-triggered communication and control. Inf. Sci. 2018, 457, 113–125. [Google Scholar] [CrossRef]
  7. Zhang, X.M.; Han, Q.L.; Zhang, B.L. An overview and deep investigation on sampled-data-based event-triggered control and filtering for networked systems. IEEE Trans. Ind. Inform. 2016, 13, 4–16. [Google Scholar] [CrossRef]
  8. Pan, Y.; Yang, G.H. Event-triggered fault detection filter design for nonlinear networked systems. IEEE Trans. Syst. Man Cybern. Syst. 2008, 48, 1851–1862. [Google Scholar] [CrossRef]
  9. Ye, C.; Song, Y. Event-Triggered Prescribed Performance Control for a Class of Unknown Nonlinear Systems. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 6576–6586. [Google Scholar]
  10. Ge, X.; Han, Q.; Zhang, X.M.; Ding, D. Dynamic event-triggered control and estimation: A survey. Int. J. Autom. Comput. 2021, 18, 857–886. [Google Scholar] [CrossRef]
  11. Ge, X.; Han, Q.L.; Ding, L.; Wang, Y.L.; Zhang, X.M. Dynamic event-triggered distributed coordination control and its applications: A survey of trends and techniques. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 3112–3125. [Google Scholar] [CrossRef]
  12. Niu, Y.; Ho, D.W.C.; Lam, J. Robust integral sliding mode control for uncertain stochastic systems with time-varying delay. Automatica 2005, 41, 873–880. [Google Scholar] [CrossRef]
  13. Roy, S.; Baldi, S.; Fridman, L.M. Adaptive sliding mode control without a priori bounded uncertainty. Automatica 2020, 111, 1–6. [Google Scholar] [CrossRef]
  14. Roy, S.; Roy, S.B.; Lee, J.; Baldi, S. Overcoming the Underestimation and Overestimation Problems in Adaptive Sliding Mode Control. IEEE/ASME Trans. Mechatron. 2019, 24, 2031–2039. [Google Scholar] [CrossRef] [Green Version]
  15. Yu, W.; Wang, H.; Cheng, F.; Yu, X.; Wen, G. Second-Order Consensus in Multiagent Systems via Distributed Sliding Mode Control. IEEE Trans. Cybern. 2017, 47, 1872–1881. [Google Scholar] [CrossRef] [PubMed]
  16. Corradini, M.L.; Cristofaro, A. Nonsingular terminal sliding-mode control of nonlinear planar systems with global fixed-time stability guarantees. Automatica 2018, 95, 561–565. [Google Scholar] [CrossRef]
  17. Li, H.; Shi, P.; Yao, D. Adaptive sliding-mode control of Markov jump nonlinear systems with actuator faults. IEEE Trans. Autom. Control 2017, 64, 1933–1939. [Google Scholar] [CrossRef]
  18. Castan˜os, F.; Fridman, L. Analysis and Design of Integral Sliding Manifolds for Systems With Unmatched Perturbations. IEEE Trans. Autom. Control 2006, 51, 853–858. [Google Scholar] [CrossRef] [Green Version]
  19. Truc, L.N.; Vu, L.A.; Thoan, T.V.; Thanh, B.T.; Nguyen, T.L. Adaptive Sliding Mode Control Anticipating Proportional Degradation of Actuator Torque in Uncertain Serial Industrial Robots. Symmetry 2022, 14, 957. [Google Scholar] [CrossRef]
  20. Xu, L.; Xiong, W.; Zhou, M.; Chen, L. A Continuous Terminal Sliding-Mode Observer-Based Anomaly Detection Approach for Industrial Communication Networks. Symmetry 2022, 14, 124. [Google Scholar] [CrossRef]
  21. Yan, H.; Zhang, H.; Zhan, X.; Wang, Y.; Chen, S.; Yang, F. Event-Triggered Sliding Mode Control of Switched Neural Networks With Mode-Dependent Average Dwell Time. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 1233–1243. [Google Scholar] [CrossRef]
  22. Zheng, B.C.; Yu, X.; Xue, Y. Quantized feedback sliding-mode control: An event-triggered approach. Automatica 2018, 91, 126–135. [Google Scholar] [CrossRef]
  23. Nair, R.R.; Behera, L.; Kumar, S. Event-triggered finite-time integral sliding mode controller for consensus-based formation of multi-robot systems with disturbances. IEEE Trans. Control Syst. Technol. 2019, 27, 39–47. [Google Scholar] [CrossRef]
  24. Zhou, K.; Doyle, J.C.; Glover, K. Robust and Optimal Control; Prentice-Hall: Englewood Cliffs, NJ, USA, 1996. [Google Scholar]
  25. Vincent, T.L. Nonlinear and Optimal Control Systems; Wiley: New York, NY, USA, 1997. [Google Scholar]
  26. Lewis, F.L.; Vrabie, D.; Syrmos, V.L. Optimal Control; Wiley: New York, NY, USA, 2012. [Google Scholar]
  27. Basar, T.; Bernard, P. H-Optimal Control and Related Minimax Design Problems; Birkhäuser: Boston, MA, USA, 1995. [Google Scholar]
  28. Vrabie, D.; Lewis, F. Neural network approach to continuous-time direct adaptive optimal control for partially unknown nonlinear systems. Neural Netw. 2009, 22, 237–246. [Google Scholar] [CrossRef] [PubMed]
  29. Bellman, R.E. Dynamic Programming; Princeton University Press: Princeton, NJ, USA, 1957. [Google Scholar]
  30. Wang, F.Y.; Zhang, H.; Liu, D. Adaptive dynamic programming: An introduction. IEEE Comput. Intell. Mag. 2009, 4, 39–47. [Google Scholar] [CrossRef]
  31. Michailidis, I.; Baldi, S.; Kosmatopoulos, E.B.; Ioannou, P.A. Adaptive Optimal Control for Large-Scale Nonlinear Systems. IEEE Trans. Autom. Control 2017, 62, 5567–5577. [Google Scholar] [CrossRef] [Green Version]
  32. Liu, D.; Xue, S.; Zhao, B.; Luo, B.; Wei, Q. Adaptive Dynamic Programming for Control: A Survey and Recent Advances. IEEE Trans. Syst. Man Cybern. Syst. 2021, 51, 142–160. [Google Scholar] [CrossRef]
  33. Fan, Q.Y.; Yang, G.H. Adaptive actor-critic design-based integral sliding-mode control for partially unknown nonlinear systems with input disturbances. IEEE Trans. Neural Netw. Learn. Syst. 2016, 27, 165–177. [Google Scholar] [CrossRef]
  34. Qu, Q.; Zhang, H.; Yu, R.; Liu, Y. Neural network-based H sliding mode control for nonlinear systems with actuator faults and unmatched disturbances. Neurocomputing 2018, 275, 2009–2018. [Google Scholar] [CrossRef]
  35. Zhang, H.; Qu, Q.; Xia, G.; Cui, Y. Optimal guaranteed cost sliding mode control for constrained-input nonlinear systems with matched and unmatched disturbances. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2112–2126. [Google Scholar] [CrossRef]
  36. Vamvoudakis, K.G. Event-triggered optimal adaptive control algorithm for continuous-time nonlinear systems. IEEE/CAA J. Autom. Sin. 2014, 1, 282–293. [Google Scholar]
  37. Zhang, Q.; Zhao, D.; Zhu, Y. Event-triggered H control for continuous-time nonlinear system via concurrent learning. IEEE Trans. Syst. Man Cybern. Syst. 2017, 47, 1071–1081. [Google Scholar] [CrossRef]
  38. Xue, S.; Luo, B.; Liu, D. Event-Triggered Adaptive Dynamic Programming for Unmatched Uncertain Nonlinear Continuous-Time Systems. IEEE Trans. Neural Netw. Learn. Syst. 2020, 99, 1–13. [Google Scholar] [CrossRef] [PubMed]
  39. Yang, D.S.; Li, T.; Xie, X.P.; Zhang, H.G.; Liu, D.R.; Li, Y.H. Event-Triggered Integral Sliding-Mode Control for Nonlinear Constrained-Input Systems With Disturbances via Adaptive Dynamic Programming. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 2168–2216. [Google Scholar] [CrossRef]
  40. Zhang, H.G.; Liang, Y.; Su, H.; Liu, C. Event-Driven Guaranteed Cost Control Design for Nonlinear Systems with Actuator Faults via Reinforcement Learning Algorithm. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 4135–4150. [Google Scholar] [CrossRef]
  41. Van, M.; Do, X.P. Optimal adaptive neural PI full-order sliding mode control for robust fault tolerant control of uncertain nonlinear system. Eur. J. Control. 2020, 54, 22–32. [Google Scholar] [CrossRef]
Figure 1. PE technique: (a,b) resulted trajectory compared with the desired trajectory with DET; (c) state trajectory.
Figure 1. PE technique: (a,b) resulted trajectory compared with the desired trajectory with DET; (c) state trajectory.
Symmetry 14 01264 g001
Figure 2. ER technique: (a,b) resulted trajectory compared with the desired trajectory with DET; (c) state trajectory of augmented systems.
Figure 2. ER technique: (a,b) resulted trajectory compared with the desired trajectory with DET; (c) state trajectory of augmented systems.
Symmetry 14 01264 g002
Figure 3. PE technique: (a) evolution of RL-based control law u 0 ; (b) evolution of SMC input u 1 ; (c) evolution of composite control law u; (d) evolution of sliding mode surface S.
Figure 3. PE technique: (a) evolution of RL-based control law u 0 ; (b) evolution of SMC input u 1 ; (c) evolution of composite control law u; (d) evolution of sliding mode surface S.
Symmetry 14 01264 g003
Figure 4. ER technique: (a) evolution of RL-based control law u 0 ; (b) evolution of SMC input u 1 ; (c) evolution of composite control law u; (d) evolution of sliding mode surface S.
Figure 4. ER technique: (a) evolution of RL-based control law u 0 ; (b) evolution of SMC input u 1 ; (c) evolution of composite control law u; (d) evolution of sliding mode surface S.
Symmetry 14 01264 g004
Figure 5. Weight curves of system states during learning stage.
Figure 5. Weight curves of system states during learning stage.
Symmetry 14 01264 g005
Figure 6. Trigger threshold | | e T | | 2 and event error | | e i | | 2 during learning stage.
Figure 6. Trigger threshold | | e T | | 2 and event error | | e i | | 2 during learning stage.
Symmetry 14 01264 g006
Figure 7. Inter-event time of four strategies during learning stage.
Figure 7. Inter-event time of four strategies during learning stage.
Symmetry 14 01264 g007
Table 1. Comparison results of four strategies.
Table 1. Comparison results of four strategies.
StrategiesSamplesAverage IntervalMinimal Interval
TT 20000.050.05
ET PE 6320.15820.1
DET PE 4760.21010.1
ET ER 5770.17330.1
DET ER 4180.23920.15
Table 2. Trigger rates of different dets based on PE.
Table 2. Trigger rates of different dets based on PE.
Triggering Rate (%) λ 0.010.10.51
θ
122.3523.7527.4529.00
317.9522.428.5530.15
515.5522.2529.2030.65
1012.5522.8530.0531.10
Table 3. Trigger rates of different dets based on ER.
Table 3. Trigger rates of different dets based on ER.
Triggering Rate (%) λ 0.010.10.51
θ
0.124.9025.1025.4525.65
0.522.1022.902525.80
117.1020.8025.2526.60
31.753.626.3027.50
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tan, W.; Yu, W.; Wang, H. Dynamic Event-Triggered Integral Sliding Mode Adaptive Optimal Tracking Control for Uncertain Nonlinear Systems. Symmetry 2022, 14, 1264. https://0-doi-org.brum.beds.ac.uk/10.3390/sym14061264

AMA Style

Tan W, Yu W, Wang H. Dynamic Event-Triggered Integral Sliding Mode Adaptive Optimal Tracking Control for Uncertain Nonlinear Systems. Symmetry. 2022; 14(6):1264. https://0-doi-org.brum.beds.ac.uk/10.3390/sym14061264

Chicago/Turabian Style

Tan, Wei, Wenwu Yu, and He Wang. 2022. "Dynamic Event-Triggered Integral Sliding Mode Adaptive Optimal Tracking Control for Uncertain Nonlinear Systems" Symmetry 14, no. 6: 1264. https://0-doi-org.brum.beds.ac.uk/10.3390/sym14061264

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop