Content area
This article addresses the problem of fault-tolerant control in nonlinear time-delay systems using adaptive dynamic programming. An adaptive neural network observer is developed to estimate unknown dynamics, system states, and actuator faults. This observer is then transformed into an augmented structure for optimal fault-tolerant control problem. The gains of this observer are determined by solving a linear matrix inequality. A new value function index is introduced to account for time-delay states, and control law is derived associated with this novel value function. The Hamilton–Jacobi–Bellman equation for this value function is solved via a critic neural network. Lyapunov functional analysis demonstrates that the closed-loop system remains uniformly ultimately bounded. Simulation results validate the proposed fault tolerant approach. The key contribution of this paper lies in incorporating time-delay states into the adaptive dynamic programming value function in the presence of actuator faults.
Introduction
In contemporary industrial applications and technological systems, control approaches must consider system safety and reliability. This is especially critical in safety-focused applications, where mission success and the protection of lives and the environment are essential. Consequently, the importance of guaranteeing system safety and reliability has reached unprecedented levels [1, 2]. Since modern systems consist of many components, such as sensors and mechanical actuators, then they may fail over time. Failures can lead to system instability and degrade control performance in systems. Actuator failures, in particular, are especially critical because they can significantly impair control performance due to their unpredictable actions. To address this issue in controller design, fault-tolerant control (FTC) strategies should be implemented [3, 4].
The primary goal of a FTC approach is to adjust the controller structure so the system can manage component failures without compromising the mission or endangering users and the public. This often requires incorporating dynamic fault estimation into the nominal control input, which presents significant challenges in ensuring system stability with the modified control input [5].
FTC methods are generally divided into two categories: passive and active. Passive FTC methods rely on robust control theory, maintaining a constant controller structure and tolerating only a limited number of predefined conditions and assumptions. While passive approaches are more conservative, they are easier to design and implement. In contrast, active FTC methods adapt to faults by re-configuring the controller based on fault detection and isolation information. These approaches are less conservative and impose fewer conditions on the system, but proving stability and achieving acceptable control performance with active methods is more challenging. Great works on both active and passive fault approaches have been published; the reader can refer to [6, 7–8] for more details.
This paper introduces an active FTC approach for nonlinear time-delay systems using adaptive dynamic programming (ADP). Various methods for addressing active FTC problems have been published, including adaptive control approaches [9, 10–11], model predictive control [12, 13], and reinforcement learning with ADP algorithms [14, 15]. Among these various techniques, ADP algorithm is particularly effective for FTC challenges in nonlinear systems. An ADP approach is a powerful method for addressing optimization and optimal control problems by utilizing the principle of optimality [16]. For example, in work [17], a novel cost function is developed to address challenges in solving optimal tracking control problems for certain nonlinear systems with known dynamics using ADP. This innovative approach introduces a specific cost function related to tracking errors and their derivatives, eliminating the need for the traditional zero equilibrium assumption and simplifying the controller design process for nonlinear systems with nonzero equilibrium. ADP, an intelligent control method, has gained significant interest in both academia and industry since its proposal [18].
In [19], the authors proposed an online fault compensation control strategy using ADP algorithm to achieve optimal control in nonlinear systems with actuator faults. This method notably avoids using an action neural network. Expanding on this, [20] introduces an observer to enhance the value function. Unlike the approach in [19], the improved value function in [20] simultaneously addresses both faults and control, providing a more integrated solution.
Building on the findings from [19] and [20], the authors in [21] developed an FTC approach that tackles both actuator and sensor faults using neuro-dynamic programming. In [22], an optimal FTC method for a class of unknown nonlinear discrete-time systems with actuator faults was introduced, employing adaptive critic design. The study in [23] utilized a sliding-mode control technique integrated with a ADP algorithm to solve the optimal FTC problem for affine nonlinear systems. However, the works in [21, 22, 14, 19, 20], and [23] did not address nonlinear systems with time delays.
Consequently, a primary motivation of our work is to develop a fault tolerant method using ADP algorithm for nonlinear systems with time delays. Nearly all the previously mentioned works depend on information from nominal systems and assume the availability of system states. Observers, particularly learning observers, are crucial and play an important role in designing control approaches. Then, another motivation is to develop a fault-tolerant method using the ADP algorithm for nonlinear time-delay systems, based on a neural network observer.
In [24], a fault identification approach for partially unknown nonlinear systems is proposed using a combination of the deterministic learning method and the adaptive high gain observer. However, this paper does not consider time delay or adaptive dynamic programming. Building on these results, [25] proposes an adaptive neural observer-based FTC approach for a class of nonlinear systems using ADP. Most works utilizing neural network observers focus on fault estimation and detection, and methods for neural network observer-based FTC are scarce.
In all the aforementioned works, nonlinear time-delay systems were not considered. Time delays are critical in many engineering systems. Industrial processes, both linear and nonlinear, often involve time delays. In control terminology, these delays are categorized as constant time delays and time-varying delays. Constant time delays, common in slow-changing processes like chemical reactors, represent the maximum tolerable delay. Time-varying delays, typical in fast-paced systems like power systems and mobile robot teams, can cause instability or reduced performance. For example, in distributed optimization problems, delays can increase the time needed for systems to converge to an optimal solution [26]. To mitigate these effects, researchers incorporate delay considerations to ensure the stability of their controllers [27]. A widely used approach for this purpose is based on Lyapunov-Krasovskii methods [28].
Addressing time-delay systems in ADP is particularly challenging due to their infinite-dimensional nature, which complicates their solution. Analyzing systems with time delays is significantly more complex than analyzing systems without delays. Currently, there is no specific method for handling nonlinear time-delay systems. For more information, see chapters 4 and 5 of [29].
Based on the discussions and to the best of our knowledge, the issue of FTC based on neural network observer for nonlinear time-delay systems using an ADP algorithm has not been previously studied. The main contributions of this paper are as follows:
An adaptive neural network observer for nonlinear time-delay systems is designed to learn the unknown system dynamics and actuator faults. This observer is integrated into a FTC approach, enhancing its adaptability to stochastic faults through active intervention.
An novel value function for neural network observer-based FTC problem using ADP algorithm: Unlike existing ADP algorithms for fault tolerant in nonlinear systems, this paper specifically targets nonlinear time-delay systems. We introduce a new value function for our ADP algorithm, which integrates state time delays into the dynamics of nonlinear systems. Compared to [19, 24, 25] and [20], our method includes time delays in the control law updates and considers their impact to ensure the stability of the proposed approach.
Stability analysis using bilinear matrix inequality (BMI): We introduce a BMI to consider the impact of time delay on the stability of the proposed method. This BMI helps analyze the effect of constant time delay on the stability of the FTC problem based on the ADP algorithm for nonlinear time-delay systems.
Notation: The symbols , , and denote the sets of all real numbers, n-dimensional real vectors, and real matrices, respectively. The notation represents the Euclidean vector norm. The functions and indicate the minimum and maximum eigenvalues, respectively.
Mathematical formulation
Consider a class of partially unknown nonlinear time-delay systems represented as follows:
1
the variables and denote the state vectors of the system at time and at a delayed time , respectively. The vector represents the measured output, while signifies the control input vector. The term denotes the unknown actuator fault. The matrices , , , and are known real and constant. The term h(t) represents the historical state of the system and d is time-delay constant. The pair is controllable, and the pair is observable. The function is an unknown nonlinear function with Lipschitz continuity.The following assumption must be defined to proceed with the proposed method in this paper.
Assumption 1
The partial derivative of the approximation error in the critic neural network, denoted as , the actuator fault , and the function are all constrained by their respective norms. Specifically, the norm of the gradient is bounded by , the norm of the function is bounded by , and the norm of the actuator fault is bounded by . Here, , , and are positive constants.
In Eq. (1), the dynamics and the fault are unknown. To handle these unknown nonlinear terms and the actuator fault, we will use a critic neural network approximation.
Radial basis function neural networks
Radial basis function (RBF) networks have the capability to approximate any function universally. This approximation can be expressed as , where represents a vector of RBF neurons, denotes a vector of network weights, and is the input vector. Let be a trajectory confined within a bounded set, and let denote the localized area along this trajectory . According to [30], by leveraging the neurons within , for any smooth nonlinear function and for any given tolerance , there exists an optimal weight vector such that:
2
where denotes a vector of RBF neurons within .Control objective
The aim of this paper is to develop a control strategy using neural network observers and the ADP algorithm for system (1). This strategy is designed to ensure stability despite actuator faults and time delays. The neural network observer allows for the estimation of system states and unknown dynamics. We consider the unknown nonlinear function and the fault as a combined term, expressed as , and assume the detectable condition is met. The control laws are formulated to incorporate the influence of time delays, guaranteeing system stability. In particular, the control law accounts for the integral of the time-delay history. We propose a new value function that includes the history of time delays.
Adaptive neural network observer
Neural network observers excel in handling nonlinear dynamics, a key advantage over traditional linear observers. While linear observers are limited to small ranges around specific operating points, neural networks can approximate complex, nonlinear functions. This allows them to learn nonlinear dynamics from data, resulting in more accurate state estimation across a wider range of operating conditions. Such capability is particularly valuable in fault-tolerant control problems, where the ability to detect and isolate faults in nonlinear systems is critical for maintaining system stability and performance [24]. In [31], fuzzy logic systems are employed to model the unknown nonlinear dynamics of the pure-feedback system. These systems are particularly well-suited for this task because they can represent complex, nonlinear relationships through a set of simple "if-then" rules derived from expert knowledge or data. However, while fuzzy logic systems are flexible, their rule base can be relatively static unless specifically designed to adapt. This static nature can limit their ability to learn complex, high-dimensional patterns compared to neural networks, which can automatically learn representations from data.
Based on (2), an adaptive neural network observer for system (1) is designed as follows:
3
and the neural network identifier as4
in which and represent the estimates of the system state z(t) and the output measurements y(t), respectively. The observer gain matrices are and . The parameters and are defined. The term , and is the estimate of . The RBF neural network is used to approximate the unknown function .Remark 1
Existing NN-based control methods rely on adaptive techniques, with NN weight update laws designed using a Lyapunov function. However, this approach only guarantees convergence of NN weights, while other NN parameters like widths and centers remain fixed and cannot be updated online. In contrast, gradient descent (GD)-based NN techniques can achieve better approximation performance by updating all NN parameters to satisfactory values through cost function minimization via the GD algorithm. In [32], a novel control strategy for nonlinear systems with unknown dynamics is introduced, utilizing a neural network-based approach that combines gradient descent (GD) with the Barzilai-Borwein (BB) method. This GD-BB-based control algorithm enhances adaptability by updating neural network parameters and learning rates online, simplifying the design process and reducing the need for manual tuning. Thus, using the proposed method in [32] for estimating unknown nonlinear dynamics could be a very interesting approach for fault-tolerant control problems.
For simplicity, the notation t is omitted in the equations that follow.
Let’s define the state observation error as and the neural network identifier observation error as . Then, based on Eqs. (1) and (3), we obtain the following.
5
The following theorem is presented to show that the state observation error (5) is uniformly ultimately bounded.Theorem 1
Given the provided observer (3) and Assumption 1, if there exist symmetric positive definite matrices E and , positive scalar , and matrix such that the following LMIs hold:
6
wherethe observer gain and are obtained as and . Consequently, the state estimation error and the identifier estimation error are uniformly ultimately bounded.Proof
The following Lyapunov function candidate is proposed:
7
where8
Given (5) and (7), we derive the following:9
10
11
In (9) and using Young’s inequality, we obtain the following:12
First, let us examine the scenario where the neural network approximation error . Then, we derive the following result:13
Utilizing Young’s inequality and Assumption 1, we derive the following result:14
here denotes the upper bound of , that is . By assuming and defining the variable15
where and is defined in (6). According to (15), if , then, it leads to .In other words, and will remain bounded over time. Thus, the proof is concluded.
Inspiring [24], Theorem 1 is obtained and the estimation errors for the states converge to zero when . Thus, the convergence of the state estimate error and the neural network weight estimate errors can be assured. By using the result of Theorem 1, the identification error systems (3) and (4) are uniformly ultimately bounded, and are also bounded.
Designing optimal controller based on neural network observer
From the designed adaptive neural network observer (5), it is clear that system (1) is replaced by (3), thus transforming (1) into:
16
where .If an optimal control policy is found for Eq. (16), it guarantees the convergence of system (1). This paper presents an augmented system to reframe the problem of fault tolerant as an optimal regulation problem. By merging Eqs. (4) and (16), the following observer system is provided as follows:
17
introduce the augmented variables and matrices as:Then, system (17) can be represented as18
Next, we will derive the control law, assuming that the control input is considered admissible.Controller formulation
The proposed control strategy leverages the ADP framework to solve optimal control problems using neural network observer. We employ a critic neural network to approximate the value function associated with the Hamilton–Jacobi–Bellman equation, training it to minimize temporal difference error. This approach, supported by Lyapunov functional analysis, ensures the convergence of the system states and neural network weights, thereby guaranteeing the stability and optimality of the control strategy.
System (1) is controllable on the set . Therefore, the goal is to find a feedback control policy , that minimizes the infinite horizon performance index function, expressed as:
19
here, the utility function is given by. and are positive definite matrices.
Consider , defined as:This function satisfies and is nonnegative for all and . The expression (19) takes into account time delays.
Next, we will prove that using the new value function (19) to design a controller can stabilize the closed-loop system.
Let be the optimal value function corresponding to , defined as follows:
20
By applying the Bellman optimality principle, we can determine the gradient of . This gradient, denoted as , satisfies the following equation:21
where is known as the Hamiltonian function [33, 34]. The Hamiltonian function associated with the cost function (19) is given by:22
Our goal is to minimize Eq. (22) by combining Eqs. (22) and (21) to derive the optimal control law, denoted as . This optimal control law can be obtained from as follows:23
From (23), a simple transformation yields:24
Equation (24) will be needed later.By substituting Eq. (23) in Eq. (22), the HJB equation can be rewritten as follows:
25
It is essential to show that the control law derived in Eq. (23) can stabilize the system described by Eq. (1). This is established in the following theorem.Theorem 2
Consider system (1), Theorem 1, and the control protocol given by Eq. (23). The control strategy outlined in Eq. (23) can ensure that nonlinear time-delay system (1) remains uniformly ultimately bounded, provided that there exist positive definite matrices and , along with free-weighting matrices and , such that the following BMI is satisfied:
26
in whichProof
The following Lyapunov function is considered:
27
Based on the definition of , we have when and when . This indicates that is positive definite, which consequently implies that is also positive definite. Following this, the derivative of the Lyapunov function (27) is evaluated along the trajectory of , yielding the following expression:28
Calling (22), we obtain:29
By substituting (29) and (23) in (28), Eq. (28) can be reformulated as follows:30
By using Jensen’s inequality [28], let us define the free-weighting matrices and . The term in (30) can be expressed as follows:31
in whichReferring to (30) and applying (31), we get:32
whereAccording to (32), if , then holds, leading to . This proves that the system (1) is asymptotically stable, thus completing the proof.It is essential to delve deeper into the discussion regarding the BMI (26). As we are aware, solving this BMI is a challenging task. To address this difficulty and analyze the permissible delays, we can employ the following expression to assess the feasibility of the BMI:
33
where34
Remark 2
The Lyapunov-Krasovskii function, , is used for analyzing the stability of time-delay systems and is incorporated into the value function (19). Its purpose in the cost function is to account for the effects of time delays on the system, allowing this information to be used to update the control law. As demonstrated in Theorem 2, integrating the Lyapunov-Krasovskii function with time delays into the cost function helps ensure the system’s stability.
According to (26), incorporates the time delay . The BMI condition helps identify the permissible delays. In other words, since the feasibility of is dependent on , we can determine the maximum allowable delays. In the subsequent stages, a single-layer neural network to tackle the task of solving the HJB equation is provided.
Neural network algorithm
We know that neural networks are great at handling complex functions. Since the performance index function (19) is usually quite complicated and does not have an exact mathematical expression, we will use a neural network to estimate it. In this paper, we use a simple one-layer neural network to do this. Then, following [35], is expressed as
35
here, consider as the activation function, where signifies the real-valued space with dimension . denotes the optimal weight vector, and represents the number of neurons in the hidden layer. stands for the approximation error of the neural network. The gradient of Eq. (35) concerning can then be expressed as:36
let denote the gradient of the activation function, and represent the gradient of the approximation error.When we substitute Eq. (36) into (21), we get:
37
So, the Hamiltonian can be stated as:38
In this context, signifies the residual error stemming from the neural network approximation. Since the ideal weight vector is unknown, we utilize the critic neural network to estimate as:39
As a result, we can formulate the gradient of as:40
hence, we can derive the approximate Hamiltonian as:41
We define the weight approximation error as . Utilizing (41) and (38), we obtain:42
We can revise the weight approximation error as:43
To update the weight vector of the critic neural network, we minimize the objective function using the normalized gradient algorithm. The weight vector should be updated as follows:44
in this expression, denotes the learning rate of the critic neural network. Therefore, considering (23) and (35), we can describe the ideal control policy as:45
and it can be estimated as:46
The control policy (46) solely relies on the critic neural network. This means that updating the weight vector of the critic neural network using (44) eliminates the need to train the action neural network. As a result, this approach is not only feasible but also computationally efficient for implementation.Theorem 3
If the weights of the critic neural network are updated according to (43) for the system described in (18), the error in weight approximation is uniformly ultimately bounded.
Proof
For this theorem, we used the following Lyapunov candidate:
47
The time derivative of (47) is48
Hence, holds when belongs to the compact set defined by , assuming , where is a positive constant. Then, we can conclude that the weight approximation error is upper uniformly bounded, thus completing the proof.The ADP algorithm based on neural network observer associated with the approach proposed in this paper is detailed in Algorithm 1.
[See PDF for image]
Algorithm 1
Fault tolerant for a class of nonlinear time delay systems using ADP
Remark 3
For the computational requirements and efficiency of using a critic neural network in real-time applications, several factors need to be considered. These include the complexity of the neural network, the hardware capabilities, and the specific requirements of the real-time system. Real-time applications require low-latency responses, so the neural network should be optimized to produce outputs within the application’s time constraints.
Simulation numerical examples
This section presents numerical simulation examples to illustrate the effectiveness of the proposed method.
The simulation results were obtained using Julia programming, and the reader can download them from this link
Example: Chemical Reaction Tank
The chemical reactor systems consist of multiple chemical reactors and stirring bars, which introduce delays into the systems. The state-space expression of the chemical reaction is given as follows:
49
The chemical reaction parameters have been set as follows: , , , , , , , , , and .According to (49):We assume that the actuator fault dynamic is as follows:In this example, the initial values are chosen as , and other parameters are set as follows: , , and . The observer’s parameters are obtained from (6) as follows:An adaptive neural network observer is designed to estimate the system states and the unknown function. The RBF network employs nodes, with neuron centers distributed across the range . The initial weight matrix is set to . The time delay is assumed to be . The weight vector in critic network learning is denoted as , with its initial values given by . The dimension of is determined by . Additionally, the activation function for the critic neural network is specified as . The network comprises neurons, and the critic weight vectors are represented by .
The simulation results are shown in Figs. 1, 2, 3, 4 and 5. Figure 1 illustrates that the states of the nonlinear system (1) are stabilized under the control policy (46). When an actuator fault happens at , there is a noticeable change in the system states, but the controller still performs well. This indicates that the system, with control policy (46), can handle actuator faults. Figure 2 displays the control input (46) during simulation time. In our theoretical development, we focused on ensuring robustness and fault tolerance, which may sometimes introduce oscillations due to the conservative nature of the design. However, in practical implementations, these oscillations can be mitigated through damping techniques, filtering, and careful parameter tuning to ensure smooth control signals.
Figure 3 shows the performance of fault estimation and unknown dynamics, demonstrating that the proposed ADP-based approach can accurately and effectively estimate faults. Furthermore, Fig. 4 presents the critic weight vectors, indicating their convergence. For plotting , we used the norm due to the large size of , which is a vector with 121 columns.
[See PDF for image]
Fig. 1
System states and observer estimations
[See PDF for image]
Fig. 2
The control input (46)
[See PDF for image]
Fig. 3
The fault and unknown dynamic estimation performance
[See PDF for image]
Fig. 4
The weights of the critic neural network
The value of the cost function (19) is shown in Fig. 5. The integral part, which is related to the time delay, is also depicted in Fig. 5, indicating that when the states of the system converge to the equilibrium point, the rate of change of the cost function becomes constant.
[See PDF for image]
Fig. 5
The value of cost function using the proposed method
Conclusion
In this paper, we have proposed a FTC method for partially unknown nonlinear systems with time delays, employing adaptive dynamic programming. The neural network observer was implemented to estimate states, unknown dynamics, and actuator fault of the system. The observer gains were determined by solving a linear matrix inequality. A new value function index was introduced to account for time-delay states, and control laws were derived based on this value function. The Hamilton–Jacobi–Bellman equation for this value function was solved using a critic neural network. Lyapunov functional analysis demonstrated that the closed-loop system is uniformly ultimately bounded.
A promising future direction would be to extend our adaptive neural network observer and time-delay inclusive ADP approach to decentralized control scenarios for interconnected nonlinear systems with multiple fault types, as investigated in [32, 36].
Acknowledgements
None.
Author Contributions
Farshad Rahimi contributed to formulation, writing, simulation, methodology, mathematics, review, and editing.
Funding
The authors claim that there are no funds, grants, or other support devoted to the preparation of this manuscript.
Data availability
The simulation results were obtained using Julia programming, and the reader can download them from [this link](https://farshad-rahimi.github.io/FarshadRahimi//Learning%20Julia/).
Declarations
Conflict of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
References
1. Ni, Q; Ding, F; Zhan, Z; Liu, J; Li, X; Zhao, Z. A fault diagnosis and measurement method for temperature measurement circuits in electric traction drive system. Measurement; 2024; 227, 114250. [DOI: https://dx.doi.org/10.1016/j.measurement.2024.114250]
2. Isapour, S; Tavakoli-Kakhki, M. Rough neural network based data-driven model-free adaptive fault-tolerant control for discrete-time nonlinear systems. Eng Appl Artif Intell; 2024; 134, 108651. [DOI: https://dx.doi.org/10.1016/j.engappai.2024.108651] 07964803
3. Gao, Z; Cecati, C; Ding, SX. A survey of fault diagnosis and fault-tolerant techniques-part i: Fault diagnosis with model-based and signal-based approaches. IEEE Trans Industr Electron; 2015; 62,
4. Li, H; Yang, C; Hu, Y; Zhao, B; Zhao, M; Chen, Z. Fault-tolerant control for current sensors of doubly fed induction generators based on an improved fault detection method. Measurement; 2014; 47, pp. 929-937. [DOI: https://dx.doi.org/10.1016/j.measurement.2013.10.021] 1313.03023
5. Amin, AA; Hasan, KM. A review of fault tolerant control systems: advancements and applications. Measurement; 2019; 143, pp. 58-68. [DOI: https://dx.doi.org/10.1016/j.measurement.2019.04.083] 1227.97065
6. Jia, F; Cao, F; He, X. Active fault-tolerant control against intermittent faults for state-constrained nonlinear systems. IEEE Trans Syst, Man, Cybern: Syst; 2024; 54, pp. 2389-2401. [DOI: https://dx.doi.org/10.1109/TSMC.2023.3344292] 07966208
7. Hu, J; Sha, Y; Yao, J. Dual neural networks based active fault-tolerant control for electromechanical systems with actuator and sensor failure. Mech Syst Signal Process; 2023; 182, 109558. [DOI: https://dx.doi.org/10.1016/j.ymssp.2022.109558] 1137.86312
8. Ke, C; Cai, K-Y; Quan, Q. Uniform passive fault-tolerant control of a quadcopter with one, two, or three rotor failure. IEEE Trans Robot; 2023; 39, pp. 4297-4311. [DOI: https://dx.doi.org/10.1109/TRO.2023.3297048] 1548.93088
9. Shen, Q; Jiang, B; Shi, P. Fault diagnosis and fault-tolerant control based on adaptive control approach; 2017; Berlin, Springer: 1417.93024
10. Zhang, X-N; Li, X-J. Adaptive fault-tolerant control for a class of stochastic nonlinear systems with multiple sensor faults. Int J Syst Sci; 2020; 51,
11. Shen, Q; Shi, P; Lim, CP. Fuzzy adaptive fault-tolerant stability control against novel actuator faults and its application to mechanical systems. IEEE Trans Fuzzy Syst; 2024; 32, pp. 2331-2340. [DOI: https://dx.doi.org/10.1109/TFUZZ.2023.3343403] 07902394
12. Camacho EF, Alamo T, Peña DM (2010) Fault-tolerant model predictive control. In: 2010 IEEE 15th Conference on Emerging Technologies & Factory Automation (ETFA 2010), IEEE, pp. 1–8
13. Bavili, RE; Khosrowjerdi, MJ et al. Active fault tolerant controller design using model predictive control. J Control Eng Appl Inf; 2015; 17,
14. Wang, L; Qi, R; Jiang, B. Adaptive fault-tolerant optimal control for hypersonic vehicles with state constrains based on adaptive dynamic programming. J Franklin Inst; 2024; 361,
15. Zhao, Y; Wang, H; Xu, N; Zong, G; Zhao, X. Reinforcement learning-based decentralized fault tolerant control for constrained interconnected nonlinear systems. Chaos, Solitons & Fractals; 2023; 167, 113034.4525496 [DOI: https://dx.doi.org/10.1016/j.chaos.2022.113034] 1452.65068
16. Bardi, M; Dolcetta, IC et al. Optimal control and viscosity solutions of Hamilton–Jacobi–Bellman equations; 1997; Berlin, Springer: [DOI: https://dx.doi.org/10.1007/978-0-8176-4755-1] 0890.49011
17. Wang, T; Wang, Y; Yang, X; Yang, J. Further results on optimal tracking control for nonlinear systems with nonzero equilibrium via adaptive dynamic programming. IEEE Trans Neural Netw Learn Syst; 2021; 34,
18. Liu, D; Xue, S; Zhao, B; Luo, B; Wei, Q. Adaptive dynamic programming for control: a survey and recent advances. IEEE Trans Syst, Man, and Cybernet: Syst; 2020; 51,
19. Zhao, B; Liu, D; Li, Y. Online fault compensation control based on policy iteration algorithm for a class of affine non-linear systems with actuator failures. IET Control Theory & Appl; 2016; 10,
20. Zhao, B; Liu, D; Li, Y. Observer based adaptive dynamic programming for fault tolerant control of a class of nonlinear systems. Inf Sci; 2017; 384, pp. 21-33. [DOI: https://dx.doi.org/10.1016/j.ins.2016.12.016] 1432.93185
21. Zeng, C; Zhao, B; Liu, D. Fault tolerant control for a class of nonlinear systems with multiple faults using neuro-dynamic programming. Neurocomputing; 2023; 553, 126502. [DOI: https://dx.doi.org/10.1016/j.neucom.2023.126502] 1432.93185
22. Wang, Z; Liu, L; Wu, Y; Zhang, H. Optimal fault-tolerant control for discrete-time nonlinear strict-feedback systems based on adaptive critic design. IEEE Trans Neural Netw Learn Syst; 2018; 29,
23. Fan, Q-Y; Yang, G-H. Adaptive fault-tolerant control for affine non-linear systems based on approximate dynamic programming. IET Control Theory & Appl; 2016; 10,
24. Chen, T; Zeng, C; Wang, C. Fault identification for a class of nonlinear systems of canonical form via deterministic learning. IEEE Trans Cybern; 2021; 52,
25. Zeng C, Zhao B, Liu D (2023) Adaptive neural network observer-based fault tolerant control for partially unknown nonlinear systems via adaptive dynamic programming. In: 2023 IEEE 13th International Conference on CYBER Technology in Automation, Control, and Intelligent Systems (CYBER), IEEE, pp. 212–216
26. Rahimi, F; Mahboobi Esfanjani, R. Estimating tolerable communication delays for distributed optimization problems in control of heterogeneous multi-agent systems. IET Control Theory & Appl; 2024; 18,
27. Zhao, W; Zhang, H; Yu, W. Distributed optimal output synchronization of heterogeneous multiagent systems with communication delays. Int J Robust Nonlinear Control; 2024; 34, pp. 7821-7836.4773657 [DOI: https://dx.doi.org/10.1002/rnc.7369] 1543.93021
28. Fridman E (2014) Introduction to time-delay systems: analysis and control. In: Systems & control: foundations & applications. Cham: Springer International Publishing. https://doi.org/10.1007/978-3-319-09393-2. Website: https://link.springer.com/book/10.1007/978-3-319-09393-2
29. Zhang, H; Liu, D; Luo, Y; Wang, D. Adaptive dynamic programming for control: algorithms and stability; 2012; Berlin, Springer: 1279.49017
30. Wang, C; Hill, DJ. Deterministic learning and rapid dynamical pattern recognition. IEEE Trans Neural Networks; 2007; 18,
31. Wang, Y; Wang, T; Li, C; Yang, J. Neural network-based optimal fault-tolerant control for interconnected nonlinear systems with actuator failures. IEEE Trans Emerg Top Comput Intell; 2024; 8, pp. 1828-1840. [DOI: https://dx.doi.org/10.1109/TETCI.2024.3358981] 1369.93680
32. Wang, Y; Wang, T; Yang, X; Yang, J. Gradient descent-Barzilai Borwein-based neural network tracking control for nonlinear systems with unknown dynamics. IEEE Trans Neural Netw Learn Syst; 2021; 34,
33. Al-Tamimi, A; Lewis, FL; Abu-Khalaf, M. Discrete-time nonlinear HJB solution using approximate dynamic programming: convergence proof. IEEE Trans Syst, Man, and Cybernet, Part B (Cybernetics); 2008; 38,
34. Lewis, FL; Vrabie, D; Syrmos, VL. Optimal control; 2012; Hoboken, John Wiley & Sons: [DOI: https://dx.doi.org/10.1002/9781118122631] 1284.49001
35. Zhang, L; Zhao, X; Zhao, N. Real-time reachable set control for neutral singular Markov jump systems with mixed delays. IEEE Trans Circuits Syst II Express Briefs; 2021; 69,
36. Rahimi, F; Rezaei, H. A distributed fault estimation approach for a class of continuous-time nonlinear networked systems subject to communication delays. IEEE Control Syst Lett; 2021; 6, pp. 295-300.4454081 [DOI: https://dx.doi.org/10.1109/LCSYS.2021.3071478] 1476.37013
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2024.