Next Article in Journal
Application of Entropy-Based Metrics to Identify Emotional Distress from Electroencephalographic Recordings
Previous Article in Journal
Empirical Laws and Foreseeing the Future of Technological Progress
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Single Neuron Stochastic Predictive PID Control Algorithm for Nonlinear and Non-Gaussian Systems Using the Survival Information Potential Criterion

1
College of Information Engineering, Taiyuan University of Technology, Taiyuan 030024, China
2
Department of Chemical Engineering, Chung-Yuan Christian University, Chung-Li, Taoyuan 32023, Taiwan, Republic of China
*
Author to whom correspondence should be addressed.
Submission received: 17 March 2016 / Revised: 14 May 2016 / Accepted: 26 May 2016 / Published: 3 June 2016
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
This paper presents a novel stochastic predictive tracking control strategy for nonlinear and non-Gaussian stochastic systems based on the single neuron controller structure in the framework of information theory. Firstly, in order to characterize the randomness of the control system, survival information potential (SIP), instead of entropy, is adopted to formulate the performance index, which is not shift-invariant, i.e., its value varies with the change of the distribution location. Then, the optimal weights of the single neuron controller can be obtained by minimizing the presented SIP based predictive control criterion. Furthermore, mean-square convergence of the proposed control algorithm is also analyzed from the energy conservation perspective. Finally, a numerical example is given to show the effectiveness of the proposed method.

Graphical Abstract

1. Introduction

Since almost all the control systems are subject to random signals (such as those originating from system parameter variations and sensor noise, etc.), stochastic systems are widely encountered in control engineering design. Minimizing the randomness in the closed-loop system is one of the important practical issues in controller design. Therefore, minimum variance control [1] has been obtained significant attentions. Its purpose is to minimize variations in the controlled system outputs or the tracking errors. Indeed, even today, most stochastic control design methods have only focused on control of the output mean and the variance of stochastic systems. In general, these developments are done mostly based on the assumptions that the system variables are of Gaussian types. Such assumptions, although strict, allow control engineers to make use of the well-established stochastic theory to perform controller design and closed-loop system analysis.
In industrial processes, product quality data can be approximated by the Gaussian probability density function (PDF) when the system operates normally. However, when abnormality occurs along the production line, the variabilities of these quality variables would not follow Gaussian distributions. In this regard, actions need to be taken so that the manipulated variables can be tuned to bring these quality variables back to certain desired ones. In fact, most industrial processes have difficulty meeting the Gaussian assumption because of the mixture of different courses with Gaussian disturbances or other factors. Moreover, the nonlinearity in stochastic systems could lead to non-Gaussian randomness even if the disturbances follow a Gaussian distribution. Thus, controlling the mean and variance of system variables may be far from sufficiently characterizing the statistical property of the stochastic processes. It is known that, in many cases, the behavior of a stochastic process can be completely characterized by the shape of its statistical distribution represented by PDF. Therefore, for analysis and design purposes, it is important to consider the entire PDF. PDF-shaping control design accounts for the issues mentioned above by selecting a certain shape for the process PDF as the goal of the control design procedure would provide an accurate and flexible control strategy that can accommodate a wide class of objectives.
In order to solve the problems existing in paper-making processes, the stochastic distribution control (SDC) theory was proposed by Wang (1996) [2]. This theory aims at controlling the shape of the output PDF instead of the mean and variance for stochastic systems [3,4,5]. After that, SDC has been used to handle the stochastic systems with non-Gaussian disturbances. Then, a linear-matrix-inequality- based convex optimization algorithm was developed for control, filter design and fault detection in non-Gaussian systems [6,7,8]. Nevertheless, the PDFs of the output are not necessarily measurable; instead, a more general measure of uncertainty, namely the entropy, has been used to characterize the uncertainty of the output tracking error. Compared with SDC (PDF shaping strategy) proposed earlier, it is more straightforward to us a minimum error entropy (MEE) based stochastic control algorithm to design a controller for tracking errors. It is more appropriate than the traditional minimum mean square error (MSE) criterion when dealing with nonlinearities and non-Gaussian disturbances. Shannon entropy is the most important and commonly used method in MEE based stochastic control [9,10,11]. A well-known generalization of Shannon entropy is Renyi entropy. When the order of Renyi entropy approaches 1, Renyi entropy will reduce and become Shannon entropy. The argument of the log in Renyi entropy is named the information potential (IP) and used as an alternative entropy criterion because of its monotonic property. Puya et al. [12] applied the minimum Renyi entropy control scheme to decrease the closed-loop randomness of the output under an iterative learning control (ILC) basis for general nonlinear and unknown non-Gaussian stochastic systems. Besides, information potential (IP) based dynamic neural networks were used to perform the modeling and control of the plant. In [13], the quadratic IP of tracking errors was employed to design controllers for nonlinear multivariate and non-Gaussian systems. ( h , ϕ ) -entropy is the most generalized definition of entropy [14]. ( h , ϕ ) -entropy has been employed in stochastic control systems [15,16]. Ren et al. [15] proposed a new tracking control algorithm for a class of networked control systems (NCSs) with non-Gaussian random disturbances and delays. Zhang et al. [16] presented an improved single neuron controller for multivariable stochastic systems with non-Gaussianities and unmodeled dynamics by minimizing ( h , ϕ ) -entropy of tracking errors.
The Shannon and order- α Renyi entropies of a continuous random variable are both defined based on the probability density function. This kind of entropy has several drawbacks. (1) The definition will be ill-suited for the case in which PDF does not exist; (2) the value can be negative; (3) the approximation using empirical distribution is impossible in general. And the IP criterion is conservative. It should be maximized to achieve smaller errors only when α > 1 . Some new definitions of entropy have been made to solve these problems. Rao et al. [17] proposed the cumulative residual entropy (CRE), which is defined based on the cumulative distribution function. In [18], Zografos and Nadarajah proposed the survival exponential and the generalized survival exponential entropies, both of which are broad entropy definitions based on the survival function and include CRE as a special case.
In this paper, survival information potential (SIP), proposed by Chen et al. [19], will be utilized to construct the performance index of stochastic control systems. Compared with the MEE criterion, adding a bias term to the tracking error [6,9,10,13] would not be necessary because of the shift-variance property of the SIP. In the previous work of MEE control [9,10,11,12,13,14,15,16], although the randomness of control input exists in practical conditions, the control input was considered as a deterministic variable for simplicity, which is unsuitable and conservative. Therefore, in this paper, the performance index with the integration of SIPs of the control input and the tracking error is proposed.
The predictive control idea was brought up as an industrial approach to process control in the 1970s. Today this technique is the most frequently applied advanced process control method in the industry. The stochastic distribution control algorithms have been extended by the advanced control algorithm to increase the control performance. So far mainly two classes of algorithms have been developed for systems affected by stochastic noise and subject to probabilistic state and/or input constraints. (1) The randomized, or scenario-based approach [20,21,22]: It is a very general methodology that can consider linear or nonlinear systems affected by noise with general distributions characterized by possibly unbounded and nonconvex support. (2) The probabilistic approximation approach [23,24,25,26,27,28]: It is based on the point-wise reformulation of probabilistic or expectation constraints in deterministic terms to be included in the MPC formulation. Reference [29] gave an overview of the main developments in the area of stochastic model predictive control (SMPC) in the past decade. It described different SMPC algorithms and the key theoretical challenges in stochastic predictive control without undue mathematical complexity. However, the above results were obtained under the assumption that the system variables obey Gaussian distribution, and only mean value and variance were considered. As presented above, PDF contains the whole characteristics of random variables. From this perspective, SMPC approaches to a class of nonlinear systems with unbounded stochastic uncertainties were proposed in [30,31]. In [30], the Fokker–Planck equation was used for describing the dynamic evolution of the states’ PDFs and the closed-loop stability was ensured by designing a stability constraint in terms of a stochastic control Lyapunov function. Polynomial chaos expansions were utilized to propagate the probabilistic parametric uncertainties through the system model in [31]. In the framework of SDC, SMPC was used to control the molecular weight distribution (MWD) [32,33] with the existence of non-Gaussian noises. As mentioned previously, considering the non-Gaussian SMPC in the statistical information framework may be an alternative and effective method.
Based on the preliminary work [19,34,35,36,37], in this paper, a single neuron stochastic predictive control method for nonlinear stochastic discrete systems affected by non-Gaussian noise is proposed. The proposed algorithm will be detailed in the rest of the sections. In Section 2, the models of the nonlinear stochastic system and single neuron controller are firstly presented. Then a new SIP-based predictive criterion, which contains both the randomness of tracking errors and the control input, is formulated. Based on the established models and the new criterion, the single neuron stochastic predictive control (SNSPC) algorithm is derived and the online computation procedure is also summarized. To analyze the convergence of the proposed control algorithm, the energy conversion principle is used in Section 3. A numerical simulation example is introduced to illustrate the efficiency of the proposed control strategy in Section 4. The last section concludes this paper.

2. Single Neuron Stochastic Predictive Control (SNSPC)

2.1. System Model and Single Neuron Controller

A general discrete, nonlinear dynamic system with process disturbances and measurement noises is considered here. It is described in the state-space form as:
{ x k + 1 = f ( x k , u k , ζ k ) y k = g ( x k , u k , ν k )
where x k n is the state, y k is the measured output, and u k is the control input. ζ k and ν k are external disturbances in the input and the measurement channels, respectively and they may not follow Gaussian distributions. f ( ) and g ( ) are known nonlinear functions that represent the system dynamics.
The goal of the control design is to find the optimal control input to make the output track the desired set-point r k as soon as possible. Due to the nonlinearity and disturbances involved in the system (1), the traditional PID control algorithm cannot have excellent performance here. In the past work, a single neuron adaptive PID controller has the advantages of neural network intelligent control and PID control. Single neuron adaptive PID controller achieves control system’s self-organization, adaptive function by adjusting weighting coefficients with the changes of the control parameters [35]. The controller is a multiple input and single output nonlinear processing unit with self-learning ability. The output of the single neuron adaptive PID controller is shown as Equation (2):
u k = u k 1 + K l = 1 3 w l k x l k / Σ k
where Σ k = l = 1 3 w l k , K > 0 is the proportional coefficient of the neuron. w l k ( l = 1 , 2 , 3 ) stands for the weight corresponding to each input. x l k ( l = 1 , 2 , 3 ) is the input of the neuron from the tracking error e k and it can be defined as:
{ x 1 k = e k x 2 k = e k e k 1 x 3 k = e k 2 e k 1 + e k 2
Remark 1. x l k ( l = 1 , 2 , 3 ) is the input of the neuron from the tracking error e k . The weights x l k ( l = 1 , 2 , 3 ) reflect the dynamic characteristics of the controlled object and the process response. The neuron weights can be updated through self-learning strategy, and the tracking error can be correspondingly eliminated and approach to the steady state effected by three kinds of control parts together.

2.2. SIP-Based Predictive Criterion

If noises ζ k and ν k are not Gaussian, the variance of the output tracking error e k = r k y k would not be sufficient to characterize the randomness of the tracking error dynamic systems. Therefore, an alternative measure of uncertainty, survival information potential (SIP), is given to construct the performance index for measuring the dispersion of the stochastic systems.
Firstly, the definition and properties of SIP are presented as follows.
Definition 1. 
Definition of SIP [19]: For a random vector X in m , SIP of order α ( α > 0 ) is defined by
S α ( X ) = + m F ¯ | X | α ( x ) d x
where F ¯ | X | ( x ) = P ( | X | > x ) = E [ I ( | X | > x ) ] is the multivariate survival function (or equivalently, the distribution function) of the random vector | X | , and + m = { x m : x = ( x 1 , , x m ) ,   x i 0 ,   i = 1 , , m } . Note that | X | > x means | X i | > x i , i = 1 , , m and I ( ) is the indicator function. Similarly, when α = 2 , SIP is called the quadratic survival information potential (QSIP).
Property 1. 
Property 1 of SIP [19]: S α ( X ) 0 , with equality if and only if P ( X = 0 ) = 1 .
Property 2. 
Property 2 of SIP [19]: Let X be an m-dimensional random vector, and let another m-dimensional random vector Y = ( Y 1 , Y 2 , Y m ) T with Y i = c i X i , c i , i = 1 , 2 , m , then S α ( Y ) = ( i = 1 m | c i | ) S α ( X ) .
Remark 2. 
SIP has some advantages over the IP: (1) It has consistent definition in the continuous and discrete domains; (2) it is not shift-invariant (i.e., its value would vary with the location of distribution); (3) it can be computed more easily from the sample data (without kernel computation and the choice of the kernel width), and the estimation asymptotically converges to the true value; (4) it is a more robust measure since the distribution function is more regular than the density function. (Note that the density is computed as the derivative of the distribution.)
Because of the nonlinearity and noises involved in the system, the tracking error is a non-Gaussian stochastic variable at a typical instant. Therefore, the control input is also non-Gaussian according to the single neuron controller Equations (2) and (3). However, in the previous work on stochastic distribution control [9,10,11,12,13,15,16], the control input is considered as a deterministic variable, which is conservative. And moreover, the multi-step predictive control strategy which can achieve better performance in industrial processes is adopted to design the optimal control input here.
The cost function of a SIP-based predictive controller for set point control is similar to that of a conventional predictive control with a continuous actor, but the SIP criterion is used in the control design instead of the MSE criterion
J = i = 1 P S α ( e k + i 1 ) + λ j = 1 M S α ( u k + j 1 )
where e k + i is the i-step ahead prediction of the system tracking error. P and M are the prediction horizon and the control horizon, respectively, and M P . It is assumed that the control variable would not change after M steps, i.e., u k + j 1 = u k + M 1    ( j > M ) .

2.3. SNSPC Algorithm

Based on the structure of the single neuron adaptive controller Equations (2) and (3) and the performance index Equation (5), the optimal weights of the single neuron can be solved by the stochastic gradient method
w k * = arg min W k 3 M   J = arg min w k 3 M { i = 1 P S α ( e k + i 1 ) + λ j = 1 M S α ( u k + j 1 ) }
where w k = [ w 1 k w 2 k w 3 k w 1 , k + 1 w 2 , k + 1 w 3 , k + 1 w 1 , k + M 1 w 2 , k + M 1 w 3 , k + M 1 ] T 3 M .
According to the definition of SIP, the optimal weights w k * can be calculated from
J w k = w k { i = 1 P + F ¯ | e k + i 1 | ( x ) d x + λ j = 1 M + F ¯ | u k + j 1 | ( z ) d z } = 0
where F ¯ | e k + i 1 | and F ¯ | u k + j 1 | are the survival functions of the tracking error e k + i 1 and the control input u k + j 1 , respectively.
In the practical application, it is generally difficult to develop a comprehensive first-principle model to describe the distribution of the error e k . Instead of the theoretical SIP, the data-driven empirical SIP used as the cost function would be useful alternative to timely online prediction of the errors. Given a sequence of error samples ( e 1 k , e 2 k , , e N ¯ k ) , assuming, without the loss of generality, that | e 1 k | | e 2 k | | e N ¯ k | , the empirical SIP would be [20]
S ^ α ( e k ) = j = 1 N ¯ μ j | e j k |
where μ j = ( N ¯ j + 1 N ¯ ) α ( N ¯ j N ¯ ) α .
Similarly, the empirical SIP of u k + j 1 , S ^ α ( u k + j 1 ) , can also be formulated. The empirical cost Equation (8) is the weighted sum of the ordered absolute errors. One drawback of this empirical SIP criterion is that it is not smooth at e j k = 0 and u j k = 0 . To address this problem, a more general cost can be defined using the empirical SIP of any mapped errors and inputs; that is,
J ¯ ( e k : k + P 1 ) = i = 1 P S ^ α ( ϕ ( e k + i 1 ) ) + λ j = 1 M S ^ α ( ϕ ( u k + j 1 ) ) = i = 1 P l = 1 N ¯ μ l ϕ ( e l , k + i 1 ) + λ j = 1 M l = 1 N ¯ σ l ϕ ( u l , k + j 1 )
where e k : k + P 1 = [ e k e k + 1 e k + P 1 ] T , μ l = ( N ¯ l + 1 N ¯ ) α ( N ¯ l N ¯ ) α . The mapping function ϕ ( ) usually satisfies
{ ( i ) positivity :        ϕ ( ) 0 ( i i ) symmetry :        ϕ ( x ) = ϕ ( x ) ( i i i ) monotonicity :     | x 1 | < | x 2 | ϕ ( x 1 ) ϕ ( x 2 )
Remark 3. As the data-driven empirical SIP-based performance index is not smooth at e j k = 0 and u j k = 0 , a proper function transformation method can be used to establish an equivalent SIP-based criterion, which can be formulated as Equation (9) and (10). In the simulation, the empirical SIP of the square errors ( e 1 k 2 , e 2 k 2 , , e N ¯ k 2 ) and the square control inputs ( u 1 k 2 , u 2 k 2 , , u N ¯ k 2 ) as an alternative adaptation cost is used here, given by
J ¯ = i = 1 P S ^ α ( e k + i 1 2 ) + λ j = 1 M S ^ α ( u k + j 1 2 ) = i = 1 P l = 1 N ¯ μ l e l + i 1 2 + λ j = 1 M l = 1 N ¯ σ l u l + j 1 2
The above cost is the weighted sum of the ordered square errors with the popular minimum mean square error (MSE) criterion as a special case (when α = 1).
Based on the general cost Equation (9), the weight update equation for adaptive system training is
w k + 1 = w k η J ¯ w k = w k η e k : k + P 1 T w k J ¯ e k : k + P 1
where η > 0 denotes the step-size (or adaptation gain).
According to the above presentation, the optimal input can be computed. The procedure below describes the steps of implementing SNSPC specifically.
Step 1: Initialize the weight vector of the single neuron w 0 . Choose the α value, step-size η , and the sliding window length N ¯ .
Step 2: Estimate the performance index using empirical SIPs of the output tracking error and the control input in (9).
Step 3: Solve the optimal weight vector w k * by (11).
Step 4: Compute the next control input
u k = u k 1 + K Σ k [ x 1 k x 2 k x 3 k ] [ I 3 × 3 0 3 × ( 3 P 3 ) ] w k *
and implement the input on the process.
Step 5: Collect the process outputs to update the SIPs of the output tracking error and the control input. Then repeat the procedure from Step 2 to Step 5 for the next time step, k = k + 1 .
Remark 4. 
The computations of the SIP-based performance index in Equation (9) are easier than the minimum error entropy (MEE) criterion, because the index is directly computed from the sample data without kernel computation and the choice of the kernel width, but the major computations of the partial derivative, e k : k + P 1 T w k , in the control algorithm Equation (12) would be done.

3. Mean-Square Stability

In order to analyze the convergence of the proposed control algorithm, the nonlinear stochastic system Equation (1) is firstly linearized as
{ Δ x k + 1 = f x | x k 1 Δ x k + f u | u k 1 Δ u k + f ζ | ζ k 1 Δ ζ k Δ y k = g x | x k 1 Δ x k + g u | u k 1 Δ u k + g ν | ν k 1 Δ ν k
where Δ u k = u k u k 1 , Δ x k = x k x k 1 = x ˜ k , Δ ζ k = ζ k ζ k 1 = ζ ˜ k , Δ y k = y k y k 1 = y ˜ k , and Δ ν k = ν k ν k 1 = ν ˜ k . It can be simply denoted as
{ x ˜ k + 1 = A k x ˜ k + B k Δ u k + N 1 k ζ ˜ k y ˜ k = C k x ˜ k + D k Δ u k + N 2 k ν ˜ k
where A k = f x | x k 1 , B k = f u | u k 1 , C k = g x | x k 1 , D k = g u | u k 1 , N 1 k = f ζ | ζ k 1 , N 2 k = g ν | ν k 1 .
Based on the state space representation, an extended state space model can be formulated as [36]
y k f = L k v m k p + L k u u k f + L k ς ς k f
where y k f = [ y ˜ k y ˜ k + M 1 ] T , u k f = [ Δ u k Δ u k + M 1 ] T , m k p = [ ( v 1 k p ) T ( v 2 k p ) T ] T , v 1 k p = [ y ˜ k M y ˜ k 1 ] T , v 2 k p = [ Δ u k M Δ u k 1 ] T . L k v , L k u and L k ς can be obtained from Equation (14); ς k f is the composite noise, including external noises ( ζ k and ν k ) and the model mismatch randomness, and it may be non-Gaussian noises.
From Equation (2), Equation (15) can be reformulated as
y k f = L k v m k p + L k u χ k w k + L k ς ς k f
where χ k is the input matrices of the single neural network, which consists of tracking errors:
χ k = [ z k 0 0 0 z k z k + 1 0 0 z k z k + 1 z k + M ]
where z k = [ K Σ k x 1 k K Σ k x 2 k K Σ k x 3 k ] and 0 = [ 0 0 0 ] .
Thus, the difference between the future predictions and the set-point trajectory is
e k f = L k v m k p + L k u χ k w k + L k ς ς k f Δ r k f
where Δ r k f = r k f r k 1 f is the change of the set-point at time point k and θ k = L k v m k p + L k ς ς k f Δ r k f , which is a new random vector; then Equation (17) can be rewritten as
e k f = L k u χ k w k + θ k
Now a priori error vector and a posteriori error vector, e k a f and e k p f , are defined as:
{ e k a f = L k u χ k w k e k p f = L k u χ k w k + 1
Obviously, e k a f and e k p f have the following relationship
e k p f = e k a f + L k u χ k ( w k + 1 w k )
By incorporating Equation (11), the following equation is gotten,
e k p f = e k a f L k u χ k η ( e k f ) T w k J ¯ e k f = e k a f + η L k u χ k χ k T L u T h ¯ ( e k f )
where h ¯ ( e k f ) = J ¯ e k f = [ J ¯ e k J ¯ e k + 1 J ¯ e k + M 1 ] T . R k = L k u χ k ( L k u χ k ) T is an M × M -dimensional symmetric matrix. Assume R k is invertible (i.e., det R k 0 ),
e k p f = e k a f + η R k h ¯ ( e k f ) R k 1 ( e k p f e k a f ) = η h ¯ ( e k f ) χ k T ( L k u ) T R k 1 ( e k p f e k a f ) = η χ k T ( L k u ) T h ¯ ( e k f ) χ k T ( L k u ) T R k 1 ( e k p f e k a f ) = ( w k + 1 w k )
And hence
w k + 1 = w k + χ k T ( L k u ) T R k 1 ( e k p f e k a f )
Both sides of Equation (23) should have the same energy; that is
w k + 1 T w k + 1 = [ w k + χ k T ( L k u ) T R k 1 ( e k p f e k a f ) ] T × [ w k + χ k T ( L k u ) T R k 1 ( e k p f e k a f ) ]            = w k T w k + 2 w k T χ k T ( L k u ) T R k 1 ( e k p f e k a f )              + ( e k p f e k a f ) T R k T ( L k u ) χ k χ k T ( L k u ) T R k 1 ( e k p f e k a f )
From R k = L k u χ k ( L k u χ k ) T , we have
w k + 1 T w k + 1 = w k T w k + 2 w k T χ k T ( L k u ) T R k 1 ( e k p f e k a f ) + ( e k p f e k a f ) T R k 1 ( e k p f e k a f )              = w k T w k + 2 w k T χ k T ( L k u ) T R k 1 ( e k p f e k a f ) + ( e k p f ) T R k 1 e k p f              2 ( e k a f ) T R k 1 e k p f + ( e k a f ) T R k 1 e k a f
Since R k 1 = ( L k u ) T χ k T χ k 1 ( L k u ) 1 , e k a f = L k u χ k w k , then
w k + 1 T w k + 1 = w k T w k + 2 w k T χ k T ( L k u ) T ( L k u ) T χ k T χ k 1 ( L k u ) 1 ( e k p f L k u χ k w k )              + ( e k p f ) T R k 1 e k p f 2 ( L k u χ k w k ) T ( L k u ) T χ k T χ k 1 ( L k u ) 1 e k p f + ( e k a f ) T R k 1 e k a f              = w k T w k + 2 w k T χ k 1 ( L k u ) 1 e k p f 2 w k T χ k 1 ( L k u ) 1 L k u χ k w k              + ( e k p f ) T R k 1 e k p f 2 w k T χ k 1 ( L k u ) 1 e k p f + ( e k a f ) T R k 1 e k a f            = w k T w k + ( e k p f ) T R k 1 e k p f + ( e k a f ) T R k 1 e k a f
Adding ( e k a f ) T R k 1 e k a f to both sides of the above equation and substituting e k a f = L k u χ k w k and R k 1 = ( L k u ) T χ k T χ k 1 ( L k u ) 1 into it, one can calculate the energy conservation relation of w k + 1 T w k + 1 and w k T w k :
   w k + 1 T w k + 1 + ( e k a f ) T R k 1 e k a f = w k T w k + ( e k p f ) T R k 1 e k p f + 2 ( L k u χ k w k ) T [ ( L k u ) T χ k T χ k 1 ( L k u ) 1 ] ( L k u χ k w k ) w k + 1 T w k + 1 + ( e k a f ) T R k 1 e k a f = w k T w k + ( e k p f ) T R k 1 e k p f w k + 1 2 + e k a f R k 1 2 = w k 2 + e k p f R k 1 2
where w k 2 = w k T w k , e k a f R k 1 2 [ e k a f ] T R k 1 e k a f , and e k p f R k 1 2 [ e k p f ] T R k 1 e k p f . To study the mean-square behavior of the algorithm, one takes expectations of both sides of (27) and write
E [ w k + 1 2 ] + E [ e k a f R k 1 2 ] = E [ w k 2 ] + E [ e k p f R k 1 2 ]
By substituting e k p f = e k a f + η R k h ¯ ( e k f ) into Equation (28), the following equation can be obtained.
E [ w k + 1 2 ] = E [ w k 2 ] + 2 η E [ ( e k a f ) T h ¯ ( e k f ) ] + η 2 E [ h ¯ T ( e k f ) R k h ¯ ( e k f ) ]
In order to evaluate the expectations E [ ( e k a f ) T h ¯ ( e k f ) ] and E [ h ¯ T ( e k f ) R k h ¯ ( e k f ) ] , the following assumptions [37] are used in this paper:
Assumption 1. 
The noise θ k is independent, identically distributed (i.i.d.), and independent of the input L k u χ k .
Assumption 2. 
The a priori error vector e k a f is jointly Gaussian distributed.
Assumption 3. 
The input vectors L k u χ k are zero-mean independent, identically distributed (i.i.d.).
Assumption 4. 
i j { 1 , , M } , R k ( i j ) is independent of { e i , e j } .
Based on the above assumptions, we have
E [ ( e k a f ) T h ¯ ( e k f ) ] = E [ ( e k a f ) T h ¯ ( e k a f + θ k ) ]                = ( 1 2 π σ k 2 ) N γ θ k ( θ k ) d θ k ×                ( e k a f ) T h ¯ ( e k a f L k ς ς k f ) i = 1 N exp ( ( e k + i 1 a ) 2 2 σ k 2 ) d e k a f
where γ θ k ( θ k ) is the density distribution of θ k and σ k = E [ ( e k + i 1 a ) 2 ] .
Therefore, the expectation E [ ( e k a f ) T h ¯ ( e k f ) ] can be expressed as a function of σ k 2 . Define the following function
h G ( σ k 2 ) E [ ( e k a f ) T h ¯ ( e k f ) ] σ k 2
and then yields
E [ ( e k a f ) T h ¯ ( e k f ) ] = h G ( σ k 2 ) σ k 2
We now evaluate the expectation E [ h ¯ T ( e k f ) R k h ¯ ( e k f ) ] .
E [ h ¯ T ( e k f ) R k h ¯ ( e k f ) ] = i = 1 N j = 1 N E [ h ¯ i ( e k f ) h ¯ j ( e k f ) R k ( i j ) ]                  = i = 1 N E [ h ¯ i 2 ( e k f ) ] E [ R k ( i i ) ]
where:
h I ( σ k 2 ) i = 1 N E [ h ¯ i 2 ( e k f ) ]
Then it is gotten:
E [ h ¯ T ( e k f ) R k h ¯ ( e k f ) ] = h I ( σ k 2 ) E [ R k ( i i ) ]
By substituting Equations (32) and (35) into Equation (29), the following equation is gotten.
E [ w k + 1 2 ] = E [ w k 2 ] + 2 η h G ( σ k 2 ) σ k 2 + η 2 h I ( σ k 2 ) E [ R k ( i i ) ]
Then the convergent condition for the sequence E [ w k 2 ] (i.e., E [ w k + 1 2 ] E [ w k 2 ] ) would be:
η 2 E [ R k ( i i ) ] sup k 0 { σ k 2 h G ( σ k 2 ) h I ( σ k 2 ) }
Remark 5. Since the nonlinear system can be approximated by a linear system at the equilibrium, the convergence of the nonlinear system Equation (1) using the linearizing method is studied here.

4. Simulation Results

In order to illustrate the efficiency of the presented SNSPC algorithm, consider a nonlinear stochastic system described by
{ x k + 1 = 1 ( 1 + x k ) 2 + 1.2 u k + ζ k y k = x k + 0.1 sin x k + v k
At each time point k , PDFs of ζ k and v k are given by
γ ζ ( x ) = { β 1 ( α 1 + 1 , λ 1 + 1 ) x α 1 ( 1 x ) λ 1 , x ( 0 , 1 ) 0 , otherwise γ v ( x ) = { [ 2 α 2 + λ 2 + 1 β ( α 1 + 1 , λ 1 + 1 ) ] 1 x α 2 ( 2 x ) λ 2 , x ( 0 , 2 ) 0 , otherwise
where β ( α + 1 , λ + 1 ) = 0 1 x α ( 1 x ) λ d x , α 1 = 1 , λ 1 = 2 , α 2 = 2 , λ 2 = 1 .
In this example, the set point of the system Equation (38) is set to be r k = { 1 ,    0 s k < 50 s 2 ,    50 s k < 100 s 3 ,    k 100 s . The prediction horizon and control horizon, P = M = 5 , are chosen. The sampling period is T = 1 . The initial control input ( u 0 = 0 ) and the state variable ( x 0 = 0 ) are selected. The step size in Equation (12) is η = 0.0001 . In the performance function Equation (9), we choose ϕ ( x ) = x 2 , α = 2 and λ = 0.5 .
In this simulation, the corresponding control inputs based on SIP and entropy criteria respectively are implemented on the system Equation (38) at each time point. Some comparative results are given to illustrate the superiority of the proposed SIP based stochastic predictive tracking control algorithm. In Figure 1, it is clear that the proposed control algorithm based on the SIP criterion has better performance, the fluctuation of the output response is smaller, and the response is quicker using the SIP method. In Figure 2, the trend of the performance index Equation (9) is presented. It is found that the performance index is overall decreasing with the progress of the time although some small variations can be recorded. The variances of the single neuron weights are presented in Figure 3. Figure 4 and Figure 5 demonstrate the PDFs of the output tracking errors using the entropy-based controller, while Figure 6 and Figure 7 illustrate the PDFs of the tracking errors using the proposed SNSPC method. Compared with Figure 4, the shape of the PDF of the tracking error in Figure 6 turns to be narrower and sharper over the control process, which indicates that the proposed SNSPC control system has achieved a smaller uncertainty in the tracking error distribution. In addition, it can also be seen that the peak of the tracking error PDF locates in the vicinity of zero using the proposed method. In Figure 8, the final PDF of the tracking error under the proposed control law is sharper and narrower than that under the entropy-based control law.

5. Conclusions

In the past, several different approaches, such as PDF shaping control, minimum entropy control, generalized minimum entropy control, etc. have been proposed to solve the control problem for non-Gaussian stochastic systems. These control strategies can achieve good performance, but there are still two main issues to be further improved: (1) The entropy value can be negative and it is shift-invariant. More suitable statistical information that describes objective functions is necessary; (2) Control input in a stochastic system is also a random variable. The randomness of control input should be considered.
In this work, a convergent SNSPC algorithm is presented for the controlled system with non-Gaussian disturbance. The proposed SNSPC is obtained by minimizing a SIP-based predictive criterion, in which the randomness of the control input is also considered besides randomness of the tracking error. Compared with the entropy or IP, the randomness measure SIP has some advantages, such as validity in a wide range of distributions, robustness, and the simplicity in computation. Moreover, the multistep predictive control strategy, rather than single step control is developed in this paper as it is more robust to disturbances and nonlinearities involved in the systems. Also, the convergent condition of the proposed SNSPC based on the energy conservation principle is proposed. The proposed control strategy is applied in a nonlinear and non-Gaussian stochastic numerical example. The simulation results confirm that this new SIP based predictive control method can achieve a good tracking performance.
Compared with the previous work in the field of stochastic distribution control, the contributions of this paper are three folds: (1) the randomness of control inputs is considered for the first time; (2) instead of the instantaneous performance index, a novel SIP-based cumulative criterion is formulated; (3) a single neuron multi-step predictive control algorithm is obtained, and it is much better than the single-step control method. However, most of the practical industrial processes have multi-inputs and multi-outputs, and there are also many constraints when the controller is designed. Future research should be focused on such problems.

Acknowledgments

This work was supported by the Natural Science Foundation of China (61503271, 61374052) and Shanxi (2014011018-2, 20140210022-7). They are gratefully acknowledged.

Author Contributions

Mifeng Ren, Xinying Xu and Lan Cheng conceived the project. Mifeng Ren, Junghui Chen and Xinying Xu searched relevant literatures. Mifeng Ren and Junghui Chen carried out the theoretical derivation. Ting Cheng and Lan Cheng provided the simulation results. Mifeng Ren, Ting Cheng and Junghui Chen analysed the simulation results and wrote the paper. Correspondence and requests for materials should be addressed to Junghui Chen. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Åström, K.J. Introduction to Stochastic Control Theory; Academic Press: New York, NY, USA, 1970. [Google Scholar]
  2. Wang, H. Bounded Dynamic Stochastic Systems: Modeling and Control; Springer: London, UK, 2000. [Google Scholar]
  3. Wang, H. Robust Control of the Output Probability Density Functions for Multivariable Stochastic Systems with Guaranteed Stability. IEEE Trans. Autom. Control 1999, 44, 2103–2107. [Google Scholar] [CrossRef]
  4. Wang, H.; Zhang, J. Bounded Stochastic Distributions Control for Pseudo-ARMAX Stochastic Systems. IEEE Trans. Autom. Control 2001, 46, 486–490. [Google Scholar] [CrossRef]
  5. Wang, H.; Yue, H. A rational spline model approximation and control of output probability density functions for dynamic stochastic systems. Trans. Inst. Meas. Control 2003, 25, 93–105. [Google Scholar] [CrossRef]
  6. Guo, L.; Wang, H. Stochastic Distribution Control System Design: A Convex Optimization Approach; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  7. Guo, L.; Wang, H. PID Controller Design for Output PDFs of Stochastic Systems Using Linear Matrix Inequalities. IEEE Trans. Syst. Man Cybern. Part B Cybern. 2005, 35, 65–71. [Google Scholar] [CrossRef]
  8. Yi, Y.; Li, T.; Guo, L.; Wang, H. Statistic tracking strategy for non-Gaussian systems based on PID controller structure and LMI approach. Dyn. Contin. Discret. Impuls. Syst. B 2008, 15, 859–872. [Google Scholar]
  9. Yue, H.; Wang, H. Minimum entropy control of closed-loop tracking errors for dynamic stochastic systems. IEEE Trans. Autom. Control 2003, 48, 118–122. [Google Scholar]
  10. Zhang, J.; Ren, M.; Wang, H. Minimum entropy control for non-linear and non-Gaussian two-input and two-output dynamic stochastic systems. IET Control Theory Appl. 2012, 6, 2434–2441. [Google Scholar] [CrossRef]
  11. Zhang, J.; Wang, H. Iterative learning-based minimum tracking error entropy controller for robotic manipulators with random communication time delays. IET Control Theory Appl. 2008, 2, 682–692. [Google Scholar] [CrossRef]
  12. Afshar, P.; Wang, H.; Chai, T. An ILC-based adaptive control for general stochastic systems with strictly decreasing entropy. IEEE Trans. Neural Netw. 2009, 20, 471–482. [Google Scholar] [CrossRef] [PubMed]
  13. Ren, M.; Zhang, J.; Wang, H. Minimized Tracking Error Randomness Control for Nonlinear Multivariate and Non-Gaussian Systems Using the Generalized Density Evolution Equation. IEEE Trans. Autom. Control 2014, 9, 2486–2490. [Google Scholar] [CrossRef]
  14. Chen, B.; Zhu, Y.; Hu, J.; Principe, J.C. System Parameter Identification: Information Criteria and Algorithms; Elsevier: Amsterdam, The Netherlands, 2013. [Google Scholar]
  15. Ren, M.; Zhang, J.; Jiang, M.; Yu, M.; Xu, J. Minimum ( h , ϕ ) - Entropy Control for Non-Gaussian Stochastic Networked Control Systems and Its Application to a Networked DC Motor Control System. IEEE Trans. Control Syst. Technol. 2015, 23, 406–411. [Google Scholar] [CrossRef]
  16. Zhang, J.; Jiang, M.; Ren, M.; Hou, G.; Xu, J. Improved single neuron controller for multivariable stochastic systems with non-gaussianities and unmodeled dynamics. ISA Trans. 2013, 52, 752–758. [Google Scholar] [CrossRef] [PubMed]
  17. Rao, M.; Chen, Y.; Vemuri, B.C.; Wang, F. Cumulative residual entropy: a new measure of information. IEEE Trans. Inf. Theory 2004, 50, 1220–1228. [Google Scholar] [CrossRef]
  18. Zografos, K.; Nadarajah, S. Survival exponential entropies. IEEE Trans. Inf. Theory 2005, 51, 1239–1246. [Google Scholar] [CrossRef]
  19. Chen, B.; Zhu, P.; Príncipe, J.C. Survival information potential: A new criterion for adaptive system training. IEEE Trans. Signal Process. 2012, 60, 1184–1194. [Google Scholar] [CrossRef]
  20. Marcello, F.; Giulioni, L.; Magni, L.; Scattolini, R. An approach to output-feedback MPC of stochastic linear discrete-time systems. Automatica 2015, 55, 140–149. [Google Scholar] [Green Version]
  21. Blackmore, L.; Ono, M.; Bektassov, A.; Williams, B.C. A probabilistic particle-control approximation of chance-constrained stochastic predictive control. IEEE Trans. Robot. 2010, 26, 502–517. [Google Scholar] [CrossRef]
  22. Calafiore, G.C.; Fagiano, L. Robust model predictive control via scenario optimization. IEEE Trans. Autom. Control 2013, 58, 219–224. [Google Scholar] [CrossRef]
  23. Zhou, Z.; Cogill, R. Reliable approximations of probability-constrained stochastic linear-quadratic control. Automatica 2013, 49, 2435–2439. [Google Scholar] [CrossRef]
  24. Bernardini, D.; Bemporad, A. Stabilizing model predictive control of stochastic constrained linear systems. IEEE Trans. Autom. Control 2012, 57, 1468–1480. [Google Scholar] [CrossRef]
  25. Korda, M.; Gondhalekar, R.; Oldewurtel, F.; Jones, C.N. Stochastic MPC framework for controlling the average constraint violation. IEEE Trans. Autom. Control 2014, 59, 1706–1721. [Google Scholar] [CrossRef]
  26. Cannon, M.; Cheng, Q.; Kouvaritakis, B.; Raković, S.V. Stochastic tube MPC with state estimation. Automatica 2012, 48, 536–541. [Google Scholar] [CrossRef]
  27. Kouvaritakis, B.; Cannon, M.; Rakovic, S.V.; Cheng, Q. Explicit use of probabilistic distributions in linear predictive control. Automatica 2010, 46, 1719–1724. [Google Scholar] [CrossRef]
  28. Cannon, M.; Kouvaritakis, B.; Wu, X. Model predictive control for systems with stochastic multiplicative uncertainty and probabilistic constraints. Automatica 2009, 45, 167–172. [Google Scholar] [CrossRef]
  29. Ali, M. Stochastic Model Predictive Control: An Overview and Perspectives for Future Research. IEEE Control Syst. Mag. 2016, in press. [Google Scholar]
  30. Edward, A.B.; Joel, A.P.; Ali, A.; Ali, M. Lyapunov-Based Stochastic Nonlinear Model Predictive Control: Shaping the State Probability Distribution Functions. In Proceedings of the 2016 American Control Conference, Boston, MA, USA, 6–8 July 2016.
  31. Mesbah, A.; Streif, S.; Findeisen, R.; Braatz, R.D. Stochastic Nonlinear Model Predictive Control with Probabilistic Constraints. In Proceedings of the 2014 American Control Conference (ACC), Portland, OR, USA, 4–6 June 2014; pp. 2413–2419.
  32. Zhang, J.; Yue, H.; Zhou, J. A new pdf modelling algorithm and predictive controller design. In Proceedings of the 10th IFAC International Symposium on Dynamics and Control of Process Systems, Mumbai, India, 18–20 December 2013; pp. 271–276.
  33. Zhang, J.; Yue, H.; Zhou, J. Predictive PDF control in shaping of molecular weight distribution based on a new modeling algorithm. J. Process Control 2015, 30, 80–89. [Google Scholar] [CrossRef]
  34. Ren, M.; Zhang, J.; Jiang, M.; Tian, Y.; Hou, G. Statistical Information Based Single Neuron Adaptive Control for Non-Gaussian Stochastic Systems. Entropy 2012, 14, 1154–1164. [Google Scholar] [CrossRef]
  35. Zan, X.; Xie, F. Switched Reluctance Generator System Based on Single Neuron adaptive PID Controller. In Proceedings of the 2011 International Conference on Advanced Mechatronic Systems, Zhengzhou, China, 11–13 August 2011; pp. 123–127.
  36. Xia, Y.; Xie, W.; Liu, B.; Wang, X. Data-driven predictive control for networked control systems. Inf. Sci. 2013, 235, 45–54. [Google Scholar] [CrossRef]
  37. Chen, B.; Zhu, Y.; Hu, J. Mean-Square Convergence Analysis of ADALINE Training With Minimum Error Entropy Criterion. IEEE Trans. Neural Netw. 2010, 21, 1168–1179. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Output responses.
Figure 1. Output responses.
Entropy 18 00218 g001
Figure 2. The performance index.
Figure 2. The performance index.
Entropy 18 00218 g002
Figure 3. Weights of a single neuron.
Figure 3. Weights of a single neuron.
Entropy 18 00218 g003
Figure 4. 3D mesh PDF using the entropy based controller.
Figure 4. 3D mesh PDF using the entropy based controller.
Entropy 18 00218 g004
Figure 5. Selected tracking error PDFs using the entropy based controller.
Figure 5. Selected tracking error PDFs using the entropy based controller.
Entropy 18 00218 g005
Figure 6. 3D mesh PDF using the proposed SIP based controller.
Figure 6. 3D mesh PDF using the proposed SIP based controller.
Entropy 18 00218 g006
Figure 7. Selected tracking error PDFs using the SIP based controller.
Figure 7. Selected tracking error PDFs using the SIP based controller.
Entropy 18 00218 g007
Figure 8. Final PDFs of the tracking error using the SIP based and entropy based controllers.
Figure 8. Final PDFs of the tracking error using the SIP based and entropy based controllers.
Entropy 18 00218 g008

Share and Cite

MDPI and ACS Style

Ren, M.; Cheng, T.; Chen, J.; Xu, X.; Cheng, L. Single Neuron Stochastic Predictive PID Control Algorithm for Nonlinear and Non-Gaussian Systems Using the Survival Information Potential Criterion. Entropy 2016, 18, 218. https://0-doi-org.brum.beds.ac.uk/10.3390/e18060218

AMA Style

Ren M, Cheng T, Chen J, Xu X, Cheng L. Single Neuron Stochastic Predictive PID Control Algorithm for Nonlinear and Non-Gaussian Systems Using the Survival Information Potential Criterion. Entropy. 2016; 18(6):218. https://0-doi-org.brum.beds.ac.uk/10.3390/e18060218

Chicago/Turabian Style

Ren, Mifeng, Ting Cheng, Junghui Chen, Xinying Xu, and Lan Cheng. 2016. "Single Neuron Stochastic Predictive PID Control Algorithm for Nonlinear and Non-Gaussian Systems Using the Survival Information Potential Criterion" Entropy 18, no. 6: 218. https://0-doi-org.brum.beds.ac.uk/10.3390/e18060218

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop