Acessibilidade / Reportar erro

Predictive control with mean derivative based neural euler integrator dynamic model

Abstracts

Neural networks can be trained to get internal working models in dynamic systems control schemes. This has usually been done designing the neural network in the form of a discrete model with delayed inputs of the NARMA type (Non-linear Auto Regressive Moving Average). In recent works the use of the neural network inside the structure of ordinary differential equations (ODE) numerical integrators has also been considered to get dynamic systems discrete models. In this paper, an extension of this latter approach, where a feed forward neural network modeling mean derivatives is used in the structure of an Euler integrator, is presented and applied in a Nonlinear Predictive Control (NPC) scheme. The use of the neural network to approximate the mean derivative function, instead of the dynamic system ODE instantaneous derivative function, allows any specified accuracy to be attained in the modeling of dynamic systems with the use of a simple Euler integrator. This makes the predictive control implementation a simpler task, since it is only necessary to deal with the linear structure of a first order integrator in the calculations of control actions. To illustrate the effectiveness of the proposed approach, results of tests in a problem of orbit transfer between Earth and Mars and in a problem of three-axis attitude control of a rigid body satellite are presented.

Neural Control; Nonlinear Predictive Control; Feed Forward Neural Nets; Dynamic Systems Neural Modeling; Ordinary Differential Equations Numerical Integrators


Redes neurais podem ser treinadas para obter o modelo de trabalho interno para esquemas de controle de sitemas dinâmicos. A forma usual adotada é projetar a rede neural na forma de um modelo discreto com entradas atrasadas do tipo NARMA (Non-linear Auto Regressive Moving Average). Em trabalhos recentes a utilização de uma rede neural inserida em uma estrutura de integração numérica tem sido também considerada para a obtenção de modelos discretos para sistemas dinâmicos. Neste trabalho, uma extensão da última abordagem é apresentada e aplicada em um esquema de controle não-linear preditivo (NPC), com uma rede feed forward modelando as derivadas médias em uma estrutura de integrador numérico de Euler. O uso de uma rede neural para aproximar a função de derivadas médias, em vez da função de derivadas instantâneas do sistema dinâmico ODE, permite que qualquer precisão desejada na modelagem discreta de sistemas dinâmicos possa ser realizada, com a utilização de um simples integrador Euler, tornando a implementação do controle preditivo uma tarefa mais simples, uma vez que ela somente necessitará lidar com a estrutura linear de um integrador de primeira ordem na determinação das ações de controle. Para ilustrar a efetividade da abordagem proposta, são apresentados resultados dos testes em um problema de transferência de órbitas Terra/Marte e em um problema de controle de atitude em três eixos de satélite comportando-se como corpo rígido.

Controle Neural; Controle Preditivo Não-Linear; Redes Feed forward; Modelagem Neural de Sistemas Dinâmicos; Integradores Numéricos de Equações Diferenciais Ordinárias


SISTEMAS DE CONTROLE

Predictive control with mean derivative based neural euler integrator dynamic model

Paulo M. Tasinaffo; Atair Rios Neto

Instituto Nacional de Pesquisas Espaciais (INPE), São José dos Campos, SP, Brasil. tasinafo@comp.ita.br, atairrn@uol.com.br

ABSTRACT

Neural networks can be trained to get internal working models in dynamic systems control schemes. This has usually been done designing the neural network in the form of a discrete model with delayed inputs of the NARMA type (Non-linear Auto Regressive Moving Average). In recent works the use of the neural network inside the structure of ordinary differential equations (ODE) numerical integrators has also been considered to get dynamic systems discrete models. In this paper, an extension of this latter approach, where a feed forward neural network modeling mean derivatives is used in the structure of an Euler integrator, is presented and applied in a Nonlinear Predictive Control (NPC) scheme. The use of the neural network to approximate the mean derivative function, instead of the dynamic system ODE instantaneous derivative function, allows any specified accuracy to be attained in the modeling of dynamic systems with the use of a simple Euler integrator. This makes the predictive control implementation a simpler task, since it is only necessary to deal with the linear structure of a first order integrator in the calculations of control actions. To illustrate the effectiveness of the proposed approach, results of tests in a problem of orbit transfer between Earth and Mars and in a problem of three-axis attitude control of a rigid body satellite are presented.

Keywords: Neural Control, Nonlinear Predictive Control, Feed Forward Neural Nets, Dynamic Systems Neural Modeling, Ordinary Differential Equations Numerical Integrators.

RESUMO

Redes neurais podem ser treinadas para obter o modelo de trabalho interno para esquemas de controle de sitemas dinâmicos. A forma usual adotada é projetar a rede neural na forma de um modelo discreto com entradas atrasadas do tipo NARMA (Non-linear Auto Regressive Moving Average). Em trabalhos recentes a utilização de uma rede neural inserida em uma estrutura de integração numérica tem sido também considerada para a obtenção de modelos discretos para sistemas dinâmicos. Neste trabalho, uma extensão da última abordagem é apresentada e aplicada em um esquema de controle não-linear preditivo (NPC), com uma rede feed forward modelando as derivadas médias em uma estrutura de integrador numérico de Euler. O uso de uma rede neural para aproximar a função de derivadas médias, em vez da função de derivadas instantâneas do sistema dinâmico ODE, permite que qualquer precisão desejada na modelagem discreta de sistemas dinâmicos possa ser realizada, com a utilização de um simples integrador Euler, tornando a implementação do controle preditivo uma tarefa mais simples, uma vez que ela somente necessitará lidar com a estrutura linear de um integrador de primeira ordem na determinação das ações de controle. Para ilustrar a efetividade da abordagem proposta, são apresentados resultados dos testes em um problema de transferência de órbitas Terra/Marte e em um problema de controle de atitude em três eixos de satélite comportando-se como corpo rígido.

Palavras-chave: Controle Neural, Controle Preditivo Não-Linear, Redes Feed forward, Modelagem Neural de Sistemas Dinâmicos, Integradores Numéricos de Equações Diferenciais Ordinárias.

1 INTRODUCTION

Multi layer feed forward artificial neural networks have the capacity of modeling nonlinear functions (e.g., Cybenko, 1988; Hornik et al, 1989). This property allows their application in control schemes, where an internal model of the dynamic system is needed, as is for example the case in predictive control (Clarke et al, 1987a, 1987b). A commonly used way of representing the internal model of the dynamics of the system has been to design the neural network to learn a system approximation in the form of a discrete model with delayed inputs of the NARMA type (Non-linear Auto Regressive Moving Average) (Leontaritis and Billings, 1985a, 1985b; Chen and Billings, 1989, 1990 and 1992; Narendra and Parthasarathy, 1990; Hunt et al, 1992; Mills et al, 1994; Liu et al 1998; Norgaard et al, 2000). The neural net designed and trained in this way has the disadvantage of needing too many neurons in the input and hidden layers.

In recent works, the use of a neural ordinary differential equation (ODE) numerical integrator as an approximate discrete model of motion, together with the use of Kalman filtering for calculations of control actions, was proposed and tested in the predictive control of dynamic systems (Rios Neto, 2001; Tasinaffo and Rios Neto, 2003). It was shown and illustrated with tests that artificial feed forward neural networks could be trained to play the role of the dynamic system derivative function in the structure of ODE numerical integrators, to get internal models in nonlinear predictive control schemes. This approach has the advantage of reducing the dimension and complexity of the neural network, and thus of facilitating its training (Wang e Lin, 1998; Rios Neto 2001). It was also shown that the stochastic nature and the good numerical performance of the Kalman filtering parameter estimator algorithm make its choice a good one, not only to train the feed forward neural network (Singhal et al, 1989; Chandran, 1994; Rios Neto, 1997), but also to estimate the predictive control actions (Rios Neto, 2000). Its use allows considering the errors in the output patterns in the supervised training of the artificial neural networks. It also allows the possibility of giving a stochastic meaning to the weight matrices present in the predictive control functional.

This paper further explores the approach of combining feed forward neural networks with the structure of ordinary differential equations (ODE) numerical integrator algorithms to get dynamic systems internal models in predictive control schemes. Instead of approximating the instantaneous derivative function in the dynamic system ODE model, the neural network is used to approximate the mean derivative function (Tasinaffo, 2003). This allows the use of an Euler structure to get a first order neural integrator. In principle this mean derivative based first order neural integrator can provide the same accuracy as that of any higher order integrator. However, it is much simpler to deal with, both in terms of the neural network training and of the implementation of the predictive control scheme.

In what follows, in Section 2 the mathematical foundation, that supports the possibility of getting discrete nonlinear dynamic system models using Euler numerical integrators with mean derivative functions, is presented. In Section 3 it is presented a summary of the method of calculating the discrete control actions in a predictive control scheme with the use of Kalman filtering. In Section 4, results of tests in a problem of orbit transfer between Earth and Mars and in a problem of three - axis attitude control of a rigid body satellite are presented to illustrate the effectiveness of the proposed approach. Finally, in Section 5 a few conclusions are drawn.

2 MEAN DERIVATIVE BASED EULER INTEGRATOR AS A DYNAMIC SYSTEM DISCRETE MODEL

2.1 Fundaments

For the sake of facilitating the understanding and of mathematically supporting the possibility of using a mean derivative based Euler integrator as a dynamic system discrete model, in this section a summary of the results obtained by Tasinaffo (2003) are presented without demonstration. With this purpose, consider the following nonlinear autonomous system of ordinary differential equations,

where,

Let, by definition, = (t), j = 1, 2,...,n be a trajectory, solution of the nonlinear ODE = f(y), starting from (to )at initial time to , that belongs to a domain of interest [(to ), (to )]n, and where (to ) and (to ) are finite. It is convenient to introduce the following vector notation to indicate possible initial condition sets and the respective solutions of (1.a):

where, i = 1, 2, ..., ¥; and ¥ is adopted to indicate that the mesh of discrete initial conditions can have as many points as desired.

To start the mathematical background, two important results (e.g., Braun, 1983) about the solution of differential equations (1.a) are considered. The first is about the existence and uniqueness of solutions and the second about the existence of stationary solutions of (1.a).

Theorem 1 (T1) Let each of the functions f1 (y1 , y2 , ..., yn ), ..., fn (y1 , y2 , ..., yn ) have continuous partial derivatives with respect to y1 , ..., yn. Then, the initial value problem = f(y), y(to ) inside a domain of interest [, ]n, j=1, 2,...,n, in to, has one and only one solution yi = y i(t), in Rn, from each yi(to ) initial state. If two solutions, y = f(t) and y = j(t), have a common point, then they must be identical.

Property 1 (P1)If y = f(t) is a solution of (1.a), then y = f(t + c) is also a solution of (1.a), where c is any real constant.

Since, in general,

i = f(y i) has not an analytical solution, it is usual to only know a discrete approximation of yi = yi(t), in an interval , through a set of discrete points, [yi(t + kDt) yi[t + (k + 1)Dt] ... yi[t + (k + ntDt)] ] º [kyi k + 1yi ... ], for a given Dt.

By definition, the secant given by two points kyi and k + 1yi of the curve yi(t) is the straight-line segment joining these two points. Thus, from the secants defined by the pair of points k and k + 1, k and k + 1, , k and k + 1 one can define the tangents:

Property 2 (P2) If kyi is a discrete solution of

i = f(y i) and Dt ¹ 0, tanDtkai exists and is unique.

Two other important theorems, which relate the values of tankDtai and tankDt

i, with the values of the mean derivatives calculated from [kyi k + 1yi ... k + nyi] and [ k
i k + 1
i ... k + n
i], respectively, are the differential and integral mean value theorems (e.g., Wilson, 1958; Munem et al, 1978; Sokolnikoff et al, 1966), enunciated in what follows.

Theorem 2 (T2) (Differential mean value theorem): If a function (t), for j =1,2,...,n,is defined and continuous in the closed interval [tk,tk+1] and is differentiable in the open interval (tk,tk+1), then there is at least one , tk < < tk+1 such that

T2 assures that given a secant of a differentiable yi(t) it is always possible to find a point between k + 1yi and kyi of the intersection of the secant with the curve in t k and tk + 1, such that the tangent to this intermediate point is parallel to the secant. The value i() is called the mean derivative of yi(t) in [tk, tk + 1 ].

Theorem 3 (T3)(Integral mean value theorem): If a function, (t), for j = 1, 2,..., n is continuous in the closed interval [tk, tk + 1 ], then there exists at least one

interior to this interval, such that

In general and are different and it is important to notice that the theorems do not tell how to determine these points.

Property 3 (P3)The mean derivative

i() of y i(t) in the closed interval [tk, tk + 1 ] is equal to tanDtkai, as an immediate consequence of the definition of mean derivatives.

Theorem 4 (T4)The pointk + 1

of the solution of the system of nonlinear differential equations
i = f(yi), for j = 1, 2,...,n, can be determined through the relation k + 1 = tankDtai · Dt + k
for a given kyi and Dt.

Corollary 1 (C1)- The solution of the system of nonlinear differential equations

i = f(yi), at a given discrete point, k + m, for j = 1, 2,...,n, can be determined, given an initial kyi, by the relation:

Corollary 2 (C2)For the system,

i = f(yi), the following relation is valid: , for j=1, 2,...,n.

Notice that for the situation where the system of Eq. (1a) is autonomous, (t1 ) = (t2 ) for i1¹ i2 and t1¹ t2 implies that (t1 ) = (t2 ). This property establishes that two trajectories of = f(y) starting from two different initial conditions, (to ) and (to ), for i1¹ i2 , will have the same derivatives only if (t1 ) = (t2 ), even when t1¹ t2.

The question remaining is if the mean derivative of the interval [ky i,k + 1y i] is also autonomous, that is, time invariant? The properties that follow answer this question.

Property 4 (P4)If (t) and (t) are solutions of = f(y) starting from (to = 0) and (to = 0), respectively, and if (to = 0) = (T) for T>0, then (Dt) = (T + Dt) for any given Dt.

Property 5 (P5)If (t1 ) = (t2 ), for i1¹ i2 and t1¹ t2, then, (t1 ) = (t2 ) for Dt > 0, that is,

is autonomous.

This result is useful since it determines that it is enough to know the values of , for i = 1, 2, ..., ¥ at to , in a region of interest [, ]n, j = 1, 2,...,n, because for t>t0 they will repeat, as long the boundaries of [, ]n are observed.

Notice also that the trajectories of the dynamic system when propagated ahead will have angles ka(i) varying only in the interval -< ka(i) < , which will thus be unique. When retro propagated, < ka(i) < , and thus ka(i) will also be unique in this case.

Theorem 5 (T5) The result of T4 is still valid when discrete values of control ku in each [tk, tk + 1 ] are used to solve the dynamic system:

Demonstration: In this case the continuous function, f(yi,u), with ku approximated as constant in [tk , tk + 1 ], can be viewed as parameterized with respect to the control variable and, thus, for any discrete interval the existence of the mean derivative i() = = tanDtkai is guaranteed and the result in Eq. (6) is still valid.

2.2 Numerical Integrators with Neural Mean Derivatives to Represent Dynamic Systems

Consider the capacity of a feed forward neural network to approximate nonlinear functions (e.g., Zurada, 1992). From the previous section, one can conclude that it is possible to have a neural network to learn the mean derivatives of a given dynamical system and use them in an ODE Euler integrator structure to get a discrete representation of this system. In the proposed approach, a first possibility was adopted as illustrated in Fig. 1, where the neural network is trained to directly learn the dynamic system mean derivative, which is then inserted in the structure of the Euler numerical integrator. In this scheme, the neural network is trained to learn the function of mean derivatives from the sampled input values of state ky i and control ku, with a previously fixed discrete interval Dt The value of the training output pattern = is generated with the help of a numerical integrator of high order used to simulate one step ahead with negligible errors k + 1yi, the solution of the system i = f(yi, u).


A second possibility that could be used, based on that adopted by Wang and Lin (1998), is depicted in Fig. 2. It is one where using the outputs of an Euler integrator the neural network is indirectly trained to learn the dynamic system mean derivative. In this case, k + 1i, the value of state estimated by the neural Euler numerical integrator, is the output value compared to the training pattern k + 1yi to generate the error signal for the supervised training. The neural network is trained to learn the function of mean derivatives º tanDtai(ky,ku) from the values of statekyi, control ku and of a previously fixed discrete intervalDt. In Fig. 2, k + 1yi is the value of the training pattern generated off line by a numerical integrator of high order used to simulate the system i = f(yi,u), and k + 1

i is the value the neural network tries to estimate for k + 1yi, in the supervised training. The relation of recurrence between kyi and k + 1yi is expressed by k + 1yi = · Dt + kyi, where @ is the mean derivative to be approximated by the neural network. It should be noticed that if k + 1yi is obtained from the Euler integration, then converges to the function of derivatives i = f(yi,u). But if k + 1y i is obtained from the use of a high order integrator or experimentally, then will converge to the function of mean derivatives.


A correspondent algorithm to get the mean derivative based first order neural integrator would then be as follows:

1. Given in to the domains of interest [(to), (to)]n1 , j = 1, 2, ..., n1, and []n2 , k=1, 2, ..., n2, generate inside these domains, m uniformly distributed random vectors to be the input p i, i=1, 2, ..., m, of training patterns to the feed forward neural network.

2. Employing a high order ODE numerical integrator, propagate ahead with step size Dt the inputs pi, i = 1, 2, ..., m , generating the state vectors , i=1, 2, ..., m, at to + Dt.

3. Calculate the vectors Ti to be used as training output patterns:

Notice that since the function is also autonomous it is only necessary propagate ahead pi , i=1, 2, ..., m, with step size Dt, to get the neural network output patterns Ti .

4. Do the supervised training of the neural network, using the patterns {(pi , Ti)}.

5. After training the neural network with a specified accuracy, there results the dynamic system discrete model, in the form of a mean derivative based first order neural integrator:

Notice that using the scheme of Fig. 1, with T i = as output patterns, avoids calculating the back propagation with k + 1yi = · Dt + kyi.

To analyze the local error of this neural Euler integrator, consider the exact value k + 1y i and the estimated value k + 1 i, respectively given by Eqs. (10) and (11).

where em is the error in the output of the neural network trained to learn the function of mean derivatives inside a domain of interest. Due to the capacity of approximation of the neural network, this error can be less than any specified value. Thus, k + 1i, in Eq. (11), can have the desired accuracy, since for a fixed Dt > 0 the neural network is approximating inside a domain of interest the function of mean derivatives that is invariant in time, and em can be made as small as specified.

Figure 3 better illustrates this situation. Consider k + 1yi, k + 1 and k + 1, respectively representing the exact value of the solution of i = f(yi,u) at tk + 1 , the approximate value of k + 1yi obtained from a high order numerical integrator, and the approximate value of k + 1yi obtained from an Euler integrator. As indicated by Fig. 3, if it is taken k + 1yi = k + 1, in the scheme of Fig. 2, then = @ f(ky i,ku), but if it is taken k + 1yi = k + 1, and if k + 1 is away from k + 1yi, then during the phase of training will approximately converge to the function of mean derivatives instead of converging to f(yi,u).


3 NEURAL PREDICTIVE CONTROL SCHEME

The neural predictive control scheme presented in what follows was proposed and demonstrated by Rios Neto (2000). In a problem of neural predictive control of a dynamic system (Mills et al, 1994), it adopts a heuristic and theoretical approach to solve the problem of minimizing a quadratic functional subject to the constraint of a neural network predictor, representing the dynamics of the system to be controlled. In the proposed scheme, the problems of training the neural network and of determining the predictive control actions are seen and treated in an integrated way, as problems of stochastic optimal linear estimation of parameters.

The problem to be solved is that of controlling a dynamical system modeled by an ODE:

It is assumed that the system to be controlled can be approximated by a discrete model. That is, for tj = t + j · Dt:

where, y(t), ..., y(t1 - ny ) and u(t - 1 ), ..., u(t1 - nu ) are the past system responses and control actions, respectively.

In the usual neural predictive control scheme, a feed forward neural network is trained to learn a discrete model as in Eq. (13). This model is then used as an internal system response model to get the smooth control actions that will track a reference response trajectory by minimizing (e.g., Clarke et al, 1987a; Clarke et al, 1987b; Liu et al, 1998) the finite horizon functional:

where, yr (tj) is the reference response trajectory at time tj; nh is the number of steps in the finite horizon of optimization; rr- 1 (tj ) and ru- 1 (tj ) positive definite weighting matrices; is the output of the feed forward neural network previously trained to approximate a discrete model of the dynamic system response.

The determination of the predictive control actions can be treated as a parameter estimation problem, if the minimization of the functional of Eq. (14) is seen as the following stochastic problem:

with, j = 1, 2, ..., nh ; where (tj ) = f[(tj - 1 ), ..., (tj - ny ); u(tj - 1 ), ..., u(tj - nu ),w] are the outputs of the neural network which is recursively used as a predictor of the dynamic system responses in the horizon of optimization and it is understood that for tj - k< t, (tj - k ) are estimations or measurement of already occurred values of outputs, in the control feedback loop; vy (tj ) and vu (tj ) are the uncorrelated noises for different values of tj.

To solve the problem of Eqs.(15) an iterative approach is needed, where in each ith iteration a perturbation is done to get a linear approximation of Eq. (15.a):

where k starts at zero, even for j>nu, as a consequence of recursively being a function of u(tj - 2 ), ..., u(t) through the successive recursions starting with (tj - 1 ), ..., (tj - ny ) (see the AppendixAppendix, for details about the recurrence relations needed in the calculations of the partial derivatives, for the proposed case where discrete nonlinear dynamic system models using Euler numerical integrators with mean derivative functions are used); 0 < a(i) < 1 is a parameter to be adjusted to guarantee the hypothesis of linear perturbation; and the partial derivatives are calculated either by numerical differentiation or by using the chain rule to account for the composed function situation, including the back propagation rule (see, e.g., Chandran (1994)) in the feed forward neural network that approximates the derivative function of the dynamic system.

The formulation as a stochastic linear estimation problem in each iteration is complete if the recursion in Eq.(15.b) is taken in account:

In a more compact notation:

where the meaning of compact variables become obvious by comparison with Eqs.(16) and (17). Applying the Kalman filtering algorithm, the following solution results in a typical iteration (Rios Neto, 2000):

where, i = 1, 2, ..., I; Ru (t), Ry (t) and R(, I) are the covariance matrices of Vu (t), Vy (t) and ((t,I) - U(t)), respectively; and Iu is an identity matrix.

A correspondent algorithm for this predictive control scheme in a typical time step t would then be as follows.

1. The control (t - 1 ) (see Eq. (18b)) is the estimated solution from the last control step. In the ith iteration: the approximated estimated value of control is U(, i) = (U(t, - 1); a(i) ¬ a (i - 1); and for i=1 estimates or extrapolations of estimates of last control step are used.

2. Calculate the partial derivatives of Eq.(16), using the expressions of Eqs. (1A) to (3A) of the AppendixAppendix. get Hu(t,i) and u(t,i), in Eq. (18c).

3. Estimate U(,i) with the Kalman filtering of Eqs. (19.a), (19.b). Notice that the Kalman filtering can be done either in batch or sequentially, by recursively processing component to component, in sequence. Increment i, and repeat steps, until the are sufficiently close to yr (tj ) according to a specified error, usually taken 3 · s of vy, and when this occurs take:

4 NUMERICAL TESTS

4.1 Tests in an Earth Mars Orbit Transfer.

This is a problem of low thrust orbit transfer where the state variables are the rocket mass m, the orbit radius r, the radial speed w and the transversal speed v, and where the control variable is the thrust steering angle q, measured from local horizontal. The ODE (e.g., Sage, 1968) of this dynamic system are:

where the variables have been normalized with: µ = 1.0, the gravitational constant; T = 0.1405, the thrust; with to = 0 and tf = 5, initial and final times, where each unit of time is equal to 58.2 days. The predictive control is used on line to produce control actions that make the spacecraft follow a reference trajectory determined off line.

Initially tests were conducted to evaluate the neural integrator capacity of giving an accurate discrete model of the dynamics. To approximate the vector mean derivative function, a multiplayer perceptron feed forward neural network, with 41 neurons (this number of neurons can be determined only empirically) with the hyperbolic tangent as activation functions (l = 2), in the hidden layer, with identity activation, in the output layer, and with input bias in the hidden and output layers, was used. This feed forward neural net was trained with a parallel processing Kalman filtering algorithm (e.g., Singhal, 1989; Rios Neto, 1997), with 3600 input - output training patterns until a mean square error of 2.4789.10 - 6 was reached and tested with 1400 patterns, reaching a mean square testing error of 2.7344.10 - 6.

This mean derivative neural network was then used in an Euler integrator structure to produce an internal model of the orbit transfer problem dynamics to be used in a predictive control scheme where the reference was defined as the optimal minimum time transfer trajectory. Results obtained are shown in Figs. 4 and 5, for a discrete step size of 0.01 of normalized time (0.582 days) and receding horizons of 1 and 5 steps ahead, respectively. These results illustrate the effectiveness of the proposed approach when applied in this kind of problem.



4.2 Tests in a Three-Axis Satellite Attitude Control

In this case, the attitude control in three axes of a rigid body satellite is considered, with the correspondent dynamic equations given as follows (Wertz, 1978; Kaplan, 1976):

where, f, q and j are the Euler angles; Ix, Iy, Iz are the principal moments of inertia; wx, wy and wz the angular velocity components, in the principal body axes; Tx, Ty and Tz the control torques. The reference trajectories are defined such as to drive the Euler angles asymptotically to the origin, using updating data from navigation (Silva, 2001):

Initially, tests were conducted to evaluate the neural integrator capacity of giving an accurate discrete model of the attitude dynamics. To approximate the vector mean derivative function, a multiplayer perceptron feed forward neural network, with 20 neurons with the hyperbolic tangent as activation functions (l = 1), in the hidden layer, neurons with identity activation, in the output layer, and with input bias in the hidden and output layers, was used. This feed forward neural net was also trained with a parallel processing Kalman filtering algorithm, with 3200 input - output training patterns until a mean square error of 8.621.10 - 5 was reached and tested with 800 patterns, reaching a mean square testing error of 8.669.10 - 5.

This mean derivative neural network was then used in an Euler integrator structure to produce the internal model of the attitude dynamics to be used in the predictive control scheme with the reference as defined in Eqs.(23) and testing data as given in Table 1. Results obtained are shown in Fig. 6.


These results illustrate the effectiveness of the proposed approach when applied to this kind of problem. The oscillatory behavior in the y component of angular velocity may be due to the fact that the reference trajectory did not include explicitly the derivative terms corresponding to the angular regulation.

5 CONCLUSIONS

A new approach of predictive control, using a mean derivative based neural Euler integrator as the internal dynamic system model, was presented. The structure of an ODE Euler numerical integrator was used to get neural discrete forward models where the neural network has only to learn and approximate the algebraic and static mean derivative function in the dynamic system ODE.

The tests indicate the effectiveness of using the mean derivative based neural model of the dynamic system as an internal model in the control scheme and reinforced the expected following characteristics:

  • It is a simpler task to train a feed forward neural network to learn the algebraic, static function of the dynamic system ODE mean derivatives (where the inputs are samples of state and control variables), than to train it to learn a NARMA type of discrete model (where the inputs are samples of delayed responses and controls).

  • The neural network in the neural ODE integrator results to be simpler, in terms of the necessary number of layers and number of neurons, since it does not have to learn the dynamic law, but only the derivative function.

The use of a Kalman filtering based approach to get the neural predictive control actions led to results where:

  • The stochastic interpretation of errors gave more realism in the treatment of the problem, and facilitated the adjustment of weight matrices in the predictive control functional.

  • The local parallel processing version of the Kalman filtering algorithm used in the control scheme exhibited efficiency and efficacy equivalent to that of the correspondent neural network training Kalman filtering algorithm. This was expected, since they are completely similar algorithms used to solve numerically equivalent parameter estimation problems.

  • Only one step ahead was sufficient in the receding horizon of control. This feature together with the efficiency and performance of the parallel processing Kalman algorithms, combined with the present on board processing capabilities, guarantees the feasibility of real time, adaptive applications.

Notice that the proposed approach can be also be applied when an ODE mathematical model is not available. This can be done as long as dynamic system input output pairs are available to be used as training information, considering the structure of the numerical integrator with a feed forward network in place of the mean derivative function. Notice also that one could directly use an ODE numerical integrator as a dynamic system discrete model to play the role of an internal model in the predictive control scheme. However, in this case one would not have the possibility of adaptive control schemes, by exploring the learning capacity of the neural network and of on line updating its training.

The application of the proposed approach is not restricted to predictive control. It can be applied to any control scheme where an internal model of the controlled system is necessary.

Further studies shall evaluate the scheme adopted by Wang and Lin (1998), and depicted in Fig. 2. It is one where using the outputs of an Euler integrator the neural network is indirectly trained to learn the dynamic system mean derivative. In this paper it was only considered the scheme where the neural network is trained to directly learn the dynamic system mean derivative, which is then inserted in the structure of the Euler numerical integrator.

REFERENCES

Braun, M. (1983). Differential equations and their applications: an introduction to applied mathematics. 3rd ed., New York: Springer-Verlag, Applied Mathematical Sciences 15.

Artigo submetido em 02/02/2006

1a. Revisão em 14/02/2007

Aceito sob recomendação do Editor Associado Prof. José Roberto Castilho Piqueira

HU(T, I)

Equations (18.a) to (18.d) can be solved only if the matrix Hu(t,i) is known. For the case where discrete nonlinear dynamic system models using Euler numerical integrators with mean derivative functions are used, it can be calculated through the following equations (Tasinaffo, 2003):

where, q=2, 3, ..., nh.; the notation is the same as in Section 2, and nh is the number of steps in the finite horizon of optimization.

It is still necessary to calculate the back propagation relative only to the feed forward network, like is showed in figure 1A, to get the matrices (2A) to (4A), in order to completely solve for Hu(t,i).


In this case, the back propagation is given by (e.g., Carrara, 1997; Tasinaffo, 2003):

where,

where the function f¢(.) is the derivate of the activation function of each neuron inside the feed forward net (Fig. 1A).

Equations 5A to 7A combined with equations (18.a) to (18.d) of Section 3 completely the problem of getting the matrix Hu(t,i).

  • Carrara, V. (1997) Redes Neurais Aplicadas ao Controle de Atitude de Satélites com Geometria Variável 202p. INPE-6384-TDI/603. Doctoral Thesis, Instituto Nacional de Pesquisas Espaciais-INPE, Săo José dos Campos, 1997.
  • Chandran, P. S. (1994). Comments on "comparative analysis of backpropagation and the extended kalman filter for training multilayer perceptrons˘˘. IEEE Transactions on Pattern Analysis and Machine Intelligence, v. 16, n. 8, pp. 862-863.
  • Chen, S., Billings, S. A., Luo, W. (1989). Orthogonal least squares methods and their application to nonlinear system identification. Int. J. Control, 50(5), pp. 1873-1896.
  • Chen, S., Billings, S. A., Cowan, C. F. N., Grant, P. M. (1990). Practical identification of NARMAX models using radial basis function. Int. J. Control, 52(6), pp. 1327-1350.
  • Chen, S.; Billings, S. A. (1992). Neural networks for nonlinear dynamic system modeling and identification. Int. J. Control, v. 56, n. 2, pp. 319-346.
  • Clarke, D.W., Mohtadi, C., Tuffs, P. S. (1987a). Generalized Predictive Control-Part I. The Basic Algorithm. The Journal of IFAC the International Federation of Automatic Control. Automatica, v. 23, n. 2, pp. 137-148.
  • Clarke, D.W., Mohtadi, C., Tuffs, P. S. (1987b). Generalized Predictive Control-Part II. Extensions and Interpretations. The Journal of IFAC the International Federation of Automatic Control Automatica, v. 23, n. 2, pp. 149-160.
  • Cybenko, G. (1988). Continuous valued networks with two hidden layers are sufficient. Technical Report, Department of Computer Science, Tufts University.
  • Hornik, K.; Stinchcombe, M.; White, H. (1989). Multilayer feedforward networks are universal approximators. Neural Networks, v. 2, n. 5, pp. 359-366.
  • Hunt, K. J.; Sbarbaro, D.; Zbikowski, R.; Gawthrop, P. J. (1992). Neural networks for control systems - A survey. Automatica, v. 28, n. 6, pp. 1083-1112.
  • Kaplan, M. H. (1976). Modern spacecraft dynamics & control New York: John Wiley & Sons.
  • Leontaritis, I. J., Billings, S. A. (1985a). Input-output parametric models for nonlinear system part I: Deterministic nonlinear systems. Int. J. Control, 41(2), pp. 303-328.
  • Leontaritis, I. J., Billings, S. A. (1985b). Input-output parametric models for nonlinear system part II: Stochastic nonlinear systems. Int. J. Control, 41(2), pp. 329-344.
  • Liu, G. P., Kadirkamanathan, V., Billings, S. A. (1998). Predictive Control for Nonlinear Systems Using Neural Networks. Int. J. Control, v. 71, n. 6, pp. 1119-1132.
  • Mills, P. M., Zomaya, A. Y., Tadé, M. O. (1994). Adaptive Model-Based Control Using Neural Networks. Int. J. Control, 60(6), pp. 1163-1192.
  • Munem, M. A., Foulis, D. J. (1978). Calculus with Analytic Geometry Volumes I and II, Worth Publishers, Inc., New York.
  • Narendra, K. S.; Parthasarathy, K. (1990). Identification and control of dynamical systems using neural networks. IEEE Transactions on Neural Networks, v. 1, n. 1, pp. 4-27.
  • Norgaard, M., Ravn, O., Poulsen, N. K., Hansen, L. K. (2000). Neural Networks for Modelling and Control of Dynamic Systems. Springer, London.
  • Rios Neto, A. (1997). Stochastic optimal linear parameter estimation and neural nets training in systems modeling. RBCM - J. of the Braz. Soc. Mechanical Sciences, v. XIX, n. 2, pp. 138-146.
  • Rios Neto, A. (2000). Design of a Kalman filtering based neural predictive control method. In: XIII CONGRESSO BRASILEIRO DE AUTOMÁTICA - CBA, 2000, UFSC (Universidade Federal de Santa Catarina), Florianópolis, Santa Catarina, Brazil, CD-ROM Proceedings, pp. 2130-2134.
  • Rios Neto, A. (2001). Dynamic systems numerical integrators control schemes. In: V CONGRESSO BRASILEIRO DE REDES NEURAIS, 2001, Rio de Janeiro, RJ, Brasil. CD-ROM Proceedings, pp. 85-88.
  • Sage A. P. (1968). Optimum systems control Englewood Cliffs, NJ: Prentice-Hall, Inc..
  • Silva, J. A. (2001). Controle preditivo utilizando redes neurais artificiais aplicado a veículos aeroespaciais 239 p. (INPE-8480-TDI/778). Doctoral Thesis, Instituto Nacional de Pesquisas Espaciais-INPE, Săo José dos Campos, Brazil.
  • Singhal, S., Wu, L. (1989). Training Multilayer Perceptrons with the Extended Kalman Algorithm. In Advances in Neural Information Processing Systems, VI, Morgan Kaufman Pub. Inc., pp 136-140.
  • Sokolnikoff, I. S.; Redheffer, R. M. (1966). Mathematics of physics and modern engineering 2nd. ed., Tokyo: McGraw-Hill Kogakusha, LTD.
  • Tasinaffo, P. M. (2003). Estruturas de integraçăo neural feedforward testadas em problemas de controle preditivo 230 p. INPE-10475-TDI/945. Doctoral Thesis, Instituto Nacional de Pesquisas Espaciais-INPE, Săo José dos Campos, Brazil.
  • Tasinaffo, P.M., Rios Neto, A. (2003). Neural Numerical Integrators in Predictive Control tested in an Orbit Transfer Problem. In: CONGRESSO TEMÁTICO DE DINÂMICA, CONTROLE E APLICAÇŐES - DINCOM, 2, Săo José dos Campos: ITA, SP, Brasil, CD-ROM Proceedings, pp. 692-702.
  • Wang, Y.-J., Lin, C.-T. (1998). Runge-Kutta Neural Network for Identification of Dynamical Systems in High Accuracy. IEEE Transactions On Neural Networks, Vol. 9, No. 2, pp. 294-307, March.
  • Wertz, J. R. (1978). Spacecraft attitude determination and control London: D. Reidel, Astrophysics and Space Science Library, v. 73.
  • Wilson, E. B. (1958). Advanced Calculus New York: Dover Publications.
  • Zurada, J. M. (1992). Introduction to Artificial Neural System St. Paul, MN, USA: West Pub. Co.

Appendix

Publication Dates

  • Publication in this collection
    25 July 2007
  • Date of issue
    Mar 2007

History

  • Received
    02 Feb 2006
  • Reviewed
    14 Feb 2007
Sociedade Brasileira de Automática Secretaria da SBA, FEEC - Unicamp, BLOCO B - LE51, Av. Albert Einstein, 400, Cidade Universitária Zeferino Vaz, Distrito de Barão Geraldo, 13083-852 - Campinas - SP - Brasil, Tel.: (55 19) 3521 3824, Fax: (55 19) 3521 3866 - Campinas - SP - Brazil
E-mail: revista_sba@fee.unicamp.br