Accessibility / Report Error

A DISCRETE DYNAMICAL SYSTEM AND ITS APPLICATIONS

ABSTRACT

The main goal of this manuscript is to introduce a discrete dynamical system defined by symmetric matrices and a real parameter. By construction, we rediscovery the Power Iteration Method from the Projected Gradient Method. Convergence of the discrete dynamical system solution is established. Finally, we consider two applications, the first one consists in find a solution of non linear equation problem and the other one consists in verifies the optimality conditions when we solve quadratic optimization problems over linear equality constraints.

Keywords:
Projected gradient type-method; Power Iteration Method; Symmetric matrix; Discrete Dynamical System

1 INTRODUCTION

Discrete dynamical system appears as a tool in order to understand differential equations from numerically view point (for more details, see Galob (20073 GALOB O. 2007. Discrete Dynamical Systems. Springer Verlag, Berlin Heidelberg.) and chapter 6 in Loneli & Rumbos (20037 LONELI OHE & RUMBOS PIB. 2003. Métodos dinámicos en Economia. Otra busqueda del tiempo perdido. International Thompson Editores, Mexico.)). The classical model, in finite dimensional space, is as follows:

x k + 1 = F ( x k ) (1)

where F : Ωnn is an operator and Ω is a subset of the domain of the operator F (the domain of F is denoted by dom(F)). According to the literature, the equation 1 is not exclusive for differential equations, for example it appears in order to find fixed points for contractive operators (remember, F is contractive if F(x)F(y)λxy, with λ(0, 1) and x, ydom(F)). For details about contractive operators, see classical books in functional analysis or general topology or fixed point theorems as for instance Brezis (19831 BREZIS H. 1983. Analyse fonctionnelle - Théorie et applications. Masson, Paris.), Istrǎţescu (19815 ISTRǍŢESCU VI. 1981. Fixed point theory. D. Reidel Publishing Co., Dordrecht.), Kelley (19556 KELLEY JL. 1955. General Topology. Van Nostrand Co.. Princeton.). Other example is the autoregressive model (for more details see Shumway & Stoffer (20179 SHUMWAY RH & STOFFER DS. 2017. Time series analysis and its applications. Springer International Publishing AG.)).

Given a symmetric matrix A and a real number λ such that λ1σ(A). We consider the following operator Tλ : SS defined by

T λ ( x ) = ( I + λ A ) x ( I + λ A ) x (2)

where I is the identity matrix, S=xn : x=1 and σ(A) is the eigenvalue set of matrix A.Note that, in this setting, I+λA is a non singular matrix. So, operator is well defined.

The focus of this manuscript is the operator defined by the equation 2, which is very interesting, because:

  1. when, either λ=0 or A is the null matrix, the operator is the identity. Here, every xS is a fixed point of .

  2. when B=I+λ A has a dominant eigenvalue (i.e there exists an eigenvalue α* such that α*>α for all eigenvalue αα* ), the operator T was used in the famous Power Iteration Method introduced by R. Von Mises and H. Pollaczek-Geiringer in 19298 MISES RV & POLLACZEK-GEIRINGER H. 1929. Praktische Verfahren der Gleichungsauflösung. ZAMM - Zeitschrift für Angewandte Mathematik und Mechanik, 9: 152-164. (see Mises & Pollaczek-Geiringer (1929)).

  3. when λ1(na, +) , where a=maxai,j :A=ai,j and n is the size of A, B=I+λA is a strong monotone operator. Moreover, each eigenvector of A belonging to S is a fixed point of T (we prove it in section 2).

1.1 The Power Iteration Method

In order to understand the Power Iteration Method, consider a function called “Rayleigh quotient” which is defined, as follows, for each x0

r ( x ) = x , B x x , x . (3)

If x is an eigenvector, then Bx=r(x)x (i.e. r(x) is the corresponding eigenvalue of x). Suppose that vii=1nis a set of eigenvectors of B which is a basis of n, Bvi=λivi and λ1>λ2···λn. Taking v00 a vector with v0=1, we have that v0=i=1n αivi.

Then,

B v 0 = α 1 λ 1 v 1 + α 2 λ 2 v 2 + · · · + α n λ n v n ,

and so

B k v 0 = λ 1 k α 1 v 1 + α 2 λ 2 λ 1 k v 2 + · · · + α n λ n λ 1 k v n .

Here, Bkv0Bkv0 converges to v 1, because limk+λiλ1k=0 i2. The Power Iteration method is elegant, simple and can be stated as follows

( P I M ) p i c k a s t a r t i n g v e c t o r x 0 w i t h x 0 = 1 F o r k = 1 , 2 , · · · L e t x k = T λ ( x k - 1 ) w h e r e A = 1 λ ( B I )

but, convergence is only guaranteed if the following two assumptions hold:

  1. Non singular matrix B has an eigenvalue that is strictly greater in absolute value than its other eigenvalues.

  2. The starting vector x0 has a nonzero component in direction of an eigenvector associated with the dominant eigenvalue.

The reader can verify that for B=0110 (matrix B is nonsingular and symmetric), the sequences generated by Power Iteration Method diverge for any stated point x (different to any eigenvector of B), because does not have a dominant eigenvalue, but B¯=I+(1/3)B is positive definite (all its eigenvalues are strictly positive) and it has a dominant eigenvalue.

1.2 The Projected Gradient Method

The Projected Gradient method was introduced by Goldstein (for more detail see Goldstein (19644 GOLDSTEIN AA. 1964. Convex programmin in Hilbert space. Bulletin of The American Mathematical Society, 70: 709-710.)) for solving the following differentiable optimization problem

( P ) m a x i m i z e f ( x ) x C (4)

where f : C is differentiable in each point of a nonempty closed subset C of n.

The essence of the Projected Gradient method is based on two facts:

  1. First order Necessary Optimality condition: If x¯ is a solution of P, then f(x), d0, dT(C, x) (Here T(C, x) is the tangent cone of C at x¯ , for more details see Crouzeix et al. (20112 CROUZEIX J, KERAGEL A & SOSA W. 2011. Programación Matematica diferenciable. Universidad Nacional de Ingenieria.))

  2. Orthogonal projection: If u is an orthogonal projection of v over C, then uvyv, yC (in short u :=PC(v) ).

Now, given a symmetric matrix A, we know that all eigenvalues of A are real numbers and we can consider n eigenvectors of matrix A as a basis of n. Moreover if we consider problem (P) with f : n defined by f(x)=x, Ax and C=xF : x, x=1, where F is a subspace of n generated by eigenvectors of matrix A. The optimal value is an eigenvalue of A and any maximizer is a normalized eigenvector associated to the optimal value.

Here, if xC, then TC, x is an hyperplane in F defined by normal vector x0 and contain vector x (from now on H(x¯, 1) :=TC(x¯)). So, the necessary condition is reduced to f(x¯), yx¯)=0, yH(x¯, 1), which is equivalent to x¯+λfx¯x¯, yx¯=0, yHx¯, 1 and λ0 fixed, which is also equivalent to x=PHx¯,1x¯+λfx¯, λ0 fixed.

So,

x k + 1 = P H x k + 1 , 1 x k + λ A x k P H x k + 1 , 1 x k + λ A x k = x k + λ A x k x k + λ A x k = T λ x k

If we define B=I+λA, then both matrices A and B have the same eigenvectors. Moreover, δ is an eigenvalue of A and u an associate eigenvector (Au=δu) if and only if δu=Au=λ-1B-Iu if and only if Bu=(1+δλ)u if and only if (1+δλ) is an eigenvalue of B and u an associate eigenvector to it. In the next section we introduce an easy result which establishes that for each symmetric matrix A and each λ0 such that 1+λδ>0 for all eigenvalue δ of A, we have that B=I+λA is a Symmetric Positive Definite (SPD) matrix. It implies that (1+λδ)>0, for all eigenvalue δ of A.

In the section 2, we introduce a discrete dynamical system defined by symmetric matrices and a real parameter λ. We show that, under some conditions on the parameter λ, any sequence generated by the discrete dynamical system converges to a fixed point of the operator which define the discrete dynamical system. Moreover, there is an equivalence between the fixed point of the operator and the eigenvector of the symmetric matrix.

In section 3 we consider two applications, the first one consists in find a solution for the non linear equation problem and the second one consists in verifies the optimality conditions when we solve quadratic optimization problems over linear equality constraints.

2 A DISCRETE DYNAMICAL SYSTEM

We start this section with two elementary results.

Lemma 1. Let A be a no null symmetric matrix with size n and a = max a i j : A = a i j . If λ with λ - 1 > n a , then 1 + λ β > 0 β σ A .

Proof. Take β¯=maxβ : βσA and consider an eigenvector x¯ such that β¯=|x¯T Ax¯|. Then,

β β ¯ = x ¯ T A x ¯ = i , j x ¯ i x ¯ j e i , A e j i , j x i x j a i j β σ A

So, we have that

β a i , j x i x j = a i x i 2

But n1/2=argmaxixi : ixi2=1 (follows directly applying optimality conditions). It implies that, βna<λ-1. And so the statement follows. □

Note that the eigenvalue set σ(A) of A exists, but its elements are unknown explicitly in the previous Lemma.

Lemma 2. If A is a symmetric matrix and β ¯ = max β : β σ A , then for all λ with λ - 1 > β ¯ , then matrix B = I + λ A is a SPD matrix. Moreover, for each δ σ B we have that B u = δ u and A u = λ 1 δ 1 u i . e . σ A = λ 1 σ B 1 .

Proof. If βσA, then

B x = I + λ A x = x + λ A x = 1 + λ β x

But, λ1>β¯β. So, 1+λβ>0 βσA. Then, the statement follows. □

From now, for each non null symmetric matrix A, define the following operator Tλ : SSS=xn : x=1 by

T λ x = I + λ A x I + λ A x = B x B x (5)

where λ\0 is such that -λ-1σA and B=I+λA.

Theorem 1.Let A be a non null symmetric matrix andλsuch that-λ-1σA. A vector xis a fixed point of Tλ if and only if there existsδσAsuch thatAx*=δx*andx*S.

Proof. If x∗ is a fixed point of Tλx*=Bx*Bx*. So, Bx*0 and x*=1. Let σB=λ1,···, λn and let u1,···, unS be an eigenvector set of B such that Bui=λiui i1,..., n. Here, u1,···, un is a basis of n, then x*=i=1n αiui=iI αiui, where I=i1,···,n : αi0. Note that I0, because x*0. On the other hand i=1n αiui=x*=Tλx*=i=1n αiλiBx*. Since u1,···, un is a basis of n, then αi=αiλiBx*i1,···,n. It implies that iI, 1=λiBx*. So, Bx*=λi>0 iI. Finally,

B x * = B i I α i u i = i I α i B u i = i I α i λ i u i = B x * i I α i u i = B x * x *

The statement follows from Lemma 2, taking δ=λ1Bx*1.

If there exists δσA such that Ax*=δx* and x*S, then Bx*=1+λδ. The statement follows because Tλx*=Bx*Bx*=1+λδx*1+λδ=x*. □

Now, we are able to find the solution of the discrete dynamical system, for each non null symmetric matrix A and λ=1na+1, where n=sizeA and a=maxai,j : A=ai,j.

Initial step Given a non null symmetric matrix A.

n = s i z e A .

a = max a i , j : A = a i , j .

λ = 1 n a + 1 .

k = 0 .

x k S .

Iterative step Calculate:

x k + 1 = T λ ( x k ) = I + λ A x k I + λ A x k = B x k B x k = B k x 0 B k x 0 k = k + 1 .

The following result establishes that the sequence generated by the discrete dynamical system (the solution of the discrete dynamical system) is asymptotically stable for any starting point x0S.

Theorem 2.Let A be a non null symmetric matrix. For eachx0S, the sequence {xk} generated by the discrete scheme converges to an eigenvector of A belonging to S and the sequenceAxk, xkconverges to its respective eigenvalue.

Proof. From Lemma 1, we have that B is a SPD matrix. It implies that Bxk0 for all k and so xk+1=BxkBxk is well defined for all k. Let λ1,···,λn be the eigenvalue set of B and u1,···, unS a respective eigenvector set. Without loss of generality consider 0<λiλi+1i1,···,n1. Since u1,···,un is an orthonormal basis of n, then x0=i=1n εiui. Here, Bkx0=i=1n εiλik ui=iI εi2λi2k1/2, where I=i1,···,n : εi0. Taking j=maxi : iI and Ij=l : λl=λj, then xk+1=iI εiλikiI εi2λi2k1/2ui. It implies that xk+1=iI εiλiλjkiI εi2λiλj2k1/2ui. Note that for any iI \ Ij we have that 0<λiλj<1. It implies that, the sequence {xk} converges to iIjεiiIjεi21/2ui. It is easy to verify that the cluster point is a normalized eigenvector of B associated to an eigenvalue λj. Since A and B have the same eigenvectors set, then the statement follows. □

3 APPLICATIONS

In this section we consider two applications.

3.1 The Non Linear Equation Problem

This problem consists in find a feasible point of a nonlinear equation defined by a function f : n (here, f is twice differentiable on n) and λ. The Mathematical Model is:

( N L E P ) { Find x such that f ( x ) = λ (6)

Take xn, the representation of Taylor around x is

f y f x + f x , y x + 1 / 2 2 f x y x , y x .

Taking y=x+td with d=1, we have that

f x + t d f x + t f x , d t 2 / 2 2 f x d , d .

Using this approach, the problem consists in find a direction d such that the function h : defined by ht= fx + td has at least one real roots.

Definition 1. Given f : n be a function twice differentiable, let x , λ n + 1 . A vector d S is called a feasible direction for the problem (NLEP), if the function h defined by h t = f x + t d = λ has at least one real root.

We point out, in the case that f be a linear function. Here fx=a, x for an \0. In

this case, ht=fx+td=a, x+ta, d. Note that d=aa is a feasible direction. Moreover, x+t¯d is the orthogonal projection of x over the hyperplane xn : a, x=λ, where t¯=λ-a, x/a.

Lemma 3. Let f : n be a quadratic function and x , λ n + 1 . The following statement follows:

  1. The matrix D = f x f x T + 2 λ f x 2 f x is symmetric.

  2. If σ D , 0 , then the problem (NLEP) has no solution.

  3. If σ D 0 , + , then the problem (NLEP) has at least one solution. Moreover, any eigenvector associated to positive eigenvector, is a feasible direction.

  4. If d is an eigenvector associated to null eigenvalue and f x , d = 0 , then d is not a feasible direction when f x λ .

  5. If d is an eigenvector associated to null eigenvalue and f x , d 0 , then d is a feasible direction.

Proof. If f is a quadratic function, then

f x + t d = f x + t f x , d + t 2 2 2 f x d , d .

So, the equation fx+td=λ has solution if the discriminant

D d , d = f x , d 2 4 f x λ 2 f x d , d 2 0 .

All items follows because the discriminant need to be non negative, in order to find real roots of the quadratic equation. □

3.2 Linearly Constrained Quadratic Programming Problems

This problem can be formulated as follows:

( L C Q P P ) minimize 1 / 2 Q x , x - a , x P x = b (7)

without loss of generality Q is a n×n non null symmetric matrix, P is a m×n non null matrix, an and bm.

For the next result, we use the following notation: λ1,···,λn is the eigenvalue set of matrix PT P, u1,···,unS an eigenvector set i.e. PTPui=λiuii, I0=i :λi=0 and spanui : iI0 is the subspace generated by ui : iI0. By convention span0=0.

Lemma 4. If P is a non null matrix with size m × n , then

K e r P = K e r B T B = s p a n u i : i I ( 0 ) .

Proof. If KerP=0, then PT P is non singular and then I(0)=0 and so spanui : iI(0)=0. If not, take hKer(P)\0, then Ph=0 and so PTPh=0. It implies that h is an eigenvector of PT P and so hspanui : iI0. Conversely, if hspanui : iI0, then h=iI0 αiui. Hence, PTPh=iI0 αiBT Bui=0. Then 0=PTPh,h=Ph2 and so Ph=0. □

Now, consider a matrix V such that V : nn is the orthogonal projection over Ker(P)=Ker(PT P). Matrix V can be calculated as follows: Apply our scheme and obtain σ(BT B)=λ1,···,λn and u1,···,un such that Qui=λiui. Then V=iIIuiuiT, where I=i : λi0. The following result is important, because we can verify the condition Qh,h0hKerP, verifying that VT QV is semi definite positive (i. e. all its eigenvalues of VQV are nonnegative real values).

Corollary 1. Q h , h 0 h K e r P if and only if VQV is symmetric semi definite positive.

For the next result, consider L=xn : Px=b and f : n defined by fx=1/2Qx,xa,x.

Theorem 3. The problem LCQPP has a solution if and only if

E = z n + m : C z = c 0 a n d Q h , h 0 h K e r B (8)

where C = Q P T B 0 and c = a b . Moreover, if z ¯ = x ¯ , y ¯ E , then x ¯ is a solution of LCQPP.

Proof. If x¯ is a solution of LCQP, then the KKT optimality conditions imply that fx¯+PT y¯=Qx¯a+PT y¯=0 and Bx¯= b. So Cz¯=c for z¯=x¯, y¯. The first order necessary optimality

condition tells us that Qx¯aT h=0 hKer(B) (because the tangent cone of L in the point x¯ is equal to kernel of P, denoted by Ker(P) ). So, for any hKerP, x=x¯+hL and fx=f(x¯+h)=fx¯+1/2Qh,hfx¯ and so Qh,h0.

Now, if E=zn+m : Cz=c0 and Qh,h0 hKerP, then taking z¯=x¯, y¯S we claim that x¯ is a solution of LCQP. Indeed, Cz¯=c implies that Px¯=b and Qx¯a=PT y¯. Defining fx=1/2xT QxaT x, we need to show that fxfx¯ for all x such that Px=b.

Note that for h=x-x¯ we have that Ph=0. So hTfx¯=hTQx¯a=hT PT y¯=0. It implies that fx=fx¯+h=fx¯+1/2hT Ahfx¯ and the claim follows. □

4 NUMERICAL EXPERIMENTS

Here, we show numerical experiments using a program code written in SciLab software. Of course, this program code is very simple and developed by an amateur in Computer Science (Wilfredo Sosa).

The following numerical experiment concern to verify optimality condition when we solve linearly constrained quadratic programming problems. If

A = 0 2 7 - 17 2 8 - 6 - 6 7 - 6 - 6 0 - 17 - 6 0 - 2 , a = 1 2 3 4 , b = 3 2 a n d B = 2 0 - 1 2 0 1 2 1 .

Running our program code in SciLab, we obtain the spectral set σBTB=9,6,0,0 and the respective eigenvector set is

- 0 . 6666667 1 . 351 ( 10 - 08 ) 0 . 3333334 - 0 . 6666667 2 . 371 ( 10 - 08 ) 0 . 4082483 0 . 8164966 0 . 4082483 - 7 . 366 ( 10 - 10 ) 0 . 9128709 - 0 . 3651484 - 0 . 1825742 0 . 7453560 0 0 . 2981424 - 0 . 5962848

By definition

V = e y e 4 , 4 C P : , 1 C P : , 1 ' e y e 4 , 4 C P : , 2 C P : , 2 '

Again applying our program code to V A V we have that

σ V A V = 19 . 018037 , 9 . 7597408 , 0 , 0

It implies that, the vector

z ¯ = 0 . 6662676 , 0 . 4182381 , 0 . 2992118 , 0 . 9833383 , 7 . 3928963 , 5 . 0168612

is solution of Cz=c, and so x¯ is solution of LCQP (z¯=x¯, y¯).

Also, we applied our scheme for find eigenvalues and eigenvectors of symmetric matrixes. We simulate symmetric matrices and then calculate their eigenvalues and eigenvectors using our program code. We build symmetric matrices as follows: Given each two matrices (data), the first one D is a diagonal matrix and the second one V is an unitary matrix (i.e. VTV=VVT=I), then we define A :=VDVT , here diagonal entries of D are the eigenvalues of A and the column vectors of V are eigenvectors of A. We generate V using Gram-Schimidt process.

  1. The first matrix was built with 10 eigenvalues equal to -30; 10 eigenvalues equal to zero; and 10 eigenvalues equal to 30.

  2. The second matrix was built with 20 eigenvalues equal to -2000; 20 eigenvalues equal to zero; and 20 eigenvalues equal to 2000.

  3. The next matrix was built with eigenvalues following the rule λi+1=λi+1 , starting with λ1=-29 until λ59=29 and λ60=0 .

  4. The next matrix was built with 20 eigenvalues equal to -3000; 10 eigenvalues equal to -3; 20 eigenvalues equal to zero; 10 eigenvalues equal to 3; and 20 eigenvalues equal to 3000.

  5. The next matrix was built with 50 eigenvalues equal to -100 and 50 eigenvalues equal to 100.

Of course, our scheme finds all eigenvalues and a eigenvector set. Unfortunately, the Power Method Iteration does not run for generated matrices, because the first four matrices has null eigenvalues and the last one has as absolute value of all eigenvalues equal to 100 (it is not dominant eigenvalue matrix).

ACKNOWLEDGEMENTS

Wilfredo Sosa was partially supported by Fundação de Apoio à Pesquisa do Distrito Federal (FAP-DF) [grant 0193.001695/2017 and 00193.00002100/2018-51 and CNPq [Grants 302074/2012-0 and 471168/2013-0]. Part of this research was carried out during visits to IMPA.

References

  • 1
    BREZIS H. 1983. Analyse fonctionnelle - Théorie et applications. Masson, Paris.
  • 2
    CROUZEIX J, KERAGEL A & SOSA W. 2011. Programación Matematica diferenciable. Universidad Nacional de Ingenieria.
  • 3
    GALOB O. 2007. Discrete Dynamical Systems. Springer Verlag, Berlin Heidelberg.
  • 4
    GOLDSTEIN AA. 1964. Convex programmin in Hilbert space. Bulletin of The American Mathematical Society, 70: 709-710.
  • 5
    ISTRǍŢESCU VI. 1981. Fixed point theory. D. Reidel Publishing Co., Dordrecht.
  • 6
    KELLEY JL. 1955. General Topology. Van Nostrand Co.. Princeton.
  • 7
    LONELI OHE & RUMBOS PIB. 2003. Métodos dinámicos en Economia. Otra busqueda del tiempo perdido. International Thompson Editores, Mexico.
  • 8
    MISES RV & POLLACZEK-GEIRINGER H. 1929. Praktische Verfahren der Gleichungsauflösung. ZAMM - Zeitschrift für Angewandte Mathematik und Mechanik, 9: 152-164.
  • 9
    SHUMWAY RH & STOFFER DS. 2017. Time series analysis and its applications. Springer International Publishing AG.
  • 1
    Mathematics Subject Classification (2000): 15A18, 90C52, 90C20

APPENDIX

In this section we present a program code of our scheme written in SciLab Software. Of course, this program code is very simple and we do not use numerical strategies in order to reduce the time of compilation or reduce the accumulation of errors. For theses reason, we do not compare our program code with others in the literature, because it is not our subject. Criterions for the program code are the following:

1. Try to find a great eigenvalue in absolute value, for do it we find j such that absAj, jAi, i ij.

2. For build a matrix B, we consider α=10 and L=maxabsAAn1/2+1.

3. If absA<108, then, we consider matrix A as a null matrix.

4. The error to find an eigenvalue will be less to 10−16.

The following function calculate an eigenvector of a symmetric matrix A.


The main part of the code is the following. Of course, it is necessarily read a matrix DC.


Publication Dates

  • Publication in this collection
    2 Dec 2019
  • Date of issue
    Sep-Dec 2019

History

  • Received
    28 Nov 2018
  • Accepted
    11 Sept 2019
Sociedade Brasileira de Pesquisa Operacional Rua Mayrink Veiga, 32 - sala 601 - Centro, 20090-050 Rio de Janeiro RJ - Brasil, Tel.: +55 21 2263-0499, Fax: +55 21 2263-0501 - Rio de Janeiro - RJ - Brazil
E-mail: sobrapo@sobrapo.org.br