## Services on Demand

## Article

## Indicators

- Cited by SciELO
- Access statistics

## Related links

- Cited by Google
- Similars in SciELO
- Similars in Google

## Share

## Computational & Applied Mathematics

*On-line version* ISSN 1807-0302

### Comput. Appl. Math. vol.31 no.2 São Carlos 2012

#### http://dx.doi.org/10.1590/S1807-03022012000200008

**Two iterative algorithms for solving coupled matrix equations over reflexive and anti-reflexive matrices **

**Mehdi Dehghan ^{I}; Masoud Hajarian^{II, *} **

^{I}Department of Applied Mathematics, Faculty of Mathematics and Computer Science, Amirkabir University of Technology, 424, Hafez Avenue, Tehran 15914, Iran. E-mails: mdehghan@aut.ac.ir / mdehghan.aut@gmail.com

^{II}Department of Mathematics, Faculty of Mathematical Sciences, Shahid Beheshti University, G.C., Tehran 19839, Iran. E-mails: m_hajarian@sbu.ac.ir / mhajarian@aut.ac.ir / masoudhajarian@gmail.com

**ABSTRACT**

An *n × n* real matrix *P* is said to be a generalized reflection matrix if *P ^{T} = P* and

*P*

^{2}=

*I*(where

*P*is the transpose of

^{T}*P*). A matrix A ∈ R

*is said to be a reflexive (anti-reflexive) matrix with respect to the generalized reflection matrix*

^{n×n}*P*if

*A*=

*P A P*(

*A*= —

*P A P*). The reflexive and anti-reflexive matrices have wide applications in many fields. In this article, two iterative algorithms are proposed to solve the coupled matrix equations

{ | A_{1} XB_{1} + C_{1}X^{T}D_{1} = M_{1}. |

A_{2} XB2 + C_{2}X^{T}D_{2} = M_{2}. |

over reflexive and anti-reflexive matrices, respectively. We prove that the first (second) algorithm converges to the reflexive (anti-reflexive) solution of the coupled matrix equations for any initial reflexive (anti-reflexive) matrix. Finally two numerical examples are used to illustrate the efficiency of the proposed algorithms.

**Mathematical subject classification:** 15A06, 15A24, 65F15, 65F20.

**Key words:** iterative algorithm, matrix equation, reflexive matrix, anti-reflexive matrix.

**1 Introduction**

In this paper we use the following notation. Let ^{m×n} be the set of all *m × n* real matrices. We use tr(*A*), *A ^{T}*,

*ρ*(

*A*),

*λ*(

*A*) and

*λ*

_{max}(

*A*) to denote the trace, the transpose, the spectral radius, the eigenvalue set and the maximum eigenvalue of the matrix

*A*respectively. We denote by

*I*and O

_{k}_{m×n}the

*k × k*identity matrix and the

*m ×n*zero matrix, respectively. We also write them as

*I*and

*O*, respectively, when the dimensions of these matrices are clear. We define an inner product as

*A, B*= tr(

*B*), then the norm of a matrix

^{T}A*A*generated by this inner product is Frobenius norm and is denoted by

*A, A*= ║

*A*║

^{2}.

An *n × n* real matrix *P* is said to be a real generalized reflection matrix if *P ^{T}* = P and

*P*

^{2}=

*I*. An

*n×n*real matrix

*A*is said to be a reflexive (anti-reflexive) matrix with respect to the generalized reflection matrix

*P*if

*A*=

*P A P*(

*A*= -

*P A P*). (

*P*) ((

*P*)) denotes the subspace reflexive (anti-reflexive) matrices with respect to the

*n×n*generalized reflection matrix

*P*. The reflexive and anti-reflexive matrices have practical applications in many areas such as the numerical solution of certain differential equations [1], pattern recognition [6], Markov processes [42], various physical and engineering problems [7] and so on (e.g. [20, 32, 43]). Chen [3] proposed three applications of reflexive and anti-reflexive matrices obtained from the altitude estimation of a level network, an electric network and structural analysis of trusses. The symmetric Toeplitz matrices, an important subclass of the class of symmetric reflexive matrices, appear naturally in digital signal processing applications and other areas [21].

The linear matrix equations, such as *AXB = C, AXB + CXD = E* and *AXB+CX ^{T}D = M*, play an important role in linear system theory therefore a large number of papers have presented several methods for solving these matrix equations [2, 9, 15, 36]. Research on solving of linear matrix equations has been actively ongoing for past years. In [5], Dai studied the linear matrix equation

over symmetric matrix *X*. By using g-inverse, Mitra [38] obtained the common solution of simultaneous matrix equations

Navarra et al. [39] studied a representation of the solution *X* to the system of matrix equations (1.2). The matrix equation

plays important roles in system theory, such as eigenstructure assignment [29], observer design [4], control of system with input constraint [28], and fault detection [30].

In [40], the necessary and sufficient condition for the existence of the solution to the matrix equation (1.3) and its solution expression was investigated by the generalized inverse matrix. In [37], Cramer's rules for some quaternion matrix equations were obtained within the framework of the theory of the column and row determinants. Kyrchei [35] considered systems of linear quaternionic equations and obtained Cramer's rules for right and left quaternionic systems of linear equations. In [44, 45, 46], the solutions of the several generalized Sylvester matrix equations were established. In [24], a family of iterative methods for linear systems is presented and a least-squares iterative solution to coupled matrix equations are studied by using the hierarchical identification principle and the star product. In [26], gradient iterative algorithms for solving Sylvester coupled matrix equations and general coupled matrix equations are studied by using the gradient search principle. In [22, 25], Ding and Chen applied a hierarchical identification principle to study solving the Sylvester and Lyapunov matrix equations. Also Ding and Chen [23] proposed a hierarchical gradient iterative algorithm and a hierarchical stochastic gradient algorithm and prove that the parameter estimation errors given by the algorithms converge to zero for any initial values under persistent excitation. In [8, 10, 11, 12, 13, 14, 17, 18], Dehghan and Hajarian introduced some efficient iterative methods for solving Sylvester and Lyapunov matrix equations.

In this paper, we introduce two iterative algorithms, respectively, for the finding reflexive and anti-reflexive solutions of the coupled matrix equations

(including the matrix equations (1.1)-(1.3) as special cases).

The rest of the paper is structured as follows. In Section 2, first we propose two iterative algorithms for solving (1.4) over reflexive and anti-reflexive matrices. Then we study the convergence properties of the iterative algorithms. Two examples verify the efficiency of the algorithms in Section 3. Section 4 concludes the paper.

**2 Main results**

In this section, first we give two systems of matrix equations equivalent to (1.4) over reflexive and anti-reflexive matrices, respectively. Then we will propose two efficient iterative algorithms for solving (1.4).

**Lemma 2.1.*** The coupled matrix equations (1.4) have the reflexive solution X* ∈ (*P*) *if and only if the system of matrix equations*

*is consistent.*

**Proof. ** First, we suppose that the coupled matrix equations (4) have the reflexive solution *X** ∈ (*P*). By using *X** = *PX* P* and *A _{i}X**

*B*+

_{i}*C*, we have

_{i}X^{*T}D_{i}= M_{i}for* i *= 1,2. It is follows from (2.2) that the reflexive matrix *X** is a solution of the system of matrix equations (2.1).

Conversely assume that the system of matrix equations (2.1) is consistent. Let *X* be a solution of the system of matrix equations (2.1). Set

Therefore ∈ (P) and we can get

for i = 1,2. Hence is a reflexive solution of the coupled matrix equations (1.4). The proof is completed. |

Similarly to the above lemma, we can obtain the following lemma.

**Lemma 2.2. ***The coupled matrix equations* (1.4) *have the anti-reflexive solution X* ∈ (*P*) (*P* ≠ *I*) *if and only if the system of matrix equations*

*is consistent.*

According to Theorem 4.3.8 and Corollary 4.3.10 in [33], the systems (2.1) and (2.5), respectively, are equivalent to

and

where *P*(*n, n*) is a permutation matrix [33]. Now by using the above results and considering

and

the following lemmas are well known [31, 33, 34].

**Lemma 2.3. ** * The coupled matrix equations (1.4) have a unique reflexive solution with respect to the generalized reflection matrix P if and only if *

rank((*Z*_{1}, vec(*M*_{1}, *M*_{2}, *M*_{1}, *M*_{2}))) = rank(*Z*_{1})

*and Z _{1} has a full column rank. In that case, the reflexive solution of (1.4) can be expressed by the following form *

*and the homogenous coupled matrix equations*

*have a unique reflexive solution X = 0. *

**Lemma 2.4. ***The coupled matrix equations *(1.4)* have a unique anti-reflexive solution with respect to the generalized reflection matrix P* ≠ *I if and only if *rank((Z_{2},vec(*M*_{1}, *M*_{2}, *M*_{1}, *M*_{2}))) = rank(*Z*_{2}) and *Z*_{2}* has a full column rank. In that case, the anti-reflexive solution of *(1.4)* can be expressed by the following form *

*and the homogenous coupled matrix equations*(2.11)

*have a unique anti-reflexive solution X = 0.*

If Lemma 2.3 (Lemma 2.4) is applied for finding the reflexive (anti-reflexive) solution of the coupled matrix equations (1.4), we need to take the inverse of the large matrix The above method may turn out to be numerically expensive and are not practical for equations of large systems. Our purpose in this paper is to obtain two iterative methods without any inverse for solving the coupled matrix equations (1.4) over reflexive and anti-reflexive matrices. We extend the idea of the Jacobi and the Gauss-Seidel iterations to solve the coupled matrix equations (1.4) over reflexive and anti-reflexive matrices.

Suppose that *A = M — N* is a splitting of the matrix *A*. The Jacobi and Gauss-Seidel procedures for solving the linear system *Ax = b* are typical members of a large family of iterations that have the form

with *M = D, N = —*(*L *+ *U*) for Jacobi and *M = D + L*, *N = — U* for Gauss-Seidel [32]. Here by extending the Jacobi and the Gauss-Seidel iterations and by applying the hierarchical identification principle [23, 25], we present two iterative methods for solving the coupled matrix equations (1.4) over reflexive and anti-reflexive matrices. These iterative methods are derived as follows:

**Algorithm 2.1.** **To solve (1.4) over reflexive matrix X**:

**Step 2.1.1.** Input matrices *A, C* ∈ ^{r×n}, *B, D* ∈ * ^{n×s}* and

*M*∈

^{r×s};

**Step 2.1.2. ** Choose arbitrary *X*(1) ∈ (*P*) where *P* is an *n*-by-*n* arbitrary generalized reflection matrix and a parameter *ω *∈ ^{+};

**Step 2.1.3** Calculate

*R _{i}*(1) =

*M*(1)

_{i }– A_{i}X*B*(1)

_{i }– C_{i}X^{T}

*D*,

_{i}*i*=1,2;

*k*:=1;

**Step 2.1.4. ** If ║*R*_{1}(*k*)║+║*R*_{2}(*k*)║ = 0, then stop; Else go to step 2.1.5;

**Step 2.1.5.**

**Step 2.1.6.** If ║*R*_{1}(*k* + 1)║+║*R*_{2}(*k* + 1)║ = 0, then stop; Else, let *k*: = *k*+1, go to step 2.

**Algorithm 2.2. ****To solve (1.4) over anti-reflexive matrix X**:

**Step 2.2.1.** Input matrices *A, C* ∈ ^{r×n}, *B*, *D* ∈ ^{n×s} and *M* ∈ ^{r×s};

**Step 2.2.2. **Choose arbitrary *X*(1) ∈ (*P*) where *P *≠ *I* is an *n*-by-*n* arbitrary generalized reflection matrix and a parameter *ω *∈ ^{+};

**Step 2.1.3.** Calculate

*R _{i}*(1) =

*M*(1)

_{i}– A_{i}X*B*(1)

_{i}– C_{i}X^{T}

*D*,

_{i}*i*=1, 2;

*k*:=1;

**Step 2.2.4. **If ║*R*_{1}(*k*)║+║*R*_{2}(*k*)║ = 0, then stop; Else go to step 2.2.5;

**Step 2.2.5. **

**Step 2.2.6.** If ║*R*_{1}(*k* + 1)║+║*R*_{2}(*k* + 1)║ = 0, then stop; Else, let *k*: = *k* + 1, go to step 2.2.5.

Now convergence properties of Algorithms 2.1 and 2.2 are presented.

**Theorem 2.1.*** If the coupled matrix equations *(1.4)* have a unique reflexive solution X, then iterative solution X(k) given by Algorithm *2.1* converges to X for any initial reflexive matrix X*(1)*, if the parameter **ω* * satisfies the inequality*

**Proof. ** We define the estimation error matrix in the form

By applying (2.15), we can get

Also it is not difficult to obtain

Now we can write

From (2.14) and (2.17), it is not difficult to get

The necessary condition of the series convergence (2.18) implies that

By considering Lemma 2.3, we have

The proof of theorem is completed. |

Similar to the proof of the above theorem, we can prove the following theorem.

**Theorem 2.2.*** If the coupled matrix equations *(1.4)* have a unique anti-reflexive solution X, then iterative solution X(k) given by Algorithm 2.2 converges to X for any initial anti-reflexive matrix X(1), if the parameter * *ω* *satisfies theinequality*

**Remark 2.1. **The convergence factor in (2.14) and (2.19) may also be taken as:

**3 Numerical examples**

In this section, we give two examples to illustrate the convergence of Algorithms 2.1 and 2.2, respectively. All the tests are performed by MATLAB.

**Example 3.1. ** As the first example we consider the linear matrix equation *AXB+CX ^{T}D = M* with

It can be verified that this matrix equation is consistent over reflexive matrices and has the reflexive solution

with

Choose arbitrary initial iterative matrix *X*(1) = 0. By Algorithm 2.1, we obtain the sequence *X*(*k*). In Figure 1, we report the obtained results with several values of *ω* where

and *r*(*k*) = log_{10} ║*M *− *AX*(*k*)*B* − *CX*(*k*)* ^{T} D*║.

It can be observed from Figure 1 that Algorithm 2.1 is effective. The effect of changing the convergence factor *ω* is illustrated in Figure 1. We see that the larger the convergence factor *ω* is, the faster the convergence the algorithm.

**Example 3.2. ** Consider a pair of matrix equations in the form of (1.2) with the following parameters:

and

We can verify the pair of matrix equations in the form of (1.2) are consistent over anti-reflexive matrix *X* and have the anti-reflexive solution

Taking *X*(1) = 0, we apply Algorithm 2.2 to compute *X*(*k*). The effect of changing the convergence factor *ω* is illustrated in Figure 2 where

and

Obviously both δ(*k*) and *r*(*k*) decrease, and converge to zero as *k* increases.

**4 Concluding remarks**

In this paper, we have considered the coupled matrix equations (1.4) over reflexive and anti-reflexive matrices. First Algorithms 2.1 and 2.2 were introduced for finding reflexive and anti-reflexive solutions of (1.4). Second the convergence theorems of the iterative algorithms were presented. The experiments are encouraging and seem to indicate that Algorithms 2.1 and 2.2 work well for numerical examples. It is interesting to develop the introduced algorithms for solving other linear matrix equations. We leave it as a topic for further research.

**Acknowledgements. ** The authors are very much indebted to an anonymous referee for his/her valuable comments and careful reading of the manuscript.

**References**

[1] A. Andrew, *Eigenvectors of certain matrices.* Linear Algebra Appl., **7** (1973), 157-162; 455-460. [ Links ]

[2] A. Ben-Israel, *A Cramer rule for least-norm solutions of consistent linear equations.* Linear Algebra Appl., **43** (1982), 223-226. [ Links ]

[3] H.C. Chen, *Generalized reflexive matrices: special properties and applications.* SIAM J. Matrix Anal. Appl., **19** (1998), 140-153. [ Links ]

[4] L. Dai, Singular Control Systems, Berlin: Springer-Vertag (1989). [ Links ]

[5] H. Dai, *On the symmetric solutions of linear matrix equations.* Linear Algebra Appl., **131** (1990), 1-7. [ Links ]

[6] L. Datta and S. Morgera, *Some results on matrix symmetries and a pattern recognition application.* IEEE Trans. Signal Process, **34** (1986), 992-994. [ Links ]

[7] L. Datta and S. Morgera, *On the reducibility of centrosymmetric matrices-applications in engineering problems.* Circuits Systems Signal Process, **8** (1989), 71-96. [ Links ]

[8] M. Dehghan and M. Hajarian, *The general coupled matrix equations over generalized bisymmetric matrices.* Linear Algebra Appl., **432** (2010), 1531-1552. [ Links ]

[9] M. Dehghan and M. Hajarian, *The reflexive and anti-reflexive solutions of a linear matrix equation and systems of matrix equations.* Rocky Mountain J. Math., **40** (2010), 1-23. [ Links ]

[10] M. Dehghan and M. Hajarian, *On the generalized bisymmetric and skew-symmetric solutions of the system of generalized Sylvester matrix equations.* Linear and Multilinear Algebra, **59** (2011), 1281-1309. [ Links ]

[11] M. Hajarian and M. Dehghan, *The generalized centro-symmetric and least squares generalized centro-symmetric solutions of the matrix equation AYB+CY ^{T}D = E.* Mathematical Methods in the Applied Sciences,

**34**(2011), 1562-1579. [ Links ]

[12] M. Dehghan and M. Hajarian, *Solving the generalized Sylvester matrix equation* Σ* ^{p}_{i = 1}A_{i}XB_{i} +*Σ

*International Journal of Control, Automation and Systems,*

^{q}_{j = 1}C_{j}YD_{j}= E over reflexive and anti-reflexive matrices.**9**(2011), 118-124. [ Links ]

[13] M. Dehghan and M. Hajarian, *Analysis of an iterative algorithm to solve the generalized coupled Sylvester matrix equations.* Applied Mathematical Modelling, **35** (2011), 3285-3300. [ Links ]

[14] M. Dehghan and M. Hajarian, *Two algorithms for the Hermitian reflexive and skew-Hermitian solutions of Sylvester matrix equations.* Applied Mathematics Letters, **24** (2011), 444-449. [ Links ]

[15] M. Dehghan and M. Hajarian, *The (R,S)-symmetric and (R,S)-skew symmetric solutions of the pair of matrix equations A _{1}XB_{1} = C_{1} and A_{2}XB_{2} = C_{2}.* Bulletin of the Iranian Mathematical Society,

**37**(2011), 269-279. [ Links ]

[16] M. Dehghan and M. Hajarian, *SSHI methods for solving general linear matrix equations.* Engineering Computations, **28** (2012), 1028-1043. [ Links ]

[17] M. Dehghan and M. Hajarian, *The generalised Sylvester matrix equations over the generalised bisymmetric and skew-symmetric matrices.* International Journal of Systems Science, **43** (2012), 1580-1590. [ Links ]

[18] M. Dehghan and M. Hajarian, *Iterative algorithms for the generalized centro-symmetric and central anti-symmetric solutions of general coupled matrix equations.* Engineering Computations, **29** (2012), 528-560. [ Links ]

[19] J. Delmas, *On Adaptive EVD asymptotic distribution of centro-symmetric covariance matrices.* IEEE Trans. Signal Process, **47** (1999), 1402-1406. [ Links ]

[20] Z.J. Bai, *The inverse eigenproblem of centrosymmetric matrices with a submatrix constraint and its approximation.* SIAM J. Matrix Anal. Appl., **26** (2005), 1100-1114. [ Links ]

[21] P. Delsarte and Y. Genin, Spectral properties of finite Toeplitz matrices, in Proceedings of the 1983 International Symposium on Mathematical Theory of Networks and Systems, Beer Sheva, Israel, 1983, Springer-Verlag, Berlin, New York (1984), 194-213. [ Links ]

[22] F. Ding and T. Chen, *Gradient based iterative algorithms for solving a class of matrix equations.* IEEE Trans. Autom. Contr., **50** (2005), 1216-1221. [ Links ]

[23] F. Ding and T. Chen, *Hierarchical gradient-based identification of multivariable discrete-time systems.* Automatica, ** 41** (2005), 315-325. [ Links ]

[24] F. Ding and T. Chen, *Iterative least squares solutions of coupled Sylvester matrix equations.* Systems Control Lett., **54** (2005), 95-107. [ Links ]

[25] F. Ding and T. Chen, *Hierarchical least squares identification methods for multivariable systems.* IEEE Trans. Autom. Contr., **50** (2005), 397-402. [ Links ]

[26] F. Ding and T. Chen, *On iterative solutions of general coupled matrix equations.* SIAM J. Control Optim., **44** (2006), 2269-2284. [ Links ]

[27] F. Ding, P.X. Liu and J. Ding, *Iterative solutions of the generalized Sylvestermatrix equations by using the hierarchical identification principle.* Appl. Math. Comput., **197** (2008), 41-50. [ Links ]

[28] G.R. Duan, *The solution to the matrix equation AV+BW = EVJ+R.* Appl. Math. Lett., **17** (2004), 1197-1202. [ Links ]

[29] L.R. Fletcher, J. Kuatsky and N.K. Nichols, *Eigenstructure assignment in descriptor systems.* IEEE Trans. Auto. Contr., **31** (1986), 1138-1141. [ Links ]

[30] P.M. Frank, *Fault diagnosis in dynamic systems using analytical and knowledge-based redundancy-a survey and some new results.* Automatica, **26** (1990), 459-474. [ Links ]

[31] J. Gilbert and L. Gilbert, Linear Algebra and Matrix Theory, 2^{nd} edition. Thomson Brooks/Cole (2005). [ Links ]

[32] G.H. Golub and C.F. Van Loan, Matrix computations, third ed., The Johns Hopkins University Press, Baltimore and London (1996). [ Links ]

[33] R.A. Horn and C.R. Johnson, Topics in Matrix Analysis, Cambridge University Press (1991), 259-260. [ Links ]

[34] A.B. Israel and T.N.E. Greville, Generalized inverses theory and applications, 2^{nd} ed., Springer-Verlag, New York (2003). [ Links ]

[35] I.I. Kyrchei, *Cramer's rule for quaternionic systems of linear equations.* Journal of Mathematical Sciences, ** 155** (2008), 839-858. [ Links ]

[36] I.I. Kyrchei, *Analogs of the adjoint matrix for generalized inverses and corresponding Cramer rules.* Linear and Multilinear Algebra, **56** (2008), 453-46. [ Links ]

[37] I.I. Kyrchei, *Cramer's rule for some quaternion matrix equations next term.* Appl. Math. Comput., **217** (2010), 2024-2030. [ Links ]

[38] S.K. Mitra, *Common solutions to a pair of linear matrix equations A _{1}XB_{1} = C_{1}, A_{2}XB_{2} = C_{2}.* Proc. Camb. Philos. Soc.,

**74**(1973), 213-216. [ Links ]

[39] A. Navarra, P.L. Odell and D.M. Young, *A representation of the general common solution to the matrix equations A _{1}XB_{1} = C_{1} and A_{2}XB_{2} = C_{2} with applications.* Comput. Math. Appl.,

**41**(2001), 929-935. [ Links ]

[40] F. Piao, Q. Zhang and Z. Wang, *The solution to matrix equation AX + X ^{T}C = B.* J. Franklin Institute,

**344**(2007), 1056-1062. [ Links ]

[41] M. Wang, X. Cheng and M. Wei, *Iterative algorithms for solving the matrix equation AXB + CX ^{T}D = E.* Appl. Math. Comput.,

**187**(2007), 622-629. [ Links ]

[42] J. Weaver, *Centrosymmetric (cross-symmetric) matrices, their basic properties, eigenvalues, and eigenvectors.* Amer. Math. Monthly, **92** (1985), 711-717. [ Links ]

[43] F.Z. Zhou, X.Y. Hu and L. Zhang, *The solvability conditions for the inverse eigenvalue problem of generalized centro-symmetric matrices.* Linear Algebra Appl., **364** (2003), 147-160. [ Links ]

[44] B. Zhou and G.R. Duan, *A new solution to the generalized Sylvester matrix equation AV — EVF = BW.* Systems Control Lett., ** 55** (2006), 193-198. [ Links ]

[45] B. Zhou and G.R. Duan, *Solutions to generalized Sylvester matrix equation by Schur decomposition.* Internat. J. Systems Sci., ** 38** (2007), 369-375. [ Links ]

[46] B. Zhou, Z.Y. Li, G.R. Duan and Y. Wang, *Weighted least squares solutions to general coupled Sylvester matrix equations.* J. Comput. Appl. Math., **224** (2009), 759-776. [ Links ]

Received: 24/VII/11.

Accepted: 05/III/12.

#CAM-394/11.

*Corresponding author.