Acessibilidade / Reportar erro

The nearest generalized doubly stochastic matrix to a real matrix with the same firstand second moments

Abstract

Let T be an arbitrary n × n matrix with real entries. We explicitly find the closest (in Frobenius norm) matrix A to T, where A is n × n with real entries, subject to the condition that A is ''generalized doubly stochastic'' (i.e. Ae = e and eT A = eT, where e = (1,1,...,1)T, although A is not necessarily nonnegative) and A has the same first moment as T (i.e. e1T Ae1 = e1T Te1). We also explicitly find the closest matrix A to T when A is generalized doubly stochastic has the same first moment as T and the same second moment as T (i.e. e1T A²e1 = e1T T²e1), when such a matrix A exists.

doubly stochastic; generalized doubly stochastic; moments; nearest matrix; closest matrix; Frobenius norm


The nearest generalized doubly stochastic matrix to a real matrix with the same firstand second moments

William GluntI; Thomas L. HaydenII; Robert ReamsIII

IDepartment of Mathematics and Computer Science, Austin Peay State University, Clarksville, TN 37044

IIDepartment of Mathematics, University of Kentucky, Lexington, KY 40506

IIIDepartment of Mathematics, Virginia Commonwealth University, 1001 West Main Street, Richmond, VA 23284 E-mails: gluntw@apsu.edu / hayden@ms.uky.edu / rbreams@vcu.edu

ABSTRACT

Let T be an arbitrary n × n matrix with real entries. We explicitly find the closest (in Frobenius norm) matrix A to T, where A is n × n with real entries, subject to the condition that A is ''generalized doubly stochastic'' (i.e. Ae = e and eT A = eT, where e = (1,1,...,1)T, although A is not necessarily nonnegative) and A has the same first moment as T (i.e. e1T Ae1 = e1T Te1). We also explicitly find the closest matrix A to T when A is generalized doubly stochastic has the same first moment as T and the same second moment as T (i.e. e1TA2e1 = e1T T2e1), when such a matrix A exists.

Mathematical subject classification: 15A51, 65K05, 90C25.

Key words: doubly stochastic, generalized doubly stochastic, moments, nearest matrix, closest matrix, Frobenius norm.

1 Introduction

Let e

n be the vector of all ones, i.e. e = (1,1,...,1)T, and let ei
n denote the vector with a 1 in the ith position and zeroes elsewhere. An n × n matrix A with real entries is said to be generalized doubly stochastic if Ae = e and eTA = eT. A generalized doubly stochastic matrix does not necessarily have nonnegative entries, unlike a doubly stochastic matrix which has all entries nonnegative. The k\textth moment of A is defined as Ake1. Let T = (tij)
n×n be an arbitrary matrix which is given. We will say that a matrix A
n×n is Mk if Ake1 = Tke1, where k is a positive integer. We use the convention that ||·|| refers to either the Frobenius matrix norm ||·||F, or the vector 2-norm ||·||2, with the context determining which is intended. Frequent use is made of the fact that if x, y
n are unit vectors we can find a Householder matrix Q
n×n such that Qx = y [5].

We will determine the closest (in Frobenius norm) matrix A to T, subject to the conditions that A is generalized doubly stochastic and has the same first and second moments as T. The motivation for this problem comes from an application in [2] where it is desired to approximate a certain matrix T, where T comes from a linear system corresponding to a large linear network, subject to the approximating matrix satisfying certain conditions. We outlined these applications in [3] and (among other things there) used a computational algorithm to find the closest matrix A to T, subject to A being generalized doubly stochastic and M1, or subject to A being doubly stochastic and M1. See the references [2] and [3] for more details about the applications. Our extended work here takes only an analytic approach to the problem, includes the second moment condition, and explicitly finds the closest matrix which is generalized doubly stochastic, M1 and M2, having dropped the requirement that the closest matrix be nonnegative. It is worth emphasizing that despite dropping the nonnegativity requirement our solution is still relevant to the original problem. Previous approaches to this problem, in both [3] and [8], did not include the second moment, although they did include the nonnegative condition. A survey of matrix nearness problems and their applications, which include areas of control theory, numerical analysis and statistics, was given by Higham [4], see also [7].

2 The closest generalized doubly stochastic matrix

Although Theorem 1 and Corollary 2 resemble results proved in [3] and [6], the results herein present a different formulation. We include them for clarity and because we will use them later.

Theorem 1. Let A

n×n and let Q
n×n be an orthogonal matrix so that Qen =
. Then A is generalized doubly stochastic if and only if

for any A1

( n -1)×( n -1) .

Proof. Ae = e if and only QT AQen = en, and eT A = eT if and only if QT AQ = .

Theorem 1 enables us to easily incorporate the condition that A is generalized doubly stochastic and in Corollary 2 find the closest such matrix to T.

Corollary 2. Let T

n×n and

where T1

(n-1)×(n-1), t2, t3
n-1, t4
, and where Q
n×n is as in Theorem 1. Then the generalized doubly stochastic matrix A
n×n given by

satisfies the inequality ||A - T|| < ||Z - T|| among all generalized doubly stochastic matrices Z

n×n.

Proof. The Frobenius norm is invariant under orthogonal similarity so ||A - T||2 = ||A1 - T1||2 + ||t2||2 + ||t3||2 + (1 - t4)2, and A is minimal among generalized doubly stochastic matrices when A1 = T1.

3 The closest generalized doubly stochastic matrix which is M1

For A

m×m and B
n×n, the direct sum of A and B, denoted A
B, is the (m + n) × (m + n) block matrix .

We now construct an orthogonal matrix Q with an additional desirable property which we need for the first and second moments. Let Q1

n×n be an orthogonal matrix such that

for some u

n-1, β . Let Q2
(n-1)×(n-1) be an orthogonal matrix such that

where α = ||u||, and let Q = Q1(Q2 1). Then Qen = Q1(Q2 1)en = Q1en = as in Section 2, and we have the additional property that

Note that if u = 0, then = βen, so e1 = βQ1en = β, which is not possible, so we must have α ≠ 0.

The proof of Theorem 3 will use the fact, as stated in the introduction, that if T = (tij)

n×n is the matrix to be approximated then saying A
n×n is M1 means Ae1 = Te1 = t11. This theorem gives the form of a matrix A
n×n that is both generalized doubly stochastic and M1.

Theorem 3. Let A

n×n and let Q
n×n be an orthogonal matrix such that

where α, β

. Then A
n×n is generalized doubly stochastic and M1 if and only if

for any A1

(n-2)×(n-2), a2, a3
n-2.

Proof. The matrix A is both generalized doubly stochastic and M1 if andonly if

where

from Theorem 1.

Theorem 1 gives us the means to now find the closest matrix to T which is both generalized doubly stochastic and M1.

Corollary 4. Let T

n×n and

where T1

(n-2)×(n-2), t2, t3,t5, t7
n-2, t4, t6, t8, t9
, and where Q
n×n is an orthogonal matrix such that

where α, β >

. Then the generalized doubly stochastic and M1 matrix A given by

is the closest matrix to T in the sense that ||A - T|| < ||Z - T|| for all generalized doubly stochastic and M1 matrices Z

n×n.

Proof. As in the proof of Corollary 2, since the Frobenius norm is invariant under orthogonal similarity, we have that

and A is minimal when A1 = T1, a2 = t2, and a3 = t3.

4 The closest generalized doubly stochastic matrix which is M1 and M2

Similar reasoning to that given in Section 3 can be used to find necessary and sufficient conditions for a matrix to be generalized doubly stochastic, M1 and M2. However, for finding a nearest point in Corollary 6, a difficulty comes from the fact that the set of M2 matrices is not a convex set, so we don't necessarily expect for there to be a unique nearest point [1]. Although, if there is such a nearest point then it will be determined by the conditions given in Corollary 6.

Theorem 5. Let A

n × n and Q
n × n be an orthogonal matrix such that

where α,β

. Then A
n×n is generalized doubly stochastic, M1 and M2 if and only if

for any A1

( n -2)×( n -2) and any x, y
n -2 such that

Proof.A is generalized doubly stochastic, M1 and M2 if and only if A both satisfies Theorem 3 and satisfies the second moment condition that T2e1 = A2e1. However,

and then substituting

T2e1 and rearranging gives the result.

Preparing for the proof of Corollary 6, and arguing similarly to the preceding corollaries, since the Frobenius norm is invariant under orthogonal similarity we calculate that

Now A is minimal among generalized doubly stochastic, M1 and M2 matrices when A1 = T1, and when x and y are such that ||x-t2||2 + ||y-t3||2 is minimized subject to the constraint

Thus we have our final corollary.

Corollary 6. Let T

n × n and

where T1

(n-2)×(n-2),t2, t3, t5, t7
n-2, t4, t6, t8, t9
, and where Q
n×n is an orthogonal matrix such that

where α, β

. Then the generalized doubly stochastic, M1 and M2 matrix A given by

satisfies the requirement that ||A - T|| < ||Z - T|| for all generalized doubly stochastic, M1 and M2 matrices Z

n×n, where x and y have been chosen (where possible) so as to minimize ||x - t2||2 + ||y - t3||2 subject to the constraint

Proof. For convenience we write c = t2, d = t3 and

It remains for us to solve the problem

for which we find the Kuhn-Tucker conditions to be x = c - μy and y = d - μx, where μ is the Lagrange multiplier. We solve simultaneously xTy = r, x + μy = c and μx + y = d. The latter two equations imply c - μd = (1 - μ2)x and d - μc = (1 - μ2)y, which imply (1 - μ2)2xTy = (c - μd)T(d - μc), and with the first equation this implies

If this quartic equation has no real solution μ then there is no minimum.

Case 1: If each real solution μ ≠ ±1 then for each such μ

and we check to see which pair x, y gives a minimum.

Case 2: If μ = 1 and c = d then we must solve x + y = c and xTy = r, i.e. xT(x-c) = -r, which by completing the square becomes

If - r < 0 there is no minimum in this case. Note that since x + y = c we have

||x - c||2 + ||y - c||2 = ||x + y||2 - 2xTy = ||c||2 - 2r.

This is a constant, so if - r > 0 we can take for any w

n, where w ≠ 0,

Case 3: If μ = -1 and c = -d then a similar calculation shows that if + r < 0 there is no minimum. Whereas if + r > 0 we can take for any w

n, where w ≠ 0,

Acknowledgements. Thomas L. Hayden received partial support from NSF grant CHE-9005960.

Received: 19/X/07. Accepted: 25/III/08.

#738/07.

  • [1] S.K. Berberian, Introduction to Hilbert Space Oxford University Press, New York, (1961).
  • [2] P. Feldmann and R.W. Freund, Efficient Linear Circuit Analysis by Padé Approximation via the Lanczos Process IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, 14(5) (1995), 639-649.
  • [3] W. Glunt, T.L. Hayden and R. Reams, The Nearest ''Doubly Stochastic'' Matrix to a Real Matrix with the same First Moment Numerical Linear Algebra with Applications, 5(1) (1999), 1-8.
  • [4] N.J. Higham, Matrix nearness problems and applications Applications of Matrix Theory, pp. 1-27, The Institute of Mathematics and its Applications Conference Series. New Series. 22 (1989), Oxford University Press, New York.
  • [5] R.A. Horn and C.R. Johnson, Matrix Analysis Cambridge University Press, Cambridge (1985).
  • [6] R.N. Khoury, Closest Matrices in the Space of Generalized Doubly Stochastic Matrices Journal of Mathematical Analysis and Applications, 222 (1998), 562-568.
  • [7] A.W. Marshall and I. Olkin, Inequalities: Theory of Majorization and its Applications Academic Press, New York (1979).
  • [8] Zheng-Jian Bai, Delin Chu and Roger C.E. Tan, Computing the nearest doubly stochastic matrix with a prescribed entry SIAM Journal on Scientific Computing, 29(2) (2007),635-655.

Publication Dates

  • Publication in this collection
    21 July 2008
  • Date of issue
    2008

History

  • Received
    19 Oct 2007
  • Accepted
    25 Mar 2008
Sociedade Brasileira de Matemática Aplicada e Computacional Sociedade Brasileira de Matemática Aplicada e Computacional - SBMAC, Rua Maestro João Seppe, nº. 900 , 16º. andar - Sala 163, 13561-120 São Carlos - SP Brasil, Tel./Fax: 55 16 3412-9752 - São Carlos - SP - Brazil
E-mail: sbmac@sbmac.org.br