Acessibilidade / Reportar erro

A New Hybrid Preconditioner for the Interior Point Method

ABSTRACT

This study aims to improve the computation of the search direction in the primal-dual Interior Point Method through preconditioned iterative methods. It is about a hybrid approach that combines the Controlled Cholesky Factorization preconditioner and the Splitting preconditioner. This approach has shown good results, however, in these preconditioners there are factors that reduce their efficiency, such as faults on the diagonal when performing the Cholesky factorization, as well as a demand for excessive memory, among others. Thus, some modifications are proposed in these preconditioners, as well as a new phase change, in order to improve the performance of the hybrid preconditioner. In the Controlled Cholesky Factorization, the parameters that control the fill-in and the correction of the faults which occur on the diagonal are modified. It considers the relationship between the components from Controlled Cholesky Factorization obtained before and after the fault on the diagonal. In the Splitting preconditioner, in turn, a sparse base is constructed through an appropriate ordering of the columns from constrained matrix optimization problem. In addition, a theoretical result is presented, which shows that, with the proposed ordering, the condition number of the preconditioned Normal Equation matrix with the Splitting preconditioner is uniformly limited by an amount that depends only on the original data of the problem and not on the iteration of the Interior Point Method. Numerical experiments with large scale problems, corroborate the robustness and computational efficiency from this approach.

Keywords:
Interior Point Method; Controlled Cholesky Factorization; Splitting preconditioner

RESUMO

Este trabalho visa melhorar o cálculo da direção de busca no Método de Pontos Interiores primal-dual usando métodos iterativos precondicionados. Trata-se de uma abordagem híbrida que combina o precondicionador Fatoração Controlada de Cholesky e o precondicionador Separador. Esta abordagem tem mostrado bons resultados, entretanto, nesses précondicioandores existem fatores que reduzem sua eficiência, como falhas na diagonal ao calcular a Fatoração Incompleta de Cholesky, assim como a demanda por memória excessiva no precondicionador Separador, entre outros. Assim, algumas modificações são propostas nestes precondicionadores, bem como uma nova mudança de fase, a fim de melhorar o desempenho da aboradagem híbrida. Na Fatoração Controlada de Cholesky, os parâmetros que controlam o preenchimento e a correção das falhas que ocorrem na diagonal são modificados, para tal considera-se a relação entre os componentes da Fatoração Controlada de Cholesky obtidos antes e depois da falha na diagonal. No precondicionador Separador, por sua vez, a base esparsa é construída usando um ordenamento apropriado das colunas da matriz do problema de otimização. Além disso, é apresentado um resultado teórico que mostra que, com a ordenação proposta, o número da condição da matriz de Equações Normais precondicionada com o Precondicionador Separador é uniformemente limitado por uma quantidade que depende apenas dos dados originais do problema e não da iteração do Método do Pontos Interiores. Experimentos numéricos com problemas de grande porte corroboram a robustez e eficiência computacional desta abordagem.

Palavras-chave:
Método de Ponto Interior; Fatoração Controlada de Cholesky; precondicionador Separador

1 INTRODUCTION

Among the Interior Point Methods (IPM) found in the literature today, the primal-dual method of infeasible points using the Mehrotra’s predictor-corrector technique results to be the most computationally efficient, see 88 J. Gondzio . Multiple centrality corrections in a primal-dual method for linear programming. Computational Optimization and Applications, 6(2) (1996), 137-156. doi:10.1007/BF00249643. URL https://doi.org/10.1007/BF00249643.
https://doi.org/10.1007/BF00249643...
), (99 J. Gondzio . Interior point methods 25 years later. European Journal of Operational Research, 218(3) (2012), 587-601.), (2121 S.J. Wrigth. “Primal-dual Interior-Point Methods:”. SIAM e-books. Society for Industrial and Applied Mathematics (SIAM) (1997).. However, the greatest computational effort in all IPM is the computation of the search direction because it results from linear systems that become ill conditioned when IPM is near to achieve the optimal solution. Additionally, this computation may require an excessive memory usage. In large scale and sparse problems, the preconditioning technique and the use of iterative methods are recommended to overcome these difficulties, see 11 G. Al-Jeiroudi, J. Gondzio & J. Hall. Preconditioning indefinite systems in interior point methods for large scale linear optimisation. Optimization Methods and Software, 23(3) (2008), 345-363. doi:10. 1080/10556780701535910. URL https://doi.org/10.1080/10556780701535910.
https://doi.org/10.1080/1055678070153591...
), (33 S. Bocanegra, F.F. Campos & A.R.L. Oliveira. Using a hybrid preconditioner for solving large-scale linear systems arising from interior point methods. Computational Optimization and Applications, 36(2-3) (2007), 149-164..

The search direction can be computed by solving both as the Augmented System (AS) that has an indefinite matrix, as the Normal Equations System (NES) that has a positive definite matrix. In this paper the NES is solved using a hybrid preconditioning approach applied to the Conjugate Gradient Method (CGM). In the early iterations of the IPM is used the Controlled Cholesky Factorization preconditioner (CCF), see 44 F.F. Campos . “Analysis of conjugate gradients-type methods for solving linear equations.”. Ph.D. thesis, University of Oxford (1995)., with the proposed modifications that will be presented in Section 4.1; the objective from the contributions of this paper is to accelerate the construction of CCF preconditioner by reducing the restarts when exist diagonal faults. It was shown in 1010 J. Gondzio . Matrix-Free Interior Point Method. Computacional Optimization and Applications, 51 (2012), 457-480. that the condition number of the NES matrix is the order Oμ-2, where µ denotes the complementarity gap of the Linear Programming (LP) problem, it means, it is inevitable that the performance of the CCF preconditioner get deteriorated when IPM is near to achieve an optimal solution since it is a generic preconditioner. Proposed by 1717 A.R.L. Oliveira & D. Sorensen. A new class of preconditioners for large-scale linear systems from interior point methods for linear programming. Linear Algebra and its applications, 394 (2005), 1-24., the Splitting Preconditioner (SP), in turn, it was exclusively made to overcome the problem of ill conditioning of linear systems from the last IPM iterations.

The CCF preconditioner is obtained by performing an Incomplete Cholesky Factorization (ICF), its fill-in in 44 F.F. Campos . “Analysis of conjugate gradients-type methods for solving linear equations.”. Ph.D. thesis, University of Oxford (1995). allows the preconditioner to vary from a diagonal matrix to another with more nonzero entries than the classical ICF matrix. It is known that any ICF is susceptible to faults on diagonal, however, if a symmetric matrix V is positive definite, it exists a constant α>0 such as an ICF of the matrix V+α diag(V) exists, see 1515 T.A. Maunteuffel. An incomplete factorization technique for positive definite linear systems. Mathematics of computation, 34(150) (1980), 473-497.. Techniques of diagonal modification in the ICF can be found in 1111 M.R. Heredia & A.R.L. Oliveira . Uma nova proposta para modificar a Fatoração Controlada de Cholesky no método dos pontos interiores. 1 (2015), 2912-2923.), (1212 M.T. Jones & P.E. Plassmann. An improved incomplete Cholesky factorization. ACM Transactions on Mathematical Software (TOMS), 21(1) (1995), 5-17.), (1313 C.J. Lin & J.J. Moré . Incomplete Cholesky factorizations with limited memory. SIAM Journal on Scientific Computing, 21(1) (1999), 24-45..

In the original CCF construction proposed in 44 F.F. Campos . “Analysis of conjugate gradients-type methods for solving linear equations.”. Ph.D. thesis, University of Oxford (1995)., the faults that occur during the factoring are corrected with an exponential increase and the computation of the elements from the preconditioner is restarted. On this study, algebraic and geometric tools are used to obtain relationships among the elements that caused the failure and the new components of the matrix obtained with the increment. In addition, it was observed that the parameter that controls the fill-in of the CCF preconditioner is related to the increase of the diagonal. Using these relations, it is proposed a modification of these parameters in order to reduce the number of factoring restarts required for its construction.

The new hybrid preconditioner is compared with the version currently used in 2020 M.I. Velazco, A.R.L. Oliveira & F.F. Campos . A note on hybrid preconditioners for large-scale normal equations arising from interior-point methods. Optimization Methods Software, 25(2) (2010), 321-332. doi:10.1080/10556780902992829.
https://doi.org/10.1080/1055678090299282...
. The computational tests show that the new proposal is more efficient and robust.

We present a criterion that evaluates the performance of the CCF preconditioner and it will indicate the moment of the preconditioner exchange that starts the second phase of hybrid preconditioning using the SP. The SP performance depends on a non singular submatrix B of the constraints matrix A, the choice of columns of B is done using ordering in the columns from matrix A. The authors from SP and, later, their collaborators 2020 M.I. Velazco, A.R.L. Oliveira & F.F. Campos . A note on hybrid preconditioners for large-scale normal equations arising from interior-point methods. Optimization Methods Software, 25(2) (2010), 321-332. doi:10.1080/10556780902992829.
https://doi.org/10.1080/1055678090299282...
, developed an efficient heuristic. However, there are problems where the approach fails or demands an excessive of computational time. Thus, in respect to SP, the objective of this paper is to study the condition number of the preconditioned NES by the SP and, from this study, to order the columns of the matrix of restrictions from LP problem in order to construct a sparse base that provides a number condition limited by an amount that is independent of the IPM iteration.

2 SEARCH DIRECTIONS IN THE INTERIOR POINT METHOD

Consider the linear programming problem

P min c T x s . t . A x = b ; x + s = u ; x , s 0 , and D max b T y - u T w s . t . A T y - w + z = c ; w , z 0 ; y m ,

where x, s, wn and Am×n. We assume that A has full row rank throughout this paper. The search direction in the infeasible Interior Point Method (IPM) is obtained by applying Newton’s method to the optimality conditions of the problem

P ' min c T x - μ i = 1 n log x i - μ i = 1 n log s i s . t . A x = b ; x + s = u ; x , s 0 ,

where the problem (P') results from applying the logarithmic barrier penalty on the nonnegativity constraints of the primal problem (P). Since (P') is a convex problem, the KKT conditions are sufficient and necessary to find the optimal solution. Consider its lagrangian R and the partial derivatives,

l x , s , y , w = c T x - μ i = 1 n log x i - μ i = 1 n log s i + y T b - A x + w T u - x - s , x l = c - μ X - 1 e - A T y - w , s l = - μ S - 1 e - w , y l = b - A x and w l = u - x - s ,

where eT=(1, . . . , 1)n, X1=diag(x1-1,..., xn-1), and S1=diag(S1-1,..., sn-1). If zn is defined as z=µX1e, the optimality conditions of the problem (P') are

A x = b ; x + s = u , x , s > 0 ; A T y + z - w = c , z , w > 0 ; S W e = μ e ; X Z e = μ e , (2.1)

the equations in (2.1) are an implicit parameterization of the central path, 2121 S.J. Wrigth. “Primal-dual Interior-Point Methods:”. SIAM e-books. Society for Industrial and Applied Mathematics (SIAM) (1997)..

In order to obtain the search direction X=(x, s, y, z, w)T, using (2.1) consider the maps F given by

F ( x , s , y , w , z ) = ( A x b , x + s u , A T y + z w c , X Z e μ e , S W e μ e ) ,

apply the Newton’s method in the maps F implies to solve the following linear system:

A 0 0 0 0 I n I n 0 0 0 0 0 A T - I n I n Z 0 0 X 0 0 W 0 0 S Δ x Δ s Δ y Δ z Δ w = r b r u r c r 1 r 2 , (2.2)

where rb=bAx, ru=uxs, rc=c+wzATy, r1=μeXZe, r2=μeSWe, eT=(1,..., 1)n, X=diag(x1,..., xn), Z=diag(z1,..., zn), S=diag(s1,...., sn), W=diag(w1,..., wn).

The predictor-corrector method modifies the right-hand side in (2.2) by

r 1 = σ μ e - X Z e , r 2 = σ μ e - S W e ,

where σ[0, 1] is known as centering parameter.

Substituting the variables

s = r u x , z = X 1 ( r 1 Z x ) and w = S 1 ( r 2 W s )

in the third equation in (2.2) we have:

A T y ( X 1 Z + S 1 W ) x = r c X 1 r 1 + S r 2 S 1 W r u . (2.3)

Considering the equation in (2.3) and the first equation in the system (2.2), we obtain the Augmented System

- Θ - 1 A T A 0 Δ x Δ y = r h , (2.4)

where Θ1=X1Z+S1W, r=rcX1r1+S1r2S1Wru and h=rb. Substituting x=ΘAT yΘr in the equation Ax=h we obtain the Normal Equation’s system

A Θ A T y = h + A Θ r , (2.5)

this system is symmetric and positive definite. In the next section, a Hybrid Preconditioner (HP) is used for the preconditioning of the matrix in (2.5).

3 PRECONDITIONER FOR THE NORMAL EQUATION SYSTEM

The preconditioning technique has as main objective to facilitate the convergence of iterative methods in order to find the solution of linear systems.

In this paper, two-sided preconditioning is used, that is, a preconditioner given by K=K1K2. The iterative method is applied to

K 1 - 1 A K 2 - 1 z = K 1 - 1 b instead of A x = b , (3.1)

where x=K2-1z. Observe that, in (3.1) the system Ax=b represents that system in (2.5) and the matrices K 1 and K 2 are the Controlled Cholesky Factorization (CCF) preconditioner in the first phase and the Splitting Preconditioner (SP) in the second phase. More precisely, in this paper, we denote as the Hybrid Preconditioner (HP) for the Interior Point Method (IPM) to the approch that computing the search direction via the Conjugate Gradient Method (CGM) in two phases, in the early iterations it uses the CCF preconditioner and after a phase change criteria the SP is used. The early iterations use the CCF preconditioner proposed in 44 F.F. Campos . “Analysis of conjugate gradients-type methods for solving linear equations.”. Ph.D. thesis, University of Oxford (1995). successfully, but its performance reduces as the IPM approaches the optimal solution. This can be justified by the fact that the condition number of the non-preconditioned Normal Equation System, see equation (2.5), is of the order O(μ2), where µ denotes the duality gap of the Linear Programming (LP) Problem, see 1010 J. Gondzio . Matrix-Free Interior Point Method. Computacional Optimization and Applications, 51 (2012), 457-480.. However, the SP uses this characteristic to your favor to contain the difficulty provided of ill conditioning. We will make a description of the preconditioners used in this paper.

3.1 Controlled Cholesky Factorization preconditioner

Preconditioners based on an Incomplete Cholesky Factorizaton (ICF) present good performance if the fill-in of the preconditioning matrix is controlled, see 1414 I.J. Lustig, R.E. Marsten & D.F. Shanno. On implementing Mehrotras s predictor-corrector interiorpoint method for linear programming. SIAM Journal on Optimization, 2(3) (1992), 435-449.. Consider the matrix AΘA T from (2.5), let be ℒ and L~, the lower triangular matrices from Cholesky factorization and Incomplete Cholesky Factorization (ICF) of matrix AΘA T , respectively, that is,

LL T = A Θ A T L ~ L ~ T + R ,

where R is the residual matrix. We define the matrix E=L-L~, then,

L ~ - 1 A Θ A T L ~ - T = I + L ~ - 1 E I + L ~ - 1 E T ,

observe that if L~L then E0 and therefore L~-1AΘATL~-TIm, this fact motivate the construction of the Controlled Cholesky Factorization (CCF) preconditioner. More precisely, the CCF preconditioner is a type of the ICF based in the minimization of the Frobenius norm of the matrix E, that is: min EF2. For this is consideraded the minimize problem:

min j = 1 m c j , where c j = i = 1 m l i j - l ~ i j 2 , (3.2)

rewriting this problem, we have:

min j = 1 m k = 1 m j + η l i k j - l ~ i k j 2 + k = m j + η + 1 m l i k j 2 , (3.3)

where m is the order of the matrix, m j is the number of nonzero entries below the diagonal in the j-th column of the matrix AΘA T and η is the extra number of nonzero entries allowed per column. Note that we want to minimize the problem in (3.3), the entries I~ikj will be higher in absolute value. We will denote by L^ the CCF matrix, that is, the matrix containing the major elements of ICF. Thus, the NES given in 3.4 preconditioned by CCF preconditioner is:

L ^ - 1 A Θ A T L ^ - T Δ y ^ = L ^ - 1 h + A Θ r , (3.4)

where Δy^=L^TΔy.

In the first iteration the number of nonzero entries allowed is given by:

η 0 = nnz A Θ A T / m , if nnz A Θ A T < 10 m ; - nnz A Θ A T / m , other case . (3.5)

As the number of CGM iterations is increased, it becomes necessary that the value of η is increased. That is, if the number of CGM iterations exceeds m/5 the value of η is increased by 10, see 33 S. Bocanegra, F.F. Campos & A.R.L. Oliveira. Using a hybrid preconditioner for solving large-scale linear systems arising from interior point methods. Computational Optimization and Applications, 36(2-3) (2007), 149-164.. When η>10 the change of phases is made in Hybrid Preconditioner (HP), see 2020 M.I. Velazco, A.R.L. Oliveira & F.F. Campos . A note on hybrid preconditioners for large-scale normal equations arising from interior-point methods. Optimization Methods Software, 25(2) (2010), 321-332. doi:10.1080/10556780902992829.
https://doi.org/10.1080/1055678090299282...
.

In the construction of the CCF preconditioner it is possible to find diagonal faults, these faults are corrected with an exponential increase. The increment value is:

α t = 5 · 10 - 4 · 2 t - 1 , (3.6)

where t=1,..., 15 represents the number of allowable restarts on the CCF, see 33 S. Bocanegra, F.F. Campos & A.R.L. Oliveira. Using a hybrid preconditioner for solving large-scale linear systems arising from interior point methods. Computational Optimization and Applications, 36(2-3) (2007), 149-164.. There are other sequences to compute an increase in the diagonal, see 1111 M.R. Heredia & A.R.L. Oliveira . Uma nova proposta para modificar a Fatoração Controlada de Cholesky no método dos pontos interiores. 1 (2015), 2912-2923.), (1212 M.T. Jones & P.E. Plassmann. An improved incomplete Cholesky factorization. ACM Transactions on Mathematical Software (TOMS), 21(1) (1995), 5-17.), (1515 T.A. Maunteuffel. An incomplete factorization technique for positive definite linear systems. Mathematics of computation, 34(150) (1980), 473-497..

In this way, every time that a diagonal fault occurs the computing of the matrix elements L~ is restarted and if the number of restarts is 15, the value of α is not a small value. In order to avoid restarts an approach is proposed in 1818 L.M. Silva & A.R.L. Oliveira . Melhoria do desempenho da fatoração controlada de Cholesky no precondicionamento de sistemas lineares oriundos dos métodos de pontos interiores. In “Proceeding Series of the Brazilian Society of Computational and Applied Mathematics”, volume 3. SBMAC (2015), pp. 1-7.. This approach was based on the paper of 22 S. Bellavia, V. Simone, D. Serafino & B. Morini. A preconditioning framework for sequences of diagonally modified linear systems arising in optimization. SIAM Journal on Numerical Analysis, 50(6) (2012), 3280-3302.. In Section 4, we apresented an approach such that the number of restarts to compute the CCF preconditioner is reduced by looking for the value of the increment in the main diagonal of matrix A to be close to the value proposed by 1515 T.A. Maunteuffel. An incomplete factorization technique for positive definite linear systems. Mathematics of computation, 34(150) (1980), 473-497..

3.2 Splitting preconditioner applied to the normal equation system

The Splitting Preconditioner is based on the complementary slackness conditions of the Linear Programming problem (P) and (D), that is,

x i z i = 0 and s i w i = 0 for all i = 1 , . . . , n . (3.7)

Note that the diagonal matrix Θ components θj=(zj/xj+wj/sj)1 given in (2.4) and (2.5) changes in each IPM iteration, particularly next to the optimal solution due to (3.7) and the nonnegativity of the variables x, z, s, w, there will be indexes j1,...n such that θj0 or θj. This feature is the reason for the good performance of the Splitting preconditioner in the last IPM iterations.

In each IPM iteration consider the ordering θσ(1)...θσ(m)...θσ(n), where σ is a permutation of the set 1,...n, this permutation changes from iteration to iteration. The sets of indexes are denoted by B=σ(1),..., σ(m) and N=σ(m+1),..., σ(n), if the A and Θ columns are reordered according to σ , the matrix (2.5) can be written as

A Θ A T = A B Θ B A B T + A N Θ N A N T . (3.8)

If the submatrix A is non-singular, the Splitting preconditioner for the normal equation’s system is given by the matrix

P = A B Θ B 1 / 2 , (3.9)

in this case, ℬ and A are known as basic indexes and base of the SP , respectively.

Preconditioning the matrix given in (3.8) for P, we obtain

P - 1 A Θ A T P - T = I m + W W T where W = Θ B - 1 / 2 A B - 1 A N Θ N 1 / 2 .

An ideal situation would occur if ΘB-1/20 and ΘN1/20 implying that W0 and, thus, P1(AΘAT )PTIm. However, nothing guarantees that this matrix A is non-singular and even if so, not all θ j with j is a large value. In fact, close to the optimal solution there are at least nm values close to zero, this implies that there will be a maximum of m not small values.

However, if =σ(1),..., σ(m) is the indexes set used in an IPM iteration for the SP construction, an advantageous property of this preconditioner is that the same indexes may be reused by several iterations making these iterations much cheaper. The papers 11 G. Al-Jeiroudi, J. Gondzio & J. Hall. Preconditioning indefinite systems in interior point methods for large scale linear optimisation. Optimization Methods and Software, 23(3) (2008), 345-363. doi:10. 1080/10556780701535910. URL https://doi.org/10.1080/10556780701535910.
https://doi.org/10.1080/1055678070153591...
), (55 L. Casacio, C. Lyra, A.R.L. Oliveira & C.O. Castro. Improving the Preconditioning of Linear Systems from Interior Point Methods. Comput. Oper. Res., 85(C) (2017), 129-138. doi:10.1016/j.cor.2017.04.005. URL https://doi.org/10.1016/j.cor.2017.04.005.
https://doi.org/10.1016/j.cor.2017.04.00...
), (99 J. Gondzio . Interior point methods 25 years later. European Journal of Operational Research, 218(3) (2012), 587-601.), (1919 P. Suñagua & A.R.L. Oliveira . A new approach for finding a basis for the splitting preconditioner for linear systems from interior point methods. Computational Optimization and Applications, 67(1) (2017), 111-127. URL https://EconPapers.repec.org/RePEc:spr:coopap:v:67:y:2017:i:1:d:10.1007_s10589-016-9887-0.
https://EconPapers.repec.org/RePEc:spr:c...
study the choice of basic indexes for a preconditioner.

In order to study the condition number of the preconditioned matrix, assume that λ and v are an eigenvalue and an eigenvector of the matrix I+WWT, that is, v+WWT v=λv, multiplying by vector v T this equation, we note that |λ|1, that is

κ P - 1 A Θ A T P - T = λ max λ min λ max .

On the other hand, we observe that: λmax(P-1AΘATP-T)=P-1AΘ1/222 and

λ max P - 1 A Θ A T P - T = P - 1 A Θ 1 / 2 F 2 = i = 1 n θ j P - 1 A j 2 2 , (3.10)

we use (3.10) to find an upper bound condition number κ(P1(AΘAT)PT) in the Section 4.3. To find the linearly independent columns of A to form the base of the SP may require an excessive memory usage because it is done trough a factorization LU of matrix A, a small pivot or zero indicates that the column corresponding is linearly dependent. The technique proposed in 1717 A.R.L. Oliveira & D. Sorensen. A new class of preconditioners for large-scale linear systems from interior point methods for linear programming. Linear Algebra and its applications, 394 (2005), 1-24. to deal the excessive fill-in is the interruption of the factorization and reordering the independent columns found thus far by the number of nonzero entries. In 1717 A.R.L. Oliveira & D. Sorensen. A new class of preconditioners for large-scale linear systems from interior point methods for linear programming. Linear Algebra and its applications, 394 (2005), 1-24., the authors of the SP suggested the choice of the first m linearly independent columns of matrix A reordered giving priority to the indexes j of the θj/Aj1 values in decreasing order. In 2020 M.I. Velazco, A.R.L. Oliveira & F.F. Campos . A note on hybrid preconditioners for large-scale normal equations arising from interior-point methods. Optimization Methods Software, 25(2) (2010), 321-332. doi:10.1080/10556780902992829.
https://doi.org/10.1080/1055678090299282...
was proposed a new column ordering according θj/Aj2 values in decreasing order, which achieved better results in the SP performance.

4 NEW PROPOSALS

4.1 Fault correction parameter

Hereafter, the matrix AΘA T is denoted by 𝒜. Suppose 𝒜 is a scaled matrix, that is, ajj=1 and aij1 for i, j=1,...m. In the construction of the CCF preconditioner, it is said that there is a diagonal fault when dj<ε for some j=1,..., m.

The proposal for the computation of the new increment α t considers the LDL T factorization of 𝒜 and 𝒜+αI. That is, if 𝒜=𝒜+αI, we look for the matrices L, D,L and D such that A=L D LT. The subscript t of α t indicates the number of attempts to correct the diagonal fault. The CCF preconditioner allows up to fifteen attempts, that is, up to fifteen restarts in its construction. From now on, for simplicity, α t is denoted only as α. Next, we establish the dependence between the entries of the matrices L and L with respect to the parameter α:

d j = a j j - k = 1 j - 1 d k l j k 2 ; (4.1a)

l i j = 1 d j a i j - k = 1 j - 1 l i k d k l j k ; (4.1b)

given that A=A+αI, we get:

d j = a j j + α - k = 1 j - 1 d k l j k 2 ; (4.2a)

l i j = 1 d j a i j - k = 1 j - 1 l i k d k l j k , (4.2b)

for j=1,...m and i=j+1,..., m. We obtain dj>ε in (4.2a) each time a value α is incremented in the diagonal of 𝒜, this would imply that the umerical value of the sumatory k=1j-1dkljk2 is decreasing. Computationally, this verified in 44 F.F. Campos . “Analysis of conjugate gradients-type methods for solving linear equations.”. Ph.D. thesis, University of Oxford (1995). and this study presented the Proposition 4.1, to justify this fact.

Proposition 4.1.Ifdj<ε, the functionsFu : +, given by

α k = 1 u - 1 d k l u k 2 , (4.3)

are decreasing, where u = 2 , . . . , j and F u 0 = k = 1 u - 1 d k l u k 2 Furthermore,

  • (i) d j > ε for all α ε d j .

  • (ii) If 0 < d j < ε , then d j > ε for all α ε .

Proof. Let’s j such that dj<ε and we assume that the function F j is decreasing and we proof this fact later. So for every α>0, we have Fj(0)>Fj(α) or

k = 1 j - 1 d k l j k 2 - d k l u k 2 > 0 . (4.4)

This inequality is used to show ((i)) and ((ii)):

  • (i) In fact, by equation (4.2a):

d j = a j j - k = 1 j - 1 d k l j k 2 + α + k = 1 j - 1 d k l j k 2 - k = 1 j - 1 d k l j k 2 = d j + α + k = 1 j - 1 d k l j k 2 - d k l j k 2 > d j + α ε , (4.5)

  • the first inequality follows from (4.4) and the second inequality from αεdj .

  • (ii) Observe that αε , we have:

d j = d j + α + k = 1 j - 1 d k l j k 2 - d k l j k 2 > d j + α d j + ε > ε , (4.6)

  • since αε and dj>0 .

In this way, the next step is to prove that the function F u defined in (4.3), is decreasing for u=2,..., j where j=2,..., m.

For every j, differentiating the function dkljk in relation to the variable α:

d k l j k ' = d k ' l j k + d k l j k ' . (4.7)

Since that Fj'α=k=1j-1dk'ljk2+2ljkdkljk' we use the Equation 4.7 to obtain:

F j ' α = k = 1 j - 1 - d k ' l j k 2 + 2 l j k d k l j k ' . (4.8)

The following will demonstrate that for every u=2,..., j1, Ft'(α)<0 and consequently F j will be decreasing. In fact, using Mathematical Induction,

  • a) Basis: when u=2 , from the equation (4.1):

d 1 ' = d 1 + α ' = 1 and d 1 l 21 ' = d 1 + l 21 ' = 0 . (4.9)

Substituting the equations from 4.9 in (4.8):

F 2 ' ( α ) = - d 1 ' l 21 2 + 2 l 21 d 1 l 21 ' = l 21 2 0 .

  • b) Inductive step: Assume for every u=2,..., j1, Fu'(α)0 . It must then be shown that Fj'(α)0 . Otherwise, assume Fj'(α)>0 , from the equation (4.8):

0 < F j ' ( α ) = k = 1 j - 1 - d k ' l j k 2 + 2 l j k d k l j k ' ,

since that 0=k=1j-1dk'ljk2, we have:

0 < k = 1 j - 1 2 l ¯ j k d k l ¯ j k ' . (4.10)

If the inequality (4.10) were true, it would imply that exist an index r, r1,..., j1, such that ljrdrljr'>0. From the equation (4.10), it would follow that the function drljr2 is increasing. In fact, for all u1,..., j1 the function du is increasing, since that we have the equation (4.1a) and the basis of the Mathematical Induction:

d u ' = 1 - k = 1 u - 1 d k l k 2 ' = 1 - F u ' α > 0 , (4.11)

for every u=1,..., j1. Thus, when α>0, we use (4.11) to obtain: 1 1 Observe that if exist diagonal fault in j-column, du>ε for all u=1,..., j−1 .

d u > d u > ε > 0 ,

that mean, dt>0 for all u1,..., j1. In order to prove that the function drljr2 is increasing, we must consider u=r to garantee dr>0. Thus, from the equation (4.10):

d r l j r 2 ' = 2 d r l j r d r l j r ' > 0 , (4.12)

that is, the function drljr2 is increasing. We use (4.12) in order to obtain a contradiction. If drljr2 is increasing, for all α>0,

d r l j r 2 < d r l j r 2 ,

and consequently from the equation (4.11), for all α>0, it would follow that:

d r l j r 2 < d r l j r 2 = d r l j r + s = 1 r - 1 l j s d s l r s - l j s d s l r s 2 .

However, when α approaches zero:

d r l j r 2 < lim α 0 d r l j r + s = 1 r - 1 l j s d s l r s - l j s d s l r s 2 = d r l j r 2 ,

that means, a contradiction.

Therefore, Fu'α0 for every u=2,..., j. Since that j is arbitrary we have that the function F j is decreasing, for every j=2,..., m. ◻

As a consequence of the Proposition 4.1, the value α=εdj could be used. However, it is necessary that α be as small as possible so that in this way AA and we have nothing to guarantee that ε-dj is a small value. Thus, it is proposed to solve the following problem:

P α min α > 0 α s . t . k = 1 j - 1 d k l j k 2 a i j + α - ε . (4.13)

This approach is a consequence of the equation (4.1) and the fact that for each α>0, dj>ε is satisfied, if, and only if, k=1j-1 dkljk2ajj+α-ε. Observe that in this case, the values dk and ljk are not known for all k=1,..., j1, because the factorization for Obtain L D LT, has not been done yet. In order to get an approximation of the solution of the problem Pα, look for a function that is equivalent to Fj : , given by αk=1j-1 dkljk2 when α approaches zero2 2 The functions f (x) and g(x) are are called equivalents when x approaches a if limx→afxgx=1 ; this is denoted as f ~x→a g . , for some j2,...m.

Using the Proposition 4.1, for every α>0, we have Fj(α)<Fj(0) or

k = 1 j - 1 d k l t k 2 < k = 1 j - 1 d k l j k 2 , (4.14)

where j=2..., m. Consider the following functions

f j : a n d g j : given by α k = 1 j - 1 d k l j k 2 d k + α α k = 1 j - 1 α d k + α d k l j k 2 , (4.15)

respectively. We use the functions f j and g j because for every α>0, we have:

f j + g j α = k = 1 j - 1 d k l j k 2

furthermore, from the Proposition 4.1, we obtain: k=1j-1 dkljk2fj+gjα.

Since that fjα~k=1j-1 dkljk2 when α approaches zero, we are looking for the solution of the following problem:

P α min α > 0 α s . t . f j α a j j - ε + α ,

in order to obtain an approximate solution of Pα.

Since the function f j is decreasing, we have that α is solution of the problem (Pα) if, and only if, fj(α)=ajjε+α. We use the Newton Raphson method for computing the numerical value of α in the Algorithm 1.

Algorithm 1:
Find the value a that solve the equation fj(α)=ajjε+α.

4.2 Modification in the fill-in of the Controlled Cholesky Factorization preconditioner

We will denote by η the fill-in parameter of the Controlled Cholesky Factorization (CCF) preconditioner. The objective in this proposal is to ensure that the number of nonzero entries (nnz) of the matrix L^ in the equation (3.4) is at most (nnz(𝒜)+3m)/2 when the CCF preconditioner is used in the IPM iterations. Thus, from (3.4) we have nnzL^=nnzL~.

In order to determine the initial parameter η, denoted by η 0, the quotient nnzA/nnzA is computed and from it we have the following cases:

  • (i) η0=1 , if 1nnz𝒜/nnzA<2 ;

  • (ii) η0=nnzA/m , otherwise.

If the number of the preconditioned CGM iterations is greater than m/5, the heuristic to determine the increment of η is given by:

  • (i) ηk=1 , when η0=1 ;

  • (ii) If for all j=1,..., m and i=j+1,..., m, lij<1/1+α in the iteration k1 , then the value η k will be incremented, ηk=ηk1/2 , if ηk1<0 .

The final η, denoted by η f , in both cases ((i)) and ((ii)) will be ηf1. Thus, the largest fill-in allowed for L^ will be (nnzA+m)/2+m.

4.3 New ordering criteria of basic indexes for Spliting Preconditioner

Based in the observations presented in the Section 3.2, the idea arises of an ordering that simultaneously considers well conditioning and sparsity for the base of the Splitting Preconditioner (SP). We denote by nnz(A j ) to the number of nonzero entries in column A j of the constrained matrix Am×n of the LP problem, where for j=1,..., n.

Observe that; 1nnz(Aj)m for all column A j of A, however, in sparse problems nnz(Aj)<<m. Define kj=θj1/2 nnz(Aj) and perform a decreasing ordering of the k j elements, with this order the Algorithm 2 is proposed.

Algorithm 2:
In order to find the basic indexes of the Splitting preconditioner

The non-increasing ordering of values k' j s is motivated by the following reason: if two columns A j1 and A j2 satisfy nnz(Aj1)nnz(Aj2), that is A j1 is more sparse than A j2 , then,

1 / nnz ( A j 1 ) 1 / nnz ( A j 2 )

Therefore, the column A j1 will have priority over A j2 if θj1θj2 . Thus, while the values θj1/2 will be used in the Theorem 4.1 to take care of well conditioning, the values nnz(A j ) will give priority to the sparse columns. The algoritm 2 and the proof of the Theorem 4.1 are based in 1616 R.D. Monteiro, J.W. O’Neal & T. Tsuchiya. Uniform boundedness of a preconditioned normal matrix used in interior-point methods. SIAM Journal on Optimization, 15(1) (2004), 96-100. adding a condition that takes into account the sparseness of the A columns. To simplify the notation we consider the permutation σ=id, where id is the identity permutation, in addition, we denote A simply by B.

Teorema 4.1. Suppose that the basic and non-basic index sets ℬ and 𝒩 of the Splitting preconditioner are obtained by thealgorithm 2. Then

  1. θj1/2ΘB-1/2B-1Aj=1 for jB ;

  2. θj1/2ΘB-1/2B-1AjnnzAjB-1Aj for jN=1,..., n\B . Also,κP-1AΘATP-TnK2B-1A2, whenK=maxnnzAj : j=1,...n .

Proof. The proof this theorem consider two cases.

Case 1. If jB, then B1Aj=ej where e j is the j-th unit vector of ℝm so,

θ j 1 / 2 Θ B - 1 / 2 B - 1 A j = θ j 1 / 2 Θ B - 1 / 2 e j = θ j 1 / 2 θ j 1 / 2 e j = 1 . (4.16)

Case 2. If jN, two cases are considered.

Case 2.1. The column A j was not considered to enter the base according to Algorithm 2, that is j>bi, thus

k b i k j for all b i B . (4.17)

Let θ01/2=minθbi1/2 : biB, if , since b m is the last basic index we have k0kbm, using (4.17) k0kj, thus

θ j 1 / 2 Θ B - 1 / 2 B - 1 A j d j 1 / 2 B - 1 A j m i n θ b i 1 / 2 : b i B = k j nnz A j k 0 nnz A 0 B - 1 A j nn z A j B - 1 A j . (4.18)

Case 2.2. The column A j was considered to be r-th column of B, however A j resulted to be linearly dependent on the colunms Ab1, Ab2,..., Abr-1, that is, Aj=Bu, 0T, where ur1, observe that u=B1Aj. Furthermore, kbikj for all i=1,..., r-1. If θ01/2=minθb11/2,..., θbr-11/2 and we define k0 :=θ01/2/nnzA0, then k0kbr-1kj thus,

θ j 1 / 2 Θ B - 1 / 2 B - 1 A j = θ j 1 / 2 i = 1 r - 1 θ b i - 1 u i 2 1 / 2 k j nnz A j k 0 nnz A 0 B - 1 A j nn z A j B - 1 A j . (4.19)

Using (3.10) we have that

λ max = Θ B - 1 / 2 B - 1 A Θ 1 / 2 2 2 j = 1 n θ j Θ B - 1 / 2 B - 1 A j 2 2 , (4.20)

substituting (4.16), (4.18) and (4.19) in (4.20) we have

λ max ( P - 1 A Θ A T P - T ) K 2 j = 1 n B - 1 A j 2 = K 2 B - 1 A F 2 m K 2 B - 1 A 2 .

Furthermore, we have

κ P - 1 A Θ A T P - T m K 2 B - 1 A 2 , (4.21)

since λmin (P-1AΘATP-T)1. ◻

Note that the condition number of the matrix preconditioned in (4.21) is uniformly bounded by amount that depends on the data of the problem and not on the interior point method iteration.

5 NUMERICAL XPERIMENTS

The PCx was originally proposed by 66 J. Czyzyk, S. Mehrotra, M. Wagner & S.J. Wright. PCx An Interior Point Code for Linear Programming. Optimization Methods & Software, 11 (1999), 397-430. and in this paper to perform the numerical experiments the PCx was modified. The direct method used in PCx to obtain the solution of linear systems was replaced by an iterative method 33 S. Bocanegra, F.F. Campos & A.R.L. Oliveira. Using a hybrid preconditioner for solving large-scale linear systems arising from interior point methods. Computational Optimization and Applications, 36(2-3) (2007), 149-164..

The tests performed compare the Hybrid Preconditioner (HP) proposed in 2020 M.I. Velazco, A.R.L. Oliveira & F.F. Campos . A note on hybrid preconditioners for large-scale normal equations arising from interior-point methods. Optimization Methods Software, 25(2) (2010), 321-332. doi:10.1080/10556780902992829.
https://doi.org/10.1080/1055678090299282...
, and the preconditioner presented in this study, denoted by HPmod. In SP, the base B can be maintained in some iterations of the IPM, this base is changed when 8*ngm, where n g denotes the number of iterations of preconditioned CGM in an IPM iteration. The problems used are in the public domain of the Netlib, QAP, and Kennington repositories.

The first two columns of Table 1 indicate the number of rows and columns of preprocessed problems. The number of CCF restarts corresponding to all iterations of the first phase, the time each problem was solved and measured in seconds, and finally, in the last two columns, the number of iterations of the IPM for solve each problem. For the comparison of approaches, the results presented in Table 1 can be summarized in performance profiles. These profiles use a logarithmic scale in the base 2, see 77 E.D. Dolan & J.J. Moré. Benchmarking optimization software with performance profiles. Mathematical programming, 91(2) (2002), 201-213..

Table 1:
Performance of the approaches.

Large scaled problems were tested and the criteria for choosing them depended on whether the number of rows or columns were greater than 5000. The most significant differences are in bold. The symbol “−” indicates that the problem has not been resolved, the symbol “‡” mean that the total number of restarts was greater than 15 in more than one iteration of the IPM.

The HPmod works well compared to the HP preconditioner when evaluating the total number of iterations, see Fig. 1. The proposal to use a hybrid approach is that the CCF and the SP did not work well on their own in most of the problems tested. According to 33 S. Bocanegra, F.F. Campos & A.R.L. Oliveira. Using a hybrid preconditioner for solving large-scale linear systems arising from interior point methods. Computational Optimization and Applications, 36(2-3) (2007), 149-164., the CCF preconditioner shows good results in the initial iterations of the IPM, however it may deteriorate in the latter, since the matrix 𝒜 becomes ill-conditioned. If the SP is used in the early iterations, it is possible that the optimal solution is not found, that happened in the problems: nug05-3rd, nug06-3rd, nug07-3rd and nug08-3rd. In particular, in these problems the HP changes phase, which does not happen in the HPmod, that is, at least in these problems it was not necessary to carry out the phase change and, therefore, the optimal solution was obtained by the CCF preconditioner. In the problems that there was a phase change, when we used the HPmod, the SP computation is performed in less time compared to SP proposed by 2020 M.I. Velazco, A.R.L. Oliveira & F.F. Campos . A note on hybrid preconditioners for large-scale normal equations arising from interior-point methods. Optimization Methods Software, 25(2) (2010), 321-332. doi:10.1080/10556780902992829.
https://doi.org/10.1080/1055678090299282...
, due to the sparse columns that were used.

Figure 1:
Performance profile for iterations of IPM.

Note that the CCF preconditioner that we propose can solve problems osa-14, osa-30 and osa60 without restarts, see Table 1. Therefore, to elaborate the performance profile presented in Fig. 2 the value 0.1 was considered instead of 0. Using the preconditioner CCF the diagonal fault correction parameter allows to increase a value lower than that computed with the CCF.

Figure 2:
Performance profile for restars to compute CCF preconditioner.

The computation of the preconditioning CCF proposed by 44 F.F. Campos . “Analysis of conjugate gradients-type methods for solving linear equations.”. Ph.D. thesis, University of Oxford (1995)., works with a diagonal matrix when the number of restarts is greater than 15. Observe that preconditioning a system in the early iterations using a diagonal matrix seems a good strategy, however in iterations that are not early it may result in an increase in time or even in not finding the optimal solution. It can be seen in Fig. 3 that the HPmod performed better than the other approaches in 32 problems. HPmod solves all the problems, since the curve of its performance profile has reached 1.

Figure 3:
Performance profile for time of IPM.

It is observed that in 14 problems the number of IPM iterations was reduced when we used the HPmod. Problems qap12, osa-14, osa-30, osa-60, nug05-3rd, nug06-3rd, nug07-3rd, and nug083rd were not solved by HP. With respect to the time used to solve the problems, HPmod was superior in 32 of the 34 problems presented. Finally, in the fifth column of Table 1, it can be seen that in the HPmod approach, less than 15 restarts were performed to compute the CCF preconditioner in all IPM iterations.

6 CONCLUSIONS

The modifications in the CCF preconditioner in both the diagonal fault correction parameter α and the fill-in parameter η reduced the number of restarts in the computation of this preconditioner. With the diagonal fault correction parameter of the original CCF preconditioner, see (3.6), up to fifteen attempts to build this preconditioner were allowed in more than one IPM iteration. With the new proposal, that did not happen in any IPM iteration. It resulted in the number of the preconditioned CGM iterations was reduced and therefore the processing time corresponding to the first phase of the HPmod also decreased.

In the SP, the computation of the base was accelerated because the sparse columns generated less fill and in addition, the proposed ordering for the SP performed well because the number of iterations of the preconditioned CGM did not increase and the effort to compute B decreased.

ACKNOWLEDGMENT

We would like to thank the agencies CNPq and FAPESP for grants which supported this research.

REFERENCES

  • 1
    G. Al-Jeiroudi, J. Gondzio & J. Hall. Preconditioning indefinite systems in interior point methods for large scale linear optimisation. Optimization Methods and Software, 23(3) (2008), 345-363. doi:10. 1080/10556780701535910. URL https://doi.org/10.1080/10556780701535910
    » https://doi.org/10. 1080/10556780701535910» https://doi.org/10.1080/10556780701535910
  • 2
    S. Bellavia, V. Simone, D. Serafino & B. Morini. A preconditioning framework for sequences of diagonally modified linear systems arising in optimization. SIAM Journal on Numerical Analysis, 50(6) (2012), 3280-3302.
  • 3
    S. Bocanegra, F.F. Campos & A.R.L. Oliveira. Using a hybrid preconditioner for solving large-scale linear systems arising from interior point methods. Computational Optimization and Applications, 36(2-3) (2007), 149-164.
  • 4
    F.F. Campos . “Analysis of conjugate gradients-type methods for solving linear equations.”. Ph.D. thesis, University of Oxford (1995).
  • 5
    L. Casacio, C. Lyra, A.R.L. Oliveira & C.O. Castro. Improving the Preconditioning of Linear Systems from Interior Point Methods. Comput. Oper. Res., 85(C) (2017), 129-138. doi:10.1016/j.cor.2017.04.005. URL https://doi.org/10.1016/j.cor.2017.04.005
    » https://doi.org/10.1016/j.cor.2017.04.005» https://doi.org/10.1016/j.cor.2017.04.005
  • 6
    J. Czyzyk, S. Mehrotra, M. Wagner & S.J. Wright. PCx An Interior Point Code for Linear Programming. Optimization Methods & Software, 11 (1999), 397-430.
  • 7
    E.D. Dolan & J.J. Moré. Benchmarking optimization software with performance profiles. Mathematical programming, 91(2) (2002), 201-213.
  • 8
    J. Gondzio . Multiple centrality corrections in a primal-dual method for linear programming. Computational Optimization and Applications, 6(2) (1996), 137-156. doi:10.1007/BF00249643. URL https://doi.org/10.1007/BF00249643
    » https://doi.org/10.1007/BF00249643» https://doi.org/10.1007/BF00249643
  • 9
    J. Gondzio . Interior point methods 25 years later. European Journal of Operational Research, 218(3) (2012), 587-601.
  • 10
    J. Gondzio . Matrix-Free Interior Point Method. Computacional Optimization and Applications, 51 (2012), 457-480.
  • 11
    M.R. Heredia & A.R.L. Oliveira . Uma nova proposta para modificar a Fatoração Controlada de Cholesky no método dos pontos interiores. 1 (2015), 2912-2923.
  • 12
    M.T. Jones & P.E. Plassmann. An improved incomplete Cholesky factorization. ACM Transactions on Mathematical Software (TOMS), 21(1) (1995), 5-17.
  • 13
    C.J. Lin & J.J. Moré . Incomplete Cholesky factorizations with limited memory. SIAM Journal on Scientific Computing, 21(1) (1999), 24-45.
  • 14
    I.J. Lustig, R.E. Marsten & D.F. Shanno. On implementing Mehrotras s predictor-corrector interiorpoint method for linear programming. SIAM Journal on Optimization, 2(3) (1992), 435-449.
  • 15
    T.A. Maunteuffel. An incomplete factorization technique for positive definite linear systems. Mathematics of computation, 34(150) (1980), 473-497.
  • 16
    R.D. Monteiro, J.W. O’Neal & T. Tsuchiya. Uniform boundedness of a preconditioned normal matrix used in interior-point methods. SIAM Journal on Optimization, 15(1) (2004), 96-100.
  • 17
    A.R.L. Oliveira & D. Sorensen. A new class of preconditioners for large-scale linear systems from interior point methods for linear programming. Linear Algebra and its applications, 394 (2005), 1-24.
  • 18
    L.M. Silva & A.R.L. Oliveira . Melhoria do desempenho da fatoração controlada de Cholesky no precondicionamento de sistemas lineares oriundos dos métodos de pontos interiores. In “Proceeding Series of the Brazilian Society of Computational and Applied Mathematics”, volume 3. SBMAC (2015), pp. 1-7.
  • 19
    P. Suñagua & A.R.L. Oliveira . A new approach for finding a basis for the splitting preconditioner for linear systems from interior point methods. Computational Optimization and Applications, 67(1) (2017), 111-127. URL https://EconPapers.repec.org/RePEc:spr:coopap:v:67:y:2017:i:1:d:10.1007_s10589-016-9887-0
    » https://EconPapers.repec.org/RePEc:spr:coopap:v:67:y:2017:i:1:d:10.1007_s10589-016-9887-0
  • 20
    M.I. Velazco, A.R.L. Oliveira & F.F. Campos . A note on hybrid preconditioners for large-scale normal equations arising from interior-point methods. Optimization Methods Software, 25(2) (2010), 321-332. doi:10.1080/10556780902992829.
    » https://doi.org/10.1080/10556780902992829
  • 21
    S.J. Wrigth. “Primal-dual Interior-Point Methods:”. SIAM e-books. Society for Industrial and Applied Mathematics (SIAM) (1997).
  • 1
    Observe that if exist diagonal fault in j-column, du>ε for all u=1,..., j1 .
  • 2
    The functions f (x) and g(x) are are called equivalents when x approaches a if limxafxgx=1 ; this is denoted as f ~xa g .

Publication Dates

  • Publication in this collection
    16 Sept 2019
  • Date of issue
    May-Aug 2019

History

  • Received
    19 Dec 2017
  • Accepted
    02 Mar 2019
Sociedade Brasileira de Matemática Aplicada e Computacional Rua Maestro João Seppe, nº. 900, 16º. andar - Sala 163 , 13561-120 São Carlos - SP, Tel. / Fax: (55 16) 3412-9752 - São Carlos - SP - Brazil
E-mail: sbmac@sbmac.org.br