Acessibilidade / Reportar erro

A CONSTRUCTIVE GLOBAL CONVERGENCE OF THE MIXED BARRIER-PENALTY METHOD FOR MATHEMATICAL OPTIMIZATION PROBLEMS

ABSTRACT

In this paper we develop a generic mixed bi-parametric barrier-penalty method based upon barrier and penalty generic algorithms for constrained nonlinear programming problems. When the feasible set is defined by equality and inequality functional constraints, it is possible to provide an explicit barrier and penalty functions. If such case, the continuity and differentiable properties of the restrictions and objective functions could be inherited to the penalized function.

The main contribution of this work is a constructive proof for the global convergence of the sequence generated by the proposed mixed method. The proof uses separately the main results of global convergence of barrier and penalty methods. Finally, for some simple nonlinear problem, we deduce explicitly the mixed barrier-penalty function and illustrate all functions defined in this work. Also we implement MATLAB code for generate iterative points for the mixed method.

Keywords:
nonlinear programming; mixed barrier-penalty methods; convergence of mixed algorithm

1 INTRODUCTION

The mathematical optimization is one of the concepts widely used to analyze many complex decision or allocation problems. In order to better use available resources, optimization techniques allow the selection of values for a certain number of interrelated variables, and with them we could measure the performance and quality of a decision by focusing on some objective functions.

Specifically, a mathematical optimization problem consists of minimizing or maximizing an objective function f (x) subject to restrictions xΩ, where f is a real valued continuous function defined on Ωn. In this work, we consider the feasible set Ω having three types of restrictions

x Ω 1 , x Ω 2 , x Ω 3 (1)

where Ω1 can be whatever restriction set that is difficult to handle, Ω2 is a robust set and Ω3 could be a simple set such as signal or boundary restrictions. The robust set means that it has a dense nonempty interior subset. In other words, the set has an interior, and it is possible to get any boundary point by approaching it from a sequence of interior points, Luenberger & Ye (200811 LUENBERGER DG & YE Y. 2008. Linear and nonlinear programming. 3th ed.. Springer New York.).

According to specifications above, we consider the following optimization problem,

min f ( x ) s . t . x Ω 1 , x Ω 2 , x Ω 3 . (2)

One of the most common nonlinear programming problems formulation is when the restrictions are characterized by equality and inequality functional constraints, Bazaraa et al. (20131 BAZARAA MS, SHERALI HD & SHETTY CM. 2013. Nonlinear programming: theory and algorithms. John Wiley & Sons.), Luenberger & Ye (200811 LUENBERGER DG & YE Y. 2008. Linear and nonlinear programming. 3th ed.. Springer New York.), Wright & Nocedal (199918 WRIGHT SJ & NOCEDAL J. 1999. Numerical optimization. vol. 2. Springer New York.), Griva et al. (20099 GRIVA I, NASH SG & SOFER A. 2009. Linear and nonlinear optimization. vol. 108. Siam.). In which given the continuous functions f : n, h : nm, g : np, the classical nonlinear optimization problem is

min f x s . t . h ( x ) = 0 g ( x ) 0 , (3)

where the restriction sets are given by Ω1=xn : h(x)=0, Ω2=xn : g(x)0 and Ω3=n.

For many decades, many authors proved some theoretical results and proposed several algorithms in order to solve nonlinear optimization problems considering penalty or barrier function methods. Luenberger & Ye (200811 LUENBERGER DG & YE Y. 2008. Linear and nonlinear programming. 3th ed.. Springer New York.), Fiacco & McCormick (19907 FIACCO AV & MCCORMICK GP. 1990. Nonlinear programming: sequential unconstrained minimization techniques. vol. 4. Siam.) state convergence for both methods, Polyak Polyak (197115 POLYAK BT. 1971. The convergence rate of the penalty function method. USSR Computational Mathematics and Mathematical Physics, 11(1): 1-12.) showed convergence rate for penalty function method within Hilbert space, Bertsekas (19762 BERTSEKAS DP. 1976. On penalty and multiplier methods for constrained minimization. SIAM Journal on Control and Optimization, 14(2): 216-235.) obtained convergence and rate of convergence results for the sequences of primal and dual variables generated on penalty and Lagrange multiplier methods, he showed that the multiplier method is faster than the pure penalty method. Fiacco & McCormick (1990) demonstrate by contradiction the global convergence for mixed penalty-barrier method, also Breitfeld & Shanno (19954 BREITFELD MG & SHANNO DF. 1995. A Globally Convergent Penalty-Barrier Algorithm for Nonlinear Programming. In: Operations Research Proceedings 1994. pp. 22-27. Springer.) proposed composite algorithm of augmented Lagrangian, modified log-barrier, and classical log-barrier methods for that they demonstrated global convergence to a first-order stationary point for the constrained problem which was based on Breitfeld & Shanno (19943 BREITFELD MG & SHANNO DF. 1994. A globally convergent penalty-barrier algorithm for nonlinear programming and its computational performance. Rutgers University. Rutgers Center for Operations Research [RUTCOR].).

In this work, we develop the mixed barrier-penalty method for solving a general nonlinear problem (2); and we provide a generic bi-parametric algorithm. The main contribution is a constructive proof of global convergence of sequence generated by that mixed method as an alternative proof to existing ones with slightly different assumptions. Suñagua & Oliveira (201716 SUÑAGUA P & OLIVEIRA AR. 2017. A new approach for finding a basis for the splitting preconditioner for linear systems from interior point methods. Computational Optimization and Applications, 67(1): 111-127.) showed that computational experiments for NETLIB problems work successfully for large scale linear optimization problems.

2 BARRIER METHODS OVERVIEW

Barrier methods are also called interior point or internal penalty methods. Some theoretical results of them were developed by Martınez & Santos (199512 MARTINEZ JM & SANTOS SA. 1995. Métodos computacionais de otimização. Colóquio Brasileiro de Matemática, Apostilas, 20. Available at: https://www.ime.unicamp.br/~martinez/mslivro.pdf.
https://www.ime.unicamp.br/~martinez/msl...
), Luenberger & Ye (200811 LUENBERGER DG & YE Y. 2008. Linear and nonlinear programming. 3th ed.. Springer New York.), Nash & Sofer (199314 NASH SG & SOFER A. 1993. A barrier method for large-scale constrained optimization. ORSA Journal on Computing, 5(1): 40-53.), and Wright (199217 WRIGHT MH. 1992. Interior methods for constrained optimization. Acta Numérica, 1: 341-407.). These methods are applicable to problems of the form

min f ( x ) s . t . x Ω (4)

where f is a continuous function and Ω is a robust restriction set. This kind of set

often arises from the inequality constraints, that is, Ω=xn : g(x)0, for which there is a point x¯Ω such that g(x¯)<0.

Barrier methods work by establishing a barrier on the boundary of the restriction set that prevents a search procedure from leaving the feasible region. A barrier function is a function B(·) defined on the interior set Int(Ω)=x : g(x)<0 of Ω such that (i) B is continuous,(ii) B(x)0, (iii) Bx as x approaches the boundary of Ω. For inequality constraints gi(x)0, i=1, 2,..., p in many practical applications, the barrier functions commonly used are the logarithmic or inverse barrier function. They are defined on Int(Ω) respectively by

B ( x ) = - i = 1 p log - g i x a n d B x = - i = 1 p 1 g i x .

Now, the problem (4) can be transformed into a penalized subproblem

( P µ ) min f ( x ) + µ B ( x ) s . t . x Int ( Ω ) (5)

where µ > 0 is called barrier parameter and we take µ small (going to zero). In this approach, the main assumption is that the original problem (4) has a global solution x . Let x(µ) be a global solution of subproblem (5). When µk0, we hope x(µ k ) converges to x .

Given ϕ(x, µ)=f(x)+µB(x), we have a generic barrier algorithm given in Algorithm 1.

Algorithm 1
Barrier Algorithm

The following Lemma gives a set of inequalities that follow directly from Algorithm 1 steps. A proof is based from Luenberger & Ye (200811 LUENBERGER DG & YE Y. 2008. Linear and nonlinear programming. 3th ed.. Springer New York.) and Martınez & Santos (199512 MARTINEZ JM & SANTOS SA. 1995. Métodos computacionais de otimização. Colóquio Brasileiro de Matemática, Apostilas, 20. Available at: https://www.ime.unicamp.br/~martinez/mslivro.pdf.
https://www.ime.unicamp.br/~martinez/msl...
).

Lemma 2.1.Let {x k } be a sequence generated byAlgorithm 1, then

  1. ϕ ( x k + 1 , µ k ) ϕ ( x k , µ k 1 )

  2. B ( x k + 1 ) B ( x k )

  3. f(xk+1)f(xk).

Proof. Since {µ k } is a monotone decreasing sequence, x k+1 is a global minimizer of (6) and recalling (ii) of barrier condition, B is non-negative function, then

ϕ ( x k + 1 , µ k ) = f ( x k + 1 ) + µ k B ( x k + 1 ) f ( x k ) + µ k B ( x k ) f ( x k ) + µ k 1 B ( x k ) = ϕ ( x k , µ k 1 ) .

For establishes the second inequality, we also have

ϕ ( x k + 1 , µ k ) = f ( x k + 1 ) + µ k B ( x k + 1 ) f ( x k ) + µ k B ( x k ) (7)

ϕ ( x k , µ k - 1 ) = f ( x k ) + µ k - 1 B ( x k ) f ( x k + 1 ) + µ k - 1 B ( x k + 1 ) , (8)

now, using (8) and (7), we get

( µ k µ k 1 ) B ( x k + 1 ) ( µ k µ k 1 ) B ( x k )

eliminating the common factor µkµk1<0, we prove the item 2.

Finally, by previous inequality,

f ( x k + 1 ) + µ k B ( x k + 1 ) f ( x k ) + µ k B ( x k ) f ( x k ) + µ k B ( x k + 1 ) ,

hence f(xk+1)f(xk). □

The global convergence of the barrier method, in the sense that any limit point of the sequence is a solution of problem (4), can be verified from the previous Lemma.

Theorem 2.1.Let {x k } be a sequence generated byAlgorithm 1, in whichµk0. Then, any limit point of the sequence is a global minimizer of problem (4).

Proof. Let fk=minϕ(x,µk) : xInt(Ω) be global minimum value of ϕ(·,µk) on Int(Ω), whose solution is x k+1 . By Lemma 2.1, fkfk+1 for all k. If f=minf(x) : xΩ, then

f 0 f 1 · · · f k f k + 1 · · · f .

First of all, we prove only the sequence {f k } converges to f , next we continue with the demonstration of convergence of some subsequence that will converge to some global minimizer.

Indeed, {f k } is a bounded below monotone decreasing sequence, hence it converges to its infimum, we say f¯. If f¯f, then f¯>f. Recalling x the global minimizer of (4) and since f is a continuous function, then there is an open ball ℬ centered at x such that, for all xInt(Ω) we have,

f ( x ) < f ¯ 1 2 ( f ¯ f ) . (9)

Since B(x)0 for all xInt(Ω), and 0<µk+1<µk , we have 0<µk+1B(x)<µkB(x), xInt(Ω). Therefore,

lim k µ k B ( x ) = 0 , x Int ( Ω ) . (10)

Thus, for any x'Int(Ω), and k large enough, we get

µ k B ( x ' ) < 1 4 ( f ¯ f ) . (11)

Then, from (9) and (11), we have

ϕ ( x ' , µ k ) < f ¯ 1 2 ( f ¯ f ) + 1 4 ( f ¯ f ) = f ¯ 1 4 ( f ¯ f ) < f ¯ ,

which contradicts to fkf¯. Therefore, f¯=f. That is

f k + 1 = ϕ ( x k + 1 , µ k ) f . (12)

Now, let x¯Ω be any subsequential limit of {x k }, more precisely, there is a subsequence xkl such that xklx¯. If x¯x with f(x¯)>f(x), then by continuity of f a subsequence f(xkl)f(x)+µklB(xkl) cannot converge to zero, which contradicts fkf0. Therefore, x¯=x or x¯x, but f(x¯)=f(x). Thus, every limit point generated by Algorithm 1 is a global solution of the problem (4). □

3 PENALTY METHODS OVERVIEW

Given f : n a continuous function, we consider the problem

( G P ) m i n f ( x ) s . t . x Ω 1 x Ω 2 . (13)

where Ω1 and Ω2 are any arbitrary subsets of ℝn . In most applications Ω1 is defined implicitly by functional restrictions as h(x)=0, where h : nm. In some cases, we assume that f and h are twice differentiable functions. A basic assumption is that problem (GP) admits global minimizer, some theoretical results were established by Polyak (197115 POLYAK BT. 1971. The convergence rate of the penalty function method. USSR Computational Mathematics and Mathematical Physics, 11(1): 1-12.), Breitfeld & Shanno (19954 BREITFELD MG & SHANNO DF. 1995. A Globally Convergent Penalty-Barrier Algorithm for Nonlinear Programming. In: Operations Research Proceedings 1994. pp. 22-27. Springer.), Nash (201013 NASH SG. 2010. Penalty and barrier methods. Wiley Encyclopedia of Operations Research and Management Science,.), and Luenberger & Ye (200811 LUENBERGER DG & YE Y. 2008. Linear and nonlinear programming. 3th ed.. Springer New York.).

Given a restriction set Ω1, a penalty function is defined as a function 𝒫 : n satisfying (i) 𝒫 is continuous, (ii) 𝒫(x)=0 if xΩ1, and (iii) 𝒫(x)>0 if xΩ1.

In order to solve the problem (13), the penalty function method solves the following penalized subproblem

Q ρ min f x + ρ P x s . t . x Ω 2 (14)

where ρ>0 is a constant called penalty parameter. For ρ large, is clear that a solution of (14) will be in a region where 𝒫 is small. Thus, when ρ is expected that the corresponding optimal points will approach the feasible set Ω1.

For C 2 class functions, h : nm and g : np, some useful penalty functions 𝒫 based on the type of restrictions h(x)=0 or g(x)0 may be

  1. 𝒫(x)=12h(x)22 , quadratic penalty,

  2. 𝒫(x)=h(x)1 ,

  3. P x = i = 1 p m a x 0 , g i x 2 ,

  4. 𝒫(x)=12h(x)22 +i=1p[max{0, gi(x)}]2,

in the first item the quadratic penalty function preserves C 2 property, but in the last three items 𝒫 is only C 1.

Given ψ(x, ρ)=f(x)+ρ𝒫(x), we have a generic penalty algorithm given in Algorithm 2 for solving the problem (13), that works iteratively updating the parameter ρ before solving the penalized subproblem (14)

Algorithm 2
Penalty Algorithm

In general, one of the suggestions to compute ρk is taking ρ0=1 and ρk+1=10ρk , Fletcher (20138 FLETCHER R. 2013. Practical methods of optimization. John Wiley & Sons, Chichester.). However, when Ω1 is the set of equality constraints h(x)=0, a basic rule that works in practice is that if h(xk)0.1h(xk1), then ρk+1=10ρk, otherwise ρ does not change. That approach was successfully tested for linear programming problems, Suñagua & Oliveira (201716 SUÑAGUA P & OLIVEIRA AR. 2017. A new approach for finding a basis for the splitting preconditioner for linear systems from interior point methods. Computational Optimization and Applications, 67(1): 111-127.).

The following Lemma gives a set of inequalities that follow directly from Algorithm 2 steps. A proof is based in Martınez & Santos (199512 MARTINEZ JM & SANTOS SA. 1995. Métodos computacionais de otimização. Colóquio Brasileiro de Matemática, Apostilas, 20. Available at: https://www.ime.unicamp.br/~martinez/mslivro.pdf.
https://www.ime.unicamp.br/~martinez/msl...
) and Luenberger & Ye (200811 LUENBERGER DG & YE Y. 2008. Linear and nonlinear programming. 3th ed.. Springer New York.).

Lemma 3.1.Let {x k } be a sequence generated byAlgorithm 2, which x k+1 is global solution of problem (Q k ). Then

  1. ψ ( x k , ρ k 1 ) ψ ( x k + 1 , ρ k )

  2. P ( x k + 1 ) P ( x k )

  3. f(xk)f(xk+1).

Proof. Since {ρ k } is a monotone increasing sequence and x k is a global minimizer of subproblem (15), then

ψ ( x k , ρ k 1 ) = f ( x k ) + ρ k 1 P ( x k ) f ( x k + 1 ) + ρ k 1 P ( x k + 1 ) f ( x k + 1 ) + ρ k P ( x k + 1 ) = ψ ( x k + 1 , ρ k ) .

To establish the second inequality, recalling the optimalities of x k and x k+1 , we have

ψ ( x k , ρ k 1 ) = f ( x k ) + ρ k 1 P ( x k ) f ( x k + 1 ) + ρ k 1 P ( x k + 1 ) (16)

ψ ( x k + 1 , ρ k ) = f ( x k + 1 ) + ρ k P ( x k + 1 ) f ( x k ) + ρ k P ( x k ) (17)

using (17) and (16), we get

( ρ k 1 ρ k ) P ( x k ) ( ρ k 1 ρ k ) P ( x k + 1 ) ,

as ρk1<ρk, then 𝒫(xk)𝒫(xk+1). Finally, using this inequality

f ( x k ) + ρ k 1 P ( x k ) f ( x k + 1 ) + ρ k 1 P ( x k + 1 ) f ( x k + 1 ) + ρ k 1 P ( x k )

hence f(xk)f(xk+1). □

Lemma 3.2.If x*is a global minimizer of (GP) then fork=0, 1, 2, ···

f ( x k ) ψ ( x k , ρ k 1 ) f ( x ) .

Consequently,xkΩ1, if and only if, xkis the global solution of (GP).

Proof. Since ρk>0 and 𝒫(x)0 xn and x k is the global minimizer of (Q k−1 ), then

f ( x k ) f ( x k ) + ρ k 1 P ( x k ) f ( x ) + ρ k 1 P ( x ) = f ( x ) ,

where P(x)=0. □

The global convergence of the penalty method, in the sense any limit point of the sequence is a solution, can be verified from the two previous Lemmas.

Theorem 3.1 (Global convergence for penalty method). Let {x k } be a sequence of global minimizers of (Qk ) generated by Algorithm 2 in which ρk+. Then, any limit point of the sequence is a global minimizer of problem (13).

Proof. With a slight change of notation, the proof is based on Martínez & Santos’ demonstration. Indeed, let xkl be a subsequence of {x k } such that xklx¯. By the continuity of f , we have

f x k l f x ¯ . (18)

Let f* be an optimal value of problem (GP). By Lemma 3.1 and Lemma 3.2, the sequence ψ(xk, ρk1) is nondecreasing and bounded above by f , then

lim l ψ ( x k l , ρ k l 1 ) = sup l 1 ψ ( x k l , ρ k l 1 ) = p f . (19)

Thus, using (18) and (19), yields

lim l ρ k l 1 P x k l = lim l f x k l + ρ k l 1 P x k l - f x k l = lim l f x k l + ρ k l 1 P x k l - lim l f x k l = p - f x ¯ .

Since Pxkl0 and ρkl, we conclude that liml Pxkl=0. Using the continuity of 𝒫, Px¯=0, thereby x¯Ω1. To prove the optimality of x¯, just note that by Lemma 3.2, fxklf, then

f ( x ¯ ) = lim l f x k l f ,

which completes the proof, because obviously ff(x¯) and then f(x¯)=f. □

Furthermore, by definition of ψ and (19)

f ( x x l ) ψ ( x x l , ρ k l 1 ) p f ( x ¯ ) p f .

Therefore f(x¯)=p=f, then

lim l ρ k l - 1 P x k l = 0 . (20)

And using (19)

lim l Ψ x k l , ρ k l - 1 = f . (21)

4 MIXED BARRIER-PENALTY METHOD

For continuous function f : n, we consider the general programming problem

( N L P ) min f ( x ) s . t . x Ω 1 , x Ω 2 , x Ω 3 . (22)

where Ω1, Ω2 and Ω3 are restriction sets that are defined in (1).

As in the previous sections, we assume that problem (22) admits a global minimizer. Now, let 𝒫 a penalty function related to Ω1 and B a barrier function related to Ω2. Then, taking the penalty parameter ρ>0 and the barrier parameter μ>0, we have the associate mixed barrier-penalty subproblem,

( B P ρ , µ ) min f ( x ) + ρ P ( x ) + µ B ( x ) s . t . x Int ( Ω 2 ) , x Ω 3 . (23)

Since the general problem (NLP) admits global minimizer, then the problem (BP ρ,µ) in (23) also admits a global solution for any feasible parameter values. Therefore, we define

Φ ( x , ρ , µ ) = f ( x ) + ρ P ( x ) + µ B ( x ) . (24)

In order to solve the general problem (22), we provide a generic algorithm given in Algorithm 3, that works iteratively updating ρ and µ parameters before solving the penalized subproblem (23).

Algorithm 3
Mixed barrier-penalty algorithm

To establish the global convergence of the Algorithm 3, firstly we can associate the additive terms in two convenient ways

Φ ( x , ρ , µ ) = [ f ( x ) + ρ P ( x ) ] + µ B ( x ) = [ f ( x ) + µ B ( x ) ] + ρ P ( x ) . (26)

Therefore, fixing respectively ρ and µ, we define Fρ(x)=f(x)+ρ𝒫(x) and Gµ(x)=f(x)+µB(x), then we associate to (NLP) the following two problems

G P ρ min F ρ x G P μ min G μ x s . t . x Ω 2 s . t . x Ω 1 , x Int Ω 2 x Ω 3 x Ω 3 . (27)

Since the problem (NLP) admits a global minimizer, both (GP ρ ) and (GP µ ) in (27) also admit global minimizers. Therefore, defining

ϕ ρ ( x , µ ) = F ρ ( x ) + µ B ( x ) and ψ µ ( x , ρ ) = G µ ( x ) + ρ P ( x ) .

We have respectively the barrier and penalty subproblems

B P ρ min ϕ ρ x , μ P P μ min ψ μ x , ρ s . t . x Int Ω 2 s . t . x Int Ω 2 x Ω 3 , x Ω 3 (28)

By fixing one of the parameters according to (27), the two problems in (28) are equivalent to (BP ρ,µ ). In fact

ϕ ρ ( x , µ ) = Φ x , ρ , μ = ψ µ ( x , ρ ) , (29)

therefore, we can apply the results obtained in the preceding two sections.

In order to understand more clearly the ideas of the mixed problem, we consider the following particular quadratic problem

min x 1 2 + x 2 2 s . t . x 2 = 2 1 - x 1 0 - 1 - x 2 0 . (30)

According to the contours of the objective function and graph of restrictions in Figure 1, the optimal point is x=(1, 2). First, if we consider the Lagrangian function (x1, x2, λ, u1, u2)=x12+x22+λ(x21)+u1(1x1)+u2(1x2), the Karush Kuhn-Tuker conditions (-Kuhn & Tucker, 195110 KUHN HW & TUCKER AW. 1951. Nonlinear Programming. In: Proceedings of the Second Berkeley Symposium on Mathematical Statistics and Probability. pp. 481-492. Berkeley, California: University of California Press. Available at: http://projecteuclid.org/euclid.bsmsp/1200500249.
http://projecteuclid.org/euclid.bsmsp/12...
) are

x 1 = 2 x 1 u 1 = 0 , x 2 = 2 x 2 + λ u 2 = 0 x 2 2 = 0 , 1 x 1 0 , 1 x 2 0 u 1 ( 1 x 1 ) = 0 , u 2 ( 1 x 2 ) = 0 u 1 0 , u 2 0

whose unique solutions for variables and Lagrangian parameters are x1*=1, x2*=2, u1*=2, u2*=0, λ*=-4.

Figure 1
Convex Problem.

Now, we associate to (30) the mixed barrier-penalty subproblem

Q P ρ , μ min Φ x , ρ , μ s . t . 1 - x 1 < 0 - 1 - x 2 < 0 . (31)

where the penalized objective function is

Φ x , ρ , μ = x 1 2 + x 2 2 + ρ 2 x 2 - 2 2 - μ log x 1 - 1 + log x 2 + 1 - log M

where M is a large enough positive number such that x1>1, x2>1 and (x11)(x2+1)<M, surely this region lies within inequality constraints 1x1<0 and 1x2<0. This condition ensures that the barrier function is non-negative in the region that contains optimal point.

It is easy to see that Φ is a smooth function, thereby from first-order necessary conditions for optimal points, we have

2 x 1 - μ x 1 - 1 = 0 , 2 x 2 + ρ x 2 - 2 - μ x 2 + 1 = 0 . (32)

Solving this nonlinear system, subject to x1>1 and x2>-1, by the substitution method, we obtain

x 1 = 1 + 1 + 2 μ 2 μ 0 1 = x 1 * x 2 = ρ - 2 + 2 - ρ 2 + 4 2 + ρ 2 ρ + μ 2 2 + ρ ρ μ 0 2 = x 2 * . (33)

Thus, for each optimal point in (33), the optimal values of problem (QP ρ,µ ) in (31) is θ(ρ, µ)=Φ(x1, x2, ρ, µ), whose graph is shown in Figure 2 with M=2.

Figure 2
θ(ρ, µ), 0<µ<2.5, 0<ρ<20.

We can see that for fixed µ, θ(ρ, µ) is an increasing function and for fixed ρ, θ(ρ, µ) is a decreasing function. This fact will be showed theoretically in Theorem 4.1.

Furthermore, using (33), when µ0 and ρ, the following gradient’s coefficients in (32) converge to optimal Lagrangian parameters

u 1 = μ x 1 - 1 = 1 + 1 + 2 μ 2 = u 1 * u 2 = μ x 2 + 1 = 2 μ 2 + ρ 2 + 3 ρ + A 0 = u 2 * λ = ρ x 2 - 2 = ρ - 10 - 3 ρ + A 2 2 + ρ - 4 = λ *

where A=4µ(2+ρ)+(2+3ρ)2. In addition, the Hessian matrix for Φ is

2 Φ = 2 + μ x 1 - 1 2 0 0 2 + ρ + μ x 2 + 1 2 = 2 + u 1 2 μ 0 0 2 + ρ + u 2 2 μ ~ 2 + 4 μ 0 0 2 + ρ .

Then ∇2Φ is a positive definite matrix, that guarantees the minimality of x 1 and x 2 in (33). Moreover, the approximate condition number of this matrix is

κ 2 Φ 2 + 4 μ 2 + ρ ,

hence ∇2Φ is ill-conditioned for very small µ and small ρ, however for large ρ, that condition number could be reduced.

Next, we have the global convergence theorem for the mixed barrier-penalty algorithm.

Theorem 4.1 (Global convergence for mixed method).Let {x k } be a sequence of global minimizers of (BP k ) problem in (25) generated by mixedAlgorithm 3in whichρk+andµk0. Then any limit point of sequence is a global minimizer of the (NLP) problem.

Proof. In order to apply the results of the preceding sections, the idea is to fix, one of the parameters in the (BP ρ,µ ) subproblem in (23) one at a time, and apply the corresponding results for each subproblems in (27).

Firstly, to fix ρ, let xkρ be the sequence generated by Algorithm 1 for solving the (GP ρ ) subproblem in (27). By applying Lemma 2.1, we get

ϕ ρ x k + 1 ρ , μ k ϕ ρ x k ρ , μ k - 1 , F ρ x k + 1 ρ F ρ x k ρ . (34)

By the monotonicity in (34) and by (12) the sequence ϕρ(xkρ, µk1) converges to global optimal value of the problem (GP ρ ) in (27), that is,

ϕ ρ ( x k ρ , µ k 1 ) inf k 1 ϕ ρ ( x k ρ , µ k 1 ) = F ρ x * ρ . (35)

In addition, from Theorem 2.1 all convergent subsequence of xkρ converges to a global minimizer of the problem (GP ρ ) in (27).

Similarly, to fix µ, let xkμ be the sequence generated by the associated Algorithm 2 for solving the (GP µ ) subproblem in (27). By applying Lemma 3.1, we get

ψ μ x k μ , ρ k - 1 ψ μ x k μ , ρ k , G μ x k μ G μ x k + 1 μ . (36)

By the monotonicity in (36) and by (21), the sequence ψμ(xkμ,ρk-1) converges to global optimal value of the (GP µ ) problem of (27), that is,

ψ μ x k μ , ρ k - 1 s u p k 1 ψ μ x k μ , ρ k - 1 = G x * μ . (37)

In addition, from Theorem 3.1 all convergent subsequence of xkμ converges to a global minimizer of the problem (GP µ ) in (27). And by Lemma 3.2, we get Gμ(xkμ)Gμ(x*μ),k.

Now let {x k } be a sequence of minimizers obtained by Algorithm 3 for a mixed problem. More precisely, let xk+1=x(ρk,μk) which also minimizes (28), because according to (29), we have

Φ ( x k + 1 , ρ k , μ k ) = f ( x k + 1 ) + ρ k P ( x k + 1 ) + μ k B ( x k + 1 ) = ϕ ρ k ( x k + 1 , μ k ) = ψ μ k ( x k + 1 , ρ k )

Since (35) and (37), we have

F ρ k ( x * ρ k ) ϕ ρ k ( x k + 1 , μ k ) = Φ ( x k + 1 , ρ k , μ k ) = ψ μ k ( x k + 1 , ρ k ) G μ k ( x * μ k ) . (38)

Giving x(ρ k , µ k ) the solution of (BP ρ ,µ ) for μ=μk and ρ=ρk. For μk<μk1, additionally we solve (BP ρ ,µ ) for μ=μk1 and ρ=ρk, which solution is called x(ρ k , µ k−1 ). Using (34)

ϕ ρ k ( x ( ρ k , μ k ) , μ k ) ϕ ρ k ( x ( ρ k , μ k 1 ) , μ k 1 ) (39)

and by (29), for x=x(ρk,μk) and y=x(ρk,μk-1), we have

ϕ ρ k ( x , μ k ) = f ( x ) + ρ k P ( x ) + μ k B ( x ) = ψ μ k ( x , ρ k ) ϕ ρ k ( y , μ k 1 ) = F ρ k ( y ) + μ k 1 B ( y ) = f ( y ) + ρ k P ( y ) + μ k 1 B ( y ) = f ( y ) + μ k 1 B ( y ) + ρ k P ( y ) = G μ k 1 ( y ) + ρ k P ( y ) = ψ μ k 1 ( y , ρ k ) (40)

Using (39) and (40), we get

ψ μ k x ρ k , μ k , ρ k ψ μ k - 1 x ρ k , μ k - 1 , ρ k G μ k x * μ k G μ k - 1 x * μ k - 1 .

Similarly, for ρk>ρk1, additionally we consider a solution of (BP ρ ,µ ) for ρ=ρk1 and μ=μk, which solution is called x(ρ k−1 k ). Using (36)

ψ μ k x ρ k - 1 , μ k , ρ k - 1 ψ μ k x ρ k , μ k , ρ k (41)

and by (29) for x=xρk,μk and z=xρk-1,μk, we have

ψ μ k x , ρ k = f x + μ k B x + ρ k P x = ϕ ρ k x , μ k ψ μ k z , ρ k - 1 = G μ k z + ρ k - 1 P z = f z + μ k B z + ρ k - 1 P z = f z + ρ k - 1 P z + μ k B z = F ρ k - 1 z + μ k B z = ϕ ρ k - 1 z , μ k (42)

Using (41) and (42), we get

ϕ ρ k - 1 x ρ k - 1 , μ k , μ k ϕ ρ k x ρ k , μ k , μ k F ρ k - 1 x * ρ k - 1 F ρ k x * ρ k .

Let x* be a global minimizer of (NLP). Recalling x*μk a solution of the problem GPμk in (27), with the additional assumption x*μkIntΩ2, we can conclude that fx*Gμkx*μk. Moreover x* is a feasible point of the problem (GP ρ ), then Fρkx*ρkfx*. Therefore, Gμkx*μk is a monotone nonincreasing sequence that is bounded below by f(x*) and using (12) this sequence converges to its infimum f(x*). Also, Fρkx*ρk is monotone nondecreasing sequence that is bounded above by f(x*) and by (21) that sequence converges to its supremum f(x*), that is,

F ρ k x * ρ k s u p k 1 F ρ k x * ρ k = f x * , G μ k x * μ k i n f k 1 G μ k x * μ k = f x * . (43)

By applying squeeze theorem1 1 formulated in modern terms by Carl Friedrich Gauss to (38) and (43), we show

l i m k Φ ( x k , ρ k 1 , μ k 1 ) = f ( x * ) . (44)

Let xkl be any subsequence of {x k } such that xklx. By the continuity of f, we get f(xkl)f(x¯). The final demonstration will be done by contradiction under the assumption xx* with f(x)>f(x*). Using (10) for the problem (GP ρ ), we have μkl1B(x)0, for any xInt(Ω2). Furthermore, using (20) for the problem (GP µ ), also we have ρkl1𝒫(xkl)0, and by continuity of f, the sequence f(xkl)f(x)+ρkl𝒫(xkl)+μklB(xkl) cannot converge to zero, which contradicts (44). □

5 APPLICATIONS

5.1 Barrier-Penalty applied to convex problem

The Algorithm 4 is an algorithm based on generic Algorithm 3 in order to solve the nonlinear problem (30).

Algorithm 4
Mixed barrier-penalty algorithm

For ρ0=1,μ0=1 and x0=(1.5,1) we write a MATLAB script for Algorithm 4 in order to compute a sequence of optimal points that approach to x*=(1,2). The iterative results is shown in Table 1.

Table 1
Iterative results.

The path following points is shown in Figure 3, where last points are close to x*. The exact results also solved by MATLAB are x*=(1.000000,2.000000) and f(x*)=5.000000.

Figure 3
Points generated by Algorithm 4.

5.2 Penalized standard linear programming problem

We consider the standard linear programming problem where several variables are upper bounded

L P m i n c T x s . t . A x = b E x u x 0 , (46)

where A is m × n matrix, c,xn,bm, and E is formed by rows of n × n identity matrix corresponding to bounded variables, thereby Ex is the vector of bounded variables for which u is the vector of upper bounds. In this case is usual to add the slack variable v such that Ex+v=u, where v ≥ 0.

In the most computational packages that implement Interior Point Methods for solving linear programming problems only barrier parameter is considered.

In order to solve the LP problem (46), by using the quadratic penalty and logarithmic barrier functions, the objective function is penalized as follow

Φ x , v , ρ , μ = c T x + ρ 2 b - A x 2 - μ j = 1 n log x j - μ j = 1 n b log v j , (47)

where µ and ρ are respectively the barrier and penalty parameters and nb is the number of bounded variables. Then the associated mixed barrier-penalty subproblem is

( L P P ρ , μ ) m i n Φ ( x , v , ρ , µ ) s . t . x , v > 0 . (48)

Since Φ(x,v,ρ,µ) is a smooth function on open set x,v>0. By applying the first-order necessary condition, we have

c A T ρ ( b A x ) µ X 1 e + µ E T V 1 e = 0 ,

Defining y=ρ(bAx),z=µX1e,w=µV1e, we get

c A T y + E T w z = 0 , E x + v = u , X Z e = µ e , V W e = µ e , y = ρ ( b A x ) .

Taking δ=1/ρ, we rewrite δy=bAx. Thus Ax+δy=b.

Therefore, the optimality conditions for subproblem (LPP ρ,µ ) on (x,v)>0 and (z,w)>0 are

A x + δ y = b E x + v = u A T y + z - E T w = c X Z e = μ e V W e = μ e . (49)

In Interior Point Methods reviewed on Suñagua & Oliveira (201716 SUÑAGUA P & OLIVEIRA AR. 2017. A new approach for finding a basis for the splitting preconditioner for linear systems from interior point methods. Computational Optimization and Applications, 67(1): 111-127.), we find a search direction by applying Newton’s Method for solving nonlinear system (49). In fact, the Newton’s directions satisfy

A 0 δ I 0 0 E I 0 0 0 0 0 A T I - E T Z 0 0 X 0 0 W 0 0 V d x d v d y d z d w = r p r u r d r c r s r p = b - A x - δ y r u = u - E x r d = c - A T y - z + E T w r c = μ e - X Z e r s = μ e - V W e , (50)

solving this block linear equations, we find up

d z = X 1 ( r c Z d x ) , d w = V 1 ( r s W d w ) , d v = r u E d x (51)

replacing this on third group of equations

A T d y D 1 d x = r d X 1 r c + E T V 1 r s E T V 1 W r u

where D1=X1Z+ETV1WE, then

d x = D ( A T d y r d + X 1 r c E T V 1 r s + E T V 1 W r u ) (52)

using (52) and first group of equations of (50), we get the normal equations

( A D A T + δ I ) d y = A D ( r d X 1 r c + E T V 1 r s E T V 1 W r u ) (53)

close to optimal point, D matrix is very bad scaled and then ADA T is also very ill-conditioned. In this case, the penalty parameter δ improves that condition number, which is helpful for solving the symmetric positive definite system by applying conjugate gradient method for instance.

Alternatively to (52) and (53), dx and dy also obtain by solving the following augmented system

- D - 1 A T A δ I d x d y = r 1 r p (54)

where r1=rdX1rc+ETV1rsETV1Wru. This system is also symmetric, indefinite and better condition number due to penalty parameter.

For computational experiments, we use the open source package PCx (Czyzyk et al., 19975 CZYZYK J, MEHROTRA S, WAGNER M & WRIGHT SJ. 1997. PCx user guide (Version 1.1). Optimization Technology Center, Northwestern University,.), that implements the Mehrotra’s Predictor-Corrector algorithm in which the barrier parameter µ is already incorporated in order to solve linear programming problems. By adding an appropriate code to PCx, we achieve to incorporate the penalty parameter δ, thus, we obtain the modified PCx called the Predictor-Corrector mixed algorithm with barrier and penalty parameters. The numerical results for several NETLIB LP problems were computed for approaches proposed in Suñagua & Oliveira (201716 SUÑAGUA P & OLIVEIRA AR. 2017. A new approach for finding a basis for the splitting preconditioner for linear systems from interior point methods. Computational Optimization and Applications, 67(1): 111-127.), where the goodness of the approaches were compared according to Dolan & More (20026 DOLAN ED & MORE JJ. 2002. Benchmarking optimization software with performance profiles. Mathematical programming, 91(2): 201-213.) performance profile criteria.

6 CONCLUSIONS

Firstly, we present a brief summary of the main concepts and results upon the barrier and penalty methods, where for each method we show the global convergence theorems in order to use these strategies in the proof of the global convergence theorem for mixed algorithm.

In the Section 4, we provide a mixed algorithm for solving mixed barrier-penalty subproblem (23), and we provide a constructive proof on global convergence theorem for mixed barrierpenalty methods as an alternative showed in Fiacco & McCormick (19907 FIACCO AV & MCCORMICK GP. 1990. Nonlinear programming: sequential unconstrained minimization techniques. vol. 4. Siam.) and Breitfeld & Shanno (19954 BREITFELD MG & SHANNO DF. 1995. A Globally Convergent Penalty-Barrier Algorithm for Nonlinear Programming. In: Operations Research Proceedings 1994. pp. 22-27. Springer.). For simple convex nonlinear problems we write MATLAB code in order to generate iterative points that illustrate penalty and barrier functions.

Finally, we develop an application for nonlinear programming problems with equality and inequality functional constraints, such as, quadratic programming problem, and a standard linear programming problem. Since the functions involved have the smooth property on an open set, then the optimality conditions for each class of problems are stated, those can be solved by applying interior point methods.

ACKNOWLEDGEMENTS

Thanks to CNPq, FAPESP (grant number 2010/06822-4) and Universidad Mayor de San Andrés (UMSA) for their financial support.

References

  • 1
    BAZARAA MS, SHERALI HD & SHETTY CM. 2013. Nonlinear programming: theory and algorithms. John Wiley & Sons.
  • 2
    BERTSEKAS DP. 1976. On penalty and multiplier methods for constrained minimization. SIAM Journal on Control and Optimization, 14(2): 216-235.
  • 3
    BREITFELD MG & SHANNO DF. 1994. A globally convergent penalty-barrier algorithm for nonlinear programming and its computational performance. Rutgers University. Rutgers Center for Operations Research [RUTCOR].
  • 4
    BREITFELD MG & SHANNO DF. 1995. A Globally Convergent Penalty-Barrier Algorithm for Nonlinear Programming. In: Operations Research Proceedings 1994. pp. 22-27. Springer.
  • 5
    CZYZYK J, MEHROTRA S, WAGNER M & WRIGHT SJ. 1997. PCx user guide (Version 1.1). Optimization Technology Center, Northwestern University,.
  • 6
    DOLAN ED & MORE JJ. 2002. Benchmarking optimization software with performance profiles. Mathematical programming, 91(2): 201-213.
  • 7
    FIACCO AV & MCCORMICK GP. 1990. Nonlinear programming: sequential unconstrained minimization techniques. vol. 4. Siam.
  • 8
    FLETCHER R. 2013. Practical methods of optimization. John Wiley & Sons, Chichester.
  • 9
    GRIVA I, NASH SG & SOFER A. 2009. Linear and nonlinear optimization. vol. 108. Siam.
  • 10
    KUHN HW & TUCKER AW. 1951. Nonlinear Programming. In: Proceedings of the Second Berkeley Symposium on Mathematical Statistics and Probability. pp. 481-492. Berkeley, California: University of California Press. Available at: http://projecteuclid.org/euclid.bsmsp/1200500249
    » http://projecteuclid.org/euclid.bsmsp/1200500249
  • 11
    LUENBERGER DG & YE Y. 2008. Linear and nonlinear programming. 3th ed.. Springer New York.
  • 12
    MARTINEZ JM & SANTOS SA. 1995. Métodos computacionais de otimização. Colóquio Brasileiro de Matemática, Apostilas, 20. Available at: https://www.ime.unicamp.br/~martinez/mslivro.pdf
    » https://www.ime.unicamp.br/~martinez/mslivro.pdf
  • 13
    NASH SG. 2010. Penalty and barrier methods. Wiley Encyclopedia of Operations Research and Management Science,.
  • 14
    NASH SG & SOFER A. 1993. A barrier method for large-scale constrained optimization. ORSA Journal on Computing, 5(1): 40-53.
  • 15
    POLYAK BT. 1971. The convergence rate of the penalty function method. USSR Computational Mathematics and Mathematical Physics, 11(1): 1-12.
  • 16
    SUÑAGUA P & OLIVEIRA AR. 2017. A new approach for finding a basis for the splitting preconditioner for linear systems from interior point methods. Computational Optimization and Applications, 67(1): 111-127.
  • 17
    WRIGHT MH. 1992. Interior methods for constrained optimization. Acta Numérica, 1: 341-407.
  • 18
    WRIGHT SJ & NOCEDAL J. 1999. Numerical optimization. vol. 2. Springer New York.
  • 1
    formulated in modern terms by Carl Friedrich Gauss

Publication Dates

  • Publication in this collection
    18 May 2020
  • Date of issue
    2020

History

  • Received
    07 Dec 2018
  • Accepted
    31 Oct 2019
Sociedade Brasileira de Pesquisa Operacional Rua Mayrink Veiga, 32 - sala 601 - Centro, 20090-050 Rio de Janeiro RJ - Brasil, Tel.: +55 21 2263-0499, Fax: +55 21 2263-0501 - Rio de Janeiro - RJ - Brazil
E-mail: sobrapo@sobrapo.org.br