ABSTRACT
Quadratically constrained quadratic programming (QCQP) problems appear in a wide range of engineering fields, including computer science, communication engineering, and finance. A key difficulty in solving these problems lies in efficiently finding global solutions, especially for large-scale and nonconvex instances. To address this, Wen & Yin (2013) proposed a method that reformulates semidefinite programming (SDP) relaxations of QCQPs into a nonlinear and nonconvex low-rank problem. This reformulated problem can be efficiently solved using a curvilinear search method combined with Barzilai-Borwein (BB) steps, known as the CSBB algorithm. In this study, we compare two approaches for solving QCQPs: the conventional convex SDP relaxation and the nonconvex low-rank reformulation introduced by Wen and Yin. We propose a set of general QCQP models that are compatible with the low-rank framework and conduct a series of numerical experiments to evaluate their performance. This study evaluates performance using several classes of NP-hard problems, including Binary Integer Quadratic (BIQ) problems, Max-Cut problems, Boolean Least Squares (BLS) problems, and 0-1 Quadratic Knapsack Problems (QKP). The results demonstrate that the low-rank approach offers competitive performance and shows strong potential for solving large-scale QCQPs more efficiently than traditional convex methods.
Keywords:
0-1 QCQP; semidefinite programming; low-rank factorization; nonconvex QCQP
1 INTRODUCTION
In a convex optimization problem, both the objective and the constraints are convex functions, whereas in a nonconvex optimization problem, either the objective or at least one of the constraints is nonconvex. The convex optimization problem guarantee a globally optimal solution provided it has a feasible solution exists. In contrast, nonconvex optimization problems do not offer such guarantees. Furthermore, a feasible convex optimization problem typically has a unique globally optimal solution, while a feasible nonconvex problem may have multiple locally optimal solutions. So, if the global solution exists for a nonconvex optimization problem, its identification is a tedious task.
Extensive research has focused on the convexification of nonconvex problems. However, obtaining a globally optimal solution for a nonconvex problem without convexification remains a significant challenge. Convexifying large-scale nonconvex problems can introduce various drawbacks. One notable issue is the increased computational time required by even high-performance computers to approach a global solution. This is often due to the looseness of the convex envelope used to approximate the original nonconvex problem.
The prime challenge in solving many engineering real-world and NP-hard problems is how fast and accurately we reach the global solution. Even a faster algorithm is sometimes not the best choice. On the other hand, one must seek an alternative, like obtaining a global solution by solving the nonconvex one, if at all possible.
In the recent past, semidefinite relaxation (SDR) has attracted many researchers as it is a powerful, computationally efficient approximation technique and moreover it is a convex reformulation for many real world optimization problems that evolve from many engineering applications. As these problems are generally NP-hard, they require relaxation, which results in suboptimal solutions. Since the class of nonconvex quadratic programming problems, especially quadratically constrained quadratic programs (QCQPs) captures many problems from Boolean least square problems (BLS), Max-cut problems, image processing problems, signal processing problems, communication problems, etc., we are attracted to solve the nonconvex QCQPs in this research.
The worst-case complexity of solving a generic SDP problem is about 𝒪(n 6.5), using the interior point method (IPM) with a matrix variable of size n×n and 𝒪(n) linear constraints. Hence, this method is not suitable for large scale problems. An SDP cut formulation for BQPs proposed by Wang et al. (2013) presented similar relaxation bounds to the conventional SDP formulations, which have the same degree of complexity as spectral methods. The method has a complexity of 𝒪(kn 3), where k denotes the number of gradient-descent steps in L-BFGS-B. But the method is impractical for large scale optimization problems, as it calculates the gradient of the dual objective function at each gradient-descent step. Note that both the methods discussed are SDP formulations, which means convex reformulation of the given optimization problem. Also, in recent years, many researchers have tried to find a better relaxed solution by adding valid linear inequality constraints, which strengthened the convex relaxation.
Sherali (2007) and Sherali & Adams (2013) first introduced the concept of reformulation linearization technique (RLT) to formulate linear programming relaxation for nonconvex problems. The RLT linearizes the product of any pairs of linear constraints, and a tight relaxation can be obtained via enhanced SDP relaxation with the RLT constraint. Anstreicher (2009) proposed a theoretical analysis of the SDP+RLT relaxation for QCQP with box constraints, showing that RLT constraints remove a large portion of the feasible region, and suggested that a combination of SDP and RLT constraints leads to a tighter bound.
The D.C. decomposition scheme by Zheng et al. (2011), and the αBB underestimators scheme by Anstreicher (2012) are some of the alternative methodologies for convexifying the quadratic form over the feasible region.
In their seminal survey, Burer & Saxena (2012) discussed a variety of methods for generating valid inequalities aimed at tightening the semidefinite programming (SDP) relaxations of quadratically constrained quadratic programs (QCQPs). Alongside these approaches, several studies have focused on developing approximation algorithms for QCQPs, particularly those involving ellipsoidal constraints. For instance, Ye (1999) extended the randomized rounding technique introduced by Goemans & Williamson (1995) to construct feasible solutions from SDP relaxations. Fu et al. (1998) presented approximation algorithms designed to produce feasible solutions that adhere to provable quality guarantees. Additionally, Tseng (2003) conducted a comprehensive analysis of the approximation bounds associated with SDP relaxations for QCQPs with more general quadratic constraints.
Jiang & Li (2019) reviewed SDP based convex relaxation for QCQP problems. Burer & Yang (2015) demonstrated that the SDP+RLT+SOC (Second-Order Cone) relaxation has no gap in an extended trust region problem of minimizing a quadratic function subject to a unit ball and multiple linear constraints, where the linear constraints do not intersect with each other in the interior of the ball. Yamada & Takeda (2018) proposed a new convex relaxation method that is computationally faster but weaker than the SDP relaxation method. Their method reformulates the QCQP as a Lagrangian dual optimization problem and successively solve subproblems while updating the Lagrangian multipliers.
Departing from the above traditional convexification techniques, Low-rank decomposition methods have been widely investigated for their applications in optimization, machine learning, and signal processing. Kaushal et al. (2023) introduced LORD, a low-rank decomposition technique aimed at compressing monolingual code language models, demonstrating its effectiveness in reducing model size while maintaining performance. Bertsimas et al. (2023) investigated sparse plus low-rank matrix decomposition through a discrete optimization approach, emphasizing its utility in high-dimensional data analysis. Wang et al. (2022) applied low-rank decomposition to time-frequency representation for diagnosing bearing faults under variable speed conditions, showcasing its robustness in industrial applications. Hu & Ye (2023) explored the linear convergence properties of an alternating polar decomposition method for low-rank orthogonal tensor approximations, providing theoretical insights into its efficiency. In the field of deep learning, Chen et al. (2021) proposed DRONE, a data-aware low-rank compression strategy designed to optimize large NLP models by balancing storage efficiency and accuracy. Furthermore Liu et al. (2022) developed a randomized quaternion singular value decomposition method to enhance low-rank matrix approximation, contributing to advancements in numerical linear algebra. These studies collectively highlight the significance of low-rank decomposition in improving computational efficiency, reducing data redundancy, and enhancing model interpretability.
The low-rank matrix decomposition method proposed by Burer & Monteiro (2003) involves factorizing the optimization variable in the SDP formulation, X, into RR ⊤. The rank of the factorization, determined by the number of columns in matrix R, is carefully chosen to enhance computational speed while maintaining equivalence with the optimal solution of the SDP.
Furthermore, Wen & Yin (2013) introduced the curvilinear search with BB-step (CSBB) algorithm, which effectively solves nonconvex nonlinear optimization problems and ensures a globally optimal solution for semidefinite programming (SDP).
1.1 Notation
In this paper, X⪰0 denotes that the symmetric matrix X is positive semidefinite, and X⪰Y means that X-Y is positive semidefinite. For two n×n matrices X and Y, the matrix inner product is written as X·Y=Trace(X ⊤ Y). For an n×n matrix X, diag(X) refers to the vector x whose elements are the diagonal entries of X, i.e., x i =X ii for i=1, ..., n. The notation x≤y is used to indicate element-wise inequality between two vectors x and y. Additionally, rank(·) represents the rank of a matrix, and e is the vector of ones.
1.2 Global optimality
In this subsection, the global optimality results for the p largest eigenvalue problem and problems corresponding to SDPs with constraints on the diagonal entries only are presented. We consider the following semidefinite programming (SDP) problem:
where C is a given n×n real symmetric matrix, and X is an n×n symmetric matrix that is required to be positive semidefinite. The primary challenge in solving this problem arises from the positive semidefiniteness constraint X⪰0, as the objective function is linear in X, which makes the semidefinite constraint the most difficult aspect to handle.
If the solution has rank p, then following the decomposition method by Burer & Monteiro (2003), X can be decomposed into V ⊤ V with V=[V 1,V 2, ...,V n ]∈ℝp×n , we obtain an equivalent problem as
The significant advantages and disadvantages of problem (2) compared to (1) are as follows:
-
Problem (2) involves fewer variables than problem (1).
-
The objective function in (2) is no longer linear; it is quadratic and, in general, nonconvex.
Although problem (2) is nonconvex, it has been shown that its local minimizer can also be a global minimizer (Wen & Yin, 2013).
Motivated by the low-rank factorization in Burer & Monteiro (2003) and the success of the CSBB algorithm (Wen & Yin, 2013), we present several QCQP formulations that fit into the framework of (2) and can be efficiently solved using the CSBB algorithm. Specifically, we consider general binary QCQPs and convert them into the SDP formulation given in (1). By utilizing the decomposition approach in (2), we model the problem as a low-rank nonconvex model and apply the CSBB algorithm to solve it. The main objective of this research is to provide a comparative study between the solutions of convex and nonconvex formulations of QCQPs.
The outline of this paper is as follows: In Section 2, we introduce the QCQPs and discuss the SDP relaxation. Section 3, presents the low-rank factorized formulations of the QCQPs described in Section 2 along with a discussion on the existence of a global optimal solution. In Section 4, we provide numerical experiments on several problems and validate our claims by comparing the results obtained through the SDP method. Finally, Section 5 concludes the paper with remarks.
2 THE PROBLEM
We consider the binary QCQP as follows:
where y∈ℝn is the optimization binary vector, P i ∈ℝn×n , q i ∈ℝn are given problem data, for i=0, 1, ..., m, r i ∈ℝ for i=1, ..., m. Here the inequality “≤” in f i (y) is element-wise. Note that y is binary means y∈{0, 1}n or y∈{-1, 1}n . Since the binary variable domain {0, 1}n can be changed to the domain {-1, 1}n by a suitable substitution y=2x-e, then in the rest of the article we consider the binary domain as {-1, 1}n . We assume that all the matrices P i ∈S n and
denotes the feasible set of problem QCQP (3). If P i ⪰0 for each i∈{0}∪{1, 2, ..., m}, then QCQP (3) is a convex programming problem and can be solved in polynomial time. But, in general the problem is NP-hard. If for some i∈{0}∪{1, 2, ..., m}, then in order to obtain a lower bound for the nonconvex QCQP (3), we have to convexify it. Since the current research is focused on solving the nonconvex QCQPs (3), we are least focused on the several approaches available for convexification with some pros and cons.
Letting , QCQP (3) can be presented as
2.1 The Semidefinite relaxation
Although SDR (in both primal and dual forms) makes general QCQPs convex, we investigate its failure in convexifying the binary QCQP stated in (4). Before applying SDR, we first reformulate QCQP (4) into an equivalent problem with a positive semidefinite variable X.
Proposition 1. The optimization problem QCQP (4) is equivalent to the following QCQP (5) , where X=xx ⊤ , and x and X represent the optimal solutions to the problems QCQP (4) and QCQP (5) , respectively.
Proof. Letting X=xx ⊤, we lift the QCQP (4) into the space of rank one matrices. Thus, the vector x∈{-1, 1}n+1 leads to X ii =1 and Rank(X)=1. Finally, for i=0, 1, ..., m, x ⊤ C i x=Tr(C i X)=C i ·X. □
In QCQPSDP(5), the constraint Rank(X)=1 is non-convex and poses a significant challenge. This constraint can either be omitted though this may compromise optimality or relaxed. Since the optimal solution of QCQPSDP(5) coincides with that of QCQP (4) when Rank(X)=1, we choose to relax the rank-one constraint rather than discard it. To this end, we introduce a convex inequality constraint, ||X||F ≤n+1, as suggested in Nayak & Mohanty (2019), where ||·||F denotes the Frobenius norm.
Based on this relaxation, we propose the following versions of the relaxed formulations of QCQPSDP(5).
Proposition 2.
Since ||X||F ≤n+1, the objective value in QCQP (6) is not larger than that of QCQP (5). When µ→0, then QCQP (6) is equivalent to QCQP (5) and for a small µ, the problem QCQP (6) approximates QCQP (5).
The second version of the relaxed problem formulation is
Proposition 3.
Since ||X||F ≤n+1 and C i ·X+r i ≤0, for i=1, 2, ..., m, the objective value in QCQP (7) is not larger than that of QCQP (5). Also, when µ→0 and λ→0, m constraints are satisfied as second and third penalty terms are zero and QCQP (7) is equivalent to QCQP (5).
The third version of the relaxed problem formulation is
Proposition 4.
Here all the m constraints are satisfied as if f i (x)≤0, then the second penalty term leads to zero. It is also clear that the second penalty term is greater than zero when all the m constraints are positive, which leads to a contradiction.
2.2. SDP relaxation with cutting planes
In this subsection, we describe SDP relaxations based on the cutting plane method such as the SDP+RLT cut for given QCQP problems. As we discussed above, any binary QCQP (3) can be presented as a particular form of QCQP (4).
Relaxation based on SDP and RLT both utilised variable X ij that replaced the product terms x i x j of the original problem. The SDP relaxation is based on the fact that since X=xx ⊤, in the actual solution of QCQP (4), one can obtain a relaxation of QCQP by imposing a convex constraint X-xx ⊤⪰0 instead of X=xx ⊤. Also, using the Schur complement, X-xx ⊤⪰0 can be replaced by
Thus, the relaxed QCQP is:
Since the vector x∈{-1, 1}n+1 is equivalent to , we replace by the constraint diag(X)=e. When the QCQP (3) is a convex problem then QCQPSDP(9) is equivalent to QCQP (3). However, When QCQP (3) is nonconvex, the SDP may be unbounded, even though all of the original variables have finite upper and lower bounds. The remedy suggested by Anstreicher (2009) adding the upper bounds to the diagonal components of X is also incorporated in our formulation. Specifically, we assume that in our QCQP formulation.
2.3 SDP with cutting planes
The SDP relaxation can further be strengthened by requiring X to satisfy additional inequalities, known as the Reformulation Linearization Technique (RLT).
2.3.1 RLT cut
The RLT relaxation of QCQP (4) is based on LP relaxation proposed by Sherali (2007); Sherali & Adams (2013). For two variables x i and x j we replace the product term x i x j with the new variable X ij . Since x∈{-1, 1}n+1 , first we relax it with a box constraint x∈[-1, 1]n+1 . As RLT relaxation utilizes products of upper bound and lower bound constraints on the original variables to generate valid linear inequality constraints on the new variable X ij , the resulting set of RLT constraints are obtained after multiplying and replacing with the new variable:
By adding the RLT constraints to QCQP (4), the resulting relaxed problem, denoted as QCQPRLT , becomes a standard linear programming (LP) problem with variables and a total of m+n(2n+3) constraints. Although this formulation is computationally efficient and applicable to large-scale LP problems, it has two primary drawbacks: increased dimensionality and relatively weak lower bounds.
To overcome these bottlenecks, many researchers (Sherali & Fraticelli, 2002; Anstreicher, 2009, 2012; Sherali & Adams, 2013) have proposed the combined SDP+RLT relaxation method, which has been proven to be much more effective than either the SDP or RLT relaxations alone. Therefore, we present an SDP+RLT relaxation by adding the RLT condition to the semidefinite relaxation of QCQP. Let
The matrix X L is obtained by replacing the quadratic term x i x j by a linear term X ij and implementing diag(X)=e in the matrix X. The resulting SDP+RLT relaxation can be written as follows:
Although the SDP+RLT relaxation produces a tighter lower bound, it has the drawback of requiring a significant amount of CPU time to reach a near-optimal solution, due to the increase in the number of constraints and variables. As a result, it is not suitable for large-scale problems.
3 LOW RANK DECOMPOSITION OF THE PROPOSED FORMULATIONS
The low rank decomposition method is a restriction of the semidefinite programming problem (SDP) in which a bound r is imposed on the rank of X, and it is well known that low rank semidefinite programming problem (LRSDPr ) (Burer & Monteiro, 2003) is equivalent to SDP if r is not too small. The local minima and optimal convergence of LRSDPr proved via the algorithm’s distinguishing feature is a nonconvex change of variables that replaces the symmetric, positive semidefinite variable X of the SDP with a rectangular variable V according to the factorization X=VV ⊤. The rank of the factorization, i.e., the number of columns of V, is chosen minimally so as to enhance computational speed while maintaining equivalence with the SDP. In this process the original problem can be transformed into one over V subject to spherical constraints ||V i ||2=1, i=1, ..., n. The problem with spherical constraints is that they are not only nonconvex but numerically expensive to preserve during iterations. Wen & Yin (2013) present a curvilinear search with BB step (CSBB) algorithm that solves the nonconvex nonlinear optimization problem and provides a global optimal solution to a SDP. We implemented CSBB algorithm to our proposed model in (6), (7), and (8) to get a global optimal solution. For this, we express these models into the form of models in (2).
The equivalent low rank formulation of (6) is
Proposition 5.
Similarly, the equivalent low rank formulation of (7) is
Proposition 6.
The equivalent low rank formulation of (8) is
Proposition 7.
We now establish the result stated in Theorem 1 for the formulations given in (11), (12), and (13). To do so, we first require the following well-known result from nonlinear optimization theory:
Lemma 1. Let x* is a solution to a general constrained optimization problem, and the linear independence constraint qualification (LICQ) is satisfied. If λ* is a Lagrange multiplier vector such that the Karush-Kuhn-Tucker (KKT) conditions hold, then the following second-order necessary condition is satisfied:
where is the Hessian of the Lagrangian with respect to x, and ℭ (x*, λ*) denotes the critical cone at x* associated with λ*.
Proof. Refer Theorem 12.5 of Nocedal & Wright (2006). □
Theorem 1. There exists such that, if , any local minimizer of (11) (or (12) , or (13) ) is globally optimal and solves (11) (or (12) , or (13) ). In particular, can be taken as (n+1)-inf{rank(C+D)):D is diagonal} or n, whichever is smaller.
Proof. As problem (11) (or (12), or (13)) satisfies the LICQ condition. Therefore, by Lemma 1, the problem also satisfies the first- and second-order necessary optimality conditions.
Therefore, the local minimizer satisfies both the first- and second-order conditions. As a result, is a global optimum of problem (11) (or (12), or (13)), even when (see Theorem 3 in Wen & Yin (2013)). □
4 NUMERICAL EXPERIMENT
We demonstrate the efficiency and effectiveness of solving the aforementioned problems using the SDP algorithm by comparing it with the nonconvex low-rank relaxation of the QCQPs. The comparison is based on the quality of lower bounds and CPU time. Since the SDP+RLT algorithm requires significantly more computation time than the SDP algorithm alone, and the focus of this research is to compare convex and nonconvex formulations, we concentrate our experiments on the SDP (Dual) formulation. The results are then compared with those obtained using the low-rank formulation.
The test problems include binary integer quadratic (BIQ) problems, max-cut problems, Boolean least squares (BLS) problems, and 0-1 quadratic knapsack problems (QKPs). All experiments were conducted using Matlab on an HP laptop with an 11th Gen Intel i5 processor (3 GHz) and 8 GB of memory.
4.1 Binary integer quadratic problems
The binary integer quadratic problem (BIQ) is described as
where A is a n×n symmetric real matrix. We can replace y∈{0, 1}n as , to get the quadratic relaxation.
Since we are interested in keeping our decision variable y in {-1, 1}, we set x=2y-e with e=(1, 1, ..., 1)⊤. Thus, BIQ can be formulated as:
The SDP relaxation is
where, .
Thus, the low rank decomposed formulation is
Table 1 presents numerical results for solving (16) and (17) and direct DNN (Kim et al., 2016). The 40 test problems are taken from the BIQMAC library (Wiegele, 2007). We present the optimal value and CPU time of each method. We also experimented with SDP+RLT (Sherali & Fraticelli, 2002) on instances with dimensions 50 and 100. We observed that SDP+RLT, although producing a tight lower bound, takes much longer time in comparison with our proposed LR-BIQ.
Table 2 presents a comparison of optimal value and CPU time among the actual optimal solutions: SDP+RLT, Direct DNN, Lagrangian-DNN (Kim et al., 2016), and SDP (IPM method), and Low-Rank (LR) method. The instances are taken from the BIQMAC library. For dimensions like 250 and higher, SDP+RLT is costly due to CPU time. It is observed that CSBB is the fastest among all and has the same optimal value as direct DNN and SDP. However, LAG-DNN is slightly better with respect to the optimal solution than LR-BIQ and inferior with respect to CPU time.
Comparison of optimal value and CPU time (in seconds) among Actual optimal value, SDP+RLT, Direct DNN, LAG-DNN, SDP relaxations, and LR-BIQ for selected problems.
4.2 Max-Cut Problem
The maximum cut (Max-Cut) problem is a very well known problem in various real-world fields, such as network design, VLSI design, statistical physics, etc. If G=(V, E) be a given undirected and connected graph, with V=1, 2, ..., n and E⊂{(i, j):1≤i<j≤n}, the max-cut problem is to find a bipartition (V 1, V 2) of V so that the sum of the weights of the edges between V 1 and V 2 is maximized. Let the edge weights w ij =w ji be given such that w ij =0 for (i, j)∉E, and in particular, let w ii =0. The max-cut problem can be formulated as a BQP as
where W=[w ij ]∈S n is the weighted adjacency matrix of the graph G. The problem (18) has the same solution as that of the following BQP
Taking X=xx ⊤ and dropping the rank(X)=1 constraint, the SDP relaxation of it is
As discussed in Section 3, the low rank formulation of it is as follows:
To test the efficiency of the proposed model, we will now execute the method on the dataset Gset, which is available in http://web.stanford.edu/~yyye/yyye/Gset/.
We consider only 35 instances of the Gset problems with dimensions ranging from 800 to 20, 000 vertices. There are graphs without weighted edges (all weights are 1) as well as graphs with weighted edges where the weights are either+1 or -1.
Table 3 presents a comparison of optimal value and CPU time among benchmark optimal solutions computed by SBM (Matsuda, 2019), SDP (20), and the proposed Low-rank decomposition model LR (21). The column dimension describes the node and edges of the graphs. The columns SDP and LR describe the optimal solutions of SDP (20) and LR (21). Similarly, the columns T-SDP and T-LR describe the CPU time taken by SDP and LR respectively. The column Diff-SDD describes the difference of SDP with SBM and the column Diff-LR describes the difference between the optimal value LR-MC and SBM. By comparing the CPU time and difference columns, it is observed that LR is not only the fastest but also provides the optimal solution close to the SBM in most of the instances.
4.3 Boolean least squares
The basic problem in digital communications, especially maximum likelihood estimation for digital signals, can be presented as an optimization problem as
and can be expressed as a nonconvex QCQP as
A brute force solution is to check all 2n possible values of x, which is usually impractical and leads to some relaxation methods. The SDP relaxation of it is
where, x∈ℝn and . By using the cyclicity of the property of Trace of the matrix, we obtain
where, X=xx ⊤. Thus, we can rewrite the SDP formulation of (22) as
where, and . Thus, the low rank decomposed formulation is
4.3.1 Experiment on BLS
In this subsection, we compare the efficiency of the low rank nonconvex formulation (24) strategy with SDP relaxation (23) with respect to lower bounds and CPU time.
The optimal value of LR and SDP-BLS is the lower bound for BLS. We have solved four instances with different m and n. The Matlab command for A and b is mentioned in column 1. Note that for a particular pair (m, n), we have run the algorithm for 50 iterations. The average of running times in seconds and optimal values are reported in Table 4.
It is observed that SDP dual shows infeasibility for large m and n, while LR-BLS is the best with respect to CPU time and lower bound.
4.4 0-1 Quadratic Knapsack Problem
The Quadratic Knapsack Problem (QKP) introduced by Gallo et al. (1980) has a wide spectrum of applications in the areas of the telecommunication industry Witzgall (1975), location selection problem with a budget constraints (Rhys, 1970) and some applications in weighted maximum b-clique problems (Dijkhuizen & Faigle, 1993; Park et al., 1996; Pisinger, 2007).
Since QKP is NP-hard (Pardalos & Vavasis, 1991), it is difficult to find a polynomial time algorithm. Therefore, finding an approximation algorithms for QKP is always the primary choice. the semidefinite relaxation method (Helmberg et al., 2000), the Lagrangian relaxation method (Michelon & Veilleux, 1996; Caprara et al., 1999; Billionnet & Soutif, 2004a; Létocart et al., 2012), the conic approximation method (Zhou et al., 2013) are some of the approximation methods that best suits for QKP.
The general 0-1 quadratic knapsack problem is given as
where, y is a vector of decision variables, P is an n×n real symmetric matrix; p, d∈ℝn and c∈ℝ. Since y∈{0, 1}n implies y 2=y i , modelling the linear constraint d ⊤ y≤c by restricting the diagonal elements of Y where, Y=yy ⊤, yielding the diagonal representation as D·Y≤c. Thus, the semidefinite relaxation is,
where, Q=P+Diag(p ⊤) and D=Diag(d ⊤). Noted that Diag(y) is a diagonal matrix with diagonal entries as vector y.
Since we need to express problem (25) to our chosen binary domain {-1, 1}, we set x=2y-e, where, e=(1, 1, ..., 1)⊤. With this, the QKP is
The Lagrangian relaxation of (27) yields as follows:
Letting z=[x·x n+1 ]⊤, with x n+1 =±1 and taking Z=zz ⊤, the problem can be rewritten as follows
where, .
Therefore, the low-rank nonconvex model of it is
4.4.1 Experiment on QKP
In this section, we compare the efficiency of the low rank nonconvex formulation (29) strategy over the SDP-QKP (28) with respect to lower bounds and CPU time.
We solved 29 QKP instances from Billionnet and Soutif (BS family) (Billionnet & Soutif, 2004a, b) available at http://cedric.cnam.fr/~soutif/QKP/QKP.html. The BS family problems have a density ranging from 25% to 100% with dimensions 100 to 300, and the results of some specific problems are available at their sites. We compare the objective and CPU time among the BS family results, SDP result, and low rank solution. The objective value and running times (in seconds) are reported in Table 5. We observed that the low rank method is the fastest among all with better solution quality.
Comparison of Objective value and CPU time (in seconds) among Lagrangian decomposition of BS family (Billionnet & Soutif, 2004a), SDP relaxations, and LR-QKP.
5 CONCLUSION
In this paper, we have presented an experimental framework for low-rank nonconvex relaxation to improve its effectiveness and efficiency. We have tested the method over BIQ, Max-cut, BLS, and 0-1 QKPs. Beside the above four problems discussed, we can apply our LR modeling to other BQP such as graph bisection problems, image segmentation with partial grouping constraints, image segmentation with histogram constraints, Image co-segmentation problems, etc. The computational efficiency of low-rank modeling on different QCQP problems suggests that the method achieves a near global solution with faster CPU time, which attracts us to use it on some engineering problems, for which we may conclude that this method may be more tested on other large scale binary quadratic programming problems.
Acknowledgements
The author(s) are grateful to the referees and editor for their helpful comments and valuable suggestions which have contributed to the final presentation of the paper.
References
- ANSTREICHER KM. 2009. Semidefinite programming versus the reformulation-linearization technique for nonconvex quadratically constrained quadratic programming. Journal of Global Optimization, 43(2): 471-484.
- ANSTREICHER KM. 2012. On convex relaxations for quadratically constrained quadratic programming. Mathematical programming, 136(2): 233-251.
- BERTSIMAS D, CORY-WRIGHT R & JOHNSON NA. 2023. Sparse plus low rank matrix decomposition: A discrete optimization approach. Journal of Machine Learning Research, 24(267): 1-51.
- BILLIONNET A & SOUTIF É. 2004a. An exact method based on Lagrangian decomposition for the 0-1 quadratic knapsack problem. European Journal of operational research, 157(3): 565-575.
- BILLIONNET A & SOUTIF E. 2004b. Using a Mixed Integer Programming Tool for Solving the 0-1 Quadratic Knapsack Problem. INFORMS J. on Computing, 16(2): 188-197.
- BURER S & MONTEIRO RDC. 2003. A nonlinear programming algorithm for solving semidefinite programs via low-rank factorization. Math. Program, 95: 329-357.
- BURER S & SAXENA A. 2012. The MILP road to MIQCP. Mixed integer nonlinear programming, pp. 373-405.
- BURER S & YANG B. 2015. The trust region subproblem with non-intersecting linear constraints. Mathematical Programming, 149(1): 253-264.
- CAPRARA A, PISINGER D & TOTH P. 1999. Exact solution of the quadratic knapsack problem. INFORMS Journal on Computing, 11(2): 125-137.
- CHEN P, YU HF, DHILLON I & HSIEH CJ. 2021. Drone: Data-aware low-rank compression for large nlp models. Advances in neural information processing systems, 34: 29321-29334.
- DIJKHUIZEN G & FAIGLE U. 1993. A cutting-plane approach to the edge-weighted maximal clique problem. European Journal of operational research, 69(1): 121-130.
- FU M, LUO ZQ & YE Y. 1998. Approximation algorithms for quadratic programming. Journal of combinatorial optimization, 2(1): 29-50.
- GALLO G, HAMMER PL & SIMEONE B. 1980. Quadratic knapsack problems. In: Combinatorial optimization. pp. 132-149. Springer.
- GOEMANS MX & WILLIAMSON DP. 1995. Improved approximation algorithms for maximum cut and satisfiability problems using semidefinite programming. Journal of the ACM (JACM), 42(6): 1115-1145.
- HELMBERG C, RENDL F & WEISMANTEL R. 2000. A semidefinite programming approach to the quadratic knapsack problem. Journal of combinatorial optimization, 4(2): 197-215.
- HU S & YE K. 2023. Linear convergence of an alternating polar decomposition method for low rank orthogonal tensor approximations. Mathematical Programming, 199(1): 1305-1364.
- JIANG R & LI D. 2019. Semidefinite Programming Based Convex Relaxation for Nonconvex Quadratically Constrained Quadratic Programming. In: World Congress on Global Optimization. pp. 213-220. Springer.
- KAUSHAL A, VAIDHYA T & RISH I. 2023. Lord: Low rank decomposition of monolingual code llms for one-shot compression. arXiv preprint arXiv: 2309.14021.
- KIM S, KOJIMA M & TOH KC. 2016. A Lagrangian-DNN relaxation: a fast method for computing tight lower bounds for a class of quadratic optimization problems. Mathematical Programming, 156(1-2): 161-187.
- LÉTOCART L, NAGIH A & PLATEAU G. 2012. Reoptimization in Lagrangian methods for the 0-1 quadratic knapsack problem. Computers & operations research, 39(1): 12-18.
- LIU Q, LING S & JIA Z. 2022. Randomized quaternion singular value decomposition for low-rank matrix approximation. SIAM Journal on Scientific Computing, 44(2): A870-A900.
- MATSUDA Y. 2019. Benchmarking the MAX-CUT problem on the Simulated Bifurcation Machine.
- MICHELON P & VEILLEUX L. 1996. Lagrangean methods for the 0-1 quadratic knapsack problem. European Journal of operational research, 92(2): 326-341.
- NAYAK RK & MOHANTY NK. 2019. Improved row-by-row method for binary quadratic optimization problems. Annals of Operations Research, 275: 587-605.
-
NOCEDAL J & WRIGHT S. 2006. Numerical optimization. 2. ed. Springer series in operations research and financial engineering. New York, NY: Springer. Available at: http://gso.gbv.de/DB=2.1/CMD?ACT=SRCHA&SRT=YOP&IKT=1016&TRM=ppn+502988711&sourceid=fbw_bibsonomy
» http://gso.gbv.de/DB=2.1/CMD?ACT=SRCHA&SRT=YOP&IKT=1016&TRM=ppn+502988711&sourceid=fbw_bibsonomy - PARDALOS PM & VAVASIS SA. 1991. Quadratic programming with one negative eigenvalue is NP-hard. Journal of Global Optimization, 1(1): 15-22.
- PARK K, LEE K & PARK S. 1996. An extended formulation approach to the edge-weighted maximal clique problem. European Journal of operational research, 95(3): 671-682.
- PISINGER D. 2007. The quadratic knapsack problem - a survey. Discrete applied mathematics, 155(5): 623-648.
- RHYS JM. 1970. A selection problem of shared fixed costs and network flows. Management Science, 17(3): 200-207.
- SHERALI HD. 2007. RLT: A unified approach for discrete and continuous nonconvex optimization. Annals of Operations Research, 149(1): 185.
- SHERALI HD & ADAMS WP. 2013. A reformulation-linearization technique for solving discrete and continuous nonconvex problems. vol. 31. Springer Science & Business Media.
- SHERALI HD & FRATICELLI BM. 2002. Enhancing RLT relaxations via a new class of semidefinite cuts. Journal of Global Optimization, 22(1): 233-261.
- TSENG P. 2003. Further results on approximating nonconvex quadratic optimization by semidefinite programming relaxation. SIAM Journal on Optimization, 14(1): 268-283.
- WANG P, SHEN C & HENGEL AVD. 2013. A Fast Semidefinite Approach to Solving Binary Quadratic Problems. Proc. IEEE Conf. Computer Vision and Pattern Recognition.
- WANG R, FANG H, YU L, YU L & CHEN J. 2022. Sparse and low-rank decomposition of the time-frequency representation for bearing fault diagnosis under variable speed conditions. ISA transactions, 128: 579-598.
- WEN Z & YIN W. 2013. A feasible method for optimization with orthogonality constraints. Math. Program, 142: 397-434.
- WIEGELE A. 2007. Biq Mac Library - A collection of Max-Cut and quadratic 0-1 programming instances of medium size. Preprint, 51: 112-127.
- WITZGALL C. 1975. Mathematical methods of site selection for Electronic Message Systems (EMS). NASA STI/Recon Technical Report N, 76: 18321.
- YAMADA S & TAKEDA A. 2018. Successive Lagrangian relaxation algorithm for nonconvex quadratic optimization. Journal of Global Optimization, 71(2): 313-339.
- YE Y. 1999. Approximating global quadratic optimization with convex quadratic constraints. Journal of Global Optimization, 15(1): 1-17.
- ZHENG X, SUN X & LI D. 2011. Nonconvex quadratically constrained quadratic programming: best DC decompositions and their SDP representations. Journal of Global Optimization, 50(4): 695-712.
- ZHOU J, CHEN D, WANG Z & XING W. 2013. A conic approximation method for the 0-1 quadratic knapsack problem. Journal of Industrial and Management Optimization, 9(3): 531-547.
Data availability
The datasets analyzed in this study are publicly available and referenced in the Numerical Experiment Section along with their respective URLs.
Publication Dates
-
Publication in this collection
08 Aug 2025 -
Date of issue
2025
History
-
Received
21 Feb 2025 -
Accepted
02 June 2025
