This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

11institutetext: Xinyue Huo 22institutetext: Nankai University
School of Statistics and Data Science, Nankai University, Tianjin 300071, China
xy [email protected]
33institutetext: Ran Gu 44institutetext: Nankai University
School of Statistics and Data Science, Nankai University, Tianjin 300071, China
[email protected]

A Vectorized Positive Semidefinite Penalty Method for Unconstrained Binary Quadratic Programming

Xinyue Huo and Ran Gu
(Received: date / Accepted: date)
Abstract

The unconstrained binary quadratic programming (UBQP) problem is a class of problems of significant importance in many practical applications, such as in combinatorial optimization, circuit design, and other fields. The positive semidefinite penalty (PSDP) method originated from research on semidefinite relaxation, where the introduction of an exact penalty function improves the efficiency and accuracy of problem solving. In this paper, we propose a vectorized PSDP method for solving the UBQP problem, which optimizes computational efficiency by vectorizing matrix variables within a PSDP framework. Algorithmic enhancements in penalty updating and initialization are implemented, along with the introduction of two algorithms that integrate the proximal point algorithm and the projection alternating BB method for subproblem resolution. Properties of the penalty function and algorithm convergence are analyzed. Numerical experiments show the superior performance of the method in providing high-quality solutions and satisfactory solution times compared to the semidefinite relaxation method and other established methods.

Keywords:
Binary quadratic programming Positive semidefinite penalty function Proximal point algorithm Projection alternating BB stepsize
MSC:
65K0590C09 90C20
journal: JOTA

1 Introduction

This study focuses on unconstrained binary quadratic programming (UBQP), a specialized form of integer programming. In this type of programming, the objective function takes a quadratic form and the variables are constrained to binary values of 0 and 1. This class of problems is considered vital and is often formulated as follows:

minx{0,1}nxTQx+bTx,\min_{x\in\{0,1\}^{n}}x^{T}Qx+b^{T}x, (1)

where x=[x1,x2,,xn]TRnx=[x_{1},x_{2},\ldots,x_{n}]^{T}\in R^{n}, QRn×nQ\in R^{n\times n}, and cRnc\in R^{n}. This problem can also be viewed as a special type of quadratically constrained quadratic programming (QCQP) problem, where the 0-1 constraint is written as xi2xi=0x_{i}^{2}-x_{i}=0.

This programming has a wide range of research significance in applications, such as combinatorial optimization Boros1991 , financial investments Cesarone_2014 , signal processing 1363998 , production scheduling TEIXEIRA2010425 and other fields. In general, this problem is NP-hard PARDALOS1992119 due to the binary constraint.

1.1 Related works

Due to the potential of the unconstrained quadratic programming problem, many authors have studied the optimality conditions and sought high-quality solution methods for the UBQP problem in recent years. Beck and Teboulle doi:10.1137/S1052623498336930 derived the sufficient and necessary global optimality conditions. Based on the work of Beck and Teboulle, Xia Xia2009NewOC obtained tighter sufficient optimality conditions and explored the relationship between the optimal solution of the problem and that of its continuous relaxation. These studies have provided us with a deeper understanding of UBQP problems, leading to the development of efficient solution strategies.

One of the main categories of methods for solving UBQP is exact solvers. Typically, this is achieved by a tree-search approach that incorporates a generalized branch-and-bound technique. In the 1980s, both Gulati et al. GULATI1984121 and Barahona et al. 1989Experiments described a branch-and-bound algorithm for UBQP. More recently, Krislock et al. Krislock2017BiqCrunchAS present BiqBin, an exact solver for binary quadratic problems. The approach is based on an exact penalty method that efficiently transforms the original problem into an instance of Max-Cut, and then solves the Max-Cut problem using a branch-and-bound algorithm. The numerical results show that BiqBin is a highly competitive solver.

Heuristic algorithms are also often used to solve UBQP problems exactly. The most common are tabu search inproceedings , quantum annealing PMID:35140264 . For example, Wang et al. 10.1007/978-3-642-29828-8_26 introduce a backbone-guided tabu search algorithm that involves a fundamental tabu search process alternating with a variable fixing/freeing phase, which is determined by the identification of strongly determined variables. Due to the NP-hardness, the optimal solution usually cannot be found in polynomial time. Therefore, it is necessary to find a fast method that is not strictly exact. Based on this, many scholars have proposed a number of inexact methods.

Among the inexact methods, there are mainly methods based on relaxation and penalty function methods. There are some penalty function methods based on vector variables. Liu et al. Liu2018ACA and Nayak et al. RePEc:spr:jcomop:v:39:y:2020:i:3:d:10.1007_s10878-019-00517-8 have adopted a continuous approach, converting binary constraints into continuous ones by using Fischer-Burmeister nonlinear complementarity problem functions, then it can be further carried on the smoothing processing by aggregate function. However, this method is relatively sensitive to the initial values, and the parameter updates are complex and may not serve as exact penalty. Furthermore, the exact penalty method is widely used in many methods, transforming the binary constraints into a set of inequality constraints. Yuan et al. yuan2017exact introduced an alternative approach by replacing the binary constraints with a Mathematical Programming with Equilibrium Constraints (MPEC). This transformation converts the UBQP problem into a biconvex problem. While this algorithm demonstrates high computational efficiency, the solution accuracy still needs to be improved for certain data sets. Luo et al. luo2019enhancing introduced a special penalty method called conditionally quasi-convex relaxation based on semidefinite relaxation and they showed that the constructed penalty function is exact. However, the subproblem of this algorithm still needs to solve an SDP subproblem, which is time-consuming.

Matrix-based relaxation methods are relatively common approaches, with typical examples including the reconstruction-linearization technique Sherali2007RLTAU and positive semidefinite relaxation (SDR) fujie1997semidefinite . SDR is a classical method for solving QCQP, which transforms quadratic terms of vector variables into matrix variables, drops the rank-one constraint, relaxes the problem to a semidefinite programming (SDP) problem, and utilizes SDP solvers for the solution. In many cases, SDR can obtain the exact solution to QCQP or prove to achieve a certain approximation ratio 2023Exact ; 5447068 . It also has wide applications in practice.

When the matrix solution of SDR is not rank one, the traditional strategies are to use projection or random sampling 5447068 . Recently, other methods have also been studied to further obtain high-quality rank one solutions from SDR solutions 10.1093/imanum/draa031 ; article . Gu et al. provided a broader perspective to SDR research by introducing a positive semidefinite-based penalty (PSDP) formulation 10.1093/imanum/draa031 . When the penalty factor is zero, it corresponds exactly to the semidefinite relaxation of QCQP problems. Moreover, if the penalty factor is sufficiently large, the penalty function problem is equivalent to QCQP. Therefore, by gradually increasing the penalty factor, the matrix solution of SDR can reach rank one. In addition, the authors used a proximal point algorithm (PPA) to solve the penalty function subproblems, which demonstrated remarkable performance when applied to UBQP, with a significant proportion yielding exact solutions for UBQP.

While the solution quality provided by PSDP is high, its significant computational cost as an SDR-based approach cannot be overlooked compared to vector methods. Therefore, a crucial research question is how to maintain the solution quality of PSDP while significantly reducing its computational cost, ideally to a form comparable to the computational efficiency of vector methods.

1.2 Our contribution

In the PSDP framework, we select a specified initial value for the penalty factor and find that the penalty function problem can depend only on the diagonal elements of the matrix. Following the penalty factor update rule of PSDP, the generated sequence of PSDP penalty problems still ensures that it is only related to the matrix diagonal elements, thus causing the matrix problem in PSDP to degrade to a vector problem, achieving the vectorization of PSDP. Furthermore, in the vectorized PSDP algorithm we propose, the algorithm for solving subproblem and penalty update rule maintain the same form as PSDP, while incorporating some improvements in the details. Therefore, the vectorized PSDP not only maintains the quality of solutions in PSDP, but also keeps the computational cost at the level of vector methods.

To further accelerate the convergence and improve the solution accuracy, we apply an algorithm accelerated by the projection alternating Barzilai-Borwein (BB) method 2005Projected . This method uses the projection step to effectively handle the 0-1 constraints, while the alternating BB method provides a robust and efficient framework for updating the search direction and step size. Additionally, we introduce a criterion to update the penalty parameter during the iterations. This ensures that the active set is appropriately updated throughout the iteration process, thereby ensuring the convergence of the algorithm. We tested our algorithm on one randomly generated dataset and five benchmark datasets, comparing its performance with four existing methods. Experimental results show that our algorithm performs well in terms of both time efficiency and accuracy.

1.3 Organization

In section 2, we will present our vectorized exact penalty function and its transformation process. In section 3, we will present our subproblem, two methods for solving the subproblem, and an approach for updating the penalty factor. Section 4 will provide the theoretical foundation of our algorithm, including the exact penalty function theory and the convergence of the subproblem algorithm. Finally, in section 5, we will demonstrate the comparison of our algorithm with several other methods across different datasets through numerical experiments.

2 Vectorization of PSDP

In this paper, we focus on the following UBQP problem:

minxTQx+bTx s.t. x{0,1}n,\begin{array}[]{cl}\min&x^{T}Qx+b^{T}x\\ \text{ s.t. }&x\in\{0,1\}^{n},\end{array} (2)

where Q𝕊n×nQ\in\mathbb{S}^{n\times n} is a real symmetric matrix, bnb\in\mathbb{R}^{n} is a vector. We can reformulate (2) as a problem with alternative equivalent equation constraints:

minxTQx+bTx s.t. xi2xi=0,i=1,,n.\begin{array}[]{cl}\min&x^{T}Qx+b^{T}x\\ \text{ s.t. }&x_{i}^{2}-x_{i}=0,\quad i=1,\ldots,n.\end{array} (3)

Therefore, UBQP can be viewed as a QCQP problem subject to specific constraints.

2.1 PSDP

For a general QCQP as in (4), PSDP 10.1093/imanum/draa031 adds a matrix variable Z0Z\succeq 0 and a penalty term PZP\cdot Z with P0P\succeq 0 to the objective, proposing the following penalty problem (5).

minxn\displaystyle\min_{x\in\mathbb{R}^{n}} (1xTxxxT)(0g0Tg0Q0)\displaystyle\left(\begin{array}[]{cc}1&x^{\mathrm{T}}\\ x&xx^{\mathrm{T}}\end{array}\right)\cdot\left(\begin{array}[]{cc}0&g_{0}^{\mathrm{T}}\\ g_{0}&Q_{0}\end{array}\right) (4)
s.t. (1xTxxxT)(cigiTgiQi)=0,i=1,,me,\displaystyle\left(\begin{array}[]{cc}1&x^{\mathrm{T}}\\ x&xx^{\mathrm{T}}\end{array}\right)\cdot\left(\begin{array}[]{cc}c_{i}&g_{i}^{\mathrm{T}}\\ g_{i}&Q_{i}\end{array}\right)=0,\quad i=1,\ldots,m_{e},
(1xTxxxT)(cigiTgiQi)0,i=me+1,,m.\displaystyle\left(\begin{array}[]{cc}1&x^{\mathrm{T}}\\ x&xx^{\mathrm{T}}\end{array}\right)\cdot\left(\begin{array}[]{cc}c_{i}&g_{i}^{\mathrm{T}}\\ g_{i}&Q_{i}\end{array}\right)\geq 0,\quad i=m_{e}+1,\ldots,m.
minxn,Z𝕊n\displaystyle\min_{x\in\mathbb{R}^{n},Z\in\mathbb{S}^{n}} (1xTxxxT+Z)(0g0Tg0Q0)+PZ\displaystyle\left(\begin{array}[]{lc}1&x^{\mathrm{T}}\\ x&xx^{\mathrm{T}}+Z\end{array}\right)\cdot\left(\begin{array}[]{cc}0&g_{0}^{\mathrm{T}}\\ g_{0}&Q_{0}\end{array}\right)+P\cdot Z (5)
s.t. (1xTxxxT+Z)(cigiTgiQi)=0,i=1,,me,\displaystyle\left(\begin{array}[]{cc}1&x^{\mathrm{T}}\\ x&xx^{\mathrm{T}}+Z\end{array}\right)\cdot\left(\begin{array}[]{ll}c_{i}&g_{i}^{\mathrm{T}}\\ g_{i}&Q_{i}\end{array}\right)=0,\quad i=1,\ldots,m_{e},
(1xTxxxT+Z)(cigiTgiQi)0,i=me+1,,m,\displaystyle\left(\begin{array}[]{cc}1&x^{\mathrm{T}}\\ x&xx^{\mathrm{T}}+Z\end{array}\right)\cdot\left(\begin{array}[]{ll}c_{i}&g_{i}^{\mathrm{T}}\\ g_{i}&Q_{i}\end{array}\right)\geq 0,\quad i=m_{e}+1,\ldots,m,
Z0.\displaystyle Z\succeq 0.

When PP is 0, (5) is equivalent to a standard SDR.

In the UBQP problem, with Q0=QQ_{0}=Q, Qi=eieiTQ_{i}=e_{i}e_{i}^{T} (eie_{i} is a vector with 1 in the ii-th position and 0 elsewhere), g0=0.5big_{0}=0.5b_{i}, gi=0.5eig_{i}=-0.5e_{i} and ci=0c_{i}=0 as in (4), the penalty problem (5) corresponds to the following problem.

minxn,Z𝕊n\displaystyle\min_{x\in\mathbb{R}^{n},Z\in\mathbb{S}^{n}} xTQx+bTx+(Q+P)Z\displaystyle x^{T}Qx+b^{T}x+(Q+P)\cdot Z (6)
s.t. xi2xi+zi=0,i=1,2,,n,\displaystyle x_{i}^{2}-x_{i}+z_{i}=0,i=1,2,\dots,n,
Z0,\displaystyle Z\succeq 0,

where ziz_{i} is the ii-th diagonal element of ZZ.

PSDP employed a PPA method to solve problem (6), with each PPA iteration involving the solution of an SDP problem, necessitating the repeated solving of SDPs. PSDP begins with P=0P=0, which is equivalent to SDR, and then updates the penalty factor PP by P=P+αZP=P+\alpha Z.

2.2 Vectorized PSDP

Solving SDP is the major computational obstacle in PSDP, so we consider what kind of problem structure can make problem (6) easy to solve.

We note that when Q+PQ+P is a diagonal matrix, the solution of ZZ depends only on its diagonal elements. This is because the first constraint involves only the diagonal elements of ZZ, and when Q+PQ+P is a diagonal matrix, the objective function also involves only the diagonal elements of ZZ. For the remaining positive semidefinite constraints on ZZ, as long as its diagonal elements are non-negative and the non-diagonal elements are set to zero, it naturally satisfies the constraints. Therefore, we only need to make Q+PQ+P a diagonal matrix so that there exists a diagonal matrix solution for ZZ. In addition, since ZZ is a diagonal matrix, Q+PQ+P actually only updates the diagonal elements when updating P=P+αZP=P+\alpha Z. Therefore, the new penalty problem still has a diagonal matrix solution for ZZ.

The construction of PP is straightforward that we simply need to set the off-diagonal elements of PP to be the negation of the off-diagonal elements of QQ. Supposing that we choose an appropriate PP such that Q+PQ+P also becomes a diagonal matrix PP^{\star}, problem (6) is thus transformed into the following form.

minxnxTQx+bTx+i=1npi(xixi2) s.t. 0xi1,i=1,,n,\begin{array}[]{cc}\min_{x\in\mathbb{R}^{n}}&x^{T}Qx+b^{T}x+\sum_{i=1}^{n}p_{i}(x_{i}-x_{i}^{2})\\ \text{ s.t. }&0\leq x_{i}\leq 1,\quad i=1,\ldots,n,\\ \end{array} (7)

where pip_{i} is the ii-th diagonal element of PP^{\star}, the inequality xi2xi0x_{i}^{2}-x_{i}\leq 0 has been replaced with 0xi10\leq x_{i}\leq 1, and ZZ has been eliminated.

Billionnet and Elloumi Billionnet2007UsingAM have discussed the form of problem (7), exploring how to construct an appropriate pp to preserve the convexity of problem (7) while also preserving equivalence with problem (2) as much as possible. In contrast, we continue to follow the PSDP approach of updating pp, allowing the quadratic problem (7) to be non-convex, and solving it using iterative algorithms.

3 Algorithms

Since problem (7) can lead to a non-convex quadratic problem, finding the global optimal solution is not easy. We present two algorithms. First, we propose a PPA based on an improvement of the PPA in PSDP. Then, we introduce an accelerated method, namely the projection alternating BB method. Other algorithmic details, including the subproblem termination criterion, penalty factor updates, and initialization, are also included in this section.

3.1 PPA

In PSDP, when applying the PPA to solve (6), the kk-th iteration is carried out in the following form.

(dk,Zk+1)=argmindn,Z𝕊nH(xk+d,Z,P)+dTPd s.t. (xk+d,Z)G1,\begin{array}[]{rl}\left(d^{k},Z^{k+1}\right)=\underset{d\in\mathbb{R}^{n},Z\in\mathbb{S}^{n}}{\arg\min}&H\left(x^{k}+d,Z,P\right)+d^{\mathrm{T}}Pd\\ \text{ s.t. }&\left(x^{k}+d,Z\right)\in G_{1},\end{array} (8)

where G1={(x,Z)|xi2xi+zi=0,i=1,2,,n,Z0}G_{1}=\{(x,Z)|x_{i}^{2}-x_{i}+z_{i}=0,i=1,2,\dots,n,Z\succeq 0\} and H(xk+d,Z,P)=xTQx+bTx+(Q+P)ZH\left(x^{k}+d,Z,P\right)=x^{T}Qx+b^{T}x+(Q+P)\cdot Z. Its proximal term is dTPdd^{T}Pd. For the vectorized problem (7), we observe the implementation of this PPA as follows:

dk=argmindnh(xk+d,p)+dTPd s.t. xk+dF1,\begin{array}[]{rl}d^{k}=\underset{d\in\mathbb{R}^{n}}{\arg\min}&h\left(x^{k}+d,p\right)+d^{T}Pd\\ \text{ s.t. }&x^{k}+d\in F_{1},\\ \end{array} (9)

where F1={x|0xi1}F_{1}=\{x|0\leq x_{i}\leq 1\}, h(x,p)=xT(QP)x+(b+p)Txh\left(x,p\right)=x^{T}(Q-P^{\star})x+(b+p)^{T}x, and pp is the vector composed of diagonal elements of PP^{\star}.

By utilizing the nature of Q+P=PQ+P=P^{\star}, it can be observed that the objective function of (9) is a linear function of dd. It is known that minimizing a linear function over box constraints leads to all components of the optimal solution being taken on the boundaries. This causes xx to reach the critical point of (7) in one iteration. However, it is typically a low-quality local solution, since problem (7) is non-convex. Therefore, we have to modify the proximal term to better solve (7) using PPA.

Building on (9), we introduce an additional quadratic term dTHpdd^{T}H_{p}d to the existing proximal term, where Hp0H_{p}\succ 0 to ensure strict convexity of the new problem and avoid linearity. Furthermore, we need to ensure that the new proximal term (Hp+P)(H_{p}+P) is positive semidefinite since we do not strictly require P0P\succeq 0 here. Lastly, it is necessary for HpH_{p} to be a diagonal matrix to facilitate the computation of the subproblems. Therefore, we define HpH_{p} in the following form.

Hp=[max(|Q1i|p1,ϵ)000max(|Q2i|p2,ϵ)000max(|Qni|pn,ϵ)],H_{p}=\begin{bmatrix}\max(\sum|Q_{1i}|-p_{1},\epsilon)&0&\cdots&0\\ 0&\max(\sum|Q_{2i}|-p_{2},\epsilon)&\cdots&0\\ \vdots&\vdots&\ddots&\vdots\\ 0&0&\cdots&\max(\sum|Q_{ni}|-p_{n},\epsilon)\end{bmatrix},

where ϵ>0\epsilon>0 is an arbitrarily small number. Thus, Hp0H_{p}\succ 0 and

Hp+P=HpQ+P[i1|Q1i|Q12Q1nQ21i2|Q2i|Q2nQn1Qn2in|Qni|]0,H_{p}+P=H_{p}-Q+P^{\star}\succeq\begin{bmatrix}\sum_{i\neq 1}|Q_{1i}|&-Q_{12}&\cdots&-Q_{1n}\\ -Q_{21}&\sum_{i\neq 2}|Q_{2i}|&\cdots&-Q_{2n}\\ \vdots&\vdots&\ddots&\vdots\\ -Q_{n1}&-Q_{n2}&\dots&\sum_{i\neq n}|Q_{ni}|\end{bmatrix}\succeq 0,

where the final step utilizes the fact that a diagonally dominant matrix is positive semidefinite.

Setting the proximal term to be (Hp+P=HpQ+P)(H_{p}+P=H_{p}-Q+P^{\star}), we obtain the kk-th iteration as follows:

dk=argmindnh(xk+d,p)+dT(HpQ+P)d s.t. xk+dF1\begin{array}[]{rl}d^{k}=\underset{d\in\mathbb{R}^{n}}{\arg\min}&h\left(x^{k}+d,p\right)+d^{T}(H_{p}-Q+P^{\star})d\\ \text{ s.t. }&x^{k}+d\in F_{1}\\ \end{array} (10)

In fact,

h(xk+d,p)+dT(HPQ+P)d=h(xk,p)+h(xk,p)Td+dTHpd.h\left(x^{k}+d,p\right)+d^{T}(H_{P}-Q+P^{\star})d=h\left(x^{k},p\right)+\nabla h\left(x^{k},p\right)^{T}d+d^{T}H_{p}d.

Thus, (10) is separable in variables, equivalent to solving n one-dimensional problems. Without considering box constraints, the optimal solution is as follows:

d~jk=h(xk,p)j2hpj,\tilde{d}^{k}_{j}=\frac{-\nabla h\left(x^{k},p\right)^{j}}{2h_{p}^{j}}, (11)

where d~jk\tilde{d}^{k}_{j} is jj-th element of d~k\tilde{d}^{k}, h(xk,p)j\nabla h\left(x^{k},p\right)^{j} is the jj-th element of h(xk,p)\nabla h\left(x^{k},p\right) and hpjh_{p}^{j} is the jj-th diagonal element of HpH_{p}. Then dkd^{k} can be obtain by projecting d~k\tilde{d}^{k} onto box set (F1xk)(F_{1}-x^{k}), that is,

dk=𝒫(F1xk)(d~k).d^{k}=\mathcal{P}_{(F_{1}-x^{k})}(\tilde{d}^{k}). (12)

Therefore, the next iterative point xk+1=xk+dkx^{k+1}=x^{k}+d^{k} is given by

xk+1=𝒫(xk+d~k),x^{k+1}=\mathcal{P}(x^{k}+\tilde{d}^{k}), (13)

where 𝒫\mathcal{P} is a projection operator defined by

𝒫(x)={0,x<0,x,0x1,1,x>1.\mathcal{P}(x)=\left\{\begin{array}[]{l}0,x<0,\\ x,0\leq x\leq 1,\\ 1,x>1.\end{array}\right.

We present the algorithmic framework of this PPA in Algorithm 1, which is used to solve the penalty problem (7) with a given penalty factor pp.

Algorithm 1 PPA for subproblem (7) with fixed pp
1:  Require:0x010\leq x_{0}\leq 1
2:  while termination criterion >1e5>1e^{-5} do
3:      Update dkd^{k} by (12)
4:     Update xk+1x^{k+1} by (13)
5:     k:=k+1k:=k+1
6:  end while
7:  return  The optimal solution 𝒙\boldsymbol{x^{\star}}.

3.2 An acceleration algorithm based on BB method

While the previously proposed algorithm is straightforward and computationally efficient, it exhibits relatively long computation time for local convergence. Consequently, we opt to use the projection alternating BB method introduced by Dai and Fletcher 2005Projected to speed up the computational process.

The projection alternating BB method strategically alternates between utilizing long and short BB step sizes, as outlined below:

αkABB={αkBB1, for odd k,αkBB2, for even k.\alpha_{k}^{ABB}=\left\{\begin{array}[]{ll}\alpha_{k}^{BB1},&\text{ for odd }k,\\ \alpha_{k}^{BB2},&\text{ for even }k.\end{array}\right. (14)

Here,

αkBB1=sk1Tsk1sk1Tyk1 and αkBB2=sk1Tyk1yk1Tyk1,\alpha_{k}^{BB1}=\frac{s_{k-1}^{T}s_{k-1}}{s_{k-1}^{T}y_{k-1}}\text{ and }\alpha_{k}^{BB2}=\frac{s_{k-1}^{T}y_{k-1}}{y_{k-1}^{T}y_{k-1}},

where sk1=xkxk1s_{k-1}=x_{k}-x_{k-1} and yk1=gkgk1y_{k-1}=g^{k}-g^{k-1}. In the context of problem (7), we have gk:=h(xk,p)g^{k}:=\nabla h\left(x^{k},p\right). Then the update of xk+1x^{k+1} is expressed as

xk+1=𝒫(xkαkABBgk).x^{k+1}=\mathcal{P}(x^{k}-\alpha_{k}^{ABB}g^{k}). (15)

We present the framework of this projection BB method in Algorithm 2.

Algorithm 2 PABB for subproblem (7) with fixed pp
1:  Require: 0x010\leq x_{0}\leq 1
2:  while termination criterion >1e5>1e^{-5} do
3:     Update αkABB\alpha_{k}^{ABB} by (14)
4:     Update xk+1x^{k+1} by (15)
5:     k:=k+1k:=k+1
6:  end while
7:  return  The optimal solution 𝒙\boldsymbol{x^{\star}}.

3.3 Termination criterion

For the optimization problem defined by (7), the termination criterion is established based on the Karush-Kuhn-Tucker (KKT) conditions.

The Lagrangian function is given by

L(x,u,v)=xT(QP)x+(b+p)TxuTx+vT(xeT),L(x,u,v)=x^{T}(Q-P^{\star})x+(b+p)^{T}x-u^{T}x+v^{T}(x-e^{T}),

where uu and vv correspond to the dual variables. Then the first-order optimality conditions can be expressed as follows:

xL=2(QP)x+\displaystyle\nabla_{x}L=2(Q-P^{\star})x+ (b+p)TuT+vT=0,\displaystyle(b+p)^{T}-u^{T}+v^{T}=0, (16)
0x\displaystyle 0\leq x 1,\displaystyle\leq 1,
uTx\displaystyle u^{T}x =0,\displaystyle=0,
vT(ex)\displaystyle v^{T}(e-x) =0,\displaystyle=0,
u0\displaystyle u\geq 0 ,v0.\displaystyle,v\geq 0.

By introducing the index sets I={i|0<xi<1}I=\{i|0<x_{i}<1\}, K={i|xi=1}K=\{i|x_{i}=1\}, J={i|xi=0}J=\{i|x_{i}=0\}, and defining Q¯=QP\bar{Q}=Q-P^{\star} and b¯=b+p\bar{b}=b+p, the conditions can be rewritten as:

cret1=2Q¯IIxI+Q¯IkeK+b¯I=0,cret2=2Q¯JIxI+Q¯JKeK+b¯J=uJT0,cret3=2Q¯KIxI+Q¯KKeK+b¯K=vKT0.\begin{array}[]{l}cret_{1}=2\bar{Q}_{II}x_{I}+\bar{Q}_{Ik}e_{K}+\bar{b}_{I}=0,\\ cret_{2}=2\bar{Q}_{JI}x_{I}+\bar{Q}_{JK}e_{K}+\bar{b}_{J}=u^{T}_{J}\geq 0,\\ cret_{3}=2\bar{Q}_{KI}x_{I}+\bar{Q}_{KK}e_{K}+\bar{b}_{K}=-v^{T}_{K}\leq 0.\end{array}

Consequently, we present the termination criterion as

cret/ρ<δ,cret/\rho<\delta, (17)

where

cret=max{max(|cret1|),max{|min(cret2,0)|},max{|max(cret3,0)|}}cret=max\{max(|cret_{1}|),max\{|min(cret_{2},0)|\},max\{|max(cret_{3},0)|\}\}
and ρ=max{1,|2(QP)|,|b+p|}.\text{and }\rho=max\{1,|2(Q-P^{\star})|,|b+p|\}.

3.4 Penalty parameter update

Due to the adoption of the penalty factor update from PSDP, specifically P=P+αZP^{\star}=P^{\star}+\alpha Z, where both PP^{\star} and ZZ are diagonal matrices in the vectorized PSDP algorithm, only the diagonal elements of PP^{\star} need to be updated, that is,

ps=ps1+αszs.p^{s}=p^{s-1}+\alpha^{s}z^{s}. (18)

Here, zs=(x1sx1s2,x2sx2s2,,xnxns2)Tz^{s}=(x_{1}^{s}-x_{1}^{s^{2}},x_{2}^{s}-x_{2}^{s^{2}},\dots,x_{n}-x_{n}^{s^{2}})^{T}, xsx^{s} denotes the best value of xx after the s1s-1 th update of pp, and psp^{s} denotes the penalty parameter after s updates.

The choice of the parameter α\alpha significantly impacts the overall efficiency of the algorithm. When α\alpha is too small, it may lead to smaller iteration steps and an increase in the number of iterations, thus increasing the runtime and slowing down convergence. Conversely, if α\alpha is too large, it can easily cause xx to rapidly approach the boundaries during the update process, potentially halting at a low-quality local solution. Therefore, selecting an appropriate value for α\alpha is crucial.

We opt for an appropriate α\alpha from the perspective of the active set. Ideally, after updating pp, we desire the new optimal solution xx of (7) to have a different active set compared to before the update, while ensuring that the change is not too drastic. In the following lemma, we provide an α\alpha that guarantees a change in the active set.

Lemma 1

In the iterations, we choose αs\alpha^{s} as (19), then the active set can be updated.

αs=λmin(zs12Q¯IIszs12).\alpha^{s}=\lambda_{min}(z^{s^{-\frac{1}{2}}}\bar{Q}_{II}^{s}z^{s^{-\frac{1}{2}}}). (19)
Proof

From the KKT condition, we have

2Q¯IIxI+Q¯IKeK+b¯I=0,2\bar{Q}_{II}x_{I}+\bar{Q}_{IK}e_{K}+\bar{b}_{I}=0,

let yI=2xIeIy_{I}=2x_{I}-e_{I}, then we have

Q¯IIyI+Q¯IKeK+Q¯IIeI+b¯I=0,\bar{Q}_{II}y_{I}+\bar{Q}_{IK}e_{K}+\bar{Q}_{II}e_{I}+\bar{b}_{I}=0,

where Q¯IIeI+b¯I\bar{Q}_{II}e_{I}+\bar{b}_{I} is constant for fixed I. Thus we have Q¯IIsyIs=Q¯IIs+1yIs+1\bar{Q}_{II}^{s}y_{I}^{s}=\bar{Q}_{II}^{s+1}y_{I}^{s+1} for fixed I, and yIsy_{I}^{s}, Q¯IIs\bar{Q}_{II}^{s} means the value of yIy_{I}, Q¯IIs\bar{Q}_{II}^{s} after the sth update of pp. We want xIs+10x_{I}^{s+1}\leq 0 or xIs+11x_{I}^{s+1}\geq 1, which means |2xIs+1eI|1|2x_{I}^{s+1}-e_{I}|\geq 1. We update pp by (18), then Q¯IIs+1=Q¯IIsαsZIIs\bar{Q}_{II}^{s+1}=\bar{Q}_{II}^{s}-\alpha^{s}Z_{II}^{s}, Z is the diagonal matrix with z as the diagonal element. We have

(Q¯IIsαsZIIs)(2xIs+1eI)=Q¯IIs(2xIseI).(\bar{Q}_{II}^{s}-\alpha^{s}Z_{II}^{s})(2x_{I}^{s+1}-e_{I})=\bar{Q}_{II}^{s}(2x_{I}^{s}-e_{I}).

Easy to see, 2xIs+1eI2x_{I}^{s+1}-e_{I}\rightarrow\infty may happens when λmin(Q¯IIsαsZIIs)0\lambda_{min}(\bar{Q}_{II}^{s}-\alpha^{s}Z_{II}^{s})\longrightarrow 0, then the active set can be updated. Therefore we take

αs=λmin(zs12Q¯IIszs12).\alpha^{s}=\lambda_{min}(z^{s^{-\frac{1}{2}}}\bar{Q}_{II}^{s}z^{s^{-\frac{1}{2}}}).

While (19) may not be the smallest αs\alpha^{s} that induces a change in the active set, we apply a slight scaling factor of 0<η<10<\eta<1 to αs\alpha^{s}, which is

αs=ηλmin(zs12Q¯IIszs12).\alpha^{s}=\eta\cdot\lambda_{min}(z^{s^{-\frac{1}{2}}}\bar{Q}_{II}^{s}z^{s^{-\frac{1}{2}}}). (20)

The above constitutes our proof.∎

3.5 Initialization

Within this section, we introduce two initialization techniques: reformation and random perturbation.

Reformation. Without loss of generality, we initialize the vectorized PSDP by setting p=0p=0. For the UBQP problem (2), leveraging the binary nature of x{0,1}nx\in\{0,1\}^{n}, we can reformulate the objective in an equivalent form by introducing an arbitrary vector γ\gamma, yielding

xT(Q+Diag(γ))x+(bγ)Tx.x^{T}(Q+Diag(\gamma))x+(b-\gamma)^{T}x.

Although the new QQ (i.e. Q+Diag(γ)Q+Diag(\gamma)) and bb (i.e. bγb-\gamma) do not change the solution of problem (2), the solution of problem (7) subject to the constraint x[0,1]nx\in[0,1]^{n} is affected by γ\gamma. Therefore, γ\gamma will affect the initial solution of xx in our algorithm.

As noted in the previous subsection, it is imperative to prevent the iterative point sequence from approaching the boundary too rapidly, as this could impede the discovery of high-quality solutions. Consequently, the selection of an appropriate γ\gamma is crucial to ensure that the initial solution for xx resides within the interior of the feasible region. Notably, this goal can be achieved by setting γ\gamma sufficiently large. This is due to the inactive nature of the boundary defined by 0xi10\leq x_{i}\leq 1, where the first-order condition is satisfied as

2(Q+Diag(γ))x=γb.2(Q+Diag(\gamma))x=\gamma-b. (21)

Analysis of (21) reveals that as γ\gamma approaches infinity, xix_{i} converges to 1/21/2. To maintain the constraint 0<xi<10<x_{i}<1, an appropriately large γ\gamma can be chosen. For example, setting γi>j=1n2|Qij|+|bi|\gamma_{i}>\sum_{j=1}^{n}2|Q_{ij}|+|b_{i}| for i=1,,ni=1,\ldots,n in the selection of γ\gamma ensures 0<xi<10<x_{i}<1, as can be deduced from (21).

Random Perturbation. For integer problems (where the elements of QQ and bb are integers), certain components of the algorithm may occasionally stall around 1/2. Therefore, following the approach of Tang and Toh tang2023feasible , we introduce a symmetric random perturbation from a normal distribution.

Qij=Qij+μij,Q_{ij}=Q_{ij}+\mu_{ij}, (22)

where μij(0,σ)\mu_{ij}\sim\mathbb{N}(0,\sigma) and μij=μji\mu_{ij}=\mu_{ji}.

Due to the continuity of problem (7), small perturbations do not alter the optimal solution. By incorporating random perturbations, the iterations no longer stagnate around 1/2, thereby enhancing the robustness and stability of the algorithm.

Having outlined all the details of the vectorized PSDP algorithm, we now present its framework in Algorithm 3.

Algorithm 3 Vectorized PSDP for UBQP
1:  Initialization:
2:  Select a sufficiently large γ\gamma, then update Q:=Q+Diag(γ)Q:=Q+Diag(\gamma) and b:=bγb:=b-\gamma
3:  Solve x0x^{0} from (21), and set p0=0p^{0}=0 and s=0s=0
4:  while 𝒫(xs)xs>1e5\|\mathcal{P}(x^{s})-x^{s}\|_{\infty}>1e^{-5} do
5:     Update ps+1p^{s+1} by (18) and (19)
6:     Solve problem (7) with p=psp=p^{s} and x0=xsx_{0}=x^{s} using either Algorithm 1 or Algorithm 2 to obtain the optimal solution xs+1x^{s+1} based on the stopping criterion (17)
7:     s:=s+1s:=s+1
8:  end while
9:  return  The optimal solution 𝒙\boldsymbol{x^{\star}}

4 Theoretical analysis

4.1 Equivalence of the models

In this subsection, we prove that the penalty problem (7) is exact, which means that its optimal solution is also an optimal solution of problem (3), when all the components of pp are larger than a certain constant.

To simplify the description of the problem, we rewrite the problem (3) as

minxf(x) s.t. xF.\begin{array}[]{cl}\min_{x}&f(x)\\ \text{ s.t. }&x\in F.\end{array} (23)

where f(x)=xTQx+bTxf(x)=x^{T}Qx+b^{T}x and F={x|x{0,1}n}F=\{x|x\in\{0,1\}^{n}\}. Also we rewrite the problem (7) as

minxh(x,p) s.t. xG.\begin{array}[]{cl}\min_{x}&h(x,p)\\ \text{ s.t. }&x\in G.\end{array} (24)

where h(x,p):=f(x)+i=1npi(xixi2)h(x,p):=f(x)+\sum_{i=1}^{n}p_{i}(x_{i}-x_{i}^{2}) and G={x|0x1}G=\{x|0\leq x\leq 1\}.

For the convenience of our proof, let us first propose two lemmas.

Lemma 2
minxGh(x,p1)minxGh(x,p2)minxFf(x),p1p2.\min_{x\in G}h\left(x,p_{1}\right)\leq\min_{x\in G}h\left(x,p_{2}\right)\leq\min_{x\in F}f(x),\forall p_{1}\leq p_{2}.

Proof. For any p we have

minxGh(x,p)minxFh(x,p)=minxFf(x).\min_{x\in G}h(x,p)\leq\min_{x\in F}h(x,p)=\min_{x\in F}f(x).

For any p1p2p_{1}\leq p_{2} we have p2p10p_{2}-p_{1}\leq 0 , which implies (p2p1)i(xixi2)0\left(p_{2}-p_{1}\right)_{i}(x_{i}-x_{i}^{2})\geq 0 for all xGx\in G . Consequently, p2i(xixi2)p1i(xixi2)p_{2i}(x_{i}-x_{i}^{2})\geq p_{1i}(x_{i}-x_{i}^{2}) . Thus we have

h(x,p1)h(x,p2),xG.h\left(x,p_{1}\right)\leq h\left(x,p_{2}\right),\quad\forall x\in G.

Then

minxGh(x,p1)minxGh(x,p2),\min_{x\in G}h\left(x,p_{1}\right)\leq\min_{x\in G}h\left(x,p_{2}\right),

which completes our proof. ∎

Lemma 3

If xx^{\star} is a local (global) optimal point of problem (24), then xFx^{\star}\in F, for any sufficiently large p.

Proof

If xFx^{\star}\notin F, there at least one xix_{i} satisfies 0<xi<10<x_{i}<1, then exists a sequence feasible direction d=eid=e_{i} leads to

dT2L(x,u,v)d=dT(QP)d=Qiipi<0, for pi>Qii,d^{T}\nabla^{2}L(x,u,v)d=d^{T}(Q-P^{\star})d=Q_{ii}-p_{i}<0,\text{ for }p_{i}>Q_{ii},

where L(x,u,v)=xT(QP)x+(b+p)TxuTx+vT(xe)L(x,u,v)=x^{T}(Q-P^{\star})x+(b+p)^{T}x-u^{T}x+v^{T}(x-e) is the Lagrangian function function of problem (24). We can observe that it is in contradiction with the second-order necessary condition.∎

By leveraging Lemma 2 and 3, we can determine the exact penalty of problem (24), as demonstrated in Theorem 4.1.

Theorem 4.1

If x¯\bar{x} is the optimal point of problem (23)\eqref{o}, it is the global optimal point of problem (24) when p>p¯p>\bar{p}; If xx^{\star} is the global optimal point of problem (24),then it is the optimal point of problem (23)\eqref{o} when p>p¯p>\bar{p}, where p¯=(Q11,Q22,,Qnn)\bar{p}=(Q_{11},Q_{22},\dots,Q_{nn}).

Proof

We have

f(x¯)=minxFf(x)=minxFh(x,p)=minxGh(x,p)=h(x,p), for any p>p¯.f(\bar{x})=\underset{x\in F}{\min}f(x)=\underset{x\in F}{\min}h(x,p)=\underset{x\in G}{\min}h(x,p)=h(x^{\star},p),\text{ for any }p>\bar{p}.

The left equality is from Lemma 2 and the right one is from Lemma 3.∎

4.2 Convergence of the algorithm

In Algorithm 1, we apply a proximal point algorithm to solve the subproblem (7), then we present the convergence of the Algorithm 1 in Lemma 4 and Theorem 4.2.

Lemma 4

Suppose that {xk}\left\{x^{k}\right\} are generated by Algorithm 1, then

h(xk+1,p)h(xk,p)dT(HpQ+P)d.h\left(x^{k+1},p\right)\leq h\left(x^{k},p\right)-d^{T}(H_{p}-Q+P^{\star})d.
Proof

Since for all kk we have xkF1x^{k}\in F_{1}, it can be obtained that

h(xk+1,p)\displaystyle h\left(x^{k+1},p\right) =(h(xk+dk,p)+dkT(HpQ+P)dk)dkT(HpQ+P)dk\displaystyle=\left(h\left(x^{k}+d^{k},p\right)+d^{k^{T}}(H_{p}-Q+P^{\star})d^{k}\right)-d^{k^{T}}(H_{p}-Q+P^{\star})d^{k}
h(xk,p)dkT(HpQ+P)dk.\displaystyle\leq h\left(x^{k},p\right)-d^{k^{T}}(H_{p}-Q+P^{\star})d^{k}.

Here, the inequality holds due to (10).∎

Theorem 4.2

Suppose that {dk},{xk}\left\{d^{k}\right\},\left\{x^{k}\right\} are generated by Algorithm 1. Since our feasible region is bounded are bounded, then {dk}\left\{d^{k}\right\} converges to 0 and any cluster point of {(xk,dk,uk,vk)}\left\{\left(x^{k},d^{k},u^{k},v^{k}\right)\right\} is a first-order stationary point of problem (7).

Proof

Based on Lemma 4, it is established that

h(xk+1,p)h(x0,p)i=1kdiT(HpQ+P)di.h\left(x^{k+1},p\right)\leq h\left(x^{0},p\right)-\sum_{i=1}^{k}d^{i^{T}}(H_{p}-Q+P^{\star})d^{i}.

Consequently, due to the boundness of function hh, we derive

k=1dkT(HpQ+P)dk<+.\sum_{k=1}^{\infty}d^{k^{T}}(H_{p}-Q+P^{\star})d^{k}<+\infty.

Hence, dk0d^{k}\rightarrow 0 as HpQ+P0H_{p}-Q+P^{\star}\succ 0.

Returning to our subproblem (10), the KKT conditions are outlined below:

2(QP)xk+(b+p)+2Hpdkuk+vk=0,uk,vk0,ukT(xk+dk)=0,vkT(exkdk)=0,0xk+dk1.\begin{array}[]{c}2(Q-P^{\star})x^{k}+(b+p)+2H_{p}d^{k}-u^{k}+v^{k}=0,\\ u^{k},v^{k}\leq 0,\\ u_{k}^{T}(x^{k}+d^{k})=0,\\ v_{k}^{T}(e-x^{k}-d^{k})=0,\\ 0\leq x^{k}+d^{k}\leq 1.\\ \end{array} (25)

We can see that, for xIk+1x_{I}^{k+1}, we have vIk=uIk=0v^{k}_{I}=u^{k}_{I}=0; for xKk+1x_{K}^{k+1}, uKk=0u^{k}_{K}=0; for xJk+1x_{J}^{k+1}, vJk=0v^{k}_{J}=0. Thus, ukTv=0u^{k^{T}}v=0, then we have uk2QP+b+p+Hp\left\|u^{k}\right\|_{\infty}\leq 2\left\|Q-P^{\star}\right\|_{\infty}+\|b+p\|_{\infty}+\|Hp\|_{\infty}, vk2Qp+b+p+Hp\left\|v^{k}\right\|_{\infty}\leq 2\left\|Q-p^{\star}\right\|_{\infty}+\|b+p\|_{\infty}+\|Hp\|_{\infty}, indicating that {(xk,dk,uk,vk)}\{(x^{k},d^{k},u^{k},v^{k})\} is bounded and possesses a convergent subsequence. Assume that {(xik,dik,uik,vik)}\{(x^{i_{k}},d^{i_{k}},u^{i_{k}},v^{i_{k}})\} is a subsequence that converges to an arbitrary cluster point (x,d,u,v)(x^{\star},d^{\star},u^{\star},v^{\star}). With dk0d^{k}\rightarrow 0, it follows that d=0d^{\star}=0. Let iki_{k}\rightarrow\infty in (25), then (x,d,u,v)(x^{\star},d^{\star},u^{\star},v^{\star}) satisfies the KKT conditions (16), which implies that it is a first-order stationary point of problem (7).∎

In Algorithm 2, an alternative BB method is adopted, which is non-monotonic, allowing for instances where h(xk+1,p)h\left(x^{k+1},p\right) may increase during certain iterations. Dai and Fletcher 2005Projected provided counterexamples illustrating that the Projected BB and Projected Alternating BB methods, without line searches, could cyclically oscillate between multiple points, and such behavior is atypical. Furthermore, Li and Huang li2023note theoretically proved that the alternative BB method without a monotone line search converges at least R-linearly with a rate of 11/κ1-1/\kappa under certain properties for convex unconstrained quadratic problems. However, we have not yet found a relevant theoretical result for such a non-convex problem. In the subsequent section, numerical experiments will be conducted to demonstrate that the PABB method does not exhibit cycling phenomena.

5 Numerical experiment

In this section, we test our proposed algorithms on some UBQP benchmark datasets and random generated datasets. Since our method is derived from the SDP, we compare it with the SDR approach based on the SDPT3 solver R2003Solving , which is used to solve subproblems in luo2019enhancing ; RePEc:spr:jcomop:v:39:y:2020:i:3:d:10.1007_s10878-019-00517-8 . Furthermore, we contrast our approach with two similar penalty function-based methods RePEc:spr:jcomop:v:39:y:2020:i:3:d:10.1007_s10878-019-00517-8 ; yuan2017exact , referred to as ALM and MPEC in this paper. The specific parameters for the five algorithms are outlined as follows:

  • PPA (Vectorized PSDP method with proximal point algorithm)
    Algorithm 1 is utilized for solving the subproblems, employing a termination criterion of 10510^{-5} or a maximum of 10000 iterations. The penalty function is updated for a maximum of 1000 iterations.

  • PABB (Vectorized PSDP method with projection alternative BB algorithm)
    Algorithm 2 is employed to address the subproblems, utilizing a termination criterion of 10510^{-5} or a maximum of 10000 iterations. The penalty function is updated for a maximum of 1000 iterations.

  • SDR (Semidefinite relaxation)
    It is implemented using the SDPT3 solver, with all termination criteria set to 10510^{-5}.

  • MPEC (Exact penalty method for binary optimization based on MPEC formulation)
    All the parameter settings remain consistent with the code provided by Yuan111https://yuangzh.github.io/.

  • ALM (The augmented Lagrangian algorithms based on continuous relaxation)
    All the parameter settings remain consistent with the original papers. We consider x0=pex_{0}=pe, where pp is a randomly chosen scalar in the interval (0,1)(0,1) and ee is the vector of all ones.

We test our algorithms on a total of six datasets, which are Random Instances, QPLIB222https://qplib.zib.de/, BE333https://biqmac.aau.at/biqmaclib.html (Billionnet and Elloumi instances), GKA444https://biqmac.aau.at/biqmaclib.html (Glover, Kochenberger and Alidaee instances), rudy555http://www-user.tu-chemnitz.de/\simhelmberg/rudy.tar.gz and the G set666https://web.stanford.edu/\simyyye/yyye/Gset/. We evaluate the computational efficiency and solution quality using CPU time and the solution GAP as metrics, where the GAP calculation formula is defined as:

 GAP =|objlblb|×100%.\text{ GAP }=\left|\frac{\mathrm{obj}-\mathrm{lb}}{\mathrm{lb}}\right|\times 100\%.

where the lower bound (lb\mathrm{lb}) is determined through the utilization of the Biq Mac Solver777https://biqmac.aau.at/.

5.1 Convergence verification

Due to the absence of any line search, it is challenging to provide the theoretical convergence analysis for Algorithm 2. Here, we validate its practical convergence through numerical experiments. We present one instance from each of the six datasets, and Fig.1 illustrates the number of iterations used by Algorithm 2 in solving the subproblems. It can be observed from the figure that the number of iterations for the subproblems is mostly within 1000. For the instance with the QPLIB ID 3565, there is a small probability that the number of iterations for the subproblems exceeds 2000. However, even in such cases, all iteration counts remain below 6000, not reaching the maximum of 10000 iterations set for the experiments. This indicates to some extent that the PABB method without line search does not exhibit cycling, confirming the convergence of Algorithm 2 in practice.

Next, we validate the overall convergence of the two vectorized PSDP algorithms (PPA and PABB) in numerical experiments. Fig.2 illustrates the convergence process of the two algorithms in reducing the number of components violating the 0-1 constraints over six test cases. While not strictly monotonically decreasing, the overall trend is toward reduction and convergence to zero. This indicates that our algorithms ultimately converge to feasible points satisfying the 0-1 constraints.

Refer to caption

(a) Random datasets

Refer to caption

(b) QPLIB datasets

Refer to caption

(c) BE datasets

Refer to caption

(d) GKA datasets

Refer to caption

(e) Rudy datasets

Refer to caption

(f) G sets

Figure 1: Histogram of the number of iterations for subproblems
Refer to caption

(a) Random datasets

Refer to caption

(b) QPLIB datasets

Refer to caption

(c) BE datasets

Refer to caption

(d) GKA datasets

Refer to caption

(e) Rudy datasets

Refer to caption

(f) G sets

Figure 2: The number of variables dissatisfied the constraint conditions

5.2 0-1 quadratic problems

The UBQP problem to be addressed is formulated as follows:

min{yTQy:y{0,1}n},\min\left\{y^{T}Qy:y\in\{0,1\}^{n}\right\},

where Q is a symmetric matrix of order nn. For this problem, we conducted experiments on four datasets, namely: random instances, QPLIB, BE, and GKA.

(Random Instances) We randomly generate the symmetric matrix QQ with dimensions of 50, 100, 150, 200, and 250. For each dimension, we create five test cases. Due to the stochastic nature of some algorithms, we run each algorithm 10 times for each test case and report the average GAP and CPU time. In cases where the runtime of the ALM algorithm exceeds 100 seconds or yields a non-optimal trivial solution of 0, we classify such instances as failures.

The detailed results are summarized in Table 1. We observe that the five algorithms do not exhibit significant differences in terms of runtime, with the SDR algorithm running relatively slower. In terms of solution quality, PPA achieves the smallest GAP with the lower bound, closely followed by PABB. However, ALM fails on average 1 to 4 times in each test case.

Table 1: Experimental results of Random Instances in dimensions of 50, 100, 150, 200, and 250
CPU Time(s) GAP(%) Fall
n PPA PABB ALM MPEC SDR PPA PABB ALM MPEC SDR ALM
50 0.13 0.05 0.13 0.33 0.42 0.51 0.53 2.09 0.80 2.19 4.40
100 0.38 0.10 0.20 0.23 0.44 0.15 0.07 2.63 0.95 3.56 3.00
150 1.14 0.20 0.30 0.26 0.56 0.21 0.23 2.72 0.79 2.98 1.80
200 3.21 0.46 0.46 0.29 0.86 0.14 0.21 2.04 0.82 4.33 1.80
250 5.07 0.65 0.63 0.34 1.17 -0.15 -0.12 1.52 0.50 3.20 1.60

Additionally, we consider some lower-dimensional matrices with dimensions of 20, 40, 60, and 80, and apply the branch-and-bound method to obtain the optimal solution as the lower bound (lb). As shown in Table 2, PABB is relatively faster in terms of time, with a smaller GAP compared to the optimal value, and even the maximum GAP is only 1.02%. PPA is more accurate, achieving an average GAP of just 0.02% when solving examples with 80 dimensions. In contrast, ALM has a probability of failure during the solving process, and even when successful, the GAP is relatively high. The MPEC solution is also accurate, but with longer run times. On the other hand, the SDR algorithm has slower solving speeds and relatively lower solution quality.

Table 2: Experimental results of Random Instances in dimensions of 20, 40, 60 and 80
CPU Time(s) GAP(%) Fall
n PPA PABB ALM MPEC SDR PPA PABB ALM MPEC SDR ALM
20 0.03 0.02 0.09 0.34 0.47 0.09 0.06 0.60 0.09 1.31 3.20
40 0.11 0.05 0.13 0.37 0.44 1.00 1.02 2.20 1.48 4.87 4.60
60 0.17 0.07 0.17 0.44 0.52 0.09 0.12 2.75 0.35 1.96 1.40
80 0.45 0.15 0.41 0.42 0.66 0.02 0.13 1.77 0.95 2.41 1.40

This observation suggests that the two vectorized PSDP methods have a clear advantage in solution quality over other approaches. PABB generally outperforms PPA in speed, effectively speeding up the process.

(QPLIB) The QPLIB is a library of quadratic programming instances, which contains 319 discrete and 134 continuous instances of different characteristics. We conduct testing on all binary instances, totaling 23 cases, with dimensions ranging from 120 to 1225 in the QPLIB dataset.

As shown in Table 3, MPEC emerges as the fastest approach, with an average of 0.49 seconds per instance. PABB and ALM follow closely with an average of about 1 second per instance. PPA and SDR take longer. In terms of solution quality, PPA and PABB maintain their superiority.

Table 3: Experimental results of QPLIB Instances
CPU Time(s) GAP(%) Fall
QPLIB n PPA PABB ALM MPEC SDR PPA PABB ALM MPEC SDR ALM
QPLIB5881 120 0.48 0.17 0.44 0.32 0.60 0.40 0.39 3.05 1.16 3.49 7
QPLIB5882 150 5.40 0.61 0.83 0.35 0.62 0.12 0.12 1.65 1.46 2.98 4
QPLIB5875 200 10.81 1.17 1.26 0.38 0.91 0.00 0.04 1.12 0.75 3.00 5
QPLIB3852 231 1.61 0.44 0.38 0.37 1.03 0.34 0.34 1.45 2.56 5.13 0
QPLIB5909 250 1.44 0.52 0.85 0.35 1.19 0.27 0.52 3.59 1.45 5.88 4
QPLIB3565 276 1.45 0.30 0.29 0.30 0.87 1.35 1.28 1.26 2.27 8.51 1
QPLIB5721 300 14.24 5.81 3.41 0.66 1.37 2.74 1.70 1.01 1.93 0.94 6
QPLIB3745 325 2.35 0.60 0.52 0.43 1.37 1.14 1.08 1.86 2.40 5.39 1
QPLIB5725 343 1.05 0.50 1.25 0.58 1.57 4.51 2.38 2.12 2.27 11.09 7
QPLIB3705 378 2.78 0.72 0.54 0.44 1.94 0.62 1.04 2.19 1.04 9.90 0
QPLIB5755 400 1.49 0.54 1.34 0.47 2.09 3.78 2.19 1.72 2.17 11.08 6
QPLIB3738 435 3.88 0.63 1.26 0.49 2.30 1.52 1.66 6.30 2.84 4.74 3
QPLIB3506 496 3.55 0.79 0.59 0.43 2.29 0.84 0.92 1.57 1.67 10.88 2
QPLIB5922 500 9.71 1.64 2.07 0.41 3.91 0.15 0.18 1.38 0.56 1.92 5
QPLIB3832 561 4.48 0.91 0.58 0.49 3.61 1.12 2.02 1.77 2.53 6.86 0
QPLIB3877 630 5.15 0.97 0.71 0.55 4.53 1.10 1.40 2.18 1.33 7.31 1
QPLIB3706 703 6.17 0.95 0.78 0.54 5.17 2.11 1.50 1.97 2.05 5.87 3
QPLIB3838 780 7.47 1.07 0.95 0.59 7.41 2.36 2.79 3.25 2.95 8.58 1
QPLIB3822 861 7.64 1.20 1.15 0.56 8.37 1.34 0.85 2.09 1.18 3.76 0
QPLIB3650 946 8.22 1.75 1.23 0.64 10.86 1.02 1.46 2.28 0.65 5.00 4
QPLIB3642 1035 7.05 1.48 1.11 0.56 10.27 1.59 1.80 2.36 1.16 8.70 1
QPLIB3693 1128 10.14 2.02 1.63 0.67 14.90 0.80 1.39 1.44 1.92 8.01 2
QPLIB3850 1225 10.41 1.61 1.91 0.75 18.35 0.96 1.50 1.29 1.53 6.72 1
Average 5.52 1.15 1.09 0.49 4.59 1.31 1.24 2.13 1.73 6.34 3

(GKA) These datasets can also be obtained from the OR-Library888http://people.brunel.ac.uk/\simmastjjb/jeb/info.html, denoted as sets a, b, c, d, e, f. The dataset parameters are generated using the Pardalos-Rodgers generators introduced by Pardalos and Rodgers Pardalos1990ComputationalAO . The dimensions of this dataset range from 40 to 500, with densities ranging from 0 to 1.

It is clear from Table 4 that for n100n\leq 100, PABB takes the least amount of computational time. However, as nn increases to 200 and 500, the computational time of PABB exceeds that of MPEC. The solution quality produced by PPA and PABB is generally superior to that of the other methods, particularly in many instances where n200n\leq 200, their GAPs are one-tenth that of MPEC.

Table 4: Experimental results of GKA Instances
CPU Time(s) GAP(%) Fall
GKA n PPA PABB ALM MPEC SDR PPA PABB ALM MPEC SDR ALM
gka1a 50 0.07 0.03 0.15 0.19 0.29 0.527 0.539 0.849 0.000 0.996 9
gka2a 60 0.04 0.02 0.19 0.22 0.36 0.000 0.000 0.099 0.000 1.039 4
gka3a 70 0.11 0.05 0.19 0.25 0.30 0.023 0.045 0.182 0.646 2.418 3
gka4a 80 0.04 0.02 0.22 0.25 0.32 0.000 0.026 0.803 0.233 0.640 1
gka5a 50 0.07 0.02 0.22 0.30 0.40 0.052 0.047 0.854 0.052 2.911 5
gka6a 30 0.04 0.02 0.15 0.28 0.37 0.000 0.018 6.583 0.000 0.176 8
gka7a 30 0.04 0.02 0.14 0.28 0.37 0.000 0.000 2.510 0.000 0.000 5
gka8a 100 0.07 0.02 0.28 0.33 0.51 0.000 0.000 0.216 0.000 0.522 3
gka1c 40 0.04 0.01 0.12 0.23 0.25 0.000 0.000 0.000 0.000 0.000 7
gka2c 50 0.07 0.02 0.19 0.29 0.31 0.000 0.000 0.000 0.306 0.998 5
gka3c 60 0.08 0.03 0.24 0.27 0.42 0.000 0.008 0.900 0.255 0.255 4
gka4c 70 0.05 0.02 0.16 0.18 0.29 0.000 0.000 0.378 0.068 0.825 8
gka5c 80 0.12 0.06 0.25 0.31 0.44 0.000 0.000 1.616 0.435 1.372 6
gka6c 90 0.10 0.05 0.27 0.30 0.48 0.000 0.000 0.017 0.343 1.185 3
gka7c 100 0.12 0.05 0.26 0.44 0.53 0.000 0.000 0.720 1.522 1.107 5
gka1d 100 0.12 0.04 0.24 0.19 0.37 0.174 0.025 1.026 2.447 2.100 3
gka2d 100 0.29 0.09 0.33 0.25 0.48 0.108 0.134 1.844 1.170 7.098 4
gka3d 100 0.30 0.10 0.34 0.37 0.52 0.173 0.080 5.129 0.173 1.447 4
gka4d 100 0.25 0.06 0.24 0.19 0.35 0.021 0.062 1.231 0.121 1.603 3
gka5d 100 0.36 0.08 0.31 0.36 0.52 0.095 0.415 1.338 3.140 4.602 4
gka6d 100 0.21 0.07 0.29 0.34 0.46 0.352 0.000 2.119 1.330 3.076 3
gka7d 100 0.40 0.09 0.28 0.35 0.51 0.015 0.089 2.344 0.000 4.255 4
gka8d 100 0.25 0.04 0.30 0.34 0.49 0.000 0.217 1.137 0.183 2.312 3
gka9d 100 0.33 0.07 0.30 0.49 0.48 0.006 0.011 1.234 1.207 2.549 5
gka10d 100 0.45 0.08 0.30 0.21 0.40 0.042 0.138 0.853 0.832 1.335 2
gka1e 200 0.85 0.23 0.68 0.25 0.81 0.081 0.165 1.252 1.039 1.761 3
gka2e 200 1.77 0.33 1.00 0.26 1.00 0.086 0.102 1.888 0.816 2.013 5
gka3e 200 1.99 0.33 0.95 0.28 1.00 0.136 0.057 2.733 0.382 2.892 1
gka4e 200 1.89 0.24 0.77 0.23 0.82 0.044 0.034 1.992 0.337 1.365 5
gka5e 200 3.18 0.35 0.85 0.39 0.94 0.071 0.132 2.402 1.155 5.402 3
gka1f 500 11.34 2.12 4.60 0.33 3.42 3.581 3.669 4.750 4.363 6.883 5
gka2f 500 20.15 1.91 4.77 0.30 3.63 4.601 4.691 6.847 5.351 6.832 4
gka3f 500 33.20 2.47 5.18 0.34 2.81 5.238 5.212 7.272 6.245 8.394 4
gka4f 500 37.94 2.37 5.07 0.45 3.58 4.945 4.942 8.309 6.032 7.902 5
gka5f 500 43.53 2.10 5.13 0.51 3.65 5.321 5.450 9.637 5.724 9.563 4
AVERAGE 4.57 0.39 1.00 0.30 0.91 0.734 0.752 2.316 1.312 2.795 4

(BE) There are a total of 8 categories, each with 10 instances in Billionnet and Elloumi instances. These instances have dimensions of 20, 40, 60, and 80, with densities ranging from 0.1 to 1.0. The diagonal coefficients of the instances fall within the range [-100, 100], while the off-diagonal coefficients range from [-50, 50]. In Table 5, we consider the solution presented in the article BENLIC20131162 as the lower bound (lb).

It is evident that both PABB and MPEC have shorter computation times compared to other methods. In the majority of cases, PABB has a shorter runtime than MPEC. Furthermore, PABB and PPA show a clear advantage in terms of GAP compared to other methods, with a GAP typically below 0.1%.

Table 5: Experimental results of BE Instances
CPU Time(s) GAP(%) Fall
n Density PPA PABB ALM MPEC SDR PPA PABB ALM MPEC SDR ALM
100 1.0 0.55 0.09 0.38 0.30 0.50 0.01 0.07 1.45 0.67 2.76 3.70
120 0.3 0.58 0.14 0.54 0.28 0.58 0.07 0.09 2.09 0.92 2.43 2.70
120 0.8 1.16 0.17 0.57 0.28 0.55 0.04 0.07 1.95 1.28 3.47 3.90
150 0.3 1.34 0.25 0.75 0.34 0.83 0.07 0.14 2.01 1.17 3.30 2.80
150 0.8 2.23 0.28 0.72 0.35 0.72 0.10 0.10 2.91 0.86 3.24 4.50
200 0.3 2.64 0.39 1.05 0.32 1.17 0.10 0.09 1.90 0.99 3.51 3.80
200 0.8 4.18 0.40 1.05 0.37 0.91 0.05 0.11 2.22 0.96 3.27 3.60
250 0.1 2.00 0.44 1.16 0.29 1.20 0.05 0.10 1.23 1.02 2.78 4.00

5.3 Max cut

The max-cut problem can be formulated as

max{xTWx:x{1,1}n},\max\left\{x^{T}Wx:x\in\{-1,1\}^{n}\right\},

where WW is the weighted adjacency matrix of the given graph G. We test max cut problem on rudy and G set datasets.

(rudy) Graphs of the following four types have been generated using rudy rinaldi1998rudy :

  • G0.5G_{0.5}, g05n.ig05_{n}.i unweighted graphs with edge probability 1/2, n=60,80,100;

  • G1/0/1G_{-1/0/1}, pm1sn.ipm1s_{n}.i, pm1dn.ipm1d_{n}.i weighted graph with edge weights chosen uniformly from {-1,0,1} and density 10% and 99% respectively, n=80,100;

  • G[10,10],wd_niG_{[-10,10]},\mathrm{w}d\_n\cdot i Graph with integer edge weights chosen from [-10,10] and density d=0.1,0.5,0.9, n=100;

  • G[0,10],pwd_n.iG_{[0,10]},\mathrm{pw}d\_n.i Graph with integer edge weights chosen from [-10,10] and density d=0.1,0.5,0.9, n=100.

All five algorithms successfully solve all instances, with each instance being run once. From Tables 6 we can see that PPA and PABB consistently have the smallest GAP, with PABB requiring less CPU time. In addition, ALM occasionally has the shortest runtime, but its solution quality is relatively inferior.

Table 6: Experimental results of rudyg05,rudyPM,rudyW,rudyPWrudy_{g05},rudy_{PM},rudy_{W},rudy_{PW}
CPU Time(s) GAP(%)
n Density PPA PABB ALM MPEC SDR PPA PABB ALM MPEC SDR
𝒓𝒖𝒅𝒚𝒈𝟎𝟓\boldsymbol{rudy_{g05}}
60 - 0.78 0.15 0.12 0.41 0.49 0.06 0.11 0.19 0.47 2.58
80 - 1.55 0.26 0.14 0.43 0.51 0.10 0.16 0.64 0.18 2.24
100 - 2.64 0.45 0.21 0.52 0.56 0.03 0.05 0.41 0.37 2.42
𝒓𝒖𝒅𝒚𝑷𝑴\boldsymbol{rudy_{PM}}
80 0.1 0.74 0.26 0.15 0.54 0.76 1.46 0.74 2.28 2.16 13.54
80 0.99 2.00 0.32 0.25 0.47 0.57 0.17 0.58 2.72 1.95 16.93
100 0.1 0.93 0.25 0.23 0.57 0.94 0.72 1.03 3.58 2.25 16.16
100 0.99 3.56 0.49 0.27 0.47 0.56 0.13 1.08 5.18 1.74 17.79
𝒓𝒖𝒅𝒚𝑾\boldsymbol{rudy_{W}}
100 0.1 0.32 0.17 0.26 0.47 0.64 0.57 0.74 1.65 1.95 14.32
100 0.5 1.49 0.27 0.31 0.44 0.55 0.44 1.20 3.78 2.88 17.10
100 0.9 2.84 0.43 0.44 0.49 0.56 1.05 1.33 6.37 2.52 16.25
𝒓𝒖𝒅𝒚𝑷𝑾\boldsymbol{rudy_{PW}}
100 0.1 0.39 0.16 0.26 0.46 0.58 0.29 0.62 0.56 1.48 4.23
100 0.5 1.87 0.41 0.35 0.50 0.52 0.12 0.06 1.04 0.38 2.64
100 0.9 4.45 0.62 0.47 0.62 0.50 0.06 0.02 4.44 0.24 1.95

(G sets) The instances in the G set are numbered in ascending order by dimension. We selected the first five instances for the 800, 1000, and 2000 dimensions. For the 3000-dimensional instances, there are only 3 total instances in the G set, so we included all of them in the test. The optimal solution (lb) is taken from the article BENLIC20131162 .

As shown in Table 7, PABB has a shorter runtime compared to PPA and SRD, similar to ALM, but longer than MPEC. Overall, the solution quality of PABB is close to that of PPA, outperforming ALM, MPEC, and SDR, especially for the 1000-dimensional instances. In the case of the 3000-dimensional instances, ALM and SDR fail, while PPA and PABB prove to be more effective in solving the problems.

Table 7: Experimental results of G sets
CPU Time(s) GAP(%)
Gset n PPA PABB ALM MPEC SDR PPA PABB ALM MPEC SDR
G16 800 4.50 1.79 1.57 0.38 6.89 1.70 1.54 1.28 2.26 3.93
G17 800 4.27 1.65 1.93 0.44 6.72 1.12 1.54 1.31 2.03 4.73
G18 800 6.78 1.33 0.99 0.40 6.37 1.51 2.62 3.33 6.96 14.42
G19 800 6.37 1.16 1.21 0.31 6.03 1.10 2.54 3.86 6.51 12.25
G20 800 8.77 1.52 1.47 0.36 6.86 1.91 1.91 3.72 4.57 16.15
G43 1000 35.78 8.18 1.21 0.83 15.59 0.42 0.32 1.22 1.47 3.50
G44 1000 35.42 8.55 1.51 0.77 11.90 0.03 0.27 1.62 1.02 4.30
G45 1000 31.90 7.39 1.94 0.63 12.06 0.27 0.30 1.55 0.90 5.03
G46 1000 24.63 9.72 1.59 0.77 12.51 0.18 0.32 1.04 0.86 3.99
G47 1000 34.50 7.64 1.47 0.59 11.86 0.20 0.15 1.19 0.69 3.59
G36 2000 19.42 7.01 4.29 0.59 62.05 1.65 1.62 1.16 3.10 4.10
G37 2000 18.20 7.64 5.05 0.66 59.30 1.87 1.38 1.11 3.10 4.03
G38 2000 21.84 7.98 4.38 0.67 59.96 1.50 1.35 1.76 2.94 4.27
G39 2000 35.33 11.41 6.06 1.32 61.18 2.28 2.70 5.90 8.06 14.87
G40 2000 27.42 8.68 4.87 0.51 61.13 1.08 3.67 3.83 10.13 15.71
G48 3000 0.99 0.61 - 0.95 - 9.40 0.00 Fall 0.00 Fall
G49 3000 8.90 0.31 - 2.20 - 0.00 0.00 Fall 1.03 Fall
G50 3000 0.59 0.61 - 0.86 - 3.95 1.43 Fall 0.07 Fall

6 Conclusion

This paper presents a vectorized PSDP method for solving the UBQP problem, where the method framework is based on PSDP and involves vectorizing matrix variables to significantly reduce computational complexity. Various enhancements have been made in algorithmic details, such as penalty factor updates and initialization. Two algorithms are proposed for solving subproblems, each incorporating proximal point method and projection alternating BB method. Properties of the penalty function and some aspects of algorithm convergence are discussed. Numerical experiments demonstrate that the proposed method not only achieves satisfactory solution times compared to SDR and other established methods but also demonstrates outstanding capability in obtaining high-quality solutions.

In future research, our goal is to delve deeper into strategies for accelerating the solution of the subproblem, with the aim of increasing the solution speed. In addition, we aim to synchronize the adjustment of the penalty parameter with the updates of the subproblem, thereby reducing the number of iterations required to solve the subproblem.

Acknowledgements.
This work was funded by National Key R&D Program of China grant No. 2022YFA1003800, NSFC grant No. 12201318 and the Fundamental Research Funds for the Central Universities No. 63243075.

References

  • (1) Azuma, G., Fukuda, M., Kim, S., Yamashita, M.: Exact sdp relaxations for quadratic programs with bipartite graph structures. Journal of global optimization (2023)
  • (2) Barahona, F., Jünger, M., Reinelt, G.: Experiments in quadratic 0–1 programming. Mathematical Programming 44(1), 127–137 (1989)
  • (3) Beck, A., Teboulle, M.: Global optimality conditions for quadratic optimization problems with binary constraints. SIAM Journal on Optimization 11(1), 179–188 (2000)
  • (4) Benlic, U., Hao, J.K.: Breakout local search for the max-cutproblem. Engineering Applications of Artificial Intelligence 26(3), 1162–1173 (2013)
  • (5) Billionnet, A., Elloumi, S.: Using a mixed integer quadratic programming solver for the unconstrained quadratic 0-1 problem. Mathematical Programming 109, 55–68 (2007)
  • (6) Boros, E., Hammer, P.L.: The max-cut problem and quadratic 0–1 optimization; polyhedral aspects, relaxations and bounds. Annals of Operations Research 33(3), 151–180 (1991)
  • (7) Cesarone, F., Scozzari, A., Tardella, F.: Linear vs. quadratic portfolio selection models with hard real-world constraints. Computational Management Science 12(3), 345–370 (2014)
  • (8) Dai, Y.H., Fletcher, R.: Projected barzilai-borwein methods for large-scale box-constrained quadratic programming. Numerische Mathematik 100(1), 21–47 (2005)
  • (9) Fujie, T., Kojima, M.: Semidefinite programming relaxation for nonconvex quadratic programs. Journal of Global optimization 10, 367–380 (1997)
  • (10) Gu, R., Du, Q., Yuan, Y.x.: Positive semidefinite penalty method for quadratically constrained quadratic programming. IMA Journal of Numerical Analysis 41(4), 2488–2515 (2020)
  • (11) Gulati, V., Gupta, S., Mittal, A.: Unconstrained quadratic bivalent programming problem. European Journal of Operational Research 15(1), 121–125 (1984)
  • (12) Krislock, N., Malick, J., Roupin, F.: Biqcrunch: A semidefinite branch-and-bound method for solving binary quadratic problems. ACM Trans. Math. Softw. 43, 32:1–32:23 (2017)
  • (13) Li, X.R., Huang, Y.K.: A note on r-linear convergence of nonmonotone gradient methods. Journal of the Operations Research Society of China pp. 1–13 (2023)
  • (14) Lin, W.H., Wang, C.: An enhanced 0-1 mixed-integer lp formulation for traffic signal control. IEEE Transactions on Intelligent Transportation Systems 5(4), 238–245 (2004)
  • (15) Liu, H., Deng, K., Liu, H., Wen, Z.: An entropy-regularized admm for binary quadratic programming. Journal of Global Optimization 87 (2022)
  • (16) Liu, Z., Yu, Z., Wang, Y.: A continuous approach to binary quadratic problems. Journal of Applied Mathematics and Physics 06, 1720–1732 (2018)
  • (17) Luo, H., Bai, X., Peng, J.: Enhancing semidefinite relaxation for quadratically constrained quadratic programming via penalty methods. Journal of Optimization Theory and Applications 180, 964–992 (2019)
  • (18) Luo, Z.q., Ma, W.k., So, A.M.c., Ye, Y., Zhang, S.: Semidefinite relaxation of quadratic optimization problems. IEEE Signal Processing Magazine 27(3), 20–34 (2010)
  • (19) Nayak, R.K., Mohanty, N.K.: Solution of boolean quadratic programming problems by two augmented lagrangian algorithms based on a continuous relaxation. Journal of Combinatorial Optimization 39(3), 792–825 (2020)
  • (20) Oshiyama, H., Ohzeki, M.: Benchmark of quantum-inspired heuristic solvers for quadratic unconstrained binary optimization. Scientific reports 12(1), 2146 (2022)
  • (21) Pardalos, P.M., Jha, S.: Complexity of uniqueness and local search in quadratic 0–1 programming. Operations Research Letters 11(2), 119–123 (1992)
  • (22) Pardalos, P.M., Rodgers, G.P.: Computational aspects of a branch and bound algorithm for quadratic zero-one programming. Computing 45, 131–144 (1990)
  • (23) Rinaldi, G.: Rudy. http://www-user.tu-chemnitz.de/ helmberg/rudy.tar.gz (1998)
  • (24) Sherali, H.D.: Rlt: A unified approach for discrete and continuous nonconvex optimization. Annals of Operations Research 149, 185–193 (2007)
  • (25) Shi, J., Zhang, Q., Derbel, B., Liefooghe, A.: A parallel tabu search for the unconstrained binary quadratic programming problem. pp. 557–564 (2017)
  • (26) Tang, T., Toh, K.C.: A feasible method for general convex low-rank sdp problems (2023)
  • (27) Teixeira, R.F., Fernandes, F.C.F., Pereira, N.A.: Binary integer programming formulations for scheduling in market-driven foundries. Computers & Industrial Engineering 59(3), 425–435 (2010)
  • (28) Tütüncü, R.H., Toh, K.C., Todd, M.J.: Solving semidefinite-quadratic-linear programs using sdpt3. Mathematical Programming 95(2), 189–217 (2003)
  • (29) Wang, Y., Lü, Z., Glover, F., Hao, J.K.: A multilevel algorithm for large unconstrained binary quadratic optimization. In: N. Beldiceanu, N. Jussien, É. Pinson (eds.) Integration of AI and OR Techniques in Contraint Programming for Combinatorial Optimzation Problems, pp. 395–408. Springer Berlin Heidelberg, Berlin, Heidelberg (2012)
  • (30) Xia, Y.: New optimality conditions for quadratic optimization problems with binary constraints. Optimization Letters 3, 253–263 (2009)
  • (31) Yuan, G., Ghanem, B.: An exact penalty method for binary optimization based on mpec formulation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 31 (2017)