This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Unified Framework for Multiclass and Multilabel Support Vector Machines

Hoda Shajari   Anand Rangarajan [email protected]
(Dept. of Computer and Information Science and Engineering
University of Florida, Gainesville, FL
)
Abstract

We propose a novel integrated formulation for multiclass and multilabel support vector machines (SVMs). A number of approaches have been proposed to extend the original binary SVM to an all-in-one multiclass SVM. However, its direct extension to a unified multilabel SVM has not been widely investigated. We propose a straightforward extension to the SVM to cope with multiclass and multilabel classification problems within a unified framework. Our framework deviates from the conventional soft margin SVM framework with its direct oppositional structure. In our formulation, class-specific weight vectors (normal vectors) are learned by maximizing their margin with respect to an origin and penalizing patterns when they get too close to this origin. As a result, each weight vector chooses an orientation and a magnitude with respect to this origin in such a way that it best represents the patterns belonging to its corresponding class. Opposition between classes is introduced into the formulation via the minimization of pairwise inner products of weight vectors. We also extend our framework to cope with nonlinear separability via standard reproducing kernel Hilbert spaces (RKHS). Biases which are closely related to the origin need to be treated properly in both the original feature space and Hilbert space. We have the flexibility to incorporate constraints into the formulation (if they better reflect the underlying geometry) and improve the performance of the classifier. To this end, specifics and technicalities such as the origin in RKHS are addressed. Results demonstrates a competitive classifier for both multiclass and multilabel classification problems.

1 Introduction

Supervised classification is of great importance in many domains and research areas and has enjoyed widespread application. In a vast majority of cases, it is usually assumed that each pattern or instance (during training and testing) only belongs to one class and therefore is assigned a single label. While this kind of single label classification has been widely studied, there are many situations which do not fit into the single label classification framework with the classification problems being inherently multilabel. In the multilabel setting, an instance can simultaneously belong to multiple classes and therefore more than one label needs to be assigned to each instance. Consider the image tagging problem for instance. Here, each image contains multiple objects, each possessing a different label. The task at hand is the assignment of all relevant labels to each image comprising the set of unique objects in the scene. Alternatively, consider the problem of identifying each instrument at every time instant of a song. Since multiple instruments—guitar, bass and drums for example—can be simultaneously heard at any given moment, instrument recognition is inherently a multilabel classification problem.

1.1 Motivation

Multilabel classification is a generalization of the well known multiclass problem in supervised learning and is consequently more challenging. Some of the challenges in the multilabel setting include, but are not limited to, exponential growth of cardinality of possible label sets, correlation among labels, structured label spaces and severely unbalanced datasets. Multilabel classification has already seen many real-world applications such as scene labeling [4], functional genomics analysis [61, 33] and text categorization [25, 38, 27, 43].

The rising number of applications has led to the development of multilabel learning as a new supervised learning paradigm [16]. Many approaches have been proposed in the literature to tackle multilabel classification problems within the support vector machine (SVM) framework. Despite the plethora of approaches and algorithms, we assert that no straightforward (natural) extension of the SVM has hitherto been proposed for multilabel classification. The SVM was first introduced as a two-class, single label (binary) classifier based on maximum margin geometry (and regularized error minimization) [3, 8, 58]. Since then, the binary SVM has been studied extensively and applied successfully in many different domains. However, its extension to multiclass and multilabel cases is still an ongoing research problem since there is neither consensus nor a majoritarian approach.

In this work, we first attempt to justify the need to revisit the two-class (binary) SVM framework. Subsequently, we investigate the multilabel problem from a unified large margin perspective. We propose a novel formulation based on a geometric reinterpretation of the binary SVM which has a natural extension resulting in a unified multiclass and multilabel classification framework. Our proposed method is also capable of dealing with the aforementioned class imbalance problem in classification tasks.

Multilabel classification approaches mainly fall into the two categories of problem transformation and algorithm adaptation. An intuitive way to handle multilabel classification tasks is to transform them into a series of binary classification problems and then use—as a method of solution—one of the existing binary classifiers (including binary SVM) to solve them. Algorithm adaptation approaches, on the other hand, solve the problem directly by extending binary classification methods in such a way that all classes are considered at once. Our method falls into the category of algorithm adaptation for multilabel (and multiclass) classification—within the SVM framework. In this paper, we therefore focus mainly on SVM-based approaches to multiclass and multilabel problems and mention other approaches (available in the literature) only as required.

1.2 Outline of the basic idea

Below, we describe the basic idea behind our multilabel SVM formulation. A more formal description is elaborated in Section 3. The main reason for presenting a précis is to quickly get to the (simple) idea behind the formulation.

If the dataset is assumed to be linearly separable, then there are infinitely many hyperplanes which can separate the patterns into two classes. The main idea of the binary SVM is to choose a hyperplane which is optimal in the sense of maximizing the margin between the two sets of patterns while being somewhat robust to outliers. Given the finite set of patterns

(x1,y1),(x2,y2),,(xN,yN)xiK,yi{1,1},(x_{1},y_{1}),(x_{2},y_{2}),\ldots,(x_{N},y_{N})\quad x_{i}\in\mathbb{R}^{K},\;y_{i}\in\{-1,1\},

subset C1(y=1)C_{1}\left(y=1\right) is linearly separable from subset C2(y=1)C_{2}\left(y=-1\right) by hyperplane vTx=cv^{T}x=c if there exist both a unit vector vv and constant c such that for all training instances, yi(vTxi)>cy_{i}(v^{T}x_{i})>c. Vapnik [58] shows that the optimal solution (v)\left(v^{*}\right) to the following optimization problem is a unit normal vector of the hyperplane with maximum margin:

maxv\displaystyle\max_{v} 12(minxiC1vTximaxxjC2vTxj)\displaystyle\quad\frac{1}{2}\left(\min_{x_{i}\in C_{1}}\;v^{T}x_{i}-\max_{x_{j}\in C_{2}}\;v^{T}x_{j}\right) (1)
s.t. v2=1,\displaystyle\quad\left\|v\right\|_{2}=1,

where the assumption is that the patterns of these two classes are linearly separable, i.e., yi(vTxi)>c,i{1,2,,N}y_{i}\left(v^{T}x_{i}\right)>c,i\in\left\{1,2,\ldots,N\right\} for some vector vv and constant cc. The optimal unit vector (v)\left(v^{*}\right) along with constant

c=12(minxiC1xiTv+maxxjC2xjTv),c^{*}=\frac{1}{2}\left(\min_{x_{i}\in C_{1}}\;x_{i}^{T}v^{*}+\max_{x_{j}\in C_{2}}\;x_{j}^{T}v^{*}\right),

determine the maximum margin hyperplane. It is also proved that the maximum margin (optimal) hyperplane is unique [58]. Furthermore, the optimization problem in (1) can be equivalently formulated as finding a pair—vector ww and bias (threshold) bb–such that ww has the smallest 2\ell_{2} norm and constraints yi(wTxi)+b1y_{i}\left(w^{T}x_{i}\right)+b\geq 1 are satisfied for all training patterns [58]:

minw,b\displaystyle\min_{w,b} 12w22\displaystyle\quad\frac{1}{2}\left\|w\right\|_{2}^{2} (2)
s.t. yi(wTxi+b)1,i{1,2,,N}.\displaystyle\quad y_{i}\left(w^{T}x_{i}+b\right)\geq 1,\quad i\in\left\{1,2,\ldots,N\right\}.

Patterns for which the inequality constraints in (2) become active (equality constraints) are of particular importance. These points locate the optimal hyperplane and are known as support vectors and hence the name support vector machine. Cortes and Vapnik [8] extended this problem to the case where the two classes are not linearly separable by relaxing the hard margin constraints in (2) and letting some patterns violate the margin:

minw,{ξi},b\displaystyle\min_{w,\left\{\xi_{i}\right\},b} 12w22+Ciξi\displaystyle\quad\frac{1}{2}\left\|w\right\|_{2}^{2}+C\sum_{i}\xi_{i} (3)
s.t. yi(wTxi+b)1ξi,i{1,2,,N}\displaystyle\quad y_{i}\left(w^{T}x_{i}+b\right)\geq 1-\xi_{i},\quad i\in\left\{1,2,\ldots,N\right\}
ξi0,i{1,2,,N}.\displaystyle\quad\xi_{i}\geq 0,\quad i\in\left\{1,2,\ldots,N\right\}.

The optimization problem in (3) is known as the soft margin SVM. Solving for {ξi}\{\xi_{i}\} in terms of the other variables, we get the following hinge loss in objective function which is an upper bound convex approximation to the zero-one misclassification loss function:

minw,b12w22+Ci[1yi(wTxi+b)]+.\min_{w,b}\quad\frac{1}{2}\left\|w\right\|_{2}^{2}+C\sum_{i}\left[1-y_{i}\left(w^{T}x_{i}+b\right)\right]_{+}.

The essential idea behind our multilabel (and multiclass) formulation was inspired by the optimization problem in (1) which can be reformulated as

maxv\displaystyle\max_{v} 12(minxiC1vT(xix0)maxxjC2vT(xjx0))\displaystyle\quad\frac{1}{2}\left(\min_{x_{i}\in C_{1}}\;v^{T}(x_{i}-x_{0})-\max_{x_{j}\in C_{2}}\;v^{T}(x_{j}-x_{0})\right)
s.t. v2=1\displaystyle\quad\left\|v\right\|_{2}=1

and furthermore, equivalently written as

maxv\displaystyle\max_{v} 12(minxiC1vT(xix0)+minxjC2(v)T(xjx0))\displaystyle\quad\frac{1}{2}\left(\min_{x_{i}\in C_{1}}\;v^{T}(x_{i}-x_{0})+\min_{x_{j}\in C_{2}}\;(-v)^{T}(x_{j}-x_{0})\right) (4)
s.t. v2=1.\displaystyle\quad\left\|v\right\|_{2}=1.

We have introduced an origin x0x_{0} and wish to point out that reconfiguring the training set with respect to this origin does not change the optimal solution to the problem in (1). In fact, Vapnik’s original formulation in (1) is relative. Therefore, adding (vTx0)\left(v^{T}x_{0}\right) to both terms in the optimization problem in (1) does not change the optimal solution. In Vapnik’s equivalent formulation (2), the origin is implicitly introduced via a bias term bb by transitioning from the relative formulation in (1). Furthermore, we are effectively considering a separate hyperplane for each class which results in the two hyperplanes becoming parallel in this formulation with the addition of a sum-to-zero constraint on the normal vectors (v+(v)=O)(v+(-v)=O). As we will see in Section 3, there is no necessity to have a sum-to-zero constraint on the normal vectors (of hyperplanes). When this is relaxed, it yields nonparallel separating hyperplanes in two-class problems. While it may seem odd to have twin, nonparallel hyperplanes in a two-class problem, as we later show, it becomes a useful construct in the multilabel setting.

Having presented the basic idea—nonparallel hyperplanes—we turn to a quick description of related work in order to forestall impressionistic notions of similarity to previous work. The twin SVM (TWSVM) also departs from parallel hyperplanes for the two-class problem [24]. However, it is neither a maximum margin formulation nor an all-in-one machine since it solves two separate optimization problems for binary classification. Mangasarian and Wild [37] proposed a generalized eigenvalue proximal SVM (GEPSVM) which deviates from a previous proximal SVM (PSVM) [15] with parallel hyperplanes. This approach also solves two separate optimization problems for obtaining the separating hyperplane of each class in binary classification. The normal vector of each of the two nonparallel proximal hyperplanes is the eigenvector corresponding to the smallest eigenvalue of a generalized eigenvalue problem.

Our approach is also different from the one-class SVM which is an unsupervised algorithm for outlier (novelty) detection [46, 54]. In the one-class SVM, training data is available for only one class and the goal is to decide whether a new pattern belongs to this class. The one-class SVM introduced in [46] is a natural extension of SVM to the case of unsupervised learning and all the patterns are separated from the origin of the feature space by a hyperplane with maximum margin. Depending on which side of the hyperplane a point falls, it is classified either as a member (of that class) or as an outlier. Whether or not this problem could be formulated for a two-class problem with outliers without the oppositional framework has not been investigated. This approach does not consider the multilabel problem (or even multiclass for that matter). Our model simultaneously addresses multiclass and multilabel problems with the origin implicitly determined by the support vectors of all classes in feature space. In particular, our problem formulation is based on a unique and novel geometry for simultaneously maximizing the margins of all classes in the multiclass and multilabel settings. In order to better demonstrate the advantages of our approach, we mainly focus on large margin formulations in the literature.

The outline of paper is as follows. In Section 2 we review previous work focused on multiclass and multilabel problems mostly within the SVM framework. In Section 3, the new formulation is presented and a quadratic majorizer for the hinge loss is used to solve the optimization problem in the primal. In Section 4, the nonlinear transformation of the input space to infinite dimensional spaces via reproducing kernel Hilbert spaces (RKHS) is discussed and the majorization algorithm extended to solve the primal kernel formulation. In Section 5, we present illustrative experiments and the results of our formulation on multiclass and multilabel datasets. We also discuss a scenario which widely used approaches are not able to handle but is easily handled in our formulation. Section 6 presents the conclusions and speculates on future possibilities and extensions.

2 Related Work

In our discussion of previous work, we mainly confine ourselves to SVM-based approaches, specifically all-in-one (single) machines for multiclass and multilabel problems. All-in-one approaches consider information of all classes simultaneously and set up a single optimization problem for learning all classifiers (discriminators) at once. The margin concept is key in binary SVM and its extensions. SVM approaches differ mainly in how they define the margin and how margin violations are formulated in the loss function. There are two notions of margin, namely, relative margin and absolute margin which basically correspond respectively to cases where the margins of the classifiers are optimized relative to each other or otherwise [12].

Vapnik [58] proposed the first all-in-one framework for the multiclass SVM based on relative margins. [60] and [5] also used this framework albeit with slightly different formulations. It can be shown that all three formulations are equivalent. In this framework, the relative margin of each pattern is maximized against all other classes to which it does not belong with a slack variable utilized for each of these pairs. Crammer and Singer [9] proposed a similar formulation for multiclass SVM with a single slack variable per pattern, essentially a penalty for the worst class. Their formulation did not take the biases into account. Hsu and Lin [23] subsequently added biases into the Crammer and Singer formulation.

Lee et al. [32] proposed another formulation for multiclass SVM with absolute margins which (while extending the Bayes decision rule for multiclass in the same fashion as for the binary case) ensures that the solution directly targets the Bayes decision rule also known as Fisher consistency. However, setting up the framework for compatibility with the Bayes decision rule eliminates the notion of support vectors for each class since the hinge loss term corresponding to the correct class gets eliminated. In other words, this approach penalizes a pattern if it gets within 1K1\frac{1}{K-1} margin distance to other classifiers (where KK is the number of classes) without encouraging a pattern to get close to its own classifier. The work proposed by [32], and [60] are theoretically sound but an efficient training algorithm has not yet been proposed [12]. Van Den Burg and Groenen [57] introduced a single machine—the generalized multiclass SVM (GenSVM)—which uses a simplex encoding to cast the multiclass problem in K1K-1 dimensions. GenSVM generalizes aforementioned methods of [32] and [9]. Liu and Yuan [35] proposed a new type of multiclass hinge loss function called reinforced hinge loss and proved that under some conditions, it is Fisher consistent. Szedmak et al. [51] introduced the multiclass maximum margin regression (MMR) framework with computational complexity no more than a binary SVM. They extend the idea of support vector regression [58] to vector label learning in an arbitrary Hilbert space. Dogan et al. [12] showed that most of the proposed all-in-one multiclass SVMs are mainly different in terms of choice of margin and margin-based loss functions and presented a unified formulation which accommodates existing single machine approaches for multiclass classification. Based on this unified view, they proposed two new multiclass formulations.

Zhang and Jordan [62] proposed a Bayesian multiclass SVM by extending the method proposed in [47]. [15] proposed multicategory proximal support vector machines (MPSVM) where instead of working in the feature space n\mathbb{R}^{n} and maximizing the margin, they concatenate the bias to the weight vector and try to maximize this extended margin in n+1\mathbb{R}^{n+1} [14]. MPSVM is a problem transformation scheme which solves KK independent nonsingular systems of linear equations to obtain the weight vectors wk,k{1,,K}w_{k},k\in\left\{1,\ldots,K\right\}. It is not clear why merging the bias with the weight vector and maximizing this extended margin should improve classification performance.

The one versus the rest (OvR) a.k.a. one versus all (OvA) approach is based on the idea of separately training binary classifiers for each class. The patterns belonging to one class form a group and all other patterns form the second group. Therefore, in the OvR approach for KK-class problems, KK binary classifiers are trained [58]. For a test pattern, each classifier outputs a real value which reflects the level of certainty of the pattern belonging to its corresponding class. Naturally, the test pattern is assigned to the class with maximum discriminant function value or confidence score (winner-take-all decision criterion). OvR does not take label correlations into account. OvR is considered to be an intuitive and simple extension to binary SVM and yet powerful in producing results as good as other classifiers [42] and can be extended to the multilabel problem by using a probabilistic scheme for obtaining confidence levels and considering a threshold for scores for which each value or probability gives rise to a label. However, as we will show in Section 5, there are cases in which this approach fails. Since a smaller number of patterns are used to train the classifier, OvR is susceptible to variance increase since partitioning data this way makes the training set unbalanced [32]. In the OvR approach, it is not clear how support vectors should be determined and their interpretation is ambiguous especially in the multilabel setting since support vectors are relative.

In the one versus one (OvO) approach a.k.a. all versus all, a binary classifier is trained for each pair of classes [22, 29]. Therefore for a KK-class problem, K(K1)/2K(K-1)/2 classifiers are trained. OvO suffers from the transitivity problem which may result in label contradiction in some areas of the input space. A test pattern is assigned to the class with maximum number of votes. When the number of classes is large, these binary-classifier-based methods may suffer from either computational cost or highly imbalanced sample sizes in training. DAGSVM was proposed to combine the result of classifiers obtained via the OvO strategy [41]. A new test pattern is classified using K1K-1 classifiers which are chosen based on which path is traversed in the graph.

Dietterich and Bakiri [11] proposed a general approach for multiclass classification based on error-correcting output codes (ECOC) for binary classifiers. A binary string of some fixed length ss known as the codeword is assigned to each class and then ss binary functions are learned. At the decoding step, a code is obtained for a test pattern in the test set by evaluating all trained binary classifiers for that pattern. The test pattern is assigned to the class with minimum Hamming distance (which counts the number of bits that differ) between the binary output code of the test pattern and base code of that class. ECOC is applied to margin-based methods including SVM where each codeword belongs to {1,0,1}s\{-1,0,1\}^{s} with 0 indicating that the way in which the corresponding classifier assigns this label is irrelevant [1]. Label prediction is based on a loss-based decoding scheme. Loss-based decoding takes the output magnitude of binary classifiers which are considered as certainty level into account and choose a label which is most consistent with the classifiers’ outputs based on the loss function. Most of the aforementioned approaches do not accommodate and cannot be extended to multilabel classification in a straightforward manner.

Approaches to multilabel learning mainly fall into the categories of algorithm adaptation and problem transformation. In general, there is no established approach for solving a multilabel classification problem [52]. OvR a.k.a binary relevance (BR) in multilabel learning is a baseline approach and usually serves as a benchmark for other multilabel approaches [36]. Boutell et al. [4] used the OvR approach with SVM as the binary classifier on the scene dataset. They use cross training which basically includes each pattern as a positive example (label 1) for each class when training the corresponding classifier. The multilabel patterns are not used as negative examples while training the classifiers. In our framework, on the other hand, no patterns are used as negative examples for training any classifier since each classifier only uses information related to its own patterns to determine the hyperplane. They have also offered an empirical (heuristic) decision criterion for testing and an evaluation metric for performance evaluation and this is mentioned in Section 5. Two approaches are proposed in [17] to improve the margin in multilabel classification within the OvR framework. The first approach eliminates patterns from the opposite class (the rest). The second approach removes all the training data in confusing classes using a confusion matrix obtained from applying a fast and relatively accurate classifier. Chen et al. [7] extended TWSVM to the multilabel setting and proposed MLTSVM in which KK classifiers are trained separately in OvR fashion.

Elisseeff and Weston [13] introduced a large margin ranking system known as Rank-SVM for multilabel classification. Rank-SVM is an algorithm adaptation method which can be considered as an extension of the Crammer and Singer multiclass SVM framework. The idea of Rank-SVM is to find KK classifiers (hyperplanes), one for each label, such that the relevant labels for each training pattern are ranked higher than irrelevant labels. To achieve this, a proposed ranking loss is minimized and relative margins are maximized. Since it is a ranking-based approach, the design of a set size predictor is needed. Formulating multilabel classification as a ranking problem imposes a quadratic number of constraints on the problem and makes optimization more challenging [21]. Rank-SVM, as its name indicates, ranks the labels for each pattern and then chooses a number of highly ranked labels for each pattern based on a probabilistic or heuristic approach for determining the cardinality of the set of labels. Therefore, Rank-SVM is not able to directly produce output labels.

Armano et al. [2] extended ECOC to multilabel text categorization. They generate a codeword for each class with only 11 and 0 bits. Then the posterior probability of each class is obtained and tt top ranking categories are selected where tt is user defined or tuned by the validation set.

Incorporating label space structure and capturing label correlations within the framework of large margin methods has been studied in multilabel learning [56, 34, 21, 49, 30, 20, 53]. There are cases in which elements of the output space are structured objects such as trees, sequences, strings or sets. The challenge here is how to effectively use label correlation information to improve accuracy in multilabel prediction. The structured SVM (SSVM) was proposed to learn a mapping from the input space to structured output spaces like a parse tree in Natural Language Parsing [56, 26].

Hariharan et al. [21] developed a maximum margin multilabel formulation (M3L) which assumes labels are densely correlated but did not include pairwise label correlation terms in the objective function. Their work is the middle ground between approaches with independent labels assumption (linear complexity) and explicitly modeling label correlations (exponential complexity). They make the assumption that prior knowledge about label correlations is available and that labels have at most linear dense correlations. This prior knowledge is incorporated into the formulation in the form of a dense correlation matrix and KK densely correlated sub-problems are solved. They show that OvR is obtained as a special case of their model.

Canonical correlation analysis (CCA) is used for dimensionality reduction while exploiting label correlations in the multilabel setting [49, 50]. In supervised dimensionality reduction via CCA, the two sets of variables are derived from the data and class labels followed by projection into a lower-dimensional space in which the patterns are maximally correlated. They incorporated a dimensionality reduction scheme into the SVM formulation and proposed a joint dimensionality reduction and multilabel classification framework. Their framework also uses opposition from other classes to learn classifiers. Liu et al. [34], modified SVM to take missing labels into consideration by extending the label set to include zero which indicates a missing class label. They also take label correlations into consideration by proposing label and class smoothness and integrating these into the objective function.

These approaches still use patterns from other classes in an oppositional fashion to obtain each classifier. Our approach is a natural extension to an all-in-one multiclass SVM [60, 9] and does not need to incorporate every possible label set prediction into model. Labels get assigned only after training—similar to multiclass classification in SVM. Furthermore, interpretation of labels is very important in multilabel classification problems since in most of the work, not having a label for a class is automatically interpreted as negative membership. Our framework takes this into account and uses available information on memberships for classification and deviates from all oppositional classification frameworks. However, it is easily extensible to negative labels by assigning a label of 1-1 to patterns for which information about negative membership is available. Our approach resides between approaches with the independent labels assumption and the ones which consider direct pairwise label correlations. From this perspective, we do not have the exponential label space complexity issue since separation in our framework is against the origin. As a matter of fact, our formulation does not explicitly explore label correlations, however, as we show, this information is implicitly taken into account. Finally, a dimensionality reduction scheme can also be easily incorporated into our formulation.

3 A Unified SVM Classification Framework

The underlying motivation for the present work originates from the desire to naturally extend the SVM formulation to a multilabel classification framework. The lack of such a unified multiclass and multilabel framework in the literature led us to revisit the original SVM formulation and suggest a new, geometrically driven approach. The basic idea is as follows. We (implicitly) learn a new origin for the feature vectors as well as the magnitude and direction of the weight vectors ({wk})(\left\{w_{k}\right\}) such that the inner product of the reconfigured patterns with respect to this new origin (xix0)(x_{i}-x_{0}) with each weight vector wkw_{k} represents the confidence (certainty) of the memberships of the patterns belonging to each class. Consequently, each class CkC_{k} chooses a magnitude and orientation of its weight vector wkw_{k} in such a way that the patterns belonging to CkC_{k} are best represented (separated) without the need of explicit opposition to other classes. Inspired by this characteristic, our approach is called one versus none (OvN). In this section, we flesh out this idea, formulate a new objective function and present an algorithm for finding the global optimum.

3.1 Problem Formulation

In Section 1, we motivated the unified framework by jettisoning the parallel hyperplane constraint which is standard in almost all binary (two-class) SVMs. Furthermore, we introduced an origin variable x0x_{0} and set up the margin constraints using this new variable. The picture that emerges from this description is of individual class-specific weight vectors ({wk})(\left\{w_{k}\right\}) which seek to satisfy margin constraints for the patterns belonging to that class. This is repeated for all classes. However, this alone does not ensure competition between classes which is enshrined using parallel hyperplanes in the standard SVM. In our formulation, we introduce competition by minimizing the inner product between all pairs of weight vectors. Note that the minimum inner product between two unit weight vectors wkw_{k} and wlw_{l} is attained when the two vectors are anti-parallel (as in the standard SVM). Later, in Section 3.5, we show that suitable constraints on a pair of weight vectors w1w_{1} and w2w_{2} in a two-class problem, make our formulation completely equivalent to the standard binary SVM.

In common with virtually all previous soft margin approaches, the objective function includes regularization and outlier rejection terms [8]. As mentioned above, in addition to the margin term for each weight vector wkw_{k}, we have the sum of all pairwise inner products between weight vectors wkw_{k} and wlw_{l} which endeavor to make the weight vectors anti-parallel. Thus far, we have not discussed the bias term in this set up. When we introduced the origin x0x_{0} as a variable in (4), we ended up with two terms that canceled each other out: vTx0v^{T}x_{0} and vTx0-v^{T}x_{0} which assumes anti-parallel hyperplanes. In a two-class multilabel SVM, these become b1w1Tx0b_{1}\equiv-w_{1}^{T}x_{0} and b2w2Tx0b_{2}\equiv-w_{2}^{T}x_{0}. We then have the option of enforcing a hard constraint b1+b2=0b_{1}+b_{2}=0 or a soft constraint term (b1+b2)2\left(b_{1}+b_{2}\right)^{2} to be mimimized as part of objective function. For problem formulation with more than two classes, the former option is selected in this work since it was practically demonstrating better performance in both linear and kernel cases and is fully demonstrated in appendices. The other case can be similarly worked out. From a conceptual perspective, a hard constraint on the biases is an extension of the anti-parallel hyperplanes constraint, whereas the soft constraint on the biases is a relaxation of the same. A candidate multilabel (and multiclass) optimization problem is

min{wk},{ξki},x0\displaystyle\min_{\left\{w_{k}\right\},\left\{\xi_{ki}\right\},x_{0}} 12k=1Kwk22+αk=1Kl=k+1KwkTwl+βk=1KiCkξki\displaystyle\quad\frac{1}{2}\sum_{k=1}^{K}\parallel w_{k}\parallel_{2}^{2}+\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}w_{k}^{T}w_{l}+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\xi_{ki} (5)
s.t. k=1K(wkTx0)=0\displaystyle\quad\sum_{k=1}^{K}\left(-w_{k}^{T}x_{0}\right)=0
wkT(xkix0)1ξki,k,iCk\displaystyle\quad w_{k}^{T}(\,x_{{}_{ki}}-x_{0})\geq 1-\xi_{ki},\quad\forall k,\quad\forall i\in C_{k}
ξki0,\displaystyle\quad\xi_{ki}\geq 0,

where the first constraint is the hard constraint on biases and the second set of constraints are margin violation constraints obtained by introducing outlier (slack) variables [8]. The first term in the objective function relates to margin maximization, the second term implicitly models pairwise label correlations and the last term penalizes deviation of patterns from their corresponding class margin. The regularization parameters, α\alpha,β>0\;\beta>0, have standard objective function semantics. We reformulate the optimization problem in (5) by eliminating all the outlier variables via hinge losses and introducing the bias variables bkwkTx0b_{k}\equiv-w_{k}^{T}x_{0}:

minW,b\displaystyle\min_{W,b} 12k=1KwkTwk+αk=1Kl=k+1KwkTwl+βk=1KiCk[1(wkTxki+bk)]+\displaystyle\quad\frac{1}{2}\sum_{k=1}^{K}w_{k}^{T}w_{k}+\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}w_{k}^{T}w_{l}+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\left[1-\left(w_{k}^{T}x_{ki}+b_{k}\right)\right]_{+} (6)
s.t. k=1Kbk=0,\displaystyle\quad\sum_{k=1}^{K}b_{k}=0,

where we define W={w1,w2,,wK}W=\left\{w_{1},w_{2},...,w_{K}\right\} as the set comprising all the normal vectors to the hyperplanes and b={b1,b2,,bK}b=\left\{b_{1},b_{2},...,b_{K}\right\} as the set of biases. The hard constraint on biases is incorporated into the objective function of (6) by the method of Lagrange multipliers and results in the following Lagrangian (where λ\lambda is a Lagrange multiplier):

(W,b,λ)=12k=1KwkTwk+αk=1Kl=k+1KwkTwl+βk=1KiCk[1(wkTxki+bk)]++λk=1Kbk.\mathcal{L}\left(W,b,\lambda\right)=\frac{1}{2}\sum_{k=1}^{K}w_{k}^{T}w_{k}+\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}w_{k}^{T}w_{l}+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\left[1-\left(w_{k}^{T}x_{ki}+b_{k}\right)\right]_{+}+\lambda\sum_{k=1}^{K}b_{k}. (7)

This is by no means the only available integrated formulation. In Section 3.4, we examine all four possibilities: the cross product space of soft and hard constraints on the hyperplanes and soft and hard constraints on the biases. We also have considered the margins as in the conventional SVM. However, it is possible to incorporate margin-related variables in the formulation by setting up the problem similar to the ν\nu-SVC [45].

3.2 Testing Scenario

After solving the optimization problem and obtaining the classifiers, a test instance is projected to each weight vector. We employ a simple decision criterion wherein each test instance gets assigned all class labels with projection greater than 11, i.e., if StS_{t} denotes the set of labels for test instance xtx_{t}, then

St={k:wkTxt+bk1}.{S_{t}=\{k:w_{k}^{T}x_{t}+b_{k}\geq 1\}}.

From a geometric point of view, we have an ambiguous region which is inhabited by patterns whose projections are all less than 1. Therefore the decisions on label assignment could include this information. In the present work however, we eschew careful examination of pattern and label test set geometry and relegate this to future work. Instead, when this situation occurs, we revert to a multiclass strategy and pick the label corresponding to the maximum projection as explained below.

In the multiclass (single label) setting, a winner-take-all strategy (as mentioned above) is applied and the test pattern is assigned to the class with maximum projection value

m=arg max𝑘{wkTxt+bk}.m=\underset{k}{\textup{arg\,max}}\>\{w_{k}^{T}x_{t}+b_{k}\}.

3.3 A Quadratic Majorizer for Optimization

Most of the existing literature on SVM optimization is focused on formulating the primal problem from which the dual formulation is extracted followed by a solution in dual space. However, primal problems can also be efficiently solved. In fact, solving the primal may have advantages for large scale optimization since—at least in the linear SVM—the objective function is linear in the training set patterns (and not quadratic). Furthermore, often in large scale optimization, an approximate solution is sought which results in a large number of support vectors. In this case, the dual may result in a solution which is not meaningful [6].

One of the most simple and effective algorithms for solving the primal is majorization-minimization (MM). The main advantage of iterative majorization for the SVM is that it results in simple linear system updates (achieved without the need for line-search parameter tuning). A secondary advantage of MM is that the objective function value is guaranteed to be non-increasing and is generally decreasing in each iteration. If the function to be optimized is strictly convex which is the case with the SVM loss function, MM converges to the global minimum. In contrast, other methods which solve the dual must often completely converge in order to obtain a meaningful and reasonable solution. Variants of MM have also been proposed which deal with large-scale optimization problems.

Consider the standard, linear, binary SVM in the primal. It comprises a margin term and the hinge loss on the patterns. An alternative to majorization is to smooth the hinge loss (resulting in a differentiable approximation like the Huber loss). However, in this case, we are minimizing a different function which is an approximation to the original function. Furthermore, the smoothed hinge loss does not result in simple update equations (achieved without having to estimate a step-size parameter). Even if we approximate the hinge loss with the generalized logistic loss gβ(x)=1βlog(1+exp{β(1x)})g_{\beta}(x)=\frac{1}{\beta}\text{log}(1+\text{exp}\{\beta(1-x)\}) (which approaches [1x]+[1-x]_{+} as β\beta\rightarrow\infty), we would still need to use nonlinear optimization on this logistic regression variant in order to obtain good solutions. Finally, the non-differentiability of the hinge loss forces us to adopt subgradient optimization which is also complex (from a line search perspective). For these reasons, we introduce a simple MM approach to minimize the SVM objective in our unified framework. Clearly, alternatives are available (including minimization in the dual) but we have relegated these to future work.

Majorization leads to a particular kind of iterative optimization algorithm. The basic idea behind majorization is to construct an auxiliary objective function which lies above the original objective function and is easier to minimize. The auxiliary objective function usually has the property of touching the original objective at the current time (iteration) instant. By descending on the auxiliary objective, we guarantee that we also descend on the original objective by virtue of the fact that the auxiliary objective is strictly greater than or equal to the original [31]. For the task at hand, note that MM can specifically deal with nondifferentiable convex objective functions (like the hinge loss) by constructing a sequence of smooth convex functions which majorize the original nonsmooth objective function. A brief formal description of MM follows. A function g(x,xk)g(x,x_{k}) majorizes f(x)f\left(x\right) at xkx_{k} if

f(xk)\displaystyle f\left(x_{k}\right) =g(xk,,xk),and\displaystyle=g\left(x_{k,},x_{k}\right),~{}\mathrm{and} (8)
f(x)\displaystyle f(x) g(x,xk),xxk,\displaystyle\leq g\left(x,x_{k}\right),\quad x\neq x_{k},

which means that g(x,xk)g(x,x_{k}) lies above f(x)f\left(x\right) everywhere and touches f(x)f(x) at least at xkx_{k} where xkx_{k} is a point in the domain of f(x)f(x) at the current iteration. Now we minimize the surrogate function g(x,xk)g\left(x,x_{k}\right) instead of f(x)f\left(x\right) in the minimization version of the MM algorithm. If xk+1x_{k+1} is the minimum of g(x,xm)g\left(x,x_{m}\right), then the descent property follows from (8). From the descent property, we see that it is possible to merely descend on gg instead of finding its global minimum:

f(xk+1)g(xk+1,xk)g(xk,xk)=f(xk).f\left(x_{k+1}\right)\leq g\left(x_{k+1},x_{k}\right)\leq g\left(x_{k},x_{k}\right)=f\left(x_{k}\right).

There exist previous approaches which solve the SVM in the primal using MM [18, 19, 39, 57]. For example, [18] have a majorization approach which considers different loss functions (hinge, quadratic, Huber). We follow [10] in our approach. [18] and [10] show that the hinge loss [1u]+:=min(0,1u)[1-u]_{+}:=\textup{$\min$}(0,1-u) is majorized by the following quadratic (in uu):

g(u,z)=14z(1u+z)2g(u,z)=\frac{1}{4z}(1-u+z)^{2}

where z>0z>0. It is shown that gg is the best quadratic majorizer for [1u]+\left[1-u\right]_{+} [10]. Theoretically, it is possible to get arbitrarily as close as possible to the hinge loss through the above quadratic majorizer. In practice however, we choose a very small positive threshold value ϵ\epsilon for zz, i.e., zϵz\geq\epsilon for the sake of numerical stability. Here zz is an auxiliary variable. Assuming (for the sake of convenience) that zz is unconstrained (instead of requiring a positivity constraint), the derivative of g(u,z)g(u,z) w.r.t. zz is

gz=14(1u)24z2.\frac{\partial g}{\partial z}=\frac{1}{4}-\frac{(1-u)^{2}}{4z^{2}}.

Setting this to zero yields z2=(1u)2z^{2}=(1-u)^{2} from which we obtain z=|1u|z=\left|1-u\right| which is non-negative and is equal to zero when u=1u=1. It can easily be shown that setting z=ϵz=\epsilon when |1u|<ϵ|1-u|<\epsilon is the global minimum for zz (when zz is constrained to be greater than or equal to ϵ\epsilon). We can therefore use the above majorization trick to obtain a quadratic objective w.r.t. uu. Note that g(u,z)g(u,z) will be replaced by a suitable function of wkw_{k} and bkb_{k} in the formulation. The majorized multiclass/multilabel SVM optimization problem can be written as

minW,b,Z\displaystyle\min_{W,b,Z} 12k=1KwkTwk+αk=1Kl=k+1KwkTwl+βk=1KiCk14zki[1+zki(wkTxki+bk)]2hinge loss majorizer\displaystyle\quad\frac{1}{2}\sum_{k=1}^{K}w_{k}^{T}w_{k}+\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}w_{k}^{T}w_{l}+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\overset{\textup{hinge\;loss\;majorizer}}{\overbrace{\frac{1}{4z_{ki}}\left[1+z_{ki}-(w_{k}^{T}x_{ki}+b_{k})\right]^{2}}} (9)
s.t. k=1Kbk=0,\displaystyle\quad\sum_{k=1}^{K}b_{k}=0,

where ZZ denotes the set of all strictly positive auxiliary variables {zki}\left\{z_{ki}\right\}. In order to simplify the notation in the optimization problem in (9), we concatenate wkw_{k} and bkb_{k} into one vector and define the following matrices:

w~k[wkbk],x~ki[xki1],D[IM×M00T0],e[001](M+1)\widetilde{w}_{k}\equiv\left[\begin{array}[]{c}w_{k}\\ b_{k}\end{array}\right],\;\widetilde{x}_{ki}\equiv\left[\begin{array}[]{c}x_{ki}\\ 1\end{array}\right],\;D\equiv\left[\begin{array}[]{cc}I_{M\times M}&\overrightarrow{0}\\ \overrightarrow{0}^{T}&0\end{array}\right],\quad e\equiv\left[\begin{array}[]{c}0\\ \vdots\\ 0\\ 1\end{array}\right]_{(M+1)} (10)

where MM is the number of features. With these changes and defining the set of augmented weight vectors W~={w~1,,w~K}\widetilde{W}=\left\{\widetilde{w}_{1},\ldots,\widetilde{w}_{K}\right\}, we may rewrite the optimization problem as

minW~,Z\displaystyle\min_{\widetilde{W},Z} k=1Kw~kTDw~k+αk=1Kl=k+1Kw~kTDw~l+βk=1KiCk14zki[1+zki(w~kTx~ki)]2\displaystyle\quad\sum_{k=1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{k}+\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{l}+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\widetilde{w}_{k}^{T}\widetilde{x}_{ki})\right]^{2} (11)
s.t. k=1KeTw~k=0,\displaystyle\quad\sum_{k=1}^{K}e^{T}\widetilde{w}_{k}=0,

and again by including the hard constraint on biases into the objective function via the method of Lagrange multipliers, the Lagrangian can be written as

(W~,Z,λ)=\displaystyle\mathcal{L}(\widetilde{W},Z,\lambda)= k=1Kw~kTDw~k+αk=1Kl=k+1Kw~kTDw~l\displaystyle\sum_{k=1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{k}+\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{l} (12)
+βk=1KiCk14zki[1+zki(w~kTx~ki)]2+λk=1KeTw~k.\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\widetilde{w}_{k}^{T}\widetilde{x}_{ki})\right]^{2}+\lambda\sum_{k=1}^{K}e^{T}\widetilde{w}_{k}.

Rather than solving for W~\widetilde{W}, we concatenate all the augmented weight vectors w~k\widetilde{w}_{k} into one vector w~=[w~1T,w~2T,,w~KT]T\widetilde{w}=\left[\widetilde{w}_{1}^{T},\widetilde{w}_{2}^{T},\ldots,\widetilde{w}_{K}^{T}\right]^{T} and alternate between least-squares solutions for w~\widetilde{w} and closed-form solutions for ZZ. The solution for ZZ follows the approach outlined above:

zki={zki|1w~kTx~ki|ϵϵ|1w~kTx~ki|<ϵ,k,iCk.z_{ki}=\begin{cases}z_{ki}&\left|1-\widetilde{w}_{k}^{T}\widetilde{x}_{ki}\right|\geq\epsilon\\ \epsilon&\left|1-\widetilde{w}_{k}^{T}\widetilde{x}_{ki}\right|<\epsilon\end{cases},\quad\forall k,\ \forall i\in C_{k}. (13)

The solution for the concatenated weight vector w~\widetilde{w} merely requires some turgid algebra and is relegated to Appendix A. The pseudo code of the algorithm is presented in Algorithm 1. We briefly mention here that the regularization coefficients for soft constraints on weight vectors (α)\left(\alpha\right) and hinge loss terms (β)\left(\beta\right) should be chosen in such a way that the Hessian obtained after concatenating the weight vectors is positive definite (as shown in Appendix A).

1:Input: x~ki,D~\widetilde{x}_{ki},\widetilde{D} (diagonal matrix with DD), D~0\widetilde{D}_{0} (off-diagonal matrix with DD)
2:Initialize zki=1k,iCkz_{ki}=1\quad\forall\,k,\ \forall\,i\in C_{k}
3:while not converged do
4:     Compute X~\widetilde{X} as a diagonal matrix with entries iCk14zkix~kix~kiT\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\widetilde{x}_{ki}\widetilde{x}_{ki}^{T}
5:     Solve w~=β2(D~+α2D~0+βX~)1x~\widetilde{w}=\frac{\beta}{2}\left(\widetilde{D}+\frac{\alpha}{2}\widetilde{D}_{0}+\beta\widetilde{X}\right)^{-1}\widetilde{x} (KKT conditions) where w~=[w~1T,w~2T,,w~KT]T\widetilde{w}=\left[\widetilde{w}_{1}^{T},\widetilde{w}_{2}^{T},\ldots,\widetilde{w}_{K}^{T}\right]^{T}
6:     Compute zki=|1w~kTx~ki|k,iCkz_{ki}=\left|1-\widetilde{w}_{k}^{T}\widetilde{x}_{ki}\right|\quad\forall\,k,\ \forall\,i\in C_{k}
7:Output: W~={w~1,w~2,,w~K}\widetilde{W}=\left\{\widetilde{w}_{1},\widetilde{w}_{2},\ldots,\widetilde{w}_{K}\right\}
Algorithm 1 Quadratic majorizer for multiclass/multilabel SVM with soft constraint on weight vectors WW and hard constraint on biases bb

3.4 Incorporating soft and hard constraints into the formulation

The formulation in the previous section imposed a soft constraint on the hyperplane normal vectors, wherein the inner product of every pair of vectors was minimized. Furthermore, a hard constraint on the biases bkb_{k} was used to make their sum equal to zero. However, depending on the problem and dataset, we could also consider hard constraints on both the normal vectors (k=1Kwk=0)\left(\sum_{k=1}^{K}w_{k}=0\right) and the biases (k=1Kbk=0)\left(\sum_{k=1}^{K}b_{k}=0\right). Table 1 provides the four different possibilities for incorporating the constraints into the formulation. For the sake of simplicity, we omit the margin terms and hinge loss terms which are common to all cases. In all these models, hyperparameters should be chosen in such a way that the Hessian of the objective function remains positive definite, ensuring that the optimization problem remains bounded from below. The details of the formulation specific to the Hessian for the case of soft ww-hard bb are relegated to Appendix A.

Constraints Appearance in objective (Lagrangian) function
soft ww-soft bb αk=1Kl=k+1KwkTwl+γ(k=1Kbk)2\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}w_{k}^{T}w_{l}+\;\gamma\left(\sum_{k=1}^{K}b_{k}\right)^{2}
soft ww-hard bb αk=1Kl=k+1KwkTwl+λk=1Kbk\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}w_{k}^{T}w_{l}+\;\lambda\sum_{k=1}^{K}b_{k}
hard ww-soft bb λk=1Kwk+γ(k=1Kbk)2\lambda\sum_{k=1}^{K}w_{k}\;+\gamma\left(\sum_{k=1}^{K}b_{k}\right)^{2}
hard ww-hard bb λk=1Kwk+ηk=1Kbk\lambda\sum_{k=1}^{K}w_{k}\;+\eta\sum_{k=1}^{K}b_{k}
Table 1: Terms in the objective function corresponding to all possible combinations of soft and hard constraints. The parameters λ\lambda and η\eta denote the Lagrange multipliers corresponding to the hard constraints.

3.5 Equivalence to the binary SVM

The original two-class soft margin SVM problem is equivalent to the minimization of the following objective function:

E(w,b)=\displaystyle E(w,b)= wTw+Ci=1N[1yi(wTxi+b)]+\displaystyle w^{T}w+C\sum_{i=1}^{N}[1-y_{i}(w^{T}x_{i}+b)]_{+} (14)
=\displaystyle= wTw+C[iC1[1(wTxi+b)]++jC2[1(1)(wTxj+b)]+]\displaystyle w^{T}w+C\left[\sum_{i\in C_{1}}[1-(w^{T}x_{i}+b)]_{+}+\sum_{j\in C_{2}}[1-(-1)(w^{T}x_{j}+b)]_{+}\right]
=\displaystyle= wTw+C[iC1[1(wTxi+b)]++jC2[1((w)Txj+(b)]+].\displaystyle w^{T}w+C\left[\sum_{i\in C_{1}}[1-(w^{T}x_{i}+b)]_{+}+\sum_{j\in C_{2}}[1-((-w)^{T}x_{j}+(-b)]_{+}\right].

Note that we have used wTww^{T}w instead of 12wTw\frac{1}{2}w^{T}w since the equivalence will be shown using two hyperplanes. From (14), we see that the existence of two hyperplanes with equal norms and anti-parallel normal vectors is implicit within the standard soft margin binary SVM formulation. The original binary SVM is equivalent to our formulation by imposing hard constraints on ww and bb:

minw1,w2,b1,b2\displaystyle\min_{w_{1},w_{2},b_{1},b_{2}} 12w1Tw1+12w2Tw2+C[iC1[1(w1Txi+b1)]++jC2[1(w2Txj+b2)]+]\displaystyle\begin{aligned} \quad&\frac{1}{2}w_{1}^{T}w_{1}+\frac{1}{2}w_{2}^{T}w_{2}+C\left[\sum_{i\in C_{1}}[1-(w_{1}^{T}x_{i}+b_{1})]_{+}+\sum_{j\in C_{2}}[1-(w_{2}^{T}x_{j}+b_{2})]_{+}\right]\end{aligned}
s.t. w1+w2=0,and\displaystyle\quad w_{1}+w_{2}=0,~{}\mathrm{and}
b1+b2=0.\displaystyle\quad b_{1}+b_{2}=0.

Incorporating these constraints into the objective function, we get

E(w1,b1)=w1Tw1+C[iC1[1(w1Txi+b1)]++jC2[1+(w1Txj+b1)]+]E(w_{1},b_{1})=w_{1}^{T}w_{1}+C\left[\sum_{i\in C_{1}}[1-(w_{1}^{T}x_{i}+b_{1})]_{+}+\sum_{j\in C_{2}}[1+(w_{1}^{T}x_{j}+b_{1})]_{+}\right]

which is identical to (14). We have shown that when hard constraints are imposed on two weight vectors (making them anti-parallel but with equal norm) and on the two biases (making them sum to zero), we recover the original soft margin SVM. Furthermore, our soft ww-hard bb model (shown in Table 1)—while providing a degree of flexibility to the weight vectors’ orientations and norms—is still capable of producing parallel hyperplanes (anti-parallel weight vectors) when the optimal solution dictates this to be the case. Consider the following example which is designed in such a way that the optimal solution must be two parallel hyperplanes. As shown in Figure 1, for any α[1,2]\alpha\in\left[1,2\right], we get parallel hyperplanes and the solution coincides with the binary SVM.

Refer to caption
(a) α=1\alpha=1
Refer to caption
(b) α=2\alpha=2
Figure 1: Soft ww-hard bb model results in parallel hyperplanes as the optimal solution. The angle between the normal vectors of the hyperplanes is 180180 degrees.

4 Generalization to RKHS

In Section 3, we presented the formulation and algorithm assuming linear separability between classes. However, linear separability is not a realistic assumption in many problems. Therefore, a more complex approach is required to extend these models to problems with nonlinear separability. In the following subsections, we give an introduction to kernels and demonstrate how the origin is formulated in this setting. We then develop the kernel formulation for unified multiclass and multilabel SVMs.

4.1 Kernels

Kernels were introduced as a tool to map a nonlinearly separable dataset into an implicit higher (or infinite) dimensional reproducing kernel Hilbert space (RKHS) \mathcal{H} where linear separability can be achieved without the need to explicitly compute the features in the transformed space [59]. In order to adapt our formulation to an RKHS, two problems have to be dealt with before we can proceed. As we will show, the origin of the RKHS needs to be properly treated and correctly interpreted in an infinite dimensional Hilbert space. Furthermore, the inclusion of a hard constraint on the weight vectors is not as straightforward as in the linearly separable case. We first briefly summarize how RKHS kernels are used in the SVM context.

Deploying a kernel function κ:X×X\kappa:X\times X\rightarrow\mathbb{R}, the inner product of feature mappings in a high dimensional RKHS is performed by function evaluation in the original space. If ϕ:X\phi:X\rightarrow\mathbb{\mathcal{H}} is a mapping from the original space to a feature space, i.e., Xxϕ(x):=κ(,x)X\ni x\mapsto\phi(x):=\kappa(\cdot,x)\in\mathbb{\mathcal{H}}, then

ϕ(xi),ϕ(xj)=κ(xi,xj).\left\langle\phi(x_{i}),\phi(x_{j})\right\rangle=\kappa(x_{i},x_{j}).

By the representer theorem [28], each weight vector wkw_{k} can be written as a linear combination of all projected patterns ϕ(xi):=κ(,xi)\phi(x_{i}):=\kappa(\cdot,x_{i}) in RKHS,

wk=i=1Nαkiϕ(xi),k{1,,K}.w_{k}=\sum_{i=1}^{N}\alpha_{ki}\phi(x_{i}),\quad k\in\{1,\ldots,K\}. (15)

If we define SWS_{W} as the subspace spanned by W={w1,w2,,wK}W=\left\{w_{1},w_{2},\cdots,w_{K}\right\}, i.e.,

SW=span{ϕ(x1),ϕ(x2),,ϕ(xN)},S_{W}=\text{span}\left\{\phi(x_{1})\mathbin{,}\phi(x_{2})\mathbin{,}\cdots\mathbin{,}\allowbreak\phi(x_{N})\right\},

we can write x0x_{0} (the origin in the infinite dimensional feature space) as a summation of elements in SWS_{W} and an element in the orthogonal complement of SWS_{W}, i.e., SWS_{W}^{\perp}

x0=k=1Kμkwk+s,sSW.x_{0}=\sum_{k=1}^{K}\mu_{k}w_{k}+s,\qquad s\in S_{W}^{\perp}.

However, in calculating wk,x0\left\langle w_{k},x_{0}\right\rangle, the inner product of all elements in SWS_{W}^{\perp} taken with wkw_{k} vanishes. Therefore, we consider only the SWS_{W} component of x0x_{0}. Replacing wkw_{k} in x0x_{0} with (15), we get

x0=k=1Kμki=1Nαkiϕ(xi).x_{0}=\sum_{k=1}^{K}\mu_{k}\sum_{i=1}^{N}\alpha_{ki}\phi(x_{i}).

In our application of RKHS principles, we need to compute the following inner products with respect to some kernel in RKHS:

wk,wl=\displaystyle\left\langle w_{k},w_{l}\right\rangle= i=1Nαkiϕ(xi),j=1Nαljϕ(xj)\displaystyle\;\left\langle\sum_{i=1}^{N}\alpha_{ki}\phi(x_{i}),\sum_{j=1}^{N}\alpha_{lj}\phi(x_{j})\right\rangle
=\displaystyle= i=1Nj=1Nαkiαljϕ(xi),ϕ(xj)\displaystyle\;\sum_{i=1}^{N}\sum_{j=1}^{N}\alpha_{ki}\alpha_{lj}\left\langle\phi(x_{i}),\phi(x_{j})\right\rangle
=\displaystyle= i=1Nj=1Nαkiαljκ(xi,xj)\displaystyle\;\sum_{i=1}^{N}\sum_{j=1}^{N}\alpha_{ki}\alpha_{lj}\kappa(x_{i},x_{j})
=\displaystyle= αkTGαl,\displaystyle\;\alpha_{k}^{T}G\alpha_{l},

where GG is the Gram matrix formed by the set of pairwise RKHS inner products of patterns. Also, we have

wk,ϕ(xi)=\displaystyle\left\langle w_{k},\phi(x_{i})\right\rangle= j=1Nαkjϕ(xj),ϕ(xi)\displaystyle\;\left\langle\sum_{j=1}^{N}\alpha_{kj}\phi(x_{j}),\phi(x_{i})\right\rangle
=\displaystyle= j=1Nαkjκ(xj,xi)\displaystyle\;\sum_{j=1}^{N}\alpha_{kj}\kappa(x_{j},x_{i})
=\displaystyle= αkTgi,\displaystyle\;\alpha_{k}^{T}g_{i},

where gig_{i} is the ithi^{\textup{th}} column of the Gram matrix GG.

4.2 A Kernel Formulation for Multiclass and Multilabel SVMs

We now extend our formulation to an RKHS. The RKHS objective function directly follows (7) with the main difference being the use of RKHS inner products:

ker(W,x0)=\displaystyle\mathcal{L}_{\mathrm{ker}}(W,x_{0})= k=1Kwk,wk+θk=1Kl=k+1Kwk,wl\displaystyle\sum_{k=1}^{K}\left\langle w_{k},w_{k}\right\rangle+\theta\sum_{k=1}^{K}\sum_{l=k+1}^{K}\left\langle w_{k},w_{l}\right\rangle
+βk=1KiCk[1wk,ϕ(xki)x0]++λk=1Kwk,x0.\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\left[1-\left\langle w_{k},\phi(x_{ki})-x_{0}\right\rangle\right]_{+}+\lambda\sum_{k=1}^{K}-\left\langle w_{k},x_{0}\right\rangle.

Again, after majorizing the hinge loss followed by a change of variables bkwk,x0b_{k}\equiv-\left\langle w_{k},x_{0}\right\rangle, we have

ker(W,b,λ)=\displaystyle\mathcal{L}_{\mathrm{ker}}(W,b,\lambda)= k=1Kwk,wk+θk=1Kl=k+1Kwk,wl\displaystyle\sum_{k=1}^{K}\left\langle w_{k},w_{k}\right\rangle+\theta\sum_{k=1}^{K}\sum_{l=k+1}^{K}\left\langle w_{k},w_{l}\right\rangle
+βk=1KiCk14zki[1+zki(wk,ϕ(xki)+bk)]2+λk=1Kbk.\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-\left(\left\langle w_{k},\phi(x_{ki})\right\rangle+b_{k}\right)\right]^{2}+\lambda\sum_{k=1}^{K}b_{k}.

Rewriting the Lagrangian using a matrix formulation of RKHS inner products, we get

ker(𝒜,b,λ,Z)=\displaystyle\mathcal{L}_{\mathrm{ker}}(\mathcal{A},b,\lambda,Z)= k=1KαkTGαk+θk=1Kl=k+1KαkTGαl\displaystyle\sum_{k=1}^{K}\alpha_{k}^{T}G\alpha_{k}+\theta\sum_{k=1}^{K}\sum_{l=k+1}^{K}\alpha_{k}^{T}G\alpha_{l}
+βk=1KiCk14zki[1+zki(αkTgki+bk)]2+λk=1Kbk,\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\alpha_{k}^{T}g_{ki}+b_{k})\right]^{2}+\lambda\sum_{k=1}^{K}b_{k},

where 𝒜={α1,α2,,αK}\mathcal{A}=\left\{\alpha_{1},\alpha_{2},...,\alpha_{K}\right\}. Again by concatenating αk\alpha_{k} and bkb_{k} into a vector α~k\widetilde{\alpha}_{k} and defining the following matrices

α~k[αkbk],g~ki[gki1],G0[G00T0],e[001](N+1),\widetilde{\alpha}_{k}\equiv\left[\begin{array}[]{c}\alpha_{k}\\ b_{k}\end{array}\right],\quad\widetilde{g}_{ki}\equiv\left[\begin{array}[]{c}g_{ki}\\ 1\end{array}\right],\quad G_{0}\equiv\left[\begin{array}[]{cc}G&\overrightarrow{0}\\ \overrightarrow{0}^{T}&0\end{array}\right],\quad e\equiv\left[\begin{array}[]{c}0\\ \vdots\\ 0\\ 1\end{array}\right]_{(N+1)},

the Lagrangian is rewritten as

ker(𝒜~,λ,Z)=\displaystyle\mathcal{L}_{\text{ker}}(\widetilde{\mathcal{A}},\lambda,Z)= k=1Kα~kTG0α~k+θk=1Kl=k+1Kα~kTG0α~l\displaystyle\sum_{k=1}^{K}\widetilde{\alpha}_{k}^{T}G_{0}\widetilde{\alpha}_{k}+\theta\sum_{k=1}^{K}\sum_{l=k+1}^{K}\widetilde{\alpha}_{k}^{T}G_{0}\widetilde{\alpha}_{l}
+βk=1KiCk14zki[1+zki(α~kTg~ki)]2+λk=1KeTα~k,\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\widetilde{\alpha}_{k}^{T}\widetilde{g}_{ki})\right]^{2}+\lambda\sum_{k=1}^{K}e^{T}\widetilde{\alpha}_{k},

where 𝒜~={α~1,α~2,,α~K}\widetilde{\mathcal{A}}=\left\{\widetilde{\alpha}_{1},\widetilde{\alpha}_{2},...,\widetilde{\alpha}_{K}\right\}.

4.3 Hard constraints on weight vectors and biases

In a manner similar to the linear setting in Section 3.4, we can also incorporate any combination of soft and hard constraints on the normal vectors and biases into the formulation in the kernel setting. As mentioned above, using the representer theorem we can write each wkw_{k} in an RKHS as a linear combination of mapped patterns. Note that we cannot directly incorporate a hard constraint on the weight vectors into the formulation as we did in the linear case since 0H0_{H} here is the origin of an infinite dimensional Hilbert space and is therefore not computable. One way to set the sum of the wkw_{k} equal to 0H0_{H} in RKHS, i.e., k=1Kwk=0H\sum_{k=1}^{K}w_{k}=0_{H}, is by adding the following constraints to the problem:

k=1Kαki=0,i{1,2,,N}.\sum_{k=1}^{K}\alpha_{ki}=0,\quad i\in\{1,2,\ldots,N\}. (16)

These constraints are derived from the fact that each wkw_{k} is a linear combination of mapped patterns

wk=i=1Nαkiϕ(xi),k{1,,K}w_{k}=\sum_{i=1}^{N}\alpha_{ki}\;\phi(x_{i}),\quad k\in\{1,\ldots,K\}

and therefore setting the sum to 0H0_{H} via

k=1Kwk=k=1Ki=1Nαkiϕ(xi)=i=1Nk=1Kαkiϕ(xi)=i=1Nϕ(xi)k=1Kαki=0H\sum_{k=1}^{K}w_{k}=\sum_{k=1}^{K}\sum_{i=1}^{N}\alpha_{ki}\;\phi(x_{i})=\sum_{i=1}^{N}\sum_{k=1}^{K}\alpha_{ki}\;\phi(x_{i})=\sum_{i=1}^{N}\phi(x_{i})\sum_{k=1}^{K}\alpha_{ki}=0_{H}

is equivalent to (16). Here, we only present the formulation for the hard constraints on both weight vectors and biases. The detailed formulation for the case of soft ww-hard bb is also presented in Appendix B. The other two cases can be obtained in similar fashion. The optimization problem is

min𝒜,b,Z\displaystyle\min_{\mathcal{A},b,Z} 12k=1KαkTGαk+βk=1KiCk14zki[1+zki(αkTgki+bk)]2\displaystyle\frac{1}{2}\sum_{k=1}^{K}\alpha_{k}^{T}G\alpha_{k}+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\alpha_{k}^{T}g_{ki}+b_{k})\right]^{2} (17)
s.t. k=1Kαki=0,i{1,2,,N}\displaystyle\sum_{k=1}^{K}\alpha_{ki}=0,\quad i\in\{1,2,\ldots,N\}
k=1Kbk=0.\displaystyle\sum_{k=1}^{K}b_{k}=0.

Similarly, by concatenating bkb_{k} to αk\alpha_{k}, we get the simplified optimization problem

min𝒜~,Z\displaystyle\min_{\widetilde{\mathcal{A}},Z} 12k=1Kα~kTG0α~k+βk=1KiCk14zki[1+zki(α~kTg~ki)]2\displaystyle\quad\frac{1}{2}\sum_{k=1}^{K}\widetilde{\alpha}_{k}^{T}G_{0}\widetilde{\alpha}_{k}+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\widetilde{\alpha}_{k}^{T}\widetilde{g}_{ki})\right]^{2} (18)
s.t. k=1Kα~k=0.\displaystyle\quad\sum_{k=1}^{K}\widetilde{\alpha}_{k}=0.

We then concatenate all α~k\widetilde{\alpha}_{k} into one vector, α~\widetilde{\alpha}, and solve for it using KKT optimality conditions. The four cases for the kernel formulation with different settings of constraints are summarized in Table 2. Similar to the linear case in Section 3, regularization terms and hinge loss terms are left out since they are common to all cases.

Constraints Appearance in objective (Lagrangian) function
soft ww-soft bb θk=1Kl=k+1KαkTGαl+γ(k=1Kbk)2\theta\sum_{k=1}^{K}\sum_{l=k+1}^{K}\alpha_{k}^{T}G\alpha_{l}+\gamma\left(\sum_{k=1}^{K}b_{k}\right)^{2}
soft ww-hard bb θk=1Kl=k+1KαkTGαl+ηk=1Kbk\theta\sum_{k=1}^{K}\sum_{l=k+1}^{K}\alpha_{k}^{T}G\alpha_{l}+\eta\sum_{k=1}^{K}b_{k}
hard ww-soft bb i=1Nλik=1Kαki+γ(k=1Kbk)2\sum_{i=1}^{N}\lambda_{i}\sum_{k=1}^{K}\alpha_{ki}+\gamma\left(\sum_{k=1}^{K}b_{k}\right)^{2}
hard ww-hard bb i=1Nλik=1Kαki+ηk=1Kbk\sum_{i=1}^{N}\lambda_{i}\sum_{k=1}^{K}\alpha_{ki}+\eta\sum_{k=1}^{K}b_{k}
Table 2: Objective functions corresponding to all combinations of soft and hard constraints in RKHS. The parameters {λi}\left\{\lambda_{i}\right\} and η\eta denote the Lagrange multipliers for hard constraints

5 Experiments

In this section, we present our results for some well known datasets using our models and compare them to the OvR approach with the soft margin two-class SVM as the base classifier—using the scikit-learn [40] implementation. In Section 5.1, we discuss one of the drawbacks of the OvR approach for a toy dataset which does not have samples exclusively drawn from each of the classes. In Section 5.2, we first present the results for a few two-class datasets and compare the performance of our model with the soft margin binary SVM. Subsequently, we present the results of our model on some famous multiclass datasets and compare them with OvR and the Crammer-Singer (CS) model for the linear case. In Section 5.3, our model is tested on benchmark multilabel datasets and the results are compared to binary relevance (BR).

Refer to caption
Figure 2: OvN determines the boundary for patterns in a class without requiring exclusive oppositional patterns from other classes.

We use 3-fold cross validation to choose the hyperparameters. We assume a finite and discrete grid of parameter values. First, the data is shuffled and divided into three equal folds. We choose the first tuple of grid points and train the model on two folds (training set). Then the trained model with this tuple is validated on the 3rd3^{\textup{rd}} fold (validation set) and the model accuracy on the validation set is recorded. We repeat this procedure three times, choosing a different fold as validation set on each occasion and the remaining two folds as the training set and record all accuracy measures. In this way, we get three accuracy measures for the first choice of hyperparameters. This procedure is repeated for each and every tuple of grid points. Once again, all accuracy measures are recorded and the average accuracy on three folds for each tuple is computed. The maximum average accuracy is reported here. We tested our models on known datasets and compare the results against frequently used classification models in the literature.

5.1 Generating unseen label set configuration

One of the drawbacks of the BR approach for multilabel problems is that we might not have data only belonging to one class in order to train its classifier. For illustration, consider the following scenario shown in Figure 2. In this two-class example, the training set patterns belong either to class 1 (label [1,0]\left[1,0\right]) or to both classes (label [1,1]\left[1,1\right]). Therefore there is no pattern which only belongs to class 2. Obviously, the OvR approach can not train a classifier for class 1 versus class 2 as there is no pattern belonging solely to the latter. Hence we need an approach for determining the hyperplanes of each class without requiring data belonging to opposing classes. In other words, we need each class to determine its classifier based on available information of its members without the need for direct opposition from other classes. This is the core idea of our approach and hence the monicker one versus none (OvN).

Scikit-learn handles this situation by just assigning the label 11 to any test set pattern and therefore only produces label set [1,0]\left[1,0\right] or [1,1]\left[1,1\right] for each training or test pattern. The cross training approach of [4] is also not able to cope with this situation. They implemented the OvR approach to train separate classifiers. If a pattern belongs to both class 1 and class 2, it is considered a class 1 (class 2) pattern while training the class 1 (class 2) classifier. However, even this strategy is not able to mitigate the situation. With the OvN approach, we are capable of producing new label sets which have not occurred during training. Figure 2 demonstrates this scenario. As we can see from Figure 2, it does not make sense to assign label set [1,1]\left[1,1\right] to patterns (4,3)\left(-4,3\right), (5,2)\left(-5,2\right) and (3,3)\left(-3,3\right) since they are far away from the points which have both labels. Our linear soft ww-hard bb model, assigns label set [0,1]\left[0,1\right] to these points.

test pattern ground truth OvN (soft ww-hard bb) OvR (scikit-learn)
(4,3)\left(-4,3\right) [0,1]\left[0,1\right] [0,1]\left[0,1\right] [1,1]×\left[1,1\right]\times
(5,2)\left(-5,2\right) [0,1]\left[0,1\right] [0,1]\left[0,1\right] [1,1]×\left[1,1\right]\times
(3,3)\left(-3,3\right) [0,1]\left[0,1\right] [0,1]\left[0,1\right] [1,1]×\left[1,1\right]\times
(2,4)\left(-2,4\right) [1,1]\left[1,1\right] [1,1]\left[1,1\right] [1,1]\left[1,1\right]
(0,3)\left(0,3\right) [1,0]\left[1,0\right] [1,0]\left[1,0\right] [1,0]\left[1,0\right]
(1,5)\left(1,5\right) [1,1]\left[1,1\right] [1,1]\left[1,1\right] [1,1]\left[1,1\right]
Table 3: OvR approach fails to obtain a classifier for patterns which only belong to class 2 since there are no such patterns in the training set. Scikit-learn assigns all test patterns to class 1 (×\times indicates wrongly assigned labels). OvN can assign correct label ([0,1]\left[0,1\right]) to these patterns.

5.2 Multiclass datasets

We first show the results of applying our model (with linear and nonlinear kernels) to 2 class problems for 2-dimensional datasets for illustrative purposes. The column SVM in Table 4 is the original two-class SVM implemented by scikit-learn. The metric for evaluating multiclass learning algorithms is the accuracy score. We also demonstrate the choice and number of support vectors for each model. In this section, we mainly use the Gaussian Radial Basis Function (GRBF) kernel which for two patterns xix_{i} and xjx_{j} is defined as κ(xi,xj)=exp(12σ2xixj22)\kappa\left(x_{i},x_{j}\right)=\text{exp}\left(-\frac{1}{2\sigma^{2}}\|x_{i}-x_{j}\|_{2}^{2}\right) where σ\sigma is a hyperparameter.

Data Kernel Accuracy No. of sv
Sww-Hbb SVM Sww-Hbb SVM
hourglass Gaussian 1 0.97 [9,10]\left[9,10\right] [17,18]\left[17,18\right]
Poly(deg.2) 1 0.99 [5,4]\left[5,4\right] [8,9]\left[8,9\right]
random Linear 0.95 0.87 [6,6]\left[6,6\right] [7,7]\left[7,7\right]
moon Gaussian 1 1 [4,6]\left[4,6\right] [7,7]\left[7,7\right]
Table 4: Performance comparison of our model with binary SVM. OvN soft ww-hard bb (Sww-Hbb) results in better accuracy with fewer support vectors

The hyperplanes corresponding to the classifiers in each model as well as the support vectors are depicted in Figure 3.

Refer to caption
(a)
Refer to caption
(b)
Refer to caption
(c)
Refer to caption
(d)
Figure 3: Applying the soft ww-hard bb model and binary SVM with linear and nonlinear kernels on different datasets. (3(a)) Hourglass with Gaussian kernel; (3(b)) Hourglass with polynomial kernel; (3(c)) Moon with Gaussian kernel; (3(d)) Two classes with linear kernel. Patterns with projection less than 1 on both weight vectors, are assigned to the class with greater projection magnitude.

Table 5 shows the results of our model, OvN, compared to OvR with binary SVM with linear and nonlinear kernels as well as the Crammer-Singer (CS) model—using the implementation in scikit-learn [40]. Three multiclass (with more than two classes) benchmark datasets are chosen. We normalized the datasets and did not use any supervised or unsupervised dimensionality reduction. In Table 5, the average accuracy and loss on left out folds in 3-fold cross validation are reported.

Dataset Kernel OvN Sww-Hbb OvR (scikit-learn) CS (scikit-learn)
AS HL AS HL AS HL
glass L 0.851 0.149 0.598 0.402 0.668 0.332
G 0.869 0.131 0.701 0.299 - -
iris L 0.987 0.013 0.960 0.04 0.967 0.033
G 0.987 0.013 0.967 0.033 - -
wine L 0.843 0.157 0.967 0.033 0.983 0.017
G 0.966 0.034 0.986 0.014 - -
Table 5: Results for multiclass with average accuracy on 3-fold cross validation with linear (L) and Gaussian (G) kernels.

5.3 Multilabel datasets

Multilabel classification is a generalization of the multiclass problem where more than one label is assigned to each pattern. There are many applications in which a multilabel framework is required to assign instances to more than one class if relevant. The scene dataset [4] is one of the most widely used multilabel classification datasets. It has 294 features, 2407 instance images and 6 classes. We implemented our unified soft ww-hard bb OvN model and compared the results to the OvR (BR) approach with the soft margin two-class SVM for multilabel classification—using the scikit-learn implementation. We also implemented our model on the emotions (music) dataset [55]. It consists of 593 songs with 6 labels and 72 features.

There are different ways of evaluating the performance of a multilabel classifier since each label set prediction can be completely incorrect, completely correct or partially correct with different levels of correctness in the multilabel setting [48]. Two main types of metrics are introduced in the literature: bipartition evaluation metrics and ranking evaluation metrics. The bipartition category in turn has two subcategories of instance-based (example-based) and label-based metrics. Our evaluation in this work is based on instance-based metrics. The Hamming loss [44]—an instance-based metric—is defined as the fraction of incorrectly predicted labels to the total number of labels averaged over all instances and normalized by the number of classes

Hamming loss1NKn=1N|PnΔYn|,\text{Hamming loss}\equiv\frac{1}{NK}\sum_{n=1}^{N}\left|P_{n}\Delta Y_{n}\right|,

where YnY_{n} is the set of actual (true) labels, PnP_{n} is the set of predicted labels and Δ\Delta denotes the symmetric difference of two sets. The ideal value of the Hamming loss is zero. Exact Match Ratio (subset accuracy) is a strict example-based metric and computes the ratio of instances whose set of predicted labels exactly matches their corresponding set of true labels, i.e., 1Nn=1NI(Pn=Yn)\frac{1}{N}\sum_{n=1}^{N}I\left(P_{n}=Y_{n}\right) where I()I\left(\cdot\right) is the indicator function. Other instance-based metrics borrowed from information retrieval are accuracy, precision, recall and F1F_{1}-score which is the harmonic mean of precision and recall [17]. These metrics account for partially correct predictions. In accuracy score, the proportion of correctly predicted labels to the total number of labels (predicted and true) for each instance is obtained and averaged over all instances [16]:

Accuracy=1Nn=1N|PnYn||PnYn|.\text{Accuracy}=\frac{1}{N}\sum_{n=1}^{N}\frac{\left|P_{n}\cap Y_{n}\right|}{\left|P_{n}\cup Y_{n}\right|}.

In our experiments, we use accuracy and Hamming loss metrics for comparison. Table 6 shows the results of implementing our model on a synthetic multilabel dataset as well as the scene and emotions dataset. As we can see, the performance is slightly better in the case of OvR. However, we need to keep two points in mind. First, our soft ww-hard bb model is a unified one which attempts to learn all classifiers in one shot. So in this sense, OvR which separately learns the classifiers has a better chance of focusing on each class. However, it completely ignores the relationships between labels. Our model on the other hand considers the interrelationships between each pair of labels through soft constraints imposed on the weight vectors. Second, the label selection criteria are very strict in our model. We only assign a label to a pattern if its projection to the corresponding weight vector is greater than or equal to 1. Designing a less strict decision criterion which assigns labels to projections less than 1—provided it meets other conditions—may improve performance. This is an exciting avenue for future research.

Dataset Kernel OvN Sww-Hbb OvR (BR)
AS HL AS HL
synthetic L 0.837 0.133 0.828 0.14
G 0.843 0.142 0.867 0.122
scene L 0.557 0.177 0.601 0.10
G 0.652 0.12 0.684 0.08
emotions L 0.494 0.235 0.530 0.195
G 0.50 0.255 0.575 0.175
Table 6: Results for multilabel datasets with average accuracy score (AS) and Hamming loss (HL) on all folds. Linear (L) and Gaussian (G) kernels are applied.

6 Conclusion

In this paper, we began by motivating the need for a new multiclass and multilabel SVM. Contrary to popular belief, a multilabel SVM is not a simple extension of previous SVMs and this impacts our multiclass SVM formulation as well—an unexpected development. At its core, the new formulation employs a separate hyperplane for each class and uses (soft or hard) constraints to incorporate class separation. From a technical perspective, the new approach results in class specific weight vectors and reformulation of the origin as biases in feature vector space. Extension to RKHS kernels requires even more careful treatment of the origin in Hilbert space. Hard and soft constraints on the weight vectors and the biases (related to the origin) can be combined and we showcased a few examples in both the multiclass and multilabel settings. The experiments clearly demonstrate a competitive classifier which eschews both the one versus all and one versus one dichotomies widely prevalent at the present time.

The core idea—separate non anti-parallel hyperplanes for each class—can be extended to other classifiers as well such as the Fisher linear discriminant. This is a possible avenue for future research. Explicit inter label dependencies were not considered in the present work and could be incorporated in future work. We can also consider a formulation similar to the ν\nu-SVM in the future and incorporate margin controlling variables as well. In terms of optimization algorithms, we developed a very simple iterative majorization approach in the primal. A more elaborate subgradient-based approach can be considered in the future. Furthermore, we have not investigated the dual formulation in our setting. This appears to be a straightforward extension. Finally, better schemes for hyperparameter estimation via nested cross-validation and the like can be worked out.

From a larger perspective, it is entirely straightforward to incorporate negative labels in our framework—theoretically speaking. The main reason why this was not developed in the present work is due to the severe lack of training sets with negative labels (such as ”I don’t know what I am but I’m definitely NOT A libertarian” in a political survey). If such information is available in future datasets, then we can flesh out the space of negative labels in our framework. This is an interesting avenue for future work.

Acknowledgement

This work is partially supported by NSF IIS 1743050 to A.R.

Appendix A Optimization scheme for linearly separable case with soft constraint on weight vectors and hard constraint on biases

Rewriting the Lagrangian in (12) with the matrices introduced in (10), we have

(W~,λ,Z)=\displaystyle\mathcal{L}(\widetilde{W},\lambda,Z)= k=1Kw~kTDw~k+αk=1Kl=k+1Kw~kTDw~l\displaystyle\sum_{k=1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{k}+\alpha\sum_{k=1}^{K}\sum_{l=k+1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{l} (19)
+βk=1KiCk14zki[1+zki(w~kTx~ki)]2+λk=1KeTw~k\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\widetilde{w}_{k}^{T}\widetilde{x}_{ki})\right]^{2}+\lambda\sum_{k=1}^{K}e^{T}\widetilde{w}_{k}
=\displaystyle= k=1Kw~kTDw~k+k=1Kl=k+1Kw~kTDw~l+λk=1KeTw~k\displaystyle\sum_{k=1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{k}+\sum_{k=1}^{K}\sum_{l=k+1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{l}+\lambda\sum_{k=1}^{K}e^{T}\widetilde{w}_{k}
+βk=1KiCk14zki[w~kTx~kix~kiTw~k2(1+zki)(w~kTx~ki)+(1+zki)2]\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[\widetilde{w}_{k}^{T}\widetilde{x}_{ki}\widetilde{x}_{ki}^{T}\widetilde{w}_{k}-2(1+z_{ki})(\widetilde{w}_{k}^{T}\widetilde{x}_{ki})+(1+z_{ki})^{2}\right]
=\displaystyle= k=1Kw~kTDw~k+k=1Kl=k+1Kw~kTDw~l+λk=1KeTw~k\displaystyle\sum_{k=1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{k}+\sum_{k=1}^{K}\sum_{l=k+1}^{K}\widetilde{w}_{k}^{T}D\widetilde{w}_{l}+\lambda\sum_{k=1}^{K}e^{T}\widetilde{w}_{k}
+β(14k=1KiCk1zkiakiw~kTx~kix~kiTw~k12k=1KiCk(1+zkizki)1+aki(w~kTx~ki)\displaystyle+\beta\Biggl{(}\frac{1}{4}\sum_{k=1}^{K}\sum_{i\in C_{k}}\overset{a_{ki}}{\overbrace{\frac{1}{z_{ki}}}}\widetilde{w}_{k}^{T}\widetilde{x}_{ki}\widetilde{x}_{ki}^{T}\widetilde{w}_{k}-\frac{1}{2}\sum_{k=1}^{K}\sum_{i\in C_{k}}\overset{1+a_{ki}}{\overbrace{(\frac{1+z_{ki}}{z_{ki}})}}(\widetilde{w}_{k}^{T}\widetilde{x}_{ki})
+k=1KiCk14zki(1+zki)2𝐶)\displaystyle\qquad\qquad\qquad\qquad\qquad\qquad\qquad\qquad+\overset{C}{\overbrace{\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}(1+z_{ki})^{2}}}\Biggr{)}

where zkiz_{ki} are variables introduced by the majorization scheme into the formulation and λ\lambda is the Lagrange multiplier. We define the following matrices:

w~[w~1w~2w~K],U[eee]L×1,D~[D000D000D]L×L,D~0[0DDD0DDD0]L×L,\widetilde{w}\equiv\left[\begin{array}[]{c}\widetilde{w}_{1}\\ \widetilde{w}_{2}\\ \vdots\\ \widetilde{w}_{K}\end{array}\right],\;U\equiv\left[\begin{array}[]{c}e\\ e\\ \vdots\\ e\end{array}\right]_{L\times 1},\;\widetilde{D}\equiv\left[\begin{array}[]{cccc}D&0&\cdots&0\\ 0&D&\cdots&0\\ \vdots&\vdots&\ddots&\vdots\\ 0&0&\cdots&D\end{array}\right]_{L\times L},\;\widetilde{D}_{0}\equiv\left[\begin{array}[]{cccc}0&D&\cdots&D\\ D&0&\cdots&D\\ \vdots&\vdots&\ddots&\vdots\\ D&D&\cdots&0\end{array}\right]_{L\times L},
X~14[iC1a1ix~1ix~1iT000iC2a2ix~2ix~2iT000iCKaKix~Kix~KiT]L×L,\widetilde{X}\equiv\frac{1}{4}\left[\begin{array}[]{cccc}\sum_{i\in C_{1}}a_{1i}\widetilde{x}_{1i}\widetilde{x}_{1i}^{T}&0&\cdots&0\\ 0&\sum_{i\in C_{2}}a_{2i}\widetilde{x}_{2i}\widetilde{x}_{2i}^{T}&\cdots&\vdots\\ \vdots&\vdots&\ddots&0\\ 0&0&\cdots&\sum_{i\in C_{K}}a_{Ki}\widetilde{x}_{Ki}\widetilde{x}_{Ki}^{T}\end{array}\right]_{L\times L},
x~12[iC1(1+a1i)x~1iiC2(1+a2i)x~2iiCK(1+aKi)x~Ki]L×1\widetilde{x}\equiv\frac{1}{2}\left[\begin{array}[]{c}\sum_{i\in C_{1}}(1+a_{1i})\widetilde{x}_{1i}\\ \sum_{i\in C_{2}}(1+a_{2i})\widetilde{x}_{2i}\\ \vdots\\ \sum_{i\in C_{K}}(1+a_{Ki})\widetilde{x}_{Ki}\end{array}\right]_{L\times 1}

where L=K(M+1)L=K(M+1), KK is the number of classes and MM is the number of features. Using these matrices, we rewrite (19) as a quadratic form

(w~,λ)=\displaystyle\mathcal{L}(\widetilde{w},\lambda)= w~TD~w~+α2w~TD~0w~+β(w~TX~w~x~Tw~+C)+λUTw~\displaystyle\widetilde{w}^{T}\widetilde{D}\widetilde{w}+\frac{\alpha}{2}\widetilde{w}^{T}\widetilde{D}_{0}\widetilde{w}+\beta\left(\widetilde{w}^{T}\widetilde{X}\widetilde{w}-\widetilde{x}^{T}\widetilde{w}+C\right)+\lambda U^{T}\widetilde{w} (20)
=\displaystyle= w~T(D~+α2D~0+βX~)w~βx~Tw~+βC+λUTw~.\displaystyle\widetilde{w}^{T}(\widetilde{D}+\frac{\alpha}{2}\widetilde{D}_{0}+\beta\widetilde{X})\widetilde{w}-\beta\widetilde{x}^{T}\widetilde{w}+\beta C+\lambda U^{T}\widetilde{w}.

Now, we differentiate 20 w.r.t. w~\widetilde{w} to get

(w~)=w~=(2D~+αD~0+2βX~)w~βx~,\nabla\mathcal{L}(\widetilde{w})=\frac{\partial\mathcal{L}}{\partial\widetilde{w}}=(2\widetilde{D}+\alpha\widetilde{D}_{0}+2\beta\widetilde{X})\widetilde{w}-\beta\widetilde{x},
w~=02(D~+α2D~0+βX~)w~=βx~,\frac{\partial\mathcal{L}}{\partial\widetilde{w}}=0\>\Longrightarrow 2\left(\widetilde{D}+\frac{\alpha}{2}\widetilde{D}_{0}+\beta\widetilde{X}\right)\widetilde{w}=\beta\widetilde{x},
w~stn=β2(D~+α2D~0+βX~)1x~.\widetilde{w}_{\mathrm{stn}}=\frac{\beta}{2}\left(\widetilde{D}+\frac{\alpha}{2}\widetilde{D}_{0}+\beta\widetilde{X}\right)^{-1}\widetilde{x}.

In order for the stationary point w~stn\widetilde{w}_{\mathrm{stn}} to be a minimum, the Hessian of \mathcal{L} should be symmetric (which it already is) and positive definite (PD), i.e.,

H(w~)=w~w~T=D~+α2D~0+βX~0.H(\widetilde{w})=\frac{\partial\mathcal{L}}{\partial\widetilde{w}\partial\widetilde{w}^{T}}=\widetilde{D}+\frac{\alpha}{2}\widetilde{D}_{0}+\beta\widetilde{X}\succ 0.

The solution for the zkiz_{ki} follows the approach outlined in Section 3.3.

Appendix B Optimization scheme for kernel case with soft constraint on weight vectors and hard constraint on biases

Similar to the previous case as explained in Appendix A, we rewrite the Lagrangian with soft constraints on the weight vectors (WW) and hard constraints on the biases (bb):

ker(W,b,λ)=\displaystyle\mathcal{L}_{\text{ker}}(W,b,\lambda)= k=1Kwk,wk+θk=1Kl=k+1Kwk,wl\displaystyle\sum_{k=1}^{K}\langle w_{k},w_{k}\rangle+\theta\sum_{k=1}^{K}\sum_{l=k+1}^{K}\langle w_{k},w_{l}\rangle
+βk=1KiCk14zki[1+zki(wk,ϕ(xki)+bk)]2+λk=1Kbk\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\langle w_{k},\phi(x_{ki})\rangle+b_{k})\right]^{2}+\lambda\sum_{k=1}^{K}b_{k}

where λ\lambda is the Lagrange multiplier corresponding to the hard constraint on the biases k=1Kbk=0\sum_{k=1}^{K}b_{k}=0. Rewriting the Lagrangian using matrices and using notation introduced in Section 4.1 results in

ker(α~,λ,z)=\displaystyle\mathcal{L}_{\text{ker}}(\widetilde{\alpha},\lambda,z)= k=1Kα~kTG0α~k+θk=1Kl=k+1Kα~kTG0α~l\displaystyle\sum_{k=1}^{K}\widetilde{\alpha}_{k}^{T}G_{0}\widetilde{\alpha}_{k}+\theta\sum_{k=1}^{K}\sum_{l=k+1}^{K}\widetilde{\alpha}_{k}^{T}G_{0}\widetilde{\alpha}_{l}
+βk=1KiCk14zki[1+zki(α~kTg~ki)]2+λk=1KeTα~k\displaystyle+\beta\sum_{k=1}^{K}\sum_{i\in C_{k}}\frac{1}{4z_{ki}}\left[1+z_{ki}-(\widetilde{\alpha}_{k}^{T}\widetilde{g}_{ki})\right]^{2}+\lambda\sum_{k=1}^{K}e^{T}\widetilde{\alpha}_{k}

which is further simplified using matrix notation:

ker(α~,λ,z)=\displaystyle\mathcal{L}_{\text{ker}}(\widetilde{\alpha},\lambda,z)= α~TG~0α~+θ2α~TG~1α~+β(α~TX~α~x~Tα~+C)+λUTα~\displaystyle\;\widetilde{\alpha}^{T}\widetilde{G}_{0}\widetilde{\alpha}+\frac{\theta}{2}\widetilde{\alpha}^{T}\widetilde{G}_{1}\widetilde{\alpha}+\beta\left(\widetilde{\alpha}^{T}\widetilde{X}\widetilde{\alpha}-\widetilde{x}^{T}\widetilde{\alpha}+C\right)+\lambda U^{T}\widetilde{\alpha}
=\displaystyle= α~T(G~0+θ2G~1+βX~)α~+(λUTβx~T)α~+βC\displaystyle\;\widetilde{\alpha}^{T}(\widetilde{G}_{0}+\frac{\theta}{2}\widetilde{G}_{1}+\beta\widetilde{X})\widetilde{\alpha}+(\lambda U^{T}-\beta\widetilde{x}^{T})\widetilde{\alpha}+\beta C

where G~0\widetilde{G}_{0} ,G~1\widetilde{G}_{1} and UU are necessary matrices for rewriting Lagrangian after concatenating the α~k\widetilde{\alpha}_{k}. From the KKT necessary conditions, we get

kerα~=02(G~0+θ2G~1+βX~)α~+λU=βx~\frac{\partial\mathcal{L}_{\text{ker}}}{\partial\widetilde{\alpha}}=0\>\Longrightarrow 2\left(\widetilde{G}_{0}+\frac{\theta}{2}\widetilde{G}_{1}+\beta\widetilde{X}\right)\widetilde{\alpha}+\lambda U=\beta\widetilde{x}
kerλ=0UTα~=0\frac{\partial\mathcal{L}_{\text{ker}}}{\partial\lambda}=0\>\Longrightarrow U^{T}\widetilde{\alpha}=0

Rewriting these equations in matrix format, we get

[2G~0+θG~1+2βX~UUT0][α~λ]=[βx~0.]\left[\begin{array}[]{cc}2\widetilde{G}_{0}+\theta\widetilde{G}_{1}+2\beta\widetilde{X}&U\\ U^{T}&0\end{array}\right]\left[\begin{array}[]{c}\widetilde{\alpha}\\ \lambda\end{array}\right]=\left[\begin{array}[]{c}\beta\widetilde{x}\\ 0.\end{array}\right]

The solution for the zkiz_{ki} follows the approach outlined in Section 3.3. After obtaining {α~1,α~2,,α~K}\left\{\widetilde{\alpha}_{1},\widetilde{\alpha}_{2},\cdots,\widetilde{\alpha}_{K}\right\}, for each class kk, we calculate the projection in RKHS for a test pattern as follows:

wk,ϕ(xt)x0=\displaystyle\langle w_{k},\phi(x_{t})-x_{0}\rangle= iSVαkiϕ(xi),ϕ(xt)+bk\displaystyle\langle\sum_{i\in SV}\alpha_{ki}\phi(x_{i}),\phi(x_{t})\rangle+b_{k}
=\displaystyle= iSVαkiϕ(xi),ϕ(xt)+bk\displaystyle\sum_{i\in SV}\alpha_{ki}\langle\phi(x_{i}),\phi(x_{t})\rangle+b_{k}
=\displaystyle= iSVαkiϕ(xi),ϕ(xt)+bk\displaystyle\sum_{i\in SV}\alpha_{ki}\langle\phi(x_{i}),\phi(x_{t})\rangle+b_{k}
=\displaystyle= iSVαkiK(xi,xt)+bk\displaystyle\sum_{i\in SV}\alpha_{ki}K(x_{i},x_{t})+b_{k}
=\displaystyle= α~kTgtTest\displaystyle\;\widetilde{\alpha}_{k}^{T}g_{t}^{\mathrm{Test}}

where gtTestg_{t}^{\mathrm{Test}} is the ttth column of the Gram matrix between training support vectors and test patterns.


References

  • [1] E. L. Allwein, R. E. Schapire, and Y. Singer. Reducing multiclass to binary: A unifying approach for margin classifiers. Journal of Machine Learning Research (JMLR), 1(Dec):113–141, 2000.
  • [2] G. Armano, C. Chira, and N. Hatami. Error-correcting output codes for multi-label text categorization. In Italian Information Retrieval Workshop(IIR), pages 26–37. Citeseer, 2012.
  • [3] B. E. Boser, I. M. Guyon, and V. N. Vapnik. A training algorithm for optimal margin classifiers. In Proceedings of the 5th Annual Workshop on Computational Learning Theory, pages 144–152. ACM, 1992.
  • [4] M. R. Boutell, J. Luo, X. Shen, and C. M. Brown. Learning multi-label scene classification. Pattern Recognition, 37(9):1757–1771, 2004.
  • [5] E. J. Bredensteiner and K. P. Bennett. Multicategory classification by support vector machines. In Computational Optimization, pages 53–79. Springer, 1999.
  • [6] O. Chapelle. Training a support vector machine in the primal. Neural Computation, 19(5):1155–1178, 2007.
  • [7] W.-J. Chen, Y.-H. Shao, C.-N. Li, and N.-Y. Deng. MLTSVM: A novel twin support vector machine to multi-label learning. Pattern Recognition, 52:61–74, 2016.
  • [8] C. Cortes and V. N. Vapnik. Support-vector networks. Machine Learning, 20(3):273–297, 1995.
  • [9] K. Crammer and Y. Singer. On the algorithmic implementation of multiclass kernel-based vector machines. Journal of Machine Learning Research (JMLR), 2(Dec):265–292, 2001.
  • [10] J. de Leeuw and K. Lange. Sharp quadratic majorization in one dimension. Computational Statistics and Data Analysis, 53(7):2471–2484, 2009.
  • [11] T. G. Dietterich and G. Bakiri. Solving multiclass learning problems via error-correcting output codes. Journal of Artificial Intelligence Research, 2:263–286, 1994.
  • [12] U. Dogan, T. Glasmachers, and C. Igel. A unified view on multi-class support vector classification. Journal of Machine Learning Research (JMLR), 17:1–32, 2016.
  • [13] A. Elisseeff and J. Weston. A kernel method for multi-labelled classification. In Advances in Neural Information Processing Systems (NIPS), pages 681–687, 2002.
  • [14] G. M. Fung and O. L. Mangasarian. Proximal support vector machine classifiers. In Proceedings of the 7th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 77–86, 2001.
  • [15] G. M. Fung and O. L. Mangasarian. Multicategory proximal support vector machine classifiers. Machine Learning, 59(1-2):77–97, 2005.
  • [16] E. Gibaja and S. Ventura. A tutorial on multilabel learning. ACM Computing Surveys (CSUR), 47(3):52, 2015.
  • [17] S. Godbole and S. Sarawagi. Discriminative methods for multi-labeled classification. In Pacific-Asia Conference on Knowledge Discovery and Data Mining, pages 22–30. Springer, 2004.
  • [18] P. J. Groenen, G. Nalbantov, and J. C. Bioch. Nonlinear support vector machines through iterative majorization and I-splines. In Advances in Data Analysis, pages 149–161. Springer, 2007.
  • [19] P. J. Groenen, G. Nalbantov, and J. C. Bioch. SVM-Maj: A majorization approach to linear support vector machines with different hinge errors. Advances in Data Analysis and Classification, 2(1):17–43, 2008.
  • [20] Y. Guo and D. Schuurmans. Adaptive large margin training for multilabel classification. In Procedeengs of 25th AAAI Conference on Artificial Intelligence, San Fransisco, California, USA, pages 374–379, 2011.
  • [21] B. Hariharan, L. Zelnik-Manor, M. Varma, and S. V. N. Vishwanathan. Large scale max-margin multi-label classification with priors. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), pages 423–430. Citeseer, 2010.
  • [22] T. Hastie and R. Tibshirani. Classification by pairwise coupling. In Advances in Neural Information Processing Systems (NIPS), pages 507–513, 1998.
  • [23] C.-W. Hsu and C.-J. Lin. A comparison of methods for multiclass support vector machines. IEEE Transactions on Neural Networks, 13(2):415–425, 2002.
  • [24] Jayadeva, R. Khemchandani, and S. Chandra. Twin support vector machines for pattern classification. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 29(5):905–910, 2007.
  • [25] T. Joachims. Text categorization with support vector machines: Learning with many relevant features. In European Conference on Machine Learning (ECML), pages 137–142. Springer, 1998.
  • [26] T. Joachims, T. Hofmann, Y. Yue, and C.-N. Yu. Predicting structured objects with support vector machines. Communications of the ACM, 52(11):97, 2009.
  • [27] H. Kazawa, T. Izumitani, H. Taira, and E. Maeda. Maximal margin labeling for multi-topic text categorization. In Advances in Neural Information Processing Systems (NIPS), pages 649–656, 2005.
  • [28] G. Kimeldorf and G. Wahba. Some results on Tchebycheffian spline functions. Journal of Mathematical Analysis and Applications, 33(1):82–95, 1971.
  • [29] U. H. Kreßel. Pairwise classification and support vector machines. In Advances in Kernel Methods, pages 255–268. MIT Press, 1999.
  • [30] C. H. Lampert. Maximum margin multi-label structured prediction. In Advances in Neural Information Processing Systems, pages 289–297, 2011.
  • [31] K. Lange. Optimization. Springer Texts in Statistics. Springer New York, 2013.
  • [32] Y. Lee, Y. Lin, and G. Wahba. Multicategory support vector machines: Theory and application to the classification of microarray data and satellite radiance data. Journal of the American Statistical Association, 99(465):67–81, 2004.
  • [33] C. Lippert, Z. Ghahramani, and K. M. Borgwardt. Gene function prediction from synthetic lethality networks via ranking on demand. Bioinformatics, 26(7):912–918, 2010.
  • [34] Y. Liu, K. Wen, Q. Gao, X. Gao, and F. Nie. SVM based multi-label learning with missing labels for image annotation. Pattern Recognition, 78:307–317, 2018.
  • [35] Y. Liu and M. Yuan. Reinforced multicategory support vector machines. Journal of Computational and Graphical Statistics, 20(4):901–919, 2011.
  • [36] O. Luaces, J. Díez, J. Barranquero, J. J. del Coz, and A. Bahamonde. Binary relevance efficacy for multilabel classification. Progress in Artificial Intelligence, 1(4):303–313, 2012.
  • [37] O. L. Mangasarian and E. W. Wild. Multisurface proximal support vector machine classification via generalized eigenvalues. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 28(1):69–74, 2005.
  • [38] A. McCallum. Multi-label text classification with a mixture model trained by EM. In AAAI workshop on Text Learning, pages 1–7, 1999.
  • [39] H. D. Nguyen and G. J. McLachlan. Iteratively-reweighted least-squares fitting of support vector machines: A majorization–minimization algorithm approach. arXiv preprint arXiv:1705.04651, 2017.
  • [40] F. Pedregosa, G. Varoquaux, A. Gramfort, V. Michel, B. Thirion, O. Grisel, M. Blondel, P. Prettenhofer, R. Weiss, and V. Dubourg. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research (JMLR), 12(Oct):2825–2830, 2011.
  • [41] J. C. Platt, N. Cristianini, and J. Shawe-Taylor. Large margin DAGs for multiclass classification. In Advances in Neural Information Processing Systems (NIPS), pages 547–553, 2000.
  • [42] R. Rifkin and A. Klautau. In defense of one-vs-all classification. Journal of Machine Learning Research (JMLR), 5(Jan):101–141, 2004.
  • [43] J. Rousu, C. Saunders, S. Szedmak, and J. Shawe-Taylor. Kernel-based learning of hierarchical multilabel classification models. Journal of Machine Learning Research (JMLR), 7(Jul):1601–1626, 2006.
  • [44] R. E. Schapire and Y. Singer. Improved boosting algorithms using confidence-rated predictions. Machine learning, 37(3):297–336, 1999.
  • [45] B. Schölkopf, A. J. Smola, R. C. Williamson, and P. L. Bartlett. New support vector algorithms. Neural computation, 12(5):1207–1245, 2000.
  • [46] B. Schölkopf, R. C. Williamson, A. J. Smola, J. Shawe-Taylor, and J. C. Platt. Support vector method for novelty detection. In Advances in Neural Information Processing Systems (NIPS), pages 582–588, 2000.
  • [47] P. Sollich. Bayesian methods for support vector machines: Evidence and predictive class probabilities. Machine Learning, 46(1-3):21–52, 2002.
  • [48] M. S. Sorower. A literature survey on algorithms for multi-label learning. Technical report, Oregon State University, Corvallis, 2010.
  • [49] L. Sun, S. Ji, and J. Ye. Canonical correlation analysis for multilabel classification: A least-squares formulation, extensions, and analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 33(1):194–200, 2011.
  • [50] L. Sun, S. Ji, and J. Ye. Multi-label dimensionality reduction. Chapman and Hall/CRC, 2016.
  • [51] S. Szedmak, J. Shawe-Taylor, and E. Parado-Hernandez. Learning via linear operators: Maximum margin regression; multiclass and multiview learning at one-class complexity. Technical report, PASKAL, Southampton, UK, 2005.
  • [52] P. Szymański and T. Kajdanowicz. A scikit-based python environment for performing multi-label classification. The Computing Research Repository (CoRR), abs/1702.01460, 2017.
  • [53] B. Taskar, C. Guestrin, and D. Koller. Max-margin Markov networks. In Advances in Neural Information Processing Systems (NIPS), pages 25–32, 2004.
  • [54] D. M. Tax and R. P. Duin. Support vector data description. Machine Learning, 54(1):45–66, 2004.
  • [55] K. Trohidis, G. Tsoumakas, G. Kalliris, and I. Vlahavas. Multi-label classification of music into emotions. In Proceedings of 9th International Conference on Music Information Retrieval (ISMIR), pages 325–330, 2008.
  • [56] I. Tsochantaridis, T. Hofmann, T. Joachims, and Y. Altun. Support vector machine learning for interdependent and structured output spaces. In Proceedings of the 21st International Conference on Machine Learning(ICML), page 104. ACM, 2004.
  • [57] G. J. Van Den Burg and P. J. Groenen. GenSVM: A generalized multiclass support vector machine. The Journal of Machine Learning Research (JMLR), 17(1):7964–8005, 2016.
  • [58] V. N. Vapnik. Statistical learning theory. Adaptive and Learning Systems for Signal Processing, communications and control series. John Wiley & Sons, New York. A Wiley-Interscience Publication, 1998.
  • [59] V. N. Vapnik. The nature of statistical learning theory. Springer Science & Business Media, 2013.
  • [60] J. Weston and C. Watkins. Support vector machines for multi-class pattern recognition. In Proceedings of the 7th European Symposium on Artificial Neural Networks (ESANN), volume 99, pages 219–224, 1999.
  • [61] M.-L. Zhang and Z.-H. Zhou. Multilabel neural networks with applications to functional genomics and text categorization. IEEE Transactions on Knowledge and Data Engineering, 18(10):1338–1351, 2006.
  • [62] Z. Zhang and M. I. Jordan. Bayesian multicategory support vector machines. In Proceedings of 22nd Conference on Uncertainty in Artificial Intelligence (UAI), 2006.