This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Deep Partial Multi-Label Learning with Graph Disambiguation

Haobo Wang 1    Shisong Yang2    Gengyu Lyu2 Corresponding author.    Weiwei Liu 3    Tianlei Hu1   
Ke Chen1
   Songhe Feng4&Gang Chen1 1Zhejiang University    2Beijing University of Technology
3Wuhan University     4Beijing Jiaotong University {wanghaobo, htl, chenk, cg}@zju.edu.cn, [email protected],
[email protected], [email protected], [email protected]
Abstract

In partial multi-label learning (PML), each data example is equipped with a candidate label set, which consists of multiple ground-truth labels and other false-positive labels. Recently, graph-based methods, which demonstrate a good ability to estimate accurate confidence scores from candidate labels, have been prevalent to deal with PML problems. However, we observe that existing graph-based PML methods typically adopt linear multi-label classifiers and thus fail to achieve superior performance. In this work, we attempt to remove several obstacles for extending them to deep models and propose a novel deep Partial multi-Label model with grAph-disambIguatioN (PLAIN). Specifically, we introduce the instance-level and label-level similarities to recover label confidences as well as exploit label dependencies. At each training epoch, labels are propagated on the instance and label graphs to produce relatively accurate pseudo-labels; then, we train the deep model to fit the numerical labels. Moreover, we provide a careful analysis of the risk functions to guarantee the robustness of the proposed model. Extensive experiments on various synthetic datasets and three real-world PML datasets demonstrate that PLAIN achieves significantly superior results to state-of-the-art methods.

1 Introduction

The rapid growth of deep learning allows us to tackle increasingly sophisticated problems. Among them, multi-label learning (MLL), which assigns multiple labels to an object, is a fundamental and intriguing task in various real-world applications such as image retrieval Lai et al. (2016), autonomous vehicles Chen et al. (2019a), and sentiment analysis Yilmaz et al. (2021). Typically, these tasks require precisely labeled data, which is expensive and time-consuming due to the complicated structure and the high volume of the output space.

To address this issue, plenty of works have studied different weakly-supervised settings of MLL, including semi-supervised MLL Shi et al. (2020), multi-label with missing labels Durand et al. (2019); Huynh and Elhamifar (2020), and single-labeled MLL Xu et al. (2022). While these learning paradigms mainly battle the rapid growth of data volume, they overlook the intrinsic difficulty and ambiguity in multi-label annotation. For example, it is hard for a human annotator to identify an Alaskan Malamute from a Siberian Husky. One potential solution is to randomly decide the binary label value. However, it may involve uncontrollable label noise that has a negative impact on the training procedure. Another strategy is to leave it missing, but then, we may lose critical information. Recently, partial multi-label learning (PML) Xie and Huang (2018) has been proposed to alleviate this problem. In PML, the annotators are allowed to provide a candidate label set that contains all potentially relevant labels as well as false-positive labels.

A straightforward approach to learning from PML data is to regard all candidate labels as valid ones, and then, off-the-shelf multi-label methods can be applied. Nevertheless, the presence of false-positive labels can highly impact predictive performance. To cope with this issue, many PML methods Xie and Huang (2018); Yu et al. (2018); He et al. (2019); Sun et al. (2019); Li et al. (2020); Lyu et al. (2020); Xie and Huang (2020); Yan et al. (2021); Xie and Huang (2022) have been developed. Amongst them, graph-based PML methods Wang et al. (2019); Li and Wang (2020); Xu et al. (2020); Lyu et al. (2020); Zhang and Fang (2021) have been popular. In practice, graph-based algorithms demonstrate a good ability to obtain relatively precise confidence of candidate labels by aggregating information from the nearest neighbors. However, we observe that most of them adopt linear classifier, which is less powerful to deal with complicated tasks and high data volumes.

Consequently, there is an urgent need for developing an effective deep model to boost the performance of graph PML methods. However, there are two main challenging issues. First, existing graph-based PML methods typically require propagating labels over the whole training dataset. It restricts their scalability when employing deep classifiers since the deep models are trained in batch mode. Second, it is well acknowledged that there exist label correlations in multi-label datasets, which also helps disambiguate the candidate labels. For example, if labels Amusement Park and Mickey Mouse are identified as true, it is very likely that label Disney is also a ground-truth label. Thus, it is required to exploit the label correlations to improve the disambiguation ability in a unified framework.

In this work, we propose a novel deep Partial multi-Label model with grAph disambIguatioN (PLAIN). In PLAIN, we involve both instance- and label-level similarities to disambiguate the candidate labels set, which simultaneously leverages instance smoothness as well as label correlations. Then, we build a deep multi-label classifier to fit the relatively credible pseudo-labels generated through disambiguation. Instead of separately training in two stages like previous works Wang et al. (2019); Zhang and Fang (2021), we propose an efficient optimization framework that iteratively propagates the labels and trains the deep model. Moreover, we give a careful analysis of the risk functions such that proper risk functions are chosen for improved robustness. Empirical results on nine synthetic datasets and three real-world PML datasets clearly verify the efficiency and effectiveness of our method. More theoretical and empirical results can be found in Appendix.

2 Related Works

Multi-Label Learning (MLL)

Liu et al. (2022); Wei et al. (2022) aims at assigning each data example multiple binary labels simultaneously. An intuitive solution, the one-versus-all method (OVA) Zhang and Zhou (2014), is to decompose the original problem into a series of single-labeled classification problems. However, it overlooks the rich semantic information and dependencies among labels, and thus fails to obtain satisfactory performance. To solve this problem, many MLL approaches are proposed, such as embedding-based methods Yu et al. (2014); Shen et al. (2018), classifier chains Read et al. (2011), tree-based algorithms Wei et al. (2021), and deep MLL models Yeh et al. (2017); Chen et al. (2019b); Bai et al. (2020). Nevertheless, these methods typically require a large amount of fully-supervised training data, which is expensive and time-consuming in MLL tasks. To this end, some works have studied different weakly-supervised MLL settings. For example, semi-supervised MLL Niu et al. (2019); Shi et al. (2020) leverages information from both fully-labeled and unlabeled data. Multi-label with missing labels Durand et al. (2019); Wei et al. (2018); Xu et al. (2022) allows providing a subset of ground-truth. In noisy MLL Veit et al. (2017), the binary labels may be flipped to the wrong one. Nevertheless, though these settings relieve the burden of multi-label annotating, they ignore the inherent difficulty of labeling, i.e. the labels can be ambiguous.

Partial Multi-Label Learning (PML)

learns from a superset of ground-truth labels, where multiple labels are possible to be relevant. To tackle the ambiguity in the labels, the pioneering work Xie and Huang (2018) estimates the confidence of each candidate label being correct via minimizing confidence-weighted ranking loss. Some works Yu et al. (2018); Sun et al. (2019); Li et al. (2020); Gong et al. (2022) recover the ground-truth by assuming the true label matrix is low-rank. Recently, graph-based methods Wang et al. (2019); Xu et al. (2020); Lyu et al. (2020); Xie and Huang (2020); Zhang and Fang (2021); Xu et al. (2021); Tan et al. (2022) have attracted much attention from the community due to the good disambiguation ability. For instance, PARTICLE Zhang and Fang (2021) identifies the credible labels through an iterative label propagation procedure and then, applies pair-wise ranking techniques to induce a multi-label predictor. Although graph-based models have shown promising results, we observe that most of them adopt linear classifiers, which significantly limits the model expressiveness. Besides, they typically separate the graph disambiguation and model training into two stages, which makes the multi-label classifier error-prone; in contrast, the proposed PLAIN is an end-to-end framework.

Some studies have also developed deep PML models, such as adversarial PML model Yan and Guo (2019) and mutual teaching networks Yan et al. (2021). However, it remains urgent to combine the graph and deep models together such that the model enjoys high expressiveness as well as good disambiguation ability. It should also be noted that the recent popular graph neural networks (GNN) model Wu et al. (2021) is not suitable for PML. The reason is that GNN aims at better representation learning (at an instance-level), while in PML, the graph structure is used to alleviate the label ambiguity.

3 Method

We denote the dd-dimensional feature space as 𝒳d\mathcal{X}\subset\mathbb{R}_{d} , and the label space as 𝒴={1,2,,L}\mathcal{Y}=\{1,2,\ldots,L\} with LL class labels. The training dataset 𝒟={(𝒙i,Si)|1in}\mathcal{D}=\{(\bm{x}_{i},S_{i})|1\leq i\leq n\} contains nn examples, where 𝒙i𝒳\bm{x}_{i}\in\mathcal{X} is the instance vector and Si𝒴S_{i}\subset\mathcal{Y} is the candidate label set. Without loss of generality, we assume that the instance vectors are normalized such that 𝒙i2=1||\bm{x}_{i}||_{2}=1. Besides, we denote S~i𝒴\tilde{S}_{i}\subset\mathcal{Y} as the ground-truth label set, which is a subset of SiS_{i}. For simplicity, we define 𝒚i\bm{y}_{i}, 𝒚~i{0,1}L\tilde{\bm{y}}_{i}\in\{0,1\}^{L} as the binary vector representations of SiS_{i} and S~i\tilde{S}_{i}. When considering the whole dataset, we also denote the candidate label matrix by 𝒀=[𝒚1,,𝒚n]n×L\bm{Y}=[\bm{y}_{1},\ldots,\bm{y}_{n}]^{\top}\in\mathbb{R}^{n\times L}.

In what follows, we first introduce the learning target of our model that integrates label correlations to boost disambiguation ability. Then, we provide an efficient optimization framework for our deep propagation network.

Refer to caption
Figure 1: The model architecture of PLAIN. At each epoch, we first train the deep model by traveling the training dataset and then propagating the labels with the help of candidate labels and model prediction; PLAIN alternates between the two operations.

3.1 Learning Objective

Our ultimate goal is to estimate a multi-label predictor f:𝒳[0,1]Lf:\mathcal{X}\mapsto[0,1]^{L} from 𝒟\mathcal{D}. In this work, we assume that its output 𝒚^=f(𝒙)\hat{\bm{y}}=f(\bm{x}) satisfies the following properties,

  1. 1.

    𝒚^\hat{\bm{y}} is close to the candidate vector 𝒚\bm{y};

  2. 2.

    𝒚^\hat{\bm{y}} satisfies the instance-level similarity. That is, if two instances are close to each other, then their label vectors are likely to be close too;

  3. 3.

    𝒚^\hat{\bm{y}} satisfies the label-level similarity. That is, if two labels co-occur in the training data frequently, then they tend to co-exist at many instances.

These properties coincide with our goals. The first property is natural since the only thing we can trust is the candidate label set. The second one has been widely adopted to disambiguate the ground-truth label from the false-positive labels. The last one exploits the label correlations to further promote the performance. In this paper, we adopt a graph-based model to achieve these goals. Generally speaking, the graph structure can be collected in various ways. One direct strategy is to use prior knowledge. For instance, for data mining tasks, social networks Qiu et al. (2018) can be used to depict instance similarity. The label similarity can also be estimated using an external knowledge graph Lee et al. (2018).

However, we notice that there is no available graph structure in existing public PML datasets. Hence, we propose to construct the instance graph 𝑨xn×n\bm{A}^{x}\in\mathbb{R}^{n\times n} and the label graph 𝑨yL×L\bm{A}^{y}\in\mathbb{R}^{L\times L} in a data-driven way. To obtain 𝑨x\bm{A}^{x}, we adopt a monomial kernel Iscen et al. (2017) as the distance metric and calculate a sparse instance affinity matrix 𝑺n×n\bm{S}\in\mathbb{R}^{n\times n} with elements,

sij:={max(𝒙i𝒙j,0)ρ,if ij and j𝒩k(𝒙i)0,otherwise.\begin{split}s_{ij}:=\begin{cases}\max(\bm{x}_{i}^{\top}\bm{x}_{j},0)^{\rho},\quad\text{if }i\neq j\text{ and }j\in\mathcal{N}_{k}(\bm{x}_{i})\\ 0,\quad\quad\quad\quad\quad\quad\text{otherwise}\end{cases}.\end{split}

Here 𝒩k(𝒙i)\mathcal{N}_{k}(\bm{x}_{i}) saves the indices of kk-nearest neighbors of 𝒙i\bm{x}_{i}. ρ>0\rho>0 controls the weights of similarity. Then, we get a symmetrized instance graph by 𝑨x=𝑺+𝑺\bm{A}^{x}=\bm{S}+\bm{S}^{\top}.

Inspired by HALE Lyu et al. (2020), we build the label graph 𝑨y\bm{A}^{y} by means of the label co-occurrence statistics from the training data, whose elements are given by,

aijy:=k=1n𝕀(yki=1ykj=1)k=1n𝕀(yki=1)+k=1n𝕀(ykj=1),\begin{split}a^{y}_{ij}:=\frac{\sum_{k=1}^{n}\mathbb{I}(y_{ki}=1\wedge y_{kj}=1)}{\sum_{k=1}^{n}\mathbb{I}(y_{ki}=1)+\sum_{k=1}^{n}\mathbb{I}(y_{kj}=1)},\end{split}

where 𝕀()\mathbb{I}(\cdot) is an indicator function returning 11 if the event is true and 0 otherwise. It is worth noting that HALE simply calculates the average co-occurrence by dividing the instance number, which overlooks the existence of false-positive labels. In our method, we penalize it by the counts of instances that are assigned the labels as the candidates. In other words, if one label is assigned to too many instances, its co-occurrence statistics are less reliable.

Next, we impose the instance-level similarity property on the model output by means of a Laplacian regularizer,

𝒥x(𝒀^)=i,j=1naijx𝒚^idix𝒚^jdjx2=tr(𝒀^𝑳x𝒀^),\begin{split}\mathcal{J}_{x}(\hat{\bm{Y}})&=\sum_{i,j=1}^{n}a_{ij}^{x}||\frac{\hat{\bm{y}}_{i}}{\sqrt{d^{x}_{i}}}-\frac{\hat{\bm{y}}_{j}}{\sqrt{d^{x}_{j}}}||^{2}=\operatorname*{tr}(\hat{\bm{Y}}^{\top}\bm{L}_{x}\hat{\bm{Y}}),\\ \end{split} (1)

where 𝒀^=[𝒚^1,,𝒚^n]\hat{\bm{Y}}=[\hat{\bm{y}}_{1},\ldots,\hat{\bm{y}}_{n}]^{\top} is the predicted label matrix. 𝑳x=𝑰𝑫x12𝑨x𝑫x12\bm{L}_{x}=\bm{I}-\bm{D}_{x}^{-\frac{1}{2}}\bm{A}^{x}\bm{D}_{x}^{-\frac{1}{2}} is the Laplacian matrix of instance graph. 𝑫x=diag[d1x,,dnx]\bm{D}_{x}=\text{diag}[d_{1}^{x},\ldots,d_{n}^{x}] is a diagonal matrix with dix=j=1naijxd_{i}^{x}=\sum_{j=1}^{n}a^{x}_{ij}. The objective in Eq. (1) enforces two data examples to have similar predicted labels if they have a large affinity score. Similarly, we can introduce the label-level similarity by,

𝒥y(𝒀^)=tr(𝒀^𝑳y𝒀^),𝑳y=𝑰𝑫y12𝑨y𝑫y12.\begin{split}\mathcal{J}_{y}(\hat{\bm{Y}})=\operatorname*{tr}(\hat{\bm{Y}}\bm{L}_{y}\hat{\bm{Y}}^{\top}),\quad\bm{L}_{y}=\bm{I}-\bm{D}_{y}^{-\frac{1}{2}}\bm{A}^{y}\bm{D}_{y}^{-\frac{1}{2}}.\\ \end{split}

Finally, we involve a candidate consistency term to get the following objective,

𝒥(𝒀^,𝒀)=η2𝒀^𝒀F2+α2𝒥x(𝒀^)+β2𝒥y(𝒀^),\begin{split}\mathcal{J}(\hat{\bm{Y}},\bm{Y})=\frac{\eta}{2}||\hat{\bm{Y}}-\bm{Y}||_{F}^{2}+\frac{\alpha}{2}\mathcal{J}_{x}(\hat{\bm{Y}})+\frac{\beta}{2}\mathcal{J}_{y}(\hat{\bm{Y}}),\end{split}

where ||||F||\cdot||_{F} is the Frobenius norm. η\eta, α\alpha, and β\beta are all positive hyperparameters.

In summary, the main goal of our objective is to utilize the instance- and label-level similarity to recover reliable label confidences and address the problem of false positives, while simultaneously enjoying an extra performance improvement by exploiting label correlations.

3.2 Deep Propagation Network

If a linear model is adopted, the parameters can be directly obtained by optimizing 𝒥(f(𝑿),𝒀)\mathcal{J}(f(\bm{X}),\bm{Y}) with gradient descent. However, in deep models, it is not the case. When we update the model using a mini-batch of data, the instance-level manifold regularizer can involve unexpected gradient updating on the nearest neighbors of these data. Then, the training time will increase rapidly. To remedy this problem, we propose an efficient optimization framework for training a deep model from PML data. Specifically, we maintain a smooth pseudo-label matrix 𝒁\bm{Z}, which minimizes the propagation error while being close to the model output 𝒀^\hat{\bm{Y}}. In the sequel, we elaborate the training procedure of our model. The training procedure is also summarized in Algorithm 1.

3.2.1 Propagation Step

At tt-th epoch, we first propagate the label information from candidates 𝒀\bm{Y} and the model output 𝒀^\hat{\bm{Y}} to the intermediate ones. In graph-based methods, the output label is required to be smooth enough. Therefore, we use the Frobenius norm as the measure to ensure the closeness of 𝒁\bm{Z} and 𝒀^\hat{\bm{Y}}. Formally, we estimate 𝒁\bm{Z} by,

min𝒁LP(𝒁)=12𝒁𝒀^F2+𝒥(𝒁,𝒀).\begin{split}\min_{\bm{Z}}\mathcal{L}_{\text{LP}}(\bm{Z})=\frac{1}{2}||\bm{Z}-\hat{\bm{Y}}||_{F}^{2}+\mathcal{J}(\bm{Z},\bm{Y}).\end{split} (2)

We can derive its gradient by,

LP𝒁=(1+η)𝒁+α𝑳x𝒁+β𝒁𝑳y(𝒀^+η𝒀).\begin{split}\frac{\partial\mathcal{L}_{\text{LP}}}{\partial\bm{Z}}=(1+\eta)\bm{Z}+\alpha\bm{L}_{x}\bm{Z}+\beta\bm{Z}\bm{L}_{y}-(\hat{\bm{Y}}+\eta\bm{Y}).\end{split} (3)

which is followed by the matrix properties that 𝒁𝒀F2𝒁=2(𝒁𝒀)\frac{\partial||\bm{Z}-\bm{Y}||_{F}^{2}}{\partial\bm{Z}}=2(\bm{Z}-\bm{Y}) and tr(𝒁𝑳x𝒁)𝒁=(𝑳x+𝑳x)𝒁\frac{\partial\text{tr}(\bm{Z}^{\top}\bm{L}_{x}\bm{Z})}{\partial\bm{Z}}=(\bm{L}_{x}+\bm{L}_{x}^{\top})\bm{Z}. We have 𝑳x+𝑳x=2𝑳x\bm{L}_{x}+\bm{L}_{x}^{\top}=2\bm{L}_{x} because 𝑳x\bm{L}_{x} is symmetric. The optimal condition LP𝒁=𝟎\frac{\partial\mathcal{L}_{\text{LP}}}{\partial\bm{Z}}=\bm{0} typically requires solving a Sylvester system, which can be time-consuming. Therefore, we solve this problem by gradient descent 𝒁=𝒁γLP(𝒁)𝒁\bm{Z}=\bm{Z}-\gamma\frac{\partial\mathcal{L}_{\text{LP}}(\bm{Z})}{\partial\bm{Z}}, where γ>0\gamma>0 is the learning rate. Note that in MLL datasets, negative labels usually dominate, which makes all the pseudo-labels tend to concentrate at a small value. To cope with this issue, we further perform column-wise min-max normalization on 𝒁t\bm{Z}_{t} to obtain a balanced output.

3.2.2 Model Updating Step

Given an instance vector 𝒙\bm{x}, we first feed it into a deep neural model 𝒚^=f(𝒙;θ)\hat{\bm{y}}=f(\bm{x};\theta). Then, we update the model parameters to fit the pseudo-labels,

minθDeep(Y^,𝒁)=1ni=1nl(𝒚^i,𝒛i),\begin{split}\min_{\theta}\mathcal{L}_{\text{Deep}}(\hat{Y},\bm{Z})=\frac{1}{n}\sum\nolimits_{i=1}^{n}l(\hat{\bm{y}}_{i},\bm{z}_{i}),\end{split} (4)

where l(,)l(\cdot,\cdot) is the loss function. Recall that in Eq. (2), we enforce 𝒁\bm{Z} to be close to 𝒀^\hat{\bm{Y}} using the Frobenius norm. Hence, a natural choice of loss is to reuse the mean squared error (MSE), which is adopted by most graph-based PML algorithms Wang et al. (2019); Xu et al. (2020); Zhang and Fang (2021). However, in PML tasks, the disambiguated labels can still be noisy and the MSE loss leads to heavy overfitting. This is largely overlooked by previous works.

To tackle this problem, we shall have a closer look at the risk function. While MSE penalize too much on prediction mistakes, other robust surrogate losses can be considered, such as mean absolute error (MAE). However, we notice that MAE violates our smooth prediction assumption. To this end, we adopt the widely-used binary cross-entropy loss (BCE) for multi-label learning,

l(𝒚^i,𝒛i)=j=1Lzijlog(σ(y^ij))(1zij)log(1σ(y^ij)),\begin{split}l(\hat{\bm{y}}_{i},\bm{z}_{i})\!=\!\sum\nolimits_{j\!=\!1}^{L}\!-\!z_{ij}\log(\sigma(\hat{y}_{ij}))\!-\!(1\!-\!z_{ij})\log(1\!-\!\sigma(\hat{y}_{ij})),\end{split}

where σ()\sigma(\cdot) is the sigmoid function. On one hand, BCE loss approximately grows linearly for largely negative values, which makes it less sensitive to outliers. On the other hand, BCE outputs relatively smooth output such that the smoothness is preserved.

In effect, we can go further on the risk. A plethora of works Ghosh et al. (2017); Feng et al. (2020) has demonstrated that BCE is still less robust since it is unbounded. Let us consider the probabilistic output σ(𝒚^i)\sigma(\hat{\bm{y}}_{i}) of the deep model. Then the BCE loss is equivalent to a negative log loss on σ(𝒚^i)\sigma(\hat{\bm{y}}_{i}). If the predicted probability of the label is close to zero, the loss can grow quickly to infinity. Previous works have suggested that bounded loss on the probabilistic output is theoretically more robust than unbounded ones. One common choice of bounded loss is the probabilistic MSE loss (PMSE), i.e. l(𝒚^i,𝒛i)=σ(𝒚^i)𝒛i22l(\hat{\bm{y}}_{i},\bm{z}_{i})=||\sigma(\hat{\bm{y}}_{i})-\bm{z}_{i}||_{2}^{2}. In that case, the training criterion is consistent with our propagation error. Empirically, both PMSE and BCE loss perform well, and thus, we instantiate our model using BCE loss when comparing it with other methods. In Section B.2, we present a comparison between these losses, which clearly verifies our conjectures.

Fast Implementation.

To efficiently build the instance graph from training data, we employ a quick similarity search python package Faiss Johnson et al. (2017) to find the nearest neighbors. The sparse matrix multiplication 𝑳x𝒁\bm{L}_{x}\bm{Z} can be efficiently implemented by the Scipy package. In Appendix A, we provide a detailed analysis of the complexity and convergence properties to show the efficacy of our method.

Algorithm 1 The pseudo-code of PLAIN
0:  The deep MLL network parameter θ\theta
0:  Training dataset 𝒟\mathcal{D}, instance graph 𝑨x\bm{A}^{x}, label graph 𝑨y\bm{A}^{y}, hyperparameters γ\gamma, α\alpha, β\beta and η\eta, iteration number TT
1:  𝒁𝒀\bm{Z}\leftarrow\bm{Y}
2:  𝑳x𝑰𝑫x12𝑨x𝑫x12\bm{L}_{x}\leftarrow\bm{I}-\bm{D}_{x}^{-\frac{1}{2}}\bm{A}^{x}\bm{D}_{x}^{-\frac{1}{2}}
3:  𝑳y𝑰𝑫y12𝑨y𝑫y12\bm{L}_{y}\leftarrow\bm{I}-\bm{D}_{y}^{-\frac{1}{2}}\bm{A}^{y}\bm{D}_{y}^{-\frac{1}{2}}
4:  for epoch t=1,2,3,t=1,2,3,\ldots do
5:     for batch j=1,2,j=1,2,\ldots in 𝒟\mathcal{D} do
6:        Sample a mini-batch data (𝑿j,𝒁j)(\bm{X}^{j},\bm{Z}^{j})
7:        𝒀^jf(𝑿j;θ)\hat{\bm{Y}}^{j}\leftarrow f(\bm{X}^{j};\theta)
8:        θOPTIMIZE(l(𝒀^j,𝒁j;θ))\theta\leftarrow\text{OPTIMIZE}(l(\hat{\bm{Y}}^{j},\bm{Z}^{j};\theta))
9:     end for
10:     𝒀^f(𝑿;θ)\hat{\bm{Y}}\leftarrow f(\bm{X};\theta)
11:     for iter i=1,2,,Ti=1,2,\ldots,T do
12:        𝒁𝒁γLP(𝒁)𝒁\bm{Z}\leftarrow\bm{Z}-\gamma\frac{\partial\mathcal{L}_{\text{LP}}(\bm{Z})}{\partial\bm{Z}}
13:     end for
14:  end for
Table 1: Comparison of PLAIN with baselines on real-world datasets, where the best results are shown in bold face.
Datasets PLAIN PML-NI HALE ML-KNN GLOCAL PML-lc PARTICLE
Ranking Loss (the lower the better)
Music-emotion 0.217±\pm0.006 0.244±\pm0.007 0.299±\pm0.008 0.365±\pm0.010 0.344±\pm0.025 0.310±\pm0.010 0.261±\pm0.007
Music-style 0.136±\pm0.005 0.138±\pm0.009 0.271±\pm0.015 0.229±\pm0.010 0.253±\pm0.007 0.269±\pm0.016 0.351±\pm0.014
Mirflickr 0.088±\pm0.006 0.123±\pm0.004 0.143±\pm0.009 0.178±\pm0.013 0.212±\pm0.003 0.206±\pm0.008 0.213±\pm0.011
Average Precision (the higher the better)
Music-emotion 0.664±\pm0.007 0.610±\pm0.011 0.543±\pm0.005 0.505±\pm0.010 0.513±\pm0.010 0.539±\pm0.013 0.626±\pm0.011
Music-style 0.742±\pm0.012 0.740±\pm0.013 0.687±\pm0.013 0.659±\pm0.014 0.645±\pm0.007 0.598±\pm0.009 0.621±\pm0.016
Mirflickr 0.847±\pm0.012 0.792±\pm0.004 0.718±\pm0.005 0.698±\pm0.013 0.679±\pm0.007 0.675±\pm0.010 0.690±\pm0.012
Hamming Loss (the lower the better)
Music-emotion 0.191±\pm0.004 0.254±\pm0.013 0.219±\pm0.003 0.364±\pm0.011 0.219±\pm0.001 0.221±\pm0.008 0.206±\pm0.003
Music-style 0.112±\pm0.003 0.157±\pm0.010 0.122±\pm0.005 0.844±\pm0.010 0.144±\pm0.001 0.162±\pm0.006 0.137±\pm0.007
Mirflickr 0.162±\pm0.003 0.223±\pm0.005 0.156±\pm0.003 0.217±\pm0.006 0.253±\pm0.001 0.193±\pm0.006 0.179±\pm0.008
Table 2: Win/tie/loss counts of PLAIN’s performance against baselines on synthetic data sets (pairwise tt-test at 0.05 significance level).
Metrics Emotions Birds Medical Image Corel5k Bibtext Eurlex-dc Eurlex-sm NUS-WIDE Sum
Ranking Loss 18/0/0 15/0/3 15/0/3 15/1/2 18/0/0 18/0/0 15/0/3 14/4/0 16/0/2 144/5/13
Average Precision 18/0/0 16/1/1 15/0/3 18/0/0 16/0/2 16/1/1 18/0/0 18/0/0 18/0/0 153/2/7
Hamming Loss 15/1/2 14/1/3 11/4/3 11/1/6 5/13/0 11/7/0 8/10/0 14/4/0 14/4/0 103/45/14
Sum 51/1/2 45/2/7 41/4/9 44/2/8 39/13/2 45/8/1 41/10/3 46/8/0 48/4/2 400/52/34

4 Experiments

In this section, we present the main empirical results of PLAIN compared with state-of-the-art baselines. More experimental results are shown in Appendix B.

4.1 Datasets

We employe a total of twelve synthetic as well as real-world datasets for comparative studies. Specifically, the synthetic ones are generated from the widely-used multi-label learning datasets. A total of nine benchmark multi-label datasets111http://mulan.sourceforge.net/datasets-mlc.html are used for synthetic PML datasets generation, including Emotions, Birds, Medical, Image, Bibtex, Corel5K, Eurlex-dc, Eurlex-sm, NUS-WIDE. Note that NUS-WIDE is a large-scale dataset, and we preprocess it as in Lyu et al. (2020). For each data example in MLL datasets, we randomly select rr irrelevant labels and aggregate them with the ground-truth labels to obtain a candidate set. For example, given an instance 𝒙i\bm{x}_{i} and its ground-truth label set S~i\tilde{S}_{i}, we select rr labels from its complementary set 𝒴S~i\mathcal{Y}-\tilde{S}_{i}. If there are less than rr irrelevant labels, i.e. |𝒴S~i|<r|\mathcal{Y}-\tilde{S}_{i}|<r, we simply set the whole training set 𝒴\mathcal{Y} as the candidate set for 𝒙i\bm{x}_{i}. Following the experimental settings in Lyu et al. (2020), we choose r{1,2,3}r\in\{1,2,3\}, resulting in 2727 (3×9)(3\times 9) synthetic datasets. Furthermore, we conducted experiments on three real-world PML datasets Zhang and Fang (2021), including Music-emotion, Music-style, Mirflickr. These three datasets are derived from the image retrieval task Huiskes and Lew (2008), where the candidate labels are collected from web users and then further verified by human annotators to determine the ground-truth labels.

4.2 Baselines and Implementation Details

We choose six benchmark methods for comparative studies, including two MLL methods ML-KNN Zhang and Zhou (2007), GLOCAL Zhu et al. (2018), and four state-of-the-art PML methods PML-NI Xie and Huang (2022), PARTICLE Zhang and Fang (2021), HALE Lyu et al. (2020), and PML-lc Xie and Huang (2018). In particular, PARTICLE and HALE are two advanced graph-based PML methods. PARTICLE Zhang and Fang (2021) adopts an instance-level label propagation algorithm to disambiguate the candidate labels. HALE Lyu et al. (2020) regards the denoising procedure as a graph-matching procedure.

The parameter setups for the used methods are as follows. We fix k=10k=10 for all kkNN-based methods, including PLAIN. For the baselines, we fix or fine-tune the hyperparameters as the suggested configurations in respective papers. For our PLAIN method, the deep model is comprised of three fully-connected layers. The hidden sizes are set as [64,64][64,64] for those datasets with less than 6464 labels, and [256,256][256,256] for those datasets with more than 6464 and less than 256256 labels. For the remaining datasets, we set the hidden sizes as [512,512][512,512]. The trade-off parameters α\alpha and β\beta are hand-tuned from {0.001,0.01,0.1}\{0.001,0.01,0.1\}. η\eta is selected from {0.1,1,10}\{0.1,1,10\}. Following Iscen et al. (2017), we set ρ=3\rho=3. We train our deep model via stochastic gradient descent and empirically set the learning rate as 0.010.01 for both propagation and deep model training procedures. The number of maximum iterations is set as T=200T=200 for small-scale datasets and T=50T=50 for the large-scale NUS-WIDE dataset. Besides, weight decay is applied with a rate of 5e55e^{-5} to avoid overfitting.

For performance measure, we use three widely-used multi-label metrics, ranking loss, average precision and hamming loss Zhang and Zhou (2014). Finally, we perform ten-fold cross-validation on each dataset and the mean metric values as well as the standard deviations are reported.

Table 3: Comparison of PLAIN with baselines on nine synthetic PML datasets (rr = 3), where the best results are shown in bold face.
Datasets PLAIN PML-NI HALE ML-KNN GLOCAL PML-lc PARTICLE
Ranking Loss (the lower the better)
Emotions 0.158±\pm0.024 0.214±\pm0.029 0.235±\pm0.037 0.241±\pm0.026 0.322±\pm0.053 0.459±\pm0.035 0.259±\pm0.019
Birds 0.205±\pm0.036 0.187±\pm0.035 0.271±\pm0.061 0.304±\pm0.048 0.302±\pm0.041 0.321±\pm0.021 0.301±\pm0.032
Medical 0.050±\pm0.011 0.039±\pm0.013 0.169±\pm0.025 0.088±\pm0.019 0.068±\pm0.005 0.056±\pm0.012 0.100±\pm0.021
Image 0.190±\pm0.015 0.289±\pm0.018 0.192±\pm0.015 0.342±\pm0.026 0.264±\pm0.010 0.467±\pm0.025 0.315±\pm0.073
Corel5K 0.120±\pm0.004 0.205±\pm0.006 0.259±\pm0.011 0.139±\pm0.007 0.164±\pm0.003 0.169±\pm0.013 0.333±\pm0.050
Bibtext 0.079±\pm0.005 0.126±\pm0.010 0.601±\pm0.009 0.232±\pm0.006 0.131±\pm0.006 0.342±\pm0.005 0.287±\pm0.010
Eurlex-dc 0.033±\pm0.002 0.030±\pm0.003 0.079±\pm0.002 0.086±\pm0.012 0.150±\pm0.003 0.071±\pm0.013 0.061±\pm0.023
Eurlex-sm 0.027±\pm0.001 0.028±\pm0.002 0.040±\pm0.002 0.043±\pm0.003 0.071±\pm0.003 0.082±\pm0.013 0.053±\pm0.002
NUS-WIDE 0.211±\pm0.003 0.221±\pm0.002 0.239±\pm0.012 0.301±\pm0.011 0.311±\pm0.020 - 0.240±\pm0.015
Average Precision (the higher the better)
Emotions 0.812±\pm0.043 0.754±\pm0.016 0.751±\pm0.035 0.741±\pm0.029 0.647±\pm0.030 0.573±\pm0.025 0.745±\pm0.024
Birds 0.606±\pm0.037 0.580±\pm0.062 0.505±\pm0.054 0.453±\pm0.052 0.390±\pm0.060 0.388±\pm0.035 0.431±\pm0.051
Medical 0.781±\pm0.039 0.849±\pm0.035 0.769±\pm0.023 0.672±\pm0.039 0.751±\pm0.009 0.713±\pm0.012 0.720±\pm0.044
Image 0.774±\pm0.020 0.668±\pm0.018 0.762±\pm0.016 0.627±\pm0.022 0.692±\pm0.008 0.523±\pm0.019 0.689±\pm0.096
Corel5K 0.319±\pm0.010 0.289±\pm0.011 0.231±\pm0.008 0.251±\pm0.007 0.284±\pm0.007 0.247±\pm0.012 0.135±\pm0.041
Bibtext 0.550±\pm0.014 0.537±\pm0.012 0.353±\pm0.010 0.306±\pm0.006 0.438±\pm0.011 0.283±\pm0.010 0.313±\pm0.012
Eurlex-dc 0.740±\pm0.008 0.700±\pm0.011 0.673±\pm0.006 0.603±\pm0.012 0.278±\pm0.005 0.602±\pm0.012 0.630±\pm0.016
Eurlex-sm 0.802±\pm0.007 0.718±\pm0.011 0.739±\pm0.013 0.761±\pm0.012 0.556±\pm0.008 0.582±\pm0.013 0.695±\pm0.012
NUS-WIDE 0.286±\pm0.004 0.274±\pm0.003 0.230±\pm0.007 0.171±\pm0.015 0.177±\pm0.024 - 0.206±\pm0.017
Hamming Loss (the lower the better)
Emotions 0.223±\pm0.033 0.455±\pm0.059 0.297±\pm0.071 0.607±\pm0.209 0.281±\pm0.015 0.437±\pm0.027 0.233±\pm0.018
Birds 0.088±\pm0.010 0.095±\pm0.013 0.096±\pm0.009 0.053±\pm0.006 0.096±\pm0.008 0.132±\pm0.012 0.142±\pm0.018
Medical 0.019±\pm0.001 0.014±\pm0.002 0.017±\pm0.001 0.022±\pm0.002 0.028±\pm0.001 0.063±\pm0.003 0.024±\pm0.002
Image 0.207±\pm0.009 0.273±\pm0.028 0.189±\pm0.017 0.753±\pm0.005 0.237±\pm0.009 0.443±\pm0.014 0.403±\pm0.042
Corel5K 0.009±\pm0.000 0.011±\pm0.000 0.014±\pm0.001 0.009±\pm0.000 0.009±\pm0.000 0.019±\pm0.002 0.009±\pm0.000
Bibtext 0.013±\pm0.000 0.015±\pm0.000 0.016±\pm0.001 0.014±\pm0.000 0.015±\pm0.001 0.021±\pm0.001 0.017±\pm0.000
Eurlex-dc 0.002±\pm0.001 0.010±\pm0.001 0.004±\pm0.003 0.009±\pm0.005 0.002±\pm0.000 0.013±\pm0.005 0.004±\pm0.000
Eurlex-sm 0.006±\pm0.002 0.011±\pm0.001 0.008±\pm0.005 0.012±\pm0.001 0.011±\pm0.000 0.019±\pm0.002 0.010±\pm0.001
NUS-WIDE 0.021±\pm0.000 0.035±\pm0.002 0.031±\pm0.008 0.049±\pm0.016 0.022±\pm0.001 - 0.290±\pm0.006
  • \sharp

    - means over long time consumption, and thus, the experimental results are not reported.

4.3 Experimental Results

Table 1 reports the experimental results on real-world datasets. Table 3 lists the results on synthetic datasets, where the parameter is configured with r=3r=3. The results of r=1,2r=1,2 are consistent with those of r=3r=3 and thus are omitted. To better understand the superiority of the proposed method, we summarize the win/tie/loss counts on all synthetic datasets in Table 2. We also report the statistical testing results in Appendix B for analyzing the relative performance amongst competing algorithms. Based on these results, we can draw the following observations:

  • The proposed PLAIN method significantly outperforms all other competing approaches. For example, on the Music-emotion and Mirflicker datasets, in terms of average precision, PLAIN achieved results superior to those of the best baselines by 6.07\bm{6.07}% and 6.94\bm{6.94}% respectively.

  • According to Table 2, out of 486 statistical comparisons, six competing algorithms, and three evaluation metrics, PLAIN beats the competing methods in 82.30\bm{82.30}% cases. In particular, PLAIN wins or ties the baselines in 95.68\bm{95.68}% cases regarding average precision.

  • ML-KNN and GLOCAL obtain inferior performance on many datasets, such as Medical and Eurlex-dc. The reason is that these tailored MLL models regard all the candidate labels as the ground-truth and thus tend to overfit the false-positive labels.

  • Two PML methods PML-NI and HALE are strong baselines. In particular, PML-NI achieved the second-best results on most datasets. However, PLAIN performs significantly better than them, since their performance is restricted to the linear classifier.

  • On large-scale dataset NUS-WIDE, PLAIN still outperforms other methods on all evaluation metrics. It demonstrates the effectiveness of PLAIN in handling high-volume of data.

4.4 Further Analysis

Refer to caption
(a) Parameter sensitivity of γ\gamma
Refer to caption
(b) Parameter sensitivity of TT
Refer to caption
(c) Analysis of risk functions
Figure 2: Analysis of parameter sensitivities. (a) Changes in the performance of PLAIN as γ\gamma changes; (b) Changes in the performance of PLAIN as TT changes; (c) Convergence curves of PLAIN with different risk functions on Music-emotion dataset.
Refer to caption
Refer to caption
Figure 3: Ablation study on three datasets. We set r=3r=3 for Image and Bibtex. Notably, the lower ranking loss is the better and so is the hamming loss. For Average Precision, the higher is the better.
Complexity Analysis

In Table 4, we report the running time of different training stages of PLAIN, including the graph building step, label propagation step (one epoch) and the deep model training step (one epoch), on datasets with various scales. First of all, we can see that the instance graph can be fast built in a few microseconds on all datasets, with the help of Faiss package Johnson et al. (2017). Moreover, by utilizing the sparse property of instance graph and the fast computation ability of GPUs, the time complexity of label propagation is at the same magnitude as that of deep model training. In particular, on those datasets with fewer labels, such as Music-emotion and Image, the propagation phase is as fast as the deep model traveling the whole training dataset. In Figure 2, (a) and (b) show the parameter sensitivities of PLAIN to TT and γ\gamma. We can observe that PLAIN is stable with varying values of TT and γ\gamma. Specifically, PLAIN obtains high average precision even with small TT and small γ\gamma. But when T=0T=0 (without label propagation), our PLAIN model degenerates to the trivial DNN model and thus, the performance degrades. It should also be noted that when TT is too large, the performance of PLAIN slightly drops because the model tends to overfit. In summary, the proposed PLAIN model requires only a few gradient steps in the label propagation phase and can be efficiently trained.

Comparison between Loss Functions

As discussed in section 3.2.2, mean square error (MSE) is error-prone to the label noise. Hence, we conducted experiments to compare the performance of different loss functions, including MSE, MAE, BCE and probabilistic MSE (PMSE). The results are shown in Figure 2 (c). We can see that MAE demonstrates inferior performance and low convergence. Despite its robustness, it tends to produce harsh predictions and is opposite to the graph techniques which generate smooth outputs. MSE obtains good performance, but underperforms BCE and PMSE, since it easily overfits the false-positive labels. Finally, both BCE and PMSE demonstrate good empirical results. These observations suggest that a robust loss function is essential in successful learning from PML data.

Table 4: Running time (in microseconds) of PLAIN on six datasets with various data scales, regarding the graph building, label propagation (one epoch), and deep model training (one epoch) stages.
Datasets Graph Building Label Propagation Model Training
Image 1.0 150.2 154.4
Medical 1.1 344.0 112.2
Music-emotion 5.6 449.6 480.7
Bibtex 26.7 2498.7 600.8
Eurlex-sm 355.7 8911.2 1902.2
NUS-WIDE 1829.3 17255.2 11251.6
Ablation Analysis

An ablation study would be informative to see how different components of the proposed PLAIN contribute to the prediction performance. Thus, we compare PLAIN with three variants: 1) a pure deep neural network model (DNN) that treats the candidate labels as valid ones; 2) the PLAIN model without considering label-level similarity (PLAIN w/o LS); 3) the PLAIN model without considering instance-level similarity (PLAIN w/o IS). Figure 3 reports the results on one real-world dataset as well as two synthetic datasets. In general, both instance- and label-level graph regularizers improve the simple DNN model on most datasets and evaluation metrics. Then, by integrating the two regularizers, our PLAIN model consistently achieves the best performance. These results clearly verify the superiority of the proposed method.

5 Conclusion

In this work, we proposed a novel deep partial multi-label learning model PLAIN that copes with several limitations in existing graph-based PML methods. At each training epoch, we first involve both the instance- and label-level similarities to generate pseudo-labels via label propagation. We then train the deep model to fit the disambiguated labels. Moreover, we analyzed the time complexity and the robustness of PLAIN from both theoretical and empirical perspectives. Comprehensive experiments on synthetic datasets as well as real-world PML datasets clearly validate the efficiency, robustness, and effectiveness of our proposed method. Our work demonstrates that combining the graph and deep models together leads to high expressiveness as well as good disambiguation ability. We hope our work will increase attention toward a broader view of graph-based deep PML methods.

Acknowledgments

This work is supported by the National Key R&D Program of China (No. 2022YFB3304100) and by the Zhejiang University-China Zheshang Bank Co., Ltd. Joint Research Center. Gengyu Lyu is supported by the National Key R&D Program of China (No. 2022YFB3103104) and the China Postdoctoral Science Foundation (No. 2022M720320).

References

  • Bai et al. [2020] Junwen Bai, Shufeng Kong, and Carla P. Gomes. Disentangled variational autoencoder based multi-label classification with covariance-aware multivariate probit model. In IJCAI, pages 4313–4321. ijcai.org, 2020.
  • Chen et al. [2019a] Long Chen, Wujing Zhan, Wei Tian, Yuhang He, and Qin Zou. Deep integration: A multi-label architecture for road scene recognition. IEEE Trans. Image Process., 28(10):4883–4898, 2019.
  • Chen et al. [2019b] Zhao-Min Chen, Xiu-Shen Wei, Peng Wang, and Yanwen Guo. Multi-label image recognition with graph convolutional networks. In CVPR, pages 5177–5186. Computer Vision Foundation / IEEE, 2019.
  • Demsar [2006] Janez Demsar. Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res., 7:1–30, 2006.
  • Durand et al. [2019] Thibaut Durand, Nazanin Mehrasa, and Greg Mori. Learning a deep convnet for multi-label classification with partial labels. In IEEE Conf. Comput. Vis. Pattern Recog., pages 647–657. IEEE, 2019.
  • Feng et al. [2020] Lei Feng, Takuo Kaneko, Bo Han, Gang Niu, Bo An, and Masashi Sugiyama. Learning with multiple complementary labels. In ICML, volume 119, pages 3072–3081. PMLR, 2020.
  • Ghosh et al. [2017] Aritra Ghosh, Himanshu Kumar, and P. S. Sastry. Robust loss functions under label noise for deep neural networks. In AAAI, pages 1919–1925. AAAI Press, 2017.
  • Gong et al. [2022] Xiuwen Gong, Dong Yuan, and Wei Bao. Partial multi-label learning via large margin nearest neighbour embeddings. In AAAI, pages 6729–6736. AAAI Press, 2022.
  • He et al. [2019] Shuo He, Ke Deng, Li Li, Senlin Shu, and Li Liu. Discriminatively relabel for partial multi-label learning. In ICDM, pages 280–288. IEEE, 2019.
  • Huiskes and Lew [2008] Mark J. Huiskes and Michael S. Lew. The MIR flickr retrieval evaluation. In SIGMM, pages 39–43. ACM, 2008.
  • Huynh and Elhamifar [2020] Dat Huynh and Ehsan Elhamifar. Interactive multi-label CNN learning with partial labels. In IEEE Conf. Comput. Vis. Pattern Recog., pages 9420–9429. IEEE, 2020.
  • Iscen et al. [2017] Ahmet Iscen, Giorgos Tolias, Yannis Avrithis, Teddy Furon, and Ondrej Chum. Efficient diffusion on region manifolds: Recovering small objects with compact CNN representations. In IEEE Conf. Comput. Vis. Pattern Recog., pages 926–935. IEEE Computer Society, 2017.
  • Johnson et al. [2017] Jeff Johnson, Matthijs Douze, and Hervé Jégou. Billion-scale similarity search with gpus. arXiv preprint arXiv:1702.08734, 2017.
  • Lai et al. [2016] Hanjiang Lai, Pan Yan, Xiangbo Shu, Yunchao Wei, and Shuicheng Yan. Instance-aware hashing for multi-label image retrieval. IEEE Trans. Image Process., 25(6):2469–2479, 2016.
  • Lee et al. [2018] Chung-Wei Lee, Wei Fang, Chih-Kuan Yeh, and Yu-Chiang Frank Wang. Multi-label zero-shot learning with structured knowledge graphs. In IEEE Conf. Comput. Vis. Pattern Recog., pages 1576–1585. IEEE, 2018.
  • Li and Wang [2020] Ximing Li and Yang Wang. Recovering accurate labeling information from partially valid data for effective multi-label learning. In IJCAI, pages 1373–1380. ijcai.org, 2020.
  • Li et al. [2020] Ziwei Li, Gengyu Lyu, and Songhe Feng. Partial multi-label learning via multi-subspace representation. In IJCAI, pages 2612–2618. ijcai.org, 2020.
  • Liu et al. [2022] Weiwei Liu, Haobo Wang, Xiaobo Shen, and Ivor W. Tsang. The emerging trends of multi-label learning. IEEE Trans. Pattern Anal. Mach. Intell., 44(11):7955–7974, 2022.
  • Lyu et al. [2020] Gengyu Lyu, Songhe Feng, and Yidong Li. Partial multi-label learning via probabilistic graph matching mechanism. In KDD. ACM, 2020.
  • Niu et al. [2019] Xuesong Niu, Hu Han, Shiguang Shan, and Xilin Chen. Multi-label co-regularization for semi-supervised facial action unit recognition. In Adv. Neural Inform. Process. Syst., pages 907–917, 2019.
  • Qiu et al. [2018] Jiezhong Qiu, Jian Tang, Hao Ma, Yuxiao Dong, Kuansan Wang, and Jie Tang. Deepinf: Social influence prediction with deep learning. In KDD. ACM, 2018.
  • Read et al. [2011] Jesse Read, Bernhard Pfahringer, Geoff Holmes, and Eibe Frank. Classifier chains for multi-label classification. Mach. Learn., 85(3):333–359, 2011.
  • Shen et al. [2018] Xiaobo Shen, Weiwei Liu, Ivor W. Tsang, Quan-Sen Sun, and Yew-Soon Ong. Multilabel prediction via cross-view search. IEEE Trans. Neural Networks Learn. Syst., 29(9):4324–4338, 2018.
  • Shi et al. [2020] Wanli Shi, Victor S. Sheng, Xiang Li, and Bin Gu. Semi-supervised multi-label learning from crowds via deep sequential generative model. In KDD, pages 1141–1149. ACM, 2020.
  • Sun et al. [2019] Lijuan Sun, Songhe Feng, Tao Wang, Congyan Lang, and Yi Jin. Partial multi-label learning by low-rank and sparse decomposition. In AAAI, pages 5016–5023. AAAI Press, 2019.
  • Tan et al. [2022] Anhui Tan, Jiye Liang, Wei-Zhi Wu, and Jia Zhang. Semi-supervised partial multi-label classification via consistency learning. Pattern Recognit., 131:108839, 2022.
  • Veit et al. [2017] Andreas Veit, Neil Alldrin, Gal Chechik, Ivan Krasin, Abhinav Gupta, and Serge J. Belongie. Learning from noisy large-scale datasets with minimal supervision. In IEEE Conf. Comput. Vis. Pattern Recog., pages 6575–6583. IEEE Computer Society, 2017.
  • Wang et al. [2019] Haobo Wang, Weiwei Liu, Yang Zhao, Chen Zhang, Tianlei Hu, and Gang Chen. Discriminative and correlative partial multi-label learning. In IJCAI, pages 3691–3697. ijcai.org, 2019.
  • Wei et al. [2018] Tong Wei, Lan-Zhe Guo, Yu-Feng Li, and Wei Gao. Learning safe multi-label prediction for weakly labeled data. Mach. Learn., 107(4):703–725, 2018.
  • Wei et al. [2021] Tong Wei, Jiang-Xin Shi, and Yu-Feng Li. Probabilistic label tree for streaming multi-label learning. In KDD, pages 1801–1811. ACM, 2021.
  • Wei et al. [2022] Tong Wei, Zhen Mao, Jiang-Xin Shi, Yu-Feng Li, and Min-Ling Zhang. A survey on extreme multi-label learning. CoRR, abs/2210.03968, 2022.
  • Wu et al. [2021] Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and Philip S. Yu. A comprehensive survey on graph neural networks. IEEE Trans. Neural Networks Learn. Syst., 32(1):4–24, 2021.
  • Xie and Huang [2018] Ming-Kun Xie and Sheng-Jun Huang. Partial multi-label learning. In AAAI, pages 4302–4309. AAAI Press, 2018.
  • Xie and Huang [2020] Ming-Kun Xie and Sheng-Jun Huang. Semi-supervised partial multi-label learning. In ICDM, pages 691–700. IEEE, 2020.
  • Xie and Huang [2022] Ming-Kun Xie and Sheng-Jun Huang. Partial multi-label learning with noisy label identification. IEEE Trans. Pattern Anal. Mach. Intell., 44(7):3676–3687, 2022.
  • Xu et al. [2020] Ning Xu, Yun-Peng Liu, and Xin Geng. Partial multi-label learning with label distribution. In AAAI, pages 6510–6517. AAAI Press, 2020.
  • Xu et al. [2021] Ning Xu, Yun-Peng Liu, Yan Zhang, and Xin Geng. Progressive enhancement of label distributions for partial multilabel learning. IEEE Trans. Neural Networks Learn. Syst., pages 1–12, 2021.
  • Xu et al. [2022] Ning Xu, Congyu Qiao, Jiaqi Lv, Xin Geng, and Min-Ling Zhang. One positive label is sufficient: Single-positive multi-label learning with label enhancement. CoRR, abs/2206.00517, 2022.
  • Yan and Guo [2019] Yan Yan and Yuhong Guo. Adversarial partial multi-label learning. CoRR, abs/1909.06717, 2019.
  • Yan et al. [2021] Yan Yan, Shining Li, and Lei Feng. Partial multi-label learning with mutual teaching. Knowl. Based Syst., 212:106624, 2021.
  • Yeh et al. [2017] Chih-Kuan Yeh, Wei-Chieh Wu, Wei-Jen Ko, and Yu-Chiang Frank Wang. Learning deep latent space for multi-label classification. In AAAI, pages 2838–2844. AAAI Press, 2017.
  • Yilmaz et al. [2021] Selim F. Yilmaz, E. Batuhan Kaynak, Aykut Koç, Hamdi Dibeklioğlu, and Suleyman Serdar Kozat. Multi-label sentiment analysis on 100 languages with dynamic weighting for label imbalance. IEEE Trans. Neural Networks Learn. Syst., pages 1–13, 2021.
  • Yu et al. [2014] Hsiang-Fu Yu, Prateek Jain, Purushottam Kar, and Inderjit S. Dhillon. Large-scale multi-label learning with missing labels. In ICML, volume 32 of JMLR Workshop and Conference Proceedings, pages 593–601. JMLR.org, 2014.
  • Yu et al. [2018] Guoxian Yu, Xia Chen, Carlotta Domeniconi, Jun Wang, Zhao Li, Zili Zhang, and Xindong Wu. Feature-induced partial multi-label learning. In ICDM, pages 1398–1403. IEEE Computer Society, 2018.
  • Zhang and Fang [2021] Min-Ling Zhang and Jun-Peng Fang. Partial multi-label learning via credible label elicitation. IEEE Trans. Pattern Anal. Mach. Intell., 43(10):3587–3599, 2021.
  • Zhang and Zhou [2007] Min-Ling Zhang and Zhi-Hua Zhou. ML-KNN: A lazy learning approach to multi-label learning. Pattern Recognit., 40(7):2038–2048, 2007.
  • Zhang and Zhou [2014] Min-Ling Zhang and Zhi-Hua Zhou. A review on multi-label learning algorithms. IEEE Trans. Knowl. Data Eng., 26(8):1819–1837, 2014.
  • Zhu et al. [2018] Yue Zhu, James T. Kwok, and Zhi-Hua Zhou. Multi-label learning with global and local label correlation. IEEE Trans. Knowl. Data Eng., 30(6):1081–1094, 2018.

Appendix A Theoretical Analysis

A.1 Complexity Analysis

It is worth noting that the propagation step can be efficient. First of all, we have disentangled the propagation step from the deep model training, which allows us to perform simple algebra operations. Moreover, since 𝒁\bm{Z} is close to both 𝒀^\hat{\bm{Y}} and 𝒀\bm{Y}, either one can be used to initialize 𝒁\bm{Z} and only a few gradient steps are required to get a satisfactory solution.

Theoretically, in Eq. (3), calculating the term 𝒁𝑳y\bm{Z}\bm{L}_{y} requires 𝒪(nL2)\mathcal{O}(nL^{2}) time. To compute 𝑳x𝒁\bm{L}_{x}\bm{Z}, since the instance graph is highly sparse, we need 𝒪(nnnz(𝑳x))=𝒪(n(n+nnz(𝑨x)))\mathcal{O}(n*\text{nnz}(\bm{L}_{x}))=\mathcal{O}(n*(n+\text{nnz}(\bm{A}_{x}))), where nnz()\text{nnz}(\cdot) denotes the number of non-zero entries in a matrix. In our data-driven graph-construction case, we have nnz(𝑨x)=nk\text{nnz}(\bm{A}_{x})=nk. Therefore, at each epoch, we require 𝒪(nT(L2+n(k+1)))\mathcal{O}(nT(L^{2}+n(k+1))) time in the graph propagation procedure, where TT is the number of gradient steps. In our implementation, we use the widely-used matrix computation package Scipy to perform sparse matrix multiplication 𝑳x𝒁\bm{L}_{x}\bm{Z}. The remaining calculations, including the deep model training, are done on GPUs using PyTorch. According to our empirical analysis, the graph propagation procedure is indeed as fast as the deep training and be efficiently computed.

A.2 Convergence Analysis

It is also of interest to see the convergence performance of PLAIN method. Take the Probabilistic MSE risk as an example, the overall objective can be formulated as follows,

(θ,𝒁)=𝒁σ(f(𝑿;θ))F2+𝒥(𝒁,𝒀)\begin{split}\mathcal{L}(\theta,\bm{Z})=||\bm{Z}-\sigma(f(\bm{X};\theta))||_{F}^{2}+\mathcal{J}(\bm{Z},\bm{Y})\end{split}

with a constant factor ignored on the deep model objective. We further assume that the gradient descent algorithm is used as the OPTIMIZE()\text{OPTIMIZE}(\cdot) function in Algorithm 1 and the learning rates are selected properly. The following Proposition ensures the convergence of our method and proves the value of the objective function is non-increasing at each updating step.

Table 5: Characteristics of the experimental datasets. The last three PML datasets are real-world datasets.
Datasets EXPs FEAs CLs M-GT A-GT DOM
Emotions 593 72 6 3 1.87 music
Birds 645 260 19 6 1.01 audio
Medical 978 1,449 45 3 1.25 text
Image 2,000 294 5 3 1.23 image
Corel5K 5,000 499 374 5 3.52 image
Bibtext 7,395 1,836 159 28 2.40 text
Eurlex-dc 19,348 5,000 412 7 1.29 text
Eurlex-sm 19,348 5,000 201 12 2.21 text
NUS-WIDE 133,441 500 81 20 1.76 image
Music-emotion 6,833 98 11 7 2.42 music
Music-style 6,839 98 10 4 1.44 music
Mirflickr 10,433 100 7 5 1.77 image
  • For each PML dataset, the number of examples (EXPs), features (FEAs), class labels (CLs), the maximum number of ground-truth labels (M-GT), the average number of ground-truth labels (A-GT) and its corresponding domain (DOM) are recorded.

  • The original number of instances is 269,648 but some of them are unlabeled w.r.t the 81 class labels, thus we only utilized the remaining 133,441 instances to conduct experiments.

Proposition 1.

The updating rules listed in Algorithm 1 guarantee finding a stationary point of the objective (θ,𝐙)\mathcal{L}(\theta,\bm{Z}).

Proof.

Let (θt,𝒁t)(\theta_{t},\bm{Z}_{t}) be the solution at the tt-th epoch and (θt,𝒁t)\mathcal{L}(\theta_{t},\bm{Z}_{t}) be the corresponding objective function value. In the propagation step, θt\theta_{t} is fixed and the remaining objective, i.e. Eq. (2), is a standard convex optimization problem. Thus, our gradient descent algorithm guarantees to decrease the objective function value. Next, in the deep model training step, 𝒁t\bm{Z}_{t} is fixed and the objective in Eq. (4) is non-convex. However, the gradient descent algorithm still decreases the objective until a first-order stationary point is achieved. We can conclude that,

(θt,𝒁t)(θt,𝒁t1)(θt1,𝒁t1).\begin{split}\mathcal{L}(\theta_{t},\bm{Z}_{t})\leq\mathcal{L}(\theta_{t},\bm{Z}_{t-1})\leq\mathcal{L}(\theta_{t-1},\bm{Z}_{t-1}).\end{split}

Consequently, the objective function (θ,𝒁)\mathcal{L}(\theta,\bm{Z}) will monotonically decrease and finally converge to a stationary point. ∎

Appendix B Additional Experiments and Implementation Details

In this section, we present additional empirical results. All the computations are carried out on a workstation with an Intel E5-2680 v4 CPU, a Quadro RTX 5000 GPU, and 500GB main memory running the Linux platform. We summarize the characteristics of the used datasets in Table 5.

Table 6: Friedman statistics τF\tau_{F} in terms of each evaluation metric.
Metrics τF\tau_{F} Critical Value
Ranking Loss 31.88 2.15 (Methods: 7, Data sets: 30)
Average Precision 49.26
Hamming Loss 2.87
Refer to caption
Figure 4: Comparison of PLAIN (control algorithm) against six comparing algorithms with the Bonferroni-Dunn test. Algorithms not connected with PLAIN in the CD diagram are considered to have a significantly different performances from the control algorithm. (CD = 1.47 at 0.05 significance level)

B.1 Statistical Tests

Following previous works Lyu et al. [2020]; Zhang and Fang [2021], to comprehensively evaluate the superiority of PLAIN, we further utilized Friedman test Demsar [2006] as the statistical test to analyze the relative performance among the competing methods. The experimental results are reported in Table 6. At a 0.05 significance level, the null hypothesis of indistinguishable performance of PLAIN among all competing methods is clearly rejected. Subsequently, we employ the Bonferroni-Dunn test as the posthoc test by regarding PLAIN as the control approach. Figure 4 reports the CD diagrams on each evaluation metric, where the average rank of the competing approaches is marked along the axis. The performance of the control method and one learning method is deemed to be significantly different if their average ranks differ by at least one CD. According to Figure 4, we can observe that PLAIN achieves highly superior results to other baseline methods.

B.2 Further Analysis

Convergence Curve

We also conducted experiments on Music-emotion to show the convergence curve of the objective function. The results in Figure 5 (a) further support our proposition. Besides, the experimental results in Figure 2 (c) also agree with our theoretical findings. When the binary cross-entropy loss is used, our deep model is trained in an asymmetric fashion, and then there is no overall objective for our algorithm. However, as shown in Figure 2 (c), our method still tends to converge to a good solution.

Refer to caption
(a) Convergence curve
Refer to caption
(b) Parameter sensitivity of TT
Figure 5: (a) The convergence curve of objective function value (θ,𝒁)\mathcal{L}(\theta,\bm{Z}) on Music-emotion dataset. (b) Changes in the performance of PLAIN as the number of nearest neighbors kk changes.
Table 7: Comparison of PLAIN with two representative deep PML methods APML and PML-MT on three real-world datasets. The best results are shown in bold face.
Datasets PLAIN APML PML-MT
Ranking Loss (the lower the better)
Music-emotion 0.215±\pm0.005 0.242±\pm0.007 0.236±\pm0.007
Music-style 0.134±\pm0.007 0.145±\pm0.006 -
Mirflickr 0.088±\pm0.008 0.124±\pm0.014 0.126±\pm0.016
Average Precision (the higher the better)
Music-emotion 0.664±\pm0.007 0.621±\pm0.013 0.627±\pm0.010
Music-style 0.743±\pm0.009 0.732±\pm0.010 -
Mirflickr 0.845±\pm0.013 0.777±\pm0.027 0.807±\pm0.034
Hamming Loss (the lower the better)
Music-emotion 0.192±\pm0.003 0.200±\pm0.004 0.207±\pm0.011
Music-style 0.114±\pm0.004 0.115±\pm0.002 -
Mirflickr 0.166±\pm0.003 0.170±\pm0.003 0.173±\pm0.011
  • \sharp

    The results of PML-MT on Music-style are not reported in the corresponding paper.

Comparing with Deep PML Methods

To demonstrate the effectiveness of PLAIN, we further compare with two state-of-the-art deep PLL methods APML Yan and Guo [2019] and PML-MT Yan et al. [2021] on three real-world datasets according to the reported results in their papers. Since their experimental settings are different from our work, we follow Yan and Guo [2019]; Yan et al. [2021] to split the datasets to 80% training and 20% testing. We then rerun the proposed PLAIN on these datasets and the experimental results are reported in Table 7. We can observe that PLAIN outperforms APML and PML-MT on all datasets and all evaluation metrics. For instance, on the Music-emotion, Music-style, and Mirflicker datasets, in terms of average precision, PLAIN achieves results superior to those of the best-competing methods by 5.90%, 1.50% and 4.71% respectively. These results demonstrate that the graph technique improves the disambiguation ability of deep PML models and further confirms the effectiveness of our proposed PLAIN.

Comparing with Two-stage Variant

To show the superiority of our end-to-end design, we further experiment with a two-stage variant of PLAIN which first propagates labels on the graph until converges and then trains the deep model to fit the pseudo-labels. The results are shown in 8. We found the two-stage method underperforms PLAIN because it fails to disambiguate labels in the first stage, making deep model overfits. In contrast, PLAIN incorporates accurate model prediction into propagation for refined pseudo-labels. We will add complete results in the revision.

Table 8: Comparison of PLAIN with two-stage learning variant.
Metrics Rloss\downarrow AP\uparrow Hloss\downarrow
Two-Stage 0.236 0.637 0.201
PLAIN 0.217 0.664 0.191
Parameter Analysis of kk

Furthermore, we study the sensitivity analysis of PLAIN with respect to the critical parameter kk. In Figure 5 (b), we show the performance of PLAIN changes as kk increases from 5 to 50 on three datasets. We can observe that PLAIN is robust in terms of the parameter kk and thus, we empirically fix k=10k=10 in our experiments.

Table 9: List of notations.
Data
nn Number of data examples
d,Ld,L Dimensions of features and labels
rr Average number of false-positive labels in datasets
𝒟\mathcal{D} Training dataset
𝒳,𝒴\mathcal{X},\mathcal{Y} Feature space and label space
S,S~S,\tilde{S} Candidate label set and true label set
𝒚,𝒚~,𝒚^\bm{y},\tilde{\bm{y}},\hat{\bm{y}} Candidate label vector, true label vector, and predicted label vector
𝒀,𝒀^\bm{Y},\hat{\bm{Y}} Candidate label matrix and predicted label matrix
𝒁\bm{Z} Pseudo-Label matrix
Algorithm
𝒩k(𝒙)\mathcal{N}_{k}(\bm{x}) Indices of kk-nearest neighbors of 𝒙\bm{x}
𝑺\bm{S} Sparse instance affinity matrix
𝑨x,𝑨y\bm{A}^{x},\bm{A}^{y} Instance and label graphs
𝑫x,𝑫x\bm{D}_{x},\bm{D}_{x} Diagonal instance and label degree matrices
𝑳x,𝑳y\bm{L}_{x},\bm{L}_{y} Instance and label Laplacian matrices
𝑰\bm{I} Identity matrix
Function
f()f(\cdot) Neural network
θ\theta Parameters of the neural network
𝒥x,𝒥y\mathcal{J}_{x},\mathcal{J}_{y} Instance and label graph regularizers
𝒥\mathcal{J} Overall propagation objective
σ()\sigma(\cdot) Sigmoid function
l()l(\cdot) Loss function
LP,Deep\mathcal{L}_{\text{LP}},\mathcal{L}_{\text{Deep}} Overall objectives of propagation and deep model training
tr()\text{tr}(\cdot) Trace of a matrix
𝕀()\mathbb{I}(\cdot) Indicator function for equivalance testing
Hyper-parameter
kk Number of nearest neighbors
ρ\rho Weights of graph similarity
η,α,β\eta,\alpha,\beta Loss weighting factors of candidate consistency, instance regularizer, and label regularizer
γ\gamma Learning rate of gradient descent for label propagation
TT Number of propagation gradient updating steps

Appendix C Notations and Terminology

The notations are summarized in Table 9.