Fairness-aware Model-agnostic Positive and Unlabeled Learning
Abstract.
With the increasing application of machine learning in high-stake decision-making problems, potential algorithmic bias towards people from certain social groups poses negative impacts on individuals and our society at large. In the real-world scenario, many such problems involve positive and unlabeled data such as medical diagnosis, criminal risk assessment and recommender systems. For instance, in medical diagnosis, only the diagnosed diseases will be recorded (positive) while others will not (unlabeled). Despite the large amount of existing work on fairness-aware machine learning in the (semi-)supervised and unsupervised settings, the fairness issue is largely under-explored in the aforementioned Positive and Unlabeled Learning (PUL) context, where it is usually more severe. In this paper, to alleviate this tension, we propose a fairness-aware PUL method named FairPUL. In particular, for binary classification over individuals from two populations, we aim to achieve similar true positive rates and false positive rates in both populations as our fairness metric. Based on the analysis of the optimal fair classifier for PUL, we design a model-agnostic post-processing framework, leveraging both the positive examples and unlabeled ones. Our framework is proven to be statistically consistent in terms of both the classification error and the fairness metric. Experiments on the synthetic and real-world data sets demonstrate that our framework outperforms state-of-the-art in both PUL and fair classification.
1. Introduction
Nowadays, machine learning systems are assisting, or in some cases even replacing, human decision making in an increasing number of application domains. Due to the profound impacts of these systems on individuals and our society at large, traditional performance metrics such as accuracy and precision are no longer the sole measure of success. In applications such as credit approval (Twala, 2010), medical diagnosis (Char et al., 2018), criminal risk assessment (Perry, 2013) and recommender systems (Sweeney, 2013; Wei and He, 2022), fairness must be carefully taken into account to ensure the absence of discrimination against certain social groups (e.g., women, blacks). Recent years have witnessed a growing interest in fairness-aware machine learning to study the fairness issue. Various metrics of fairness for a predictive model have been studied in the literature (Chouldechova and Roth, 2018), including group fairness (Chouldechova, 2017; Zafar et al., 2017; Hardt et al., 2016), individual fairness (Jung et al., 2019; Dwork et al., 2012, 2020) and causal fairness (Kusner et al., 2017; Wu et al., 2019). Depending on the amount of label information available during training, researchers have developed a variety of algorithms to address the unfairness issue. For example, (Zafar et al., 2017; Hardt et al., 2016; Zemel et al., 2013; Song et al., 2019) focused on the traditional supervised learning setting where the learning algorithms have access to the class labels of all training examples; (Chierichetti et al., 2017; Kleindessner et al., 2019; Chen et al., 2019) studied the unsupervised learning setting where no label information is available in the training set; and (Chzhen et al., 2019; Zhang et al., 2020) mainly designed algorithms for the semi-supervised learning setting where only a small portion of the training examples contain class labels.
In many real-world applications, the training data consists of only positive labeled examples and unlabeled ones, which cannot naturally fit in any of the aforementioned learning settings. For example, medical records usually only list which diseases a patient has been diagnosed with (i.e., positive samples) and they usually do not include which diseases a patient does not have. However, the absence of a diagnosis does not mean that the patient does not have the disease. Another example is scoring recidivism. When predicting criminal defendants’ likelihood of re-offending, it is easy for us to collect part of the positive samples from the criminal records. For those who have not yet appeared in the criminal records, it is wrong to assume that they will not re-offend. They should therefore not be treated as negative examples but as unlabeled ones. In general, positive and unlabeled learning (PUL) (Li and Liu, 2005) attempts to learn a classifier from this type of data. Existing (semi-)supervised and unsupervised methods can neither deal with positive-only labeled data nor make a full use of unlabeled data, and thus different PUL methods have been proposed (Bekker and Davis, 2020).
Despite the large amount of existing work on PUL (Elkan and Noto, 2008; du Plessis et al., 2014; Kiryo et al., 2018; Dhurandhar and Gurumoorthy, 2020), the fairness issue in this setting has been largely under-explored. In applications such as medical diagnosis and recidivism scoring mentioned before, existing techniques can easily lead to inequality among different groups (e.g., black and white, female and male). The difference between fairness in PUL and in the other settings can be summarized in the following three folds: (1) Intuitively, the label imbalance will exacerbate the unfairness problem in PUL. (2) Empirically, methods proposed from supervised and semi-supervised settings cannot handle the unfairness issue in PUL as shown in Table 1. (3) Theoretically, the properties (e.g. consistency) of other methods no longer hold in PUL due to the lack of negatively labeled data. Therefore, in this paper, we aim to bridge this gap and study the fairness-aware binary classification problem in PUL. To be more specific, we seek a classifier which minimizes the misclassification risk in PUL while satisfying the Equalized Odds / Equal Opportunity (Hardt et al., 2016) fairness constraint. We derive the optimal fair classifier via recalibration of the Bayes regressor. This theoretical result motivates us to devise a generic post-processing framework named FairPUL. Based on both positive and unlabeled examples, FairPUL estimates the regression function and the unknown threshold to achieve the fairness criterion in the PUL setting. It enjoys the consistency property where it asymptotically satisfies the fairness criterion and its risk converges to the one of the theoretical optimal fair classifier. Extensive experiments on both the synthetic and the real-world data sets demonstrate FairPUL’s effectiveness. Our main contributions can be summarized as follows:
-
•
To the best of our knowledge, this is the first work to systematically study the fairness issue in the positive and unlabeled learning setting.
-
•
We derive the optimal fair classifier in PUL and propose a model-agnostic post-processing framework, which can accommodate different base models and enjoys the consistency property.
-
•
Experiments on the synthetic and real-world data sets show that our framework performs favorably against state-of-the-art in both PUL and fair classification.
2. Related Work
2.1. Positive and Unlabeled Learning
Positive and Unlabeled Learning is a variant of the classical classification setup where the training data consists of only positive and unlabeled (PU) examples. It fits within the long-standing interest in developing learning algorithms that do not require fully supervised data. The PU data can originate from two scenarios. One is called the single-training-set scenario (Elkan and Noto, 2008) where the data come from one single training set; the other is called the case-control scenario (Ward et al., 2009) where the data come from two independently drawn datasets, one with all positive examples and one with all unlabeled examples. The single-training-set scenario has received substantially more attention in the literature (Bekker and Davis, 2020). Most PUL methods can be divided into the following three categories. Two-step techniques (He et al., 2018; Ienco and Pensa, 2016) first identify reliable negative examples and then conduct classical classification. Biased learning (Shao et al., 2015; Hsieh et al., 2015; Claesen et al., 2015) considers PU data as fully labeled data with class label noise for the negative class. Class prior incorporation (Elkan and Noto, 2008; Kiryo et al., 2018; du Plessis et al., 2014) modifies standard learning methods using the class prior and learns weights for all examples. However, none of the existing work in PUL explores the fairness issue. As shown in our experiments (Subsection 5.2), current PUL methods exhibit unfairness against subgroups, which will result in potential discrimination in real-world applications.
2.2. Fairness in Classification
The existing studies on fairness in classification have focused on two key issues: how to formalize the fairness metric in classification, and how to design efficient algorithms that strike a desirable trade-off between classification performance and fairness. To seek equality between different populations, a number of works have been proposed to control group fairness. Equal Opportunity (Hardt et al., 2016) requires the same true positive rates among groups and Equalized Odds (Berk et al., 2018) simultaneously considers false positive rates. Equalizing Disincentives (Feldman et al., 2015) requires the difference of two metrics to be equal across the groups. Using these metrics, a variety of fairness-aware algorithms in classification have been proposed under different learning settings. (Zafar et al., 2017; Hardt et al., 2016; Zemel et al., 2013; Agarwal et al., 2018; Song et al., 2019) focused on the supervised learning setting where the learning algorithms have access to the class labels of all training examples. Since they highly rely on supervised data, even if we adjust their decision boundary with a standard PUL technique, the unlabeled data still cannot contribute to algorithmic fairness explicitly. In a more realistic scenario where both labeled and unlabeled examples are given, (Bird et al., 2020) devised a linear transformation pre-processing technique to remove the underlying discrimination. (Zhang et al., 2020) incorporated the fairness constraint into the original training process. (Chzhen et al., 2019) post-processed the output conditional probability to improve fairness by recalibration with the unlabeled data. Although these methods do not rely on labeled data as much as the fully supervised methods, both positive and negative labeled samples are necessary for their algorithms. The most related work to us is (Chzhen et al., 2019) where they limit the fairness criterion to Equal Opportunity. However, the unlabeled examples can only be used to estimate the fairness threshold in their method. How to properly use the positive-only and unlabeled data in fair classification is still under-explored. In this paper, we instead explore how to make the best use of PU data to boost both classification performance and algorithmic fairness.
3. Problem Definition
We consider a fairness-aware binary classification task in the single-training-set scenario (Elkan and Noto, 2008) as explained in Subsection 2.1. The training examples consist of tuples where is a feature vector, is a binary sensitive attribute, and represents whether the example is labeled () or not. Notice that we focus on one binary sensitive attribute in this paper, although the proposed technique can be naturally extended to multiple multi-class sensitive attributes (details in Subsection 4.2). The training examples are drawn randomly from distribution , where is the ground-truth label. But for each tuple that is drawn, only is observed. A classifier receives a pair as input, and outputs a binary prediction for the label. The set of all such functions from to is denoted by . For any classifier , we denote its associated misclassification risk as . An optimal fair classifier is then defined as:
Various definitions of fairness have been proposed so far (Chouldechova and Roth, 2018), but there is no consensus regarding which definition is universally the most appropriate. In this work, we employ the following group fairness metrics Equalized Odds and Equal Opportunity introduced in (Hardt et al., 2016):
Definition 0 (Equalized Odds (EO)).
We say that a binary classifier satisfies equalized odds with respect to and if .
Definition 0 (Equal Opportunity (EOP)).
We say that a binary classifier satisfies equal opportunity with respect to and if .
In the two fairness metrics, EO requires the same true positive rates and the same false positive rates across the sensitive groups. Compared to EO, EOP is a weaker notion which only requires the same true positive rates. The two fairness metrics have been used extensively in the literature either as a post-processing step (Hardt et al., 2016) on a learned classifier or directly during training (Donini et al., 2018). The motivation of these metrics and more discussion regarding the comparison with other fairness metrics can be found in (Hardt et al., 2016; Agarwal et al., 2018; Menon and Williamson, 2018).
4. The Proposed FairPUL Framework
In this section, we introduce the model-agnostic post-processing framework FairPUL for fairness-aware PUL.
4.1. Labeling Mechanism in PUL
As stated before, the unlabeled samples in PUL mainly come from the following two sources: (1) It is truly a negative example; (2) It is a positive example, but simply was not selected by the labeling mechanism. In order to enable learning with PU data, it is necessary (Bekker and Davis, 2020) to make assumptions about either the labeling mechanism, the class distributions in the data, or both. We base our work on the most frequently used assumption (Lee and Liu, 2003) in PUL:
Assumption 1 (Selected Completely At Random (SCAR)).
Labeled examples are selected completely at random, independent from their attributes, from the positive distribution. Formally speaking, , where is called label frequency.
Under this assumption, the set of labeled examples are i.i.d. samples from the positive distribution. The label frequency plays an important role in the single-training-set scenario (Elkan and Noto, 2008) we consider here. Define the regression function , i.e., the probability of an example being labeled. Then we have the following lemma regarding the constant :
Lemma 0.
Suppose that the SCAR assumption holds. Then .
The proof can be found in Appendix A. This lemma shows how we can obtain the probability of an example being positive from . Next, we further incorporate fairness into the classification problem and show how to get the optimal fair classifier theoretically and empirically.
4.2. Optimal Fair Classifier in PUL
To obtain the optimal fair classifier, we study the following problem:
(1) |
For EO, . When , the constraint requires the same true positive rates (TPR), that is , where we use the superscript to identify the sensitive attribute . Similarly, when , the constraint requires the same false positive rates (FPR), that is . For EOP, and it only requires . In our later derivation, we will focus on EO while discussing how our results can easily accommodate EOP.
Using the misclassification risk and based on Lemma 1, we can solve problem (1) and get the following result:
Lemma 0.
The minimizer for every Lagrange multipliers is:
(2) | ||||
Here, is the Lagrange multiplier for the same TPR constraint, and is for the same FPR constraint. Since , we can rewrite the Lagrange function of Problem (1) into expressions of and then get the above equations. The detailed derivation is shown in Appendix B. Note that by incorporating explicitly, our designed minimizer can boost performance and meanwhile reduce discrimination for PU data.
Remark. We consider one binary sensitive attribute in this problem for derivation simplicity. This can be naturally extended to accommodate multiple multi-class sensitive attributes (such as White, Black, Asian for race). We can break down the fairness into pairs of equations, add more equality constraints of EOD/EOP with respect to all sensitive attributes to the optimization problem in Equation (1), and obtain a similar solution as Equation (2).
With Lemma 2, the problem of finding the optimal fair classifier in PUL is equivalent to obtaining the optimal value of in Equation (2). To do this, we introduce an assumption on the regression function , i.e., the probability of an example being positive. For notation simplicity, we abbreviate it as .
Assumption 2.
The mapping is continuous on .
This Assumption requires that the random variable does not have atoms for each . This is proved achievable by many distributions (Yan et al., 2018; Sadinle et al., 2019; Denis and Hebiri, 2020). Based on this assumption, we introduce the optimal :
Lemma 0.
The optimal for satisfies
(3) | ||||
Proof Sketch.
Remark. The results in Lemma 2 are the minimizer under the fairness metric of EO, and Equations (3) show the constraints for the optimal under EO. To accommodate EOP, since we do not need to consider TNR, we can simply let , and the optimal only needs to satisfy the first equation in Equations (3). Notice that since for EOP, the first equation is only the constraint for the optimal .
4.3. FairPUL for Fairness-aware PUL
Based on the theoretical optimal classifier in Equations (2) and (3), we are now ready to introduce the proposed FairPUL framework for empirically estimating the optimal fair classifier. Given an existing classifier, FairPUL will modify its output for PUL data while reducing unfairness, and output the fair version of the classifier with a low misclassification risk. In the next subsections, any notation (e.g., ) with a hat denotes its estimated counterpart (e.g., ).
In the PUL setting, the training data consists of two parts, the labeled (positive) data set and the unlabeled data set . A classifier can be constructed to estimate the probability of an example being labeled by treating as positive examples and as negative ones. Let be the validation set that is drawn from the overall distribution in the same manner as the training set and be the subset of examples in that are labeled (and positive). We can estimate the label frequency as follows:
(4) |
where is the cardinality of . If is trained well enough, provides a precise estimate of labeling probability. Since every example in is positive, we have based on Total Probability Theorem and SCAR assumption. Although this ideal case hardly holds in practice, the average estimation is a good choice for estimating the label frequency with low variance.
Since the ideal should satisfy Equations (3), we minimize the difference between its two sides. Based on the rule of conditional probability, this is equivalent to minimizing unfairness defined as follows:
Definition 0 (Unfairness).
For a binary classifier , its unfairness under different fairness metrics can be defined as:
Under EO:
Under EOP:
We can get their empirical versions by substituting all the unknown terms with their empirical estimators:
(5) | ||||
Then we can obtain the estimator of as:
(6) |
where can be either or to satisfy EO or EOP respectively. is the estimated optimal fair classifier defined in a similar way as Equation (2) by replacing all the unknown terms with their empirical estimators.:
(7) | ||||
We adopt the Simulated Annealing strategy to search for the optimal .
Alg. 1 summarizes our proposed FairPUL framework. It takes the training and validation sets as well as a base model as input, and outputs a fairness-aware classifier with a low misclassification risk. The base regressor is trained with and . It can be any classifier that outputs the conditional probability of an example being labeled. First, an estimator of can be learned on the validation set with in Step 1. Then we use Simulated Annealing to search for the best to minimize unfairness for PUL data in Step 2 and 3. We finally compute the empirical optimal fair classifier accordingly in Step 4.
4.4. Consistency of FairPUL
Consistency is a desired property for a classifier in asymptotic theory. It guarantees that with the increase in the amount of data, the estimation will converge to the true value. In this subsection, we show that the proposed framework FairPUL is consistent, that is, it asymptotically satisfies the fairness criterion and its risk converges to the one of the theoretical optimal fair classifier.
First of all, following (Chzhen et al., 2019), we make the following realistic assumptions on the estimator of :
Assumption 3.
The estimator satisfies that, ,
-
•
as , where and denote the number of examples in and respectively.
-
•
There exists a sequence satisfying and such that almost surely.
-
•
The mapping is continuous on almost surely.
The first part of the assumption requires the estimator to be consistent in norm. This can be achieved by a variety of estimations for different regression functions as shown in the literature (Devroye, 1978; Audibert et al., 2007; Van de Geer et al., 2008). The second part means that is lower bounded by a certain positive term which vanishes as and go to infinity. This can be easily achieved by slightly modifying any existing consistent estimator. The last part is similar to Assumption 2.
Based on these realistic assumptions on , next we establish the statistical consistency of FairPUL.
Theorem 5 (Asymptotic properties).
FairPUL satisfies:
Proof Sketch.
To prove asymptotic optimality (the second part), we introduce an intermediate estimator and use it to upper bound the excess risk with triangle inequality. The upper bound converges to zero based on the first two parts in Assumption 3. For asymptotic fairness (the first part), we upper bound the unfairness with its empirical version which converges to zero. Detailed proof can be found in Appendix D. Recall that . Its consistency as is vital to the consistency of our framework in PUL. ∎
5. Experiments
To evaluate the effectiveness of FairPUL, we conduct extensive experiments to answer the following research questions:
-
•
RQ1: How does FairPUL perform compared with state-of-the-art in fair classification and PUL?
-
•
RQ2: How does FairPUL for post-processing compare with in-processing and pre-processing methods?
-
•
RQ3: How do unlabeled examples affect FairPUL?
5.1. Experiments on Synthetic Data.
5.1.1. Experimental setup
The aim of the synthetic experiment is to study the behavior of FairPUL in comparison with other methods with the base model of linear logistic regression (Lin.LR), in terms of both classification performance and fairness. To this end, we generate a synthetic binary classification data set with two sensitive groups following Donini et al. (Donini et al., 2018). For each group in the class 0 and for the group in the class 1, we generate 1,000 examples for training and the same number for testing. For the group in the class 1, we generate 200 examples for training and the same number for testing. Each set of examples is sampled from a 2-dimensional isotropic Gaussian distribution with different mean and variance : (i) Group , Label 1: = (-1, -1), = 0.8; (ii) Group , Label 0: = (1, 1), = 0.8; (iii) Group , Label 1: = (-0.5, -0.5), = 0.5; (iv) Group , Label 0: = (0.5, 0.5), = 0.5. When a standard machine learning method is applied to this synthetic data set, the generated model is unfair with respect to the group , in that the classifier tends to negatively classify the examples in this group. We search in for the best regularization parameter . We generate the validation set from the training set via holdout validation and the holdout ratio is set to 0.2.
5.1.2. Baselines
To study and compare the performance of FairPUL, we first adopt two classic methods as baselines:
-
•
Oracle: A method using the fully labeled training set (all examples are labeled either positive or negative). Its results correspond to the fully supervised setting.
-
•
Naïve: A method using the labeled examples as positive ones and treating unlabeled examples as negative.
Since this is the first work on the fairness issue in the PUL setting, we further compare FairPUL with the following two types of existing work. For fairness work:
-
•
Agarwal (Agarwal et al., 2018): An in-processing method reducing fair classification to cost-sensitive classification problems and yielding a randomized classifier with the lowest error subject to the desired constraints.
-
•
Hardt (Hardt et al., 2016): A post-processing method which takes as input an existing classifier and the sensitive feature, and derives a monotone transformation of the prediction to enforce the specified fairness constraints.
-
•
Chzhen (Chzhen et al., 2019): A post-processing method which recalibrates the Bayes classifier by a group-dependent threshold to minimize unfairness.
All the above fairness baselines adopt the similar fairness metric as we do. For PUL work:
-
•
uPU/wPU (Elkan and Noto, 2008): Unbiased PUL methods under the SCAR assumption by reweighting the examples. Both unweighted (uPU) and weighted (wPU) versions were proposed.
-
•
nnPU (Kiryo et al., 2018): A designed non-negative risk estimator for PUL which can be trained on flexible neural networks when minimized.
-
•
BaggingSVM (Mordelet and Vert, 2014): A bootstrap-bagging-based model which iteratively trains multiple classifiers to discriminate the known positive examples from random subsamples of the unlabeled set, and averages their predictions.
5.1.3. Results (RQ1)
We compare the test classification error and the fairness metrics under different labeling rates in of the baselines and our proposed FairPUL. For FairPUL with EO constraint, the fairness metric is AOD. For FairPUL with EOP constraint, the fairness metric is EOD. The base model for all the methods is Lin.LR. We report the average results of misclassification error, AOD and EOD of 5 independent trials in Fig. 1(b). The closer the dot is to the origin, the more fair and accurate the model is. From the figures we can see that for all the methods, their classification performances drop as the labeling rate decreases. Under high labeling rates (), our framework achieves much lower AOD and EOD (i.e., higher level of fairness) while maintaining a good level of accuracy; under low labeling rates, our framework achieves much better performance in both classification accuracy and fairness. Note that even though some methods achieve zero AOD/EOD when the labeling rate is 0.2, it is not an ideal model we are seeking. In this case, the model simply predicts every example to be negative, showing ”fake” fairness. What we are seeking is maintaining task-specific performance and reducing discrimination simultaneously.


Test classification error, AOD and EOD under different labeling rates.
5.2. Experiments on Real Data.
5.2.1. Data sets and experimental setup
To further study how FairPUL performs, we conduct experiments on three publicly available real-world data sets. In all the experiments, to obtain reliable estimates of classification performance and fairness, we repeatedly randomly split each data set into training (70%) and test (30%) sets 10 times, and report the averages and standard deviations of the metrics over different independent runs. Following (Gong et al., 2019), to realize the PUL setting, for each training set, we randomly select 90% positive examples as labeled and leave the remaining 10% positive examples as well as all negative examples as unlabeled. We split up the training set for holdout validation in FairPUL and the holdout ratio is set to 0.2.
-
•
COMPAS recidivism data (Angwin et al., 2016). It includes 5278 records with 47% positive examples. The task is to predict recidivism from someone’s criminal history, jail and prison time, demographics, and COMPAS risk scores, with race as the protected sensitive attribute restricted to black (about 40%) and white defendants.
-
•
German Credit (Lichman, 2013). It includes 1000 examples where 30% are positive. The task is to classify people as good or bad credit risks by features related to the economical situation, with gender as the sensitive attribute restricted to female (about 31%) and male.
-
•
Drug (Fehrman et al., 2017). It comprises 1885 records of human subjects, and for each subject, it provides five demographic features, seven features measuring personality traits and 18 features each of which describes the subject’s last use of a certain drug. We choose the use of heroin here, where 80% of the subjects have never used heroin. We restrict the sensitive attribute race to black (about 9%) and white.
Average results and standard deviations on real data of 10 runs. Models German COMPAS Drug F1 AOD EOD F1 AOD EOD F1 AOD EOD Oracle 0.5650.051 0.0530.033 0.0610.042 0.6040.018 0.1940.020 0.2190.032 0.7770.021 0.1300.024 0.0410.036 Naïve 0.4710.048 0.0680.024 0.0990.052 0.5290.058 0.1800.031 0.2370.045 0.7530.019 0.1420.031 0.0860.049 Agarwal 0.4700.033 0.0530.026 0.0770.051 0.4670.042 0.0320.021 0.0400.036 0.7430.031 0.1230.019 0.0650.033 Hardt 0.4490.058 0.0590.025 0.0930.046 0.4090.062 0.0340.023 0.0340.029 0.7560.022 0.1110.033 0.0750.042 Chzhen 0.3930.069 0.0480.018 0.0530.028 0.4860.047 0.0360.012 0.0320.020 0.7550.020 0.1080.024 0.0410.027 uPU 0.5740.022 0.0970.045 0.0540.027 0.6440.012 0.0310.024 0.0160.009 0.8710.026 0.1430.060 0.0660.043 wPU 0.5340.131 0.0500.022 0.0660.039 0.6120.071 0.1880.035 0.2450.048 0.7660.027 0.1470.016 0.0740.049 Bagging 0.5440.056 0.0540.034 0.0620.052 0.6490.013 0.2150.031 0.2520.038 0.8130.020 0.1960.041 0.0740.051 FairPUL-EO 0.5780.023 0.0450.018 0.0560.031 0.6460.012 0.0050.003 0.0040.002 0.8710.021 0.0560.023 0.0240.013 FairPUL-EOP 0.5800.020 0.0420.029 0.0340.027 0.6480.013 0.0060.004 0.0030.002 0.8970.023 0.0610.048 0.0220.010
We compare FairPUL with the baselines described in Subsection 5.1.2 with the base model of Linear Support Vector Machine (Lin.SVM). Since previous fairness works cannot be naturally adapted to the PUL setting, the fairness baselines we use here all follow the naïve method to transform PUL to the traditional learning setting. The hyper-parameters of every algorithm have been carefully tuned to achieve the best classification performance, and the details can be found in Appendix E.
5.2.2. Metrics
We compare our framework with the baselines using the metrics of F1 score for evaluating the model performance, and AOD and EOD for evaluating the fairness level. It is worth noting that most previous works on fairness simply use classification accuracy to evaluate the models’ performance. However, higher accuracy does not necessarily mean a better classification ability, especially on imbalanced data sets like German Credit and drug. Therefore, we use the F1 score instead.
5.2.3. Results (RQ1)
The comparison of baselines and FairPUL is summarized in Table 1. From the results we have the following observations:
-
•
The naïve and PUL baselines show high AOD and EOD, and most often higher than the oracle method. It demonstrates the unfairness problem in the PUL setting, which is usually more severe than in the supervised learning setting as we discussed before.
-
•
Fair classification methods often obtain lower AOD and EOD, and PUL methods often obtain higher F1 scores. However, none of them can perform well on both metrics. That is, existing fairness works cannot ensure model performance in the PUL setting, while existing PUL methods cannot ensure fairness.
-
•
FairPUL methods achieve both high F1 scores and low AOD and EOD. This demonstrates that our framework strikes a good trade-off between classification performance and fairness. FairPUL even beats the oracle model. This may be due to the imbalance of the data sets, where FairPUL resolves it via proper estimation of the label frequency.
-
•
FairPUL-EOP often achieves higher F1 scores than FairPUL-EO under comparable AOD and EOD. This is because EOP is a weaker fairness constraint compared with EO, and thus it typically allows for stronger task performance.
We report the results of the methods with different base models on COMPAS in Table 4 in Appendix F. Similar observations can be drawn, which demonstrate that our proposed FairPUL can always strike a good trade-off between classification performance and fairness, and generalize well to different base models.
Post-processing vs In-processing/Pre-processing. Rates Models F1 AOD EOD 100% Lin.SVM Naïve 0.6040.018 0.1940.020 0.2190.032 In 0.5880.014 0.0300.026 0.0390.021 Pre 0.5980.015 0.0760.025 0.0980.027 FairPUL-EO 0.6490.013 0.0040.003 0.0050.004 FairPUL-EOP 0.6500.014 0.0080.004 0.0040.002 Lin.LR Naïve 0.6210.015 0.2360.033 0.2930.045 In 0.5950.011 0.0420.022 0.0380.026 Pre 0.5900.016 0.0540.029 0.0470.027 FairPUL-EO 0.6570.024 0.0140.010 0.0130.006 FairPUL-EOP 0.6600.013 0.0180.005 0.0120.009 90% Lin.SVM Naïve 0.5290.058 0.1800.031 0.2370.045 In 0.4670.042 0.0320.021 0.0400.036 Pre 0.4290.054 0.0280.016 0.0380.029 FairPUL-EO 0.6460.012 0.0050.003 0.0040.002 FairPUL-EOP 0.6480.013 0.0060.004 0.0030.002 Lin.LR Naïve 0.5520.023 0.2400.048 0.3110.054 In 0.5070.019 0.0320.021 0.0370.036 Pre 0.4950.019 0.0490.030 0.0530.024 FairPUL-EO 0.6560.016 0.0150.008 0.0130.010 FairPUL-EOP 0.6580.012 0.0160.007 0.0090.008
Comparison results of different labeling rates. Rates Models F1 AOD EOD 100% Lin.SVM Naïve 0.6040.018 0.1940.020 0.2190.032 FairPUL 0.6500.014 0.0080.004 0.0040.002 Lin.LR Naïve 0.6210.015 0.2360.033 0.2930.045 FairPUL 0.6600.013 0.0180.005 0.0120.009 90% Lin.SVM Naïve 0.5290.058 0.1800.031 0.2370.045 FairPUL 0.6480.013 0.0060.004 0.0030.002 Lin.LR Naïve 0.5520.023 0.2400.048 0.3110.054 FairPUL 0.6580.012 0.0160.007 0.0090.008 80% Lin.SVM Naïve 0.2740.024 0.0900.011 0.1260.023 FairPUL 0.6460.012 0.0060.005 0.0040.003 Lin.LR Naïve 0.4010.018 0.1420.021 0.2050.031 FairPUL 0.6550.011 0.0150.007 0.0070.007 50% Lin.SVM Naïve 0.0000.000 0.0000.000 0.0000.000 FairPUL 0.6440.012 0.0090.007 0.0060.005 Lin.LR Naïve 0.0680.015 0.0220.009 0.0370.016 FairPUL 0.6520.014 0.0170.007 0.0080.004
5.3. Post-processing vs. In-/Pre-processing (RQ2)
In this section, we compare the post-processing FairPUL with the in-processing method proposed by (Agarwal et al., 2018) and the pre-processing method in (Bird et al., 2020). The in-processing method is introduced in Section 5.2. The pre-processing method is a widely-used technique which transforms the non-sensitive features to remove their correlation with the sensitive feature while retaining as much information as possible. We report the average results of 10 independent runs on COMPAS data set under labeling rates of 1.0 and 0.9 in Table 2. Results on German can be found in Appendix F. In the fully labeled setting, the in-processing and pre-processing methods sacrifice much classification performance for better fairness. FairPUL achieves a good balance instead. Under labeling rate 90%, FairPUL significantly outperforms all the baselines in both metrics. Compared with the in-processing method, FairPUL only needs black-box access to the predictions and sensitive attribute information without requiring access to the actual algorithms and ML models. While the pre-processing method only needs to transform the data set before the actual model takes effect, it leads to classifiers that still exhibit substantial unfairness in practice. So our proposed FairPUL is more flexible and also effective in fairness-aware PUL.
5.4. In-depth Study
5.4.1. Effect of labeling rates (RQ3)
We test labeling rates of [100%, 90%, 80%, 50%] on the base models Lin.SVM and Lin.LR. Changing the labeling rates affects the number of positive and unlabeled examples simultaneously. As we have observed before, FairPUL-EOP often achieves a better trade-off between fairness and classification compared with FairPUL-EO, so we only report the results of FairPUL-EOP here. Average results of 10 runs on COMPAS data set are shown in Table 3. Additional results on German are in Appendix F. From the table we can see that FairPUL always achieves higher F1 scores and lower AOD and EOD under different labeling rates compared with the naïve method. Although FairPUL’s classification performance drops with the decrease in labeling rates, it shows much more stable performance and manages to always maintain a good level of fairness. Even under the labeling rate as low as 50%, where the naïve method fails and simply predicts every example as negative, FairPUL can achieve even comparable performance with naïve method in the fully labeled setting. This means that FairPUL can achieve satisfying performance with much fewer labeled examples. It is a very desirable property in practice where labeled data is often time-consuming and expensive to obtain.

F1 scores, AOD and EOD using different numbers of unlabeled examples.
5.4.2. Effect of unlabeled samples (RQ3)
To further study how unlabeled examples affect our framework, we fix the number of positive examples and compare FairPUL’s performance with different numbers of unlabeled examples. Since the benchmark data sets are not provided with additional unlabeled data, we deploy the following data generation procedure: we randomly select 50% examples in the original training set. The remaining 50% will serve as a pool of unlabeled examples. We test our model leveraging examples in the pool. Note that to ensure the SCAR assumption holds, we randomly label examples of the fixed number from the positive distribution in the current training set. Average results of 10 runs of FairPUL-EOP on COMPAS are shown in Fig. 2. We can see that with more unlabeled examples, our framework achieves better F1 scores and lower AOD and EOD. From the perspective of our framework design, unlabeled examples are leveraged in two aspects: to help predict the labeling probability for building the traditional classifier; to help reduce unfairness in constructing the empirical optimal fair classifier. Adding more unlabeled examples to the training set will benefit both aspects, which is proved by this experiment. This demonstrates FairPUL’s advantage in improving both classification performance and fairness simply with unlabeled examples, which may be otherwise useless in other methods.
6. Conclusion
In this paper, motivated by real applications such as medical diagnosis and recidivism scoring, we study the important and yet less studied problem of the optimal fair binary classifier in the positive and unlabeled learning using the notion of equalized odds and equal opportunity. In particular, we bridge the gap between fair classification and PUL by first providing the theoretical analysis, and then designing a model-agnostic post-processing framework which preserves favorable consistency properties under mild assumptions. We highlight our framework’s flexibility of being easily generalized to any base classifier which outputs conditional labeling probabilities. Extensive experiments demonstrate that our framework outperforms state-of-the-art in both PUL and fair classification.
Acknowledgements.
This work is supported by National Science Foundation under Award No. IIS-1947203, IIS-2117902, and IIS-2002540. The views and conclusions are those of the authors and should not be interpreted as representing the official policies of the funding agencies or the government.References
- (1)
- Agarwal et al. (2018) Alekh Agarwal, Alina Beygelzimer, Miroslav Dudík, John Langford, and Hanna Wallach. 2018. A reductions approach to fair classification. In ICML. PMLR, 60–69.
- Angwin et al. (2016) Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner. 2016. Machine Bias. https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
- Audibert et al. (2007) Jean-Yves Audibert, Alexandre B Tsybakov, et al. 2007. Fast learning rates for plug-in classifiers. The Annals of statistics 35, 2 (2007), 608–633.
- Bekker and Davis (2020) Jessa Bekker and Jesse Davis. 2020. Learning from positive and unlabeled data: A survey. Machine Learning (2020), 719–760.
- Berk et al. (2018) Richard Berk, Hoda Heidari, Shahin Jabbari, Michael Kearns, and Aaron Roth. 2018. Fairness in criminal justice risk assessments: The state of the art. Sociological Methods & Research (2018), 0049124118782533.
- Bird et al. (2020) Sarah Bird, Miro Dudík, Richard Edgar, Brandon Horn, Roman Lutz, Vanessa Milan, Mehrnoosh Sameki, Hanna Wallach, and Kathleen Walker. 2020. Fairlearn: A toolkit for assessing and improving fairness in AI. Technical Report MSR-TR-2020-32. Microsoft. https://www.microsoft.com/en-us/research/publication/fairlearn-a-toolkit-for-assessing-and-improving-fairness-in-ai/
- Char et al. (2018) Danton S Char, Nigam H Shah, and David Magnus. 2018. Implementing machine learning in health care—addressing ethical challenges. The New England journal of medicine 378, 11 (2018), 981.
- Chen et al. (2019) Xingyu Chen, Brandon Fain, Liang Lyu, and Kamesh Munagala. 2019. Proportionally fair clustering. In ICML. PMLR, 1032–1041.
- Chierichetti et al. (2017) Flavio Chierichetti, Ravi Kumar, Silvio Lattanzi, and Sergei Vassilvitskii. 2017. Fair Clustering Through Fairlets. In NeurIPS. 5029–5037.
- Chouldechova (2017) Alexandra Chouldechova. 2017. Fair prediction with disparate impact: A study of bias in recidivism prediction instruments. Big data 5, 2 (2017), 153–163.
- Chouldechova and Roth (2018) Alexandra Chouldechova and Aaron Roth. 2018. The frontiers of fairness in machine learning. arXiv preprint arXiv:1810.08810 (2018).
- Chzhen et al. (2019) Evgenii Chzhen, Christophe Denis, Mohamed Hebiri, Luca Oneto, and Massimiliano Pontil. 2019. Leveraging Labeled and Unlabeled Data for Consistent Fair Binary Classification. In NeurIPS, Vol. 32.
- Claesen et al. (2015) Marc Claesen, Frank De Smet, Johan AK Suykens, and Bart De Moor. 2015. A robust ensemble approach to learn from positive and unlabeled data using SVM base models. Neurocomputing 160 (2015), 73–84.
- Denis and Hebiri (2020) Christophe Denis and Mohamed Hebiri. 2020. Consistency of plug-in confidence sets for classification in semi-supervised learning. Journal of Nonparametric Statistics 32, 1 (2020), 42–72.
- Devroye (1978) Luc Devroye. 1978. The uniform convergence of nearest neighbor regression function estimators and their application in optimization. IEEE Transactions on Information Theory 24, 2 (1978), 142–151.
- Dhurandhar and Gurumoorthy (2020) Amit Dhurandhar and Karthik S Gurumoorthy. 2020. Classifier Invariant Approach to Learn from Positive-Unlabeled Data. In ICDM. IEEE, 102–111.
- Donini et al. (2018) Michele Donini, Luca Oneto, Shai Ben-David, John S Shawe-Taylor, and Massimiliano Pontil. 2018. Empirical risk minimization under fairness constraints. In NeurIPS. 2791–2801.
- du Plessis et al. (2014) Marthinus Christoffel du Plessis, Gang Niu, and Masashi Sugiyama. 2014. Analysis of Learning from Positive and Unlabeled Data. In NeurIPS.
- Dwork et al. (2012) Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. 2012. Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference. 214–226.
- Dwork et al. (2020) Cynthia Dwork, Christina Ilvento, and Meena Jagadeesan. 2020. Individual fairness in pipelines. arXiv preprint arXiv:2004.05167 (2020).
- Elkan and Noto (2008) Charles Elkan and Keith Noto. 2008. Learning classifiers from only positive and unlabeled data. In SIGKDD. 213–220.
- Fehrman et al. (2017) Elaine Fehrman, Awaz K Muhammad, Evgeny M Mirkes, Vincent Egan, and Alexander N Gorban. 2017. The five factor model of personality and evaluation of drug consumption risk. In Data science. Springer, 231–242.
- Feldman et al. (2015) Michael Feldman, Sorelle A Friedler, John Moeller, Carlos Scheidegger, and Suresh Venkatasubramanian. 2015. Certifying and removing disparate impact. In SIGKDD. 259–268.
- Gong et al. (2019) Chen Gong, Tongliang Liu, Jian Yang, and Dacheng Tao. 2019. Large-margin label-calibrated support vector machines for positive and unlabeled learning. IEEE transactions on neural networks and learning systems 30, 11 (2019), 3471–3483.
- Hardt et al. (2016) Moritz Hardt, Eric Price, Eric Price, and Nati Srebro. 2016. Equality of Opportunity in Supervised Learning. In NeurIPS.
- He et al. (2018) Fengxiang He, Tongliang Liu, Geoffrey I Webb, and Dacheng Tao. 2018. Instance-dependent pu learning by bayesian optimal relabeling. arXiv preprint arXiv:1808.02180 (2018).
- Hsieh et al. (2015) Cho-Jui Hsieh, Nagarajan Natarajan, and Inderjit Dhillon. 2015. PU learning for matrix completion. In ICML. PMLR, 2445–2453.
- Ienco and Pensa (2016) Dino Ienco and Ruggero G Pensa. 2016. Positive and unlabeled learning in categorical data. Neurocomputing 196 (2016), 113–124.
- Jung et al. (2019) Christopher Jung, Michael Kearns, Seth Neel, Aaron Roth, Logan Stapleton, and Zhiwei Steven Wu. 2019. Eliciting and enforcing subjective individual fairness. arXiv preprint arXiv:1905.10660 (2019).
- Kiryo et al. (2018) Ryuichi Kiryo, Gang Niu, Marthinus Christoffel du Plessis, and Masashi Sugiyama. 2018. Positive-Unlabeled Learning with Non-Negative Risk Estimator. In NeurIPS.
- Kleindessner et al. (2019) Matthäus Kleindessner, Pranjal Awasthi, and Jamie Morgenstern. 2019. Fair k-center clustering for data summarization. In ICML. PMLR, 3448–3457.
- Kusner et al. (2017) Matt J Kusner, Joshua Loftus, Chris Russell, and Ricardo Silva. 2017. Counterfactual Fairness. In NeurIPS.
- Lee and Liu (2003) Wee Sun Lee and Bing Liu. 2003. Learning with positive and unlabeled examples using weighted logistic regression. In ICML, Vol. 3. 448–455.
- Li and Liu (2005) Xiao-Li Li and Bing Liu. 2005. Learning from positive and unlabeled examples with different data distributions. In European conference on machine learning. Springer, 218–229.
- Lichman (2013) M. Lichman. 2013. UCI Machine Learning Repository. http://archive.ics.uci.edu/ml
- Menon and Williamson (2018) Aditya Krishna Menon and Robert C Williamson. 2018. The cost of fairness in binary classification. In FAccT. PMLR, 107–118.
- Mordelet and Vert (2014) Fantine Mordelet and J-P Vert. 2014. A bagging SVM to learn from positive and unlabeled examples. Pattern Recognition Letters 37 (2014), 201–209.
- Perry (2013) Walt L Perry. 2013. Predictive policing: The role of crime forecasting in law enforcement operations. Rand Corporation.
- Pollard (1990) David Pollard. 1990. Empirical processes: theory and applications. In NSF-CBMS regional conference series in probability and statistics. JSTOR, i–86.
- Sadinle et al. (2019) Mauricio Sadinle, Jing Lei, and Larry Wasserman. 2019. Least ambiguous set-valued classifiers with bounded error levels. J. Amer. Statist. Assoc. 114, 525 (2019), 223–234.
- Shao et al. (2015) Yuan-Hai Shao, Wei-Jie Chen, Li-Ming Liu, and Nai-Yang Deng. 2015. Laplacian unit-hyperplane learning from positive and unlabeled examples. Information Sciences 314 (2015), 152–168.
- Song et al. (2019) Jiaming Song, Pratyusha Kalluri, Aditya Grover, Shengjia Zhao, and Stefano Ermon. 2019. Learning controllable fair representations. In The 22nd International Conference on Artificial Intelligence and Statistics. PMLR, 2164–2173.
- Sweeney (2013) Latanya Sweeney. 2013. Discrimination in online ad delivery. Commun. ACM 56, 5 (2013), 44–54.
- Twala (2010) Bhekisipho Twala. 2010. Multiple classifier application to credit risk assessment. Expert Systems with Applications 37, 4 (2010), 3326–3336.
- Van de Geer et al. (2008) Sara A Van de Geer et al. 2008. High-dimensional generalized linear models and the lasso. Annals of Statistics 36, 2 (2008), 614–645.
- Ward et al. (2009) Gill Ward, Trevor Hastie, Simon Barry, Jane Elith, and John R Leathwick. 2009. Presence-only data and the EM algorithm. Biometrics 65, 2 (2009), 554–563.
- Wei and He (2022) Tianxin Wei and Jingrui He. 2022. Comprehensive Fair Meta-learned Recommender System. arXiv preprint arXiv:2206.04789 (2022).
- Wu et al. (2019) Yongkai Wu, Lu Zhang, Xintao Wu, and Hanghang Tong. 2019. PC-Fairness: A Unified Framework for Measuring Causality-based Fairness. In NeurIPS, Vol. 32.
- Yan et al. (2018) Bowei Yan, Sanmi Koyejo, Kai Zhong, and Pradeep Ravikumar. 2018. Binary classification with karmic, threshold-quasi-concave metrics. In ICML. PMLR, 5531–5540.
- Zafar et al. (2017) Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rogriguez, and Krishna P Gummadi. 2017. Fairness constraints: Mechanisms for fair classification. In Artificial Intelligence and Statistics. PMLR, 962–970.
- Zemel et al. (2013) Rich Zemel, Yu Wu, Kevin Swersky, Toni Pitassi, and Cynthia Dwork. 2013. Learning fair representations. In ICML. PMLR, 325–333.
- Zhang et al. (2020) Tao Zhang, Tianqing Zhu, Mengde Han, Jing Li, Wanlei Zhou, and Philip S Yu. 2020. Fairness Constraints in Semi-supervised Learning. arXiv preprint arXiv:2009.06190 (2020).
Appendix A Proof of Lemma 1
Proof.
Using the conditional probabilities, we have
The result follows by dividing both sides with . ∎
Appendix B Derivation of the Optimal Fair Classifier
For EO, using weak duality we can write Equation (1) as:
(8) | ||||
In the PUL setting, based on Lemma 1, we can write:
(9) | ||||
The risk function is:
So the objective function can be simplified as:
Since , we can get the minimizer :
(10) | ||||
Appendix C Proof of Lemma 2
Substituting the minimizer into the Lagrange function, we could see that the mappings of and are convex. We can write the first order optimality conditions of the objective function as:
Based on Assumption 2, this subgradient is reduced to the gradient almost surely. So we have Equations (3):
Combining it with Equation (9), we can see that . In other words, satisfies Equalized Odds with respect to and is thus fair. Therefore, we have because is defined as the optimal fair classifier to minimize the risk. Furthermore, since is a solution to the dual problem, we have according to Equation (8). Therefore, we can conclude that .
Appendix D Proof of Theorem 1
Proof.
Following the strategy of (Denis and Hebiri, 2020; Chzhen et al., 2019), we first introduce an intermediate pseudo-estimator as follows:
(11) | ||||
where satisfies:
(12) | ||||
Comparing this pseudo-estimator with the theoretical ideal in Equation (2) and our estimator in Equation (7), we can see that knows the marginal distribution of . That is, it has precise information on the distributions and . It can be seen as a nearly-idealized version of where the uncertainty in it is only induced by the estimator .
To demonstrate our proposed method is asymptotically optimal, we can upper bound the excess risk by expressing it as a sum of two terms, . The first term can be bounded by the distance between and . Based on the first part of Assumption 3 that is consistent as , it converges to zero. For the second term, comparing the upper bound on and the lower bound on , we can upper bound the difference and show that based on the law of large numbers and the second part of Assumption 3.
To demonstrate our proposed framework is asymptotically fair, we can first upper bound the unfairness with the triangle inequality by considering and respectively and their estimators:
Appendix E Implementation Details
The three data sets we use can be obtained as follows:
-
•
COMPAS is available at https://github.com/propublica/
compas-analysis.
-
•
German Credit is available at https://archive.ics.uci.edu/
ml/machine-learning-databases/statlog/german/.
-
•
Drug is available at https://archive.ics.uci.edu/ml/
datasets/Drug+Review+Dataset+%28Drugs.com%29
The hyper-parameters are found by grid search and we report the best results in terms of classification performance. More details of different models (Lin.LR, Support Vector Machine with polynomial kernel (SVM) and Multilayer Perceptron (MLP)) are shown as follows:
-
•
For SVM, = 0.1, degree = 2, = 2.
-
•
For Lin.SVM, = 10, tolerance = 1e-4.
-
•
For Lin.LR, = 1, solver = ’lbfgs’, max iteration=1000.
-
•
For MLP, we use the ReLU activation function and Adam optimizer. For hyper-parameters, regularization term parameter=1e-4, learning rate=1e-3. The hidden layers sizes in MLP are set to for COMPAS, for German Credit and for Drug.
For baselines, we use the code provided by original authors and Fairlearn (https://fairlearn.org/), and carefully tune all the hyper-parameters for the best classification performance.
Appendix F Auxiliary Experimental Results
Average results and standard deviation on COMPAS of 10 runs. Models F1 AOD EOD Lin.LR Oracle 0.6210.015 0.2360.033 0.2930.045 Naïve 0.5520.023 0.2400.048 0.3110.054 +Agarwal 0.5070.019 0.0320.021 0.0370.036 +Hardt 0.3570.042 0.0220.012 0.0270.019 +Chzhen 0.4910.021 0.0180.015 0.0140.010 uPU 0.6350.010 0.0750.048 0.0960.028 wPU 0.5880.036 0.2780.164 0.3060.051 Bagging 0.5590.020 0.2930.223 0.3350.037 FairPUL-EO 0.6560.016 0.0150.008 0.0130.010 FairPUL-EOP 0.6580.012 0.0160.007 0.0090.008 SVM Oracle 0.6280.019 0.2870.127 0.2460.038 Naïve 0.5530.022 0.2990.102 0.3240.073 +Agarwal 0.4910.037 0.0460.032 0.0350.032 +Hardt 0.3580.047 0.0280.021 0.0300.023 +Chzhen 0.4780.033 0.0320.012 0.0400.034 uPU 0.6440.012 0.1860.086 0.1340.023 wPU 0.5650.027 0.2980.076 0.3150.055 Bagging 0.6480.013 0.2490.048 0.2860.035 FairPUL-EO 0.6460.013 0.0120.006 0.0150.013 FairPUL-EOP 0.6480.012 0.0160.012 0.0120.004 MLP Oracle 0.6250.008 0.2390.068 0.2640.051 Naïve 0.5920.018 0.2480.098 0.2990.066 +Hardt 0.4570.038 0.0360.016 0.0290.024 +Chzhen 0.5480.027 0.0320.029 0.0240.017 uPU 0.6600.012 0.1980.087 0.2300.067 wPU 0.6060.021 0.2670.138 0.2950.064 nnPU 0.6590.023 0.2480.243 0.2890.062 FairPUL-EO 0.6630.014 0.0320.017 0.0400.034 FairPUL-EOP 0.6660.023 0.0360.018 0.0340.012
Post-processing FairPUL-EOP vs In-processing/Pre-processing. Rates Models F1 EOD 100% Lin.SVM Naïve 0.5650.051 0.0610.042 In 0.5410.050 0.0510.044 Pre 0.5200.057 0.0930.069 FairPUL 0.5910.035 0.0560.049 Lin.LR Naïve 0.5570.054 0.0640.053 In 0.5400.057 0.0710.064 Pre 0.5330.057 0.0620.036 FairPUL 0.6090.038 0.0510.041 90% Lin.SVM Naïve 0.4710.048 0.0990.052 In 0.4700.033 0.0770.051 Pre 0.4140.062 0.0780.074 FairPUL 0.5800.020 0.0340.027 Lin.LR Naïve 0.5070.045 0.1320.037 In 0.4860.057 0.0690.044 Pre 0.4790.052 0.1250.072 FairPUL 0.6060.034 0.0620.047
Comparison results of different labeling rates on German. Rates Models F1 EOD 100% Lin.SVM Naïve 0.5650.051 0.0610.042 FairPUL 0.5910.035 0.0560.049 Lin.LR Naïve 0.5570.054 0.0640.053 FairPUL 0.6090.038 0.0510.041 90% Lin.SVM Naïve 0.5570.054 0.0640.053 FairPUL 0.5800.020 0.0340.027 Lin.LR Naïve 0.5070.045 0.1320.037 FairPUL 0.6060.034 0.0620.047 80% Lin.SVM Naïve 0.1930.136 0.0440.051 FairPUL 0.5270.047 0.0330.029 Lin.LR Naïve 0.4010.046 0.0830.055 FairPUL 0.6050.027 0.0560.040 50% Lin.SVM Naïve 0.0000.000 0.0000.000 FairPUL 0.4810.042 0.0190.018 Lin.LR Naïve 0.1060.036 0.0530.026 FairPUL 0.5930.036 0.0440.019