Muffin: A Framework Toward Multi-Dimension AI Fairness by Uniting Off-the-Shelf Models
Abstract
Model fairness (a.k.a., bias) has become one of the most critical problems in a wide range of AI applications. An unfair model in autonomous driving may cause a traffic accident if corner cases (e.g., extreme weather) cannot be fairly regarded; or it will incur healthcare disparities if the AI model misdiagnoses a certain group of people (e.g., brown and black skin). In recent years, there are emerging research works on addressing unfairness, and they mainly focus on a single unfair attribute, like skin tone; however, real-world data commonly have multiple attributes, among which unfairness can exist in more than one attribute, called “multi-dimensional fairness”. In this paper, we first reveal a strong correlation between the different unfair attributes, i.e., optimizing fairness on one attribute will lead to the collapse of others. Then, we propose a novel Multi-Dimension Fairness framework, namely Muffin, which includes an automatic tool to unite off-the-shelf models to improve the fairness on multiple attributes simultaneously. Case studies on dermatology datasets with two unfair attributes show that the existing approach can achieve 21.05% fairness improvement on the first attribute while it makes the second attribute unfair by 1.85%. On the other hand, the proposed Muffin can unite multiple models to achieve simultaneously 26.32% and 20.37% fairness improvement on both attributes; meanwhile, it obtains 5.58% accuracy gain.
Index Terms:
multi-dimensional fairness, model fusing, reinforcement learning, parametersI Introduction
With the continuous progress of AI democratization, we have witnessed the employment of machine learning (ML) in various applications [1, 2, 3, 4, 5, 6, 7, 8], such as autonomous driving [9], language translation [10], dermatology assistant [11], and vital signs monitoring [12]. Along with such progress, the fairness issue in ML models has emerged and become critical in applications. An unfair ML model performs well in most cases, but its performance may crash when corner cases occur. This may lead to accidents in autonomous driving [13], linguistic discrimination in language translation [14], and even threatening life by misdiagnosis in healthcare [15]. Thus, it is high time to take fairness into consideration in designing the ML system.
In recent years, a notable number of research works aim to address the AI fairness issue from an algorithm’s perspective [16]; however, there are very few works taking this problem from a systematic view that can provide an automatic tool to generate a fair ML system. In the existing works, there are commonly two directions: (1) Revising training algorithm by using adversarial training [17], domain discriminative training [18], or training with a fair loss function [19]; and (2) Improving data balance by generating an unprivileged group of data using GAN [20]. These methods have achieved a great improvement in the fairness of a single unfair attribute; however, almost all of these works ignore the important fact that one dataset will have several fairness-related attributes, called “multi-dimensional fairness”. For example in the dermatology dataset, the unfairness may not only occur at the skin tone attribute, but also at age, gender, or disease site. It seems straightforward to extend the existing approach; however, by examining the state-of-the-art solutions that optimize fairness from the algorithm’s perspective [16], we observe that there exists a correlation between unfair attributes, that is, improving fairness on one attribute will decrease that on the other one.
The correlation among unfair attributes creates a challenge of achieving multi-dimensional fairness by optimizing a single model. In response to this, we take an alternative way to leverage and unite multiple existing models to improve fairness. This is inspired by our key observation that two models with similar accuracy on an unprivileged group may have a disagreement on specified data, that is, for one data, only one model makes the correct classification. As a result, if we can correct the results by smartly selecting the model to be used when a disagreement happens, then we can improve the fairness of an unprivileged group. Although promising, it is still unclear which models should be paired, how can we correct classification. when a disagreement happens, and how to improve the fairness of multiple attributes.
In this paper, we propose a novel framework, namely “Muffin”, to address all of the above issues in a systematic way. Given a dataset with multiple unfair attributes and a set of off-the-shelf models, the proposed framework can automatically select multiple models to maximize the fairness of all attributes simultaneously. To achieve this goal, we first develop a model-fusing structure, where a multi-layer perception (MLP) will take the output of the selected models as input. Then, we will create a proxy dataset by assigning weights to data with multiple attributes in unprivileged groups. On top of this, we employ the reinforcement learning approach to identify the model-fusing structure (i.e., select the models to be involved and search for the best MLP architecture). The identified architecture will be trained on the proxy dataset to achieve multi-dimension fairness.
The main contributions of this paper are as follows.
-
•
Problem. To the best of our knowledge, this is the first work to identify the multi-dimensional fairness problems in the ML domain. We reveal the correlation among different unfair attributes, which inhibits us from directly applying existing single-dimensional fairness techniques.
-
•
Framework. Given a dataset with multiple unfair attributes and a set of ML models, a novel framework, namely Muffin, is proposed to unite off-the-shelf models to improve multiple unfair attributes simultaneously.
-
•
Automated Tool. A reinforcement learning-based algorithm coupled with a fairness-aware proxy dataset is developed to enable the optimizations to be performed automatically.
Experimental results on the two dermatology datasets evaluate the performance of Muffin. First, with small models like ShuffleNet_V2_X1_0, and MobileNet_V3_Small, Muffin can improve the overall accuracy to 80.55% and 81.77% from 77.21% and 76.19%, respectively. For ShuffleNet_V2_X1_0, Muffin can achieve 19.44% and 2.22% improvements on two unfair attributes; while the figures are 26.32% and 20.37% for MobileNet_V3_Small. With large models, the overall accuracy can be maintained by Muffin; meanwhile, it can consistently improve unfair attributes by 7.69% and 9.30% for ResNet-18.
In the rest of the paper, Section 2 reviews the related work and provides motivation; Section 3 defines the problem and presents Muffin. Experimental results are reported in Section 4 and concluding remarks are given in Section 5.

II Related Work and Motivation
This section provides several key observations on AI fairness and discusses the existing works, which gives us the motivation to research the fairness issue on multiple attributes.
Observation 1: Unfairness exists on multiple attributes.
On the dermatology dataset ISIC2019 [21], there are three attributes (age, gender, and disease site). We examine the fairness of these attributes by using a set of commonly used neural architectures, including MobileNet [22], ResNet [23], DenseNet [24], and ShuffleNet [25]. After training these architectures on ISIC2019 dataset, Figure 1 shows the unfairness score of the trained model on different attributes. In Figure 1(a)-(b), we compare the fairness of different models between the attributes of age/site and gender. We can see that the maximum unfairness score of all models on the gender attribute is less than 0.12, which reflects the low accuracy gap between males and females (i.e., merely 3%). On the other hand, in Figure 1(c), we observe that both attributes age and site have high unfairness scores (i.e., over 0.4), which are led by 45.04% and 36.27% accuracy gaps between different groups in site and age attributes, respectively. More importantly, the unfairness scores of these two attributes on different models are not positively correlated. For example, regarding the unfairness score of site, DenseNet121 (denoted D121) has the best performance; however, when it comes to age, ResNet-18 (denoted R18) has the lowest unfairness score. Results demonstrated that there is no neural architecture can take over both age and site.
Related work: Fairness on multiple attributes has not been well studied in existing research works.
Most AI fairness research efforts target a single attribute. Regarding to the technique, there are commonly two directions to improve fairness: (1) balancing data, and (2) improving training. First, data unbalancing can easily lead to unfairness; as a result, the most straightforward way is to balance data by applying data augmentation (such as flip, rotation, and scale) on the unprivileged groups [26, 27, 28]. Second, since the unfairness is measured by the accuracy of different groups of data and the accuracy is directly affected by training, the methods to improve training for high fairness were proposed [19, 29, 30]. The intuition is that fairness can be improved if we can add a regularization term in the loss function with the consideration of bias. Besides data and training, authors in [2] also point out that neural network architecture can also affect model fairness.
All the above works aimed at optimizing fairness on a single attribute, when it comes to multiple attributes, it is unclear whether we can directly apply these approaches to optimize the model with several unfair attributes simultaneously.

Observation 2: Fairness attributes are entangled and models encounter bottlenecks in improving fairness.
To answer the above question, we conduct a set of experiments using both balancing data () and improving training by a fair loss function () on the ISIC2019 dataset. For the neural network, we select three networks, including R18 and D121 on the Pareto frontier in Figure 1(c), and M_V2. Results are reported in Figure 2.
From Figure 2(a), we observe that two fairness-related attributes (i.e., age and site) are entangled with each other. Specifically, when we optimize age, the unfairness score of the site will increase (see ). Correspondingly, if we apply the optimization method to the site, the model on the age attribute will become unfairer (see and ). It’s like a seesaw game, we can only press down one end but not both ends at the same time.
From Figure 2(b)-(c), we further observe that it is difficult to improve the model with a low fairness score. For example, DenseNet121 in Figure 2(b) has a low unfairness score on site; we cannot further reduce this score, even if we use and on the site attribute. Similar behavior occurs on the age attribute of the ResNet-18 model, as shown in Figure 2(c). In consequence, neural network models will encounter bottlenecks when doing the single attribute optimization. The above results demonstrate the limits of the single attribute optimization and motivate us to find a solution to solve multi-dimensional fairness.

Observation 3: Models are complementary on fairness. From the previous example, it seems that optimizing fairness on multiple attributes is unattainable. In this paper, we consider such a multi-dimensional fairness problem from another angle: Instead of optimizing one model, can we leverage multiple models to improve fairness? We breakdown the classification results of R18 and D121 in Figure 3. As shown in Figure 3(a), two middle bars indicate the probability of one model is correct and another one is incorrect, of which the sum probability reaches 15.93%. This inspires us to leverage the correct result of multiple models (called unite multiple models), and then we can improve the accuracy of the unprivileged group. Figure 3(b) further shows that if we can unite two models for the unprivileged group, its accuracy can be even higher than the accuracy of privileged groups on both models.
Motivation: Unite off-the-shelf models to improve fairness on multiple attributes at the same time.
With the insight that multiple models can improve fairness, given a set of off-the-shelf models (called model pool), we propose an automatic tool named “Muffin” to design the system with the objective of maximum fairness on multiple unfair attributes. Specifically, the proposed structure contains two components: (1) the “muffin body” will be made by selecting models from the model pool; (2) the “muffin head” is based on a multi-layer perceptron to process the results of multiple models, which are expected make a correct classification when multiple models have a disagreement. We will formally formulate the problem and provide the detailed design of the Muffin framework in the next section.
III Muffin Framework
In this paper, we study the multi-dimensional fairness for classification tasks in computer vision. This section will formally define the problem of “Optimization of multi-dimensional fair classification” and gives out framework of Muffin.
3.1 Problem Definition
Classification. Given a dataset , we define as a set of classes, where each data belongs to a class . That is, there exists a mapping function : . A network is building it from to . With training data, will learn a function to approximate . If , it is a correct classification on data . The accuracy describes the ratio of data in getting the correct classification by model .
Attribute. For a dataset , we define as a single sensitive attribute. For each , the whole dataset has groups . Each data has the sensitive attribute set . That is the dataset can be divided according to different attributes .
Multi-Dimensional Fairness. Fairness reflects the model’s performance of diverse groups for the single attribute . For a model , the accuracy for one group is . Based on the accuracy, we define the unfairness score of a model for attribute on the dataset based on L1-norm,which is . When it comes to the multi-dimensional fairness score , one can extend the fairness from one single attribute to multiple attributes by formulating the objective as:
(1) |
Problem Formulation. Based on the above definitions, we can formally define the problem as follows: Given a dataset which has an attribute set with classes and off-the-shelf models, our objective is to automatically optimize several unfair attributes at one time, it means we need to minimize the unfairness score of each unfair attribute and the multi-dimensional unfairness score , meanwhile overall accuracy meet the requirement.

3.2 Framework Overview
Figure 4 illustrates the overview of our proposed Muffin framework. It is composed of four components: ➀ model-fusing structure based on a model pool and a backbone structure of MLP, ➁ a fairness proxy dataset and training, ➂ a multi-fairness reward, and ➃ a recurrent neural network (RNN) based controller.
These components will be iteratively executed for optimization. A model-fusing structure will be identified from the search space defined in component ➀. Then, it will be sent to component ➁ for training. To enable the optimization of fairness on multiple unfair attributes, we will build a fairness proxy dataset, which will be utilized in training. Next, the trained model will be used to generate a multi-fairness reward in component ➂, and ➃ controller will be updated based on the reward in each iteration. In the following text, we will introduce these components in detail.
➀ Model-Fusing Structure: It contains two parts: (1) “Muffin body” includes a set of off-the-shelf models, and these models can be selected by the controller to be utilized in the final structure; (2) in “Muffin head”, an MLP backbone specifies a set of hyperparameters, including the number of layers, the number of parameters in each layer, and different activation functions. The generated model-fusing structure will be composed of the selected models in the muffin body and the identified MLP from the muffin head, where the MLP is expected to be trained such that it can correct select results when there has a disagreement on multiple selected models.
➁ Fairness Proxy Dataset and Training: In order to train the model-fusing structure, there are two fundamental questions: which parameters will be tuned, and what dataset will be used? To maximize the search efficiency, we will freeze the parameters in the pre-trained off-the-shelf models in the model-fusing structure and train parameters in MLP only. For the second question, we will only involve unprivileged group data for training. This is based on the Observation 3 in section II, where the proposed technique is not going to change the output if all models reached consensus. In our experiments, we further observe that models in privileged group data have a low rate of disagreement. Therefore, we will not involve the privileged group data for model-fusing structure training.
The next question is how to consider multiple unfair attributes in training. As pointed out by [31], one of the major challenges for fairness is how to improve the impact of biased data which are difficult to recognize in training. Based on this concept, we carry out a one-time pre-processing to consider multiple unfair attributes. More specifically, we will assign a weight for each group, and it will be used to calculate the loss in training. In the following algorithm, we give the pseudo code to generate the weights.
In Algorithm 1, we consider the impact of the same data on the different unprivileged groups, and we associate the data with a higher weight if it appears in the groups under multiple unfair attributes. In this way, the multiple unfair attributes can be considered in a holistic optimization. Based on the obtained weights, we formulate the loss function in our fairness-aware training as follows:
(2) |
➂ Multi-Fairness Reward: After the training is finished, our next target is to generate a reward to assess the identified model. To this end, we will perform inference on the original dataset using the trained model-fusing structure to generate overall accuracy and unfairness score for each unfair attribute. Then, the reward is generated based on these metrics, which can be formulated as below.
(3) |
where is the number of unfair attributes, and is a specific attribute. A larger reward indicates the model has higher accuracy and lower unfairness on average.
➃ RNN Controller: The controller will iteratively generate a model-fusing structure by determining the hyperparameters using a recurrent neural network (RNN). In each step of RNN, a fully connected (FC) layer with parameters will generate one hyperparameter. And the parameters in RNN and FC layers will be updated by using the reward. More specifically, we apply the Monte Carlo policy gradient algorithm [32]:
(4) |
where , , , and indicate the batch size, the number of steps in each episode, an exponential factor, and an average exponential moving of rewards, respectively. In every episode, the reward is discounted by and .
Villina(vil.) | Method [33] | Method [34] | Muffin | ||||||||||||||
Model | Age | Site | Acc. | Age | Site | Acc. | Age | Site | Acc. | MLP | Paired Models | Age | vs.Vil | Site | vs.Vil. | Acc. | Acc.Imp. |
0.28 | 0.49 | 78.32% | 0.31 | 0.57 | 74.24% | ||||||||||||
ShuffleNet_V2_X1_0 | 0.36 | 0.45 | 77.21% | 0.30 | 0.68 | 77.49% | 0.46 | 0.66 | 76.70% | [16,18,12,8] | DenseNet121 | 0.29 | 19.44% | 0.44 | 2.22% | 80.55% | 3.34% |
0.30 | 0.55 | 77.90% | 0.33 | 0.69 | 76.09% | ||||||||||||
MobileNet_V3_Small | 0.38 | 0.54 | 76.19% | 0.40 | 0.63 | 76.46% | 0.44 | 0.64 | 73.09% | [16,10,10,8] | ResNet-34 | 0.28 | 26.32% | 0.43 | 20.37% | 81.77% | 5.58% |
0.30 | 0.48 | 82.44% | 0.26 | 0.44 | 81.65% | ||||||||||||
DenseNet121 | 0.31 | 0.36 | 81.83% | 0.34 | 0.54 | 83.31% | 0.34 | 0.41 | 80.19% | [16,10,10,8] | DenseNet201 | 0.26 | 16.13% | 0.35 | 2.78% | 81.93% | 0.10% |
0.27 | 0.43 | 80.84% | 0.27 | 0.49 | 80.69% | ||||||||||||
ResNet-18 | 0.26 | 0.43 | 81.28% | 0.38 | 0.41 | 81.97% | 0.35 | 0.34 | 80.00% | [16,10,16,8] | DenseNet121 | 0.24 | 7.69% | 0.39 | 9.30% | 82.10% | 0.82% |
-
•
Green indicates successful optimization of the unfairness score. Red indicates failed optimization of the unfairness score.
IV Experiments
Muffin is evaluated on two datasets for diagnosing dermatological disease. Results show the superiority of Muffin-Nets over the existing neural architectures since it can improve multi-dimensional fairness without compromising accuracy.
4.1. Experimental Setup
A. Dataset: Two datasets ISIC2019 [21] and Fitzpatrick17K [35] are built based on patients’ images collected from the open-access datasets. The images from ISIC2019 are utilized for a classification task with 8 dermatology diseases and Fitzpatrick17K is used for 9 classes of classification.
B. Muffin settings: During the whole process, datasets were split into three sets: (1) training set with 64% images; (2) validation set with 16% images; and (3) test set with the rest 20% images. The number of episodes for reinforcement learning is set to 500.
C. Competitors and training settings: To evaluate Muffin-Nets, we select a set of state-of-the-art neural networks for comparison, including (1) the manually designed ResNet [23], and (2) the AutoML identified MobileNetV3 [36], DenseNet [24] and ShuffleNet [37]. For a fair comparison, all neural networks are trained from scratch with the same hyperparameters on a GPU cluster with 48 RTX 3080: (1) learning rate starts from 0.1 with a decay of 0.9 in 20 steps, (2) 64 for the batch size, and (3) 500 epochs for training.
4.2 Comparison of Muffin with existing fairness Techniques for different models
In Table I, we compare Muffin-Nets against basedlines with 2 optimization methods (Data-based (D) and Loss-function based (L)). Considering the model size, we presented 4 different model types from the smallest one ShuffleNet_V2_X1_0 to the biggest one ResNet-18.
We have several observations from Table I: (1) Both existing optimization methods are not consistent for the models listed here. If we optimize one attribute, it will lead to an increase in unfairness for another attribute. The data method can improve the accuracy with some architectures. But when comes to the loss function method, it declines the accuracy to a certain degree. (2) Compared with the above methods, Muffin can significantly improve the accuracy of small architectures. The parameters of ShuffleNet_v2_X1_0 and MobileNet_V3_Samll is 1261804 and 1526056 respectively, the overall accuracy is below 80% even with the optimization methods, when it unites another model by Muffin framework, the accuracy improvement can reach 3.34% and 5.58%. As the big models like DenseNet121 and ResNet-18, Muffin still has a slight improvement of overall accuracy during the process. (3) Muffin can optimize two attributes at a time. It means we don’t need to do the tradeoff between several fairness-related attributes. The improvement for fairness of age and site can reach 19.44%, 2.22% for ShuffleNet_v2_X1_0, 26.32%, and 20.37% for MobileNet_V3_Samll. When comes to DenseNet121 which meets the bottleneck of fairness for site, the enhancement can reach 2.78% of site. Similarly, ResNet-18 which encounters the bottleneck of fairness for age, the improvement is 7.69%. (4) Compared with age, site is a much more difficult attribute to be optimized. With Muffin, the unfairness score of age can reach 0.22 but the lowest unfairness score of site is only 0.29. There are 6 subgroups for the age but 9 subgroups for the site, as the number of subgroups increases, it will be harder to make every subgroup at the same level.

4.3. Exploration by Muffin
In the second set of experiments, we will demonstrate that Muffin-Nets can significantly push forward the Pareto frontiers among multiple attributes and accuracy for dataset.
A. Unfairness score of sites vs. Unfairness score of age: In Figure 5(a) the x-axis is the unfairness score of age and the y-axis is the unfairness score of site. The ideal solution is located in the left corner, denoted as a star. Each circled point corresponds to an existing network like Figure 1 and the red triangle is Muffin-Net which is generated by our framework.
In Figure 5(a), The red and black lines plot the Pareto frontiers of Muffin-Nets and existing networks, respectively. We observe that Muffin-Age dominates all the existing neural networks in terms of unfairness score of age, it is only 0.2171; while Muffin-Sites located in the right corner achieves the highest fairness of the site. These figures clearly show that Muffin can significantly push forward the Pareto frontiers considering the fairness of age and site.
B. Accuracy vs. Unfairness. We further investigate the Pareto frontier between overall fairness and accuracy by integrating the unfairness score of age and site from Figure 5(a). Results in Figure 5(b) consistently show that Muffin can also push forward the Pareto frontier considering accuracy and overall fairness. More specifically, Muffin-Net is the only architecture whose accuracy can exceed 82%. Muffin-Age and Muffin-Sites which lies in the Pareto frontier in Figure 5(a) achieve accuracy of 81.71% and 81.24%, respectively.

4.4 Insights from Muffin-Site
Figure 6 provided the accuracy of each sub-group, meanwhile, it gives out detailed composition of accuracy and error rate.
Muffin-site unites the ResNet-50 and MobileNet_V3_Large from the model pool. In the Figure 6(a), when we look at the privileged groups which are not indicated by bolded, Muffin has a slight rise compared with the highest accuracy from paired models. What’s more, for the unprivileged groups which are bolded, the accuracy improvement of Muffin-site is more remarkable. As the accuracy gap between these two groups becoming smaller, it will lead to fairness improvement.
In Figure 6(b), for all unprivileged groups considering the site, the accuracy is increasing especially the group lateraltorso. The composition of accuracy and error rate for each unprivileged are presented in Figure 6(c). The green bar which indicates that both models are correct is the main source of the accuracy for Muffin-Site. The yellow bar which represents MobileNet_V3_Large is correct and ResNet-50 is incorrect has the opposite meaning as the red bar.
To illustrate the situation even further, if we look at the site group lateraltorso, the 4th pillar from the bottom, Muffin-Site can fully leverage paired models. All images which can be judged correctly by either of the two models can be classified by Muffin-Site. There is no yellow or red part in the error rate composition. Another group named oral/Genital, Muffin-Site has done some incorrect classification thus there is green in the error composition. Nevertheless, all correct determination from ResNet-50 are kept by Muffin-Site and a portion of the right classification from MobileNet_V3_Large are also reserved by it. If we add these two parts, it will be far more than the green part which is moved to the error rate. As we hoped in Observation 3, Muffin can unite both models perfectly.
4.5 Validation By Fitzpatrick17K
Based on dataset Fitzpatrick17K which also has the problem of multi-dimensional fairness, a model pool that has ResNet, ShuffleNet and MobileNet has been built. In Figure 7(a), Muffin improves the unfairness score of skin tones and type significantly. Muffin-Balacen which lies in the Pareto frontiers considering the unfairness score of skin tones and type will be used for further illustration in Figure 8. In Figure 7(b), Muffin also pushes forward Pareto frontiers for overall unfairness and accuracy. This is good proof that Muffin is also useful for other datasets by leveraging models.
Fitzpatrick scale [38] has 6 types of skin color, from light to black. In Figure 8, we have compared the accuracy for each sub-group of ResNet-18 which is located in the Pareto frontiers and Muffin. For the group white and medium, Muffin achieves accuracy gain but for black, Muffin has some loss of accuracy. In such a complementary way, the overall accuracy can be unaffected, while the model can become much fairer.


4.6 Ablation Studies
In Figure 9(a), we compare the effect of training with the original dataset and the weighted dataset without unknown group for the same MLP structure. The paired model is optimized DesNet121 and original ResNet-18 and the MLP structure is 16,16,16,8. It is obvious that with the weighted dataset, regardless age or site, the unfairness score will decline, and meanwhile, the overall accuracy can be kept.

In Figure 9(b), for each number of paired models, two architectures that have the best performance will be selected to plot. When it is equal to 1, ResNet-18 and DenseNet121 which locate in the Pareto frontier are selected. For Muffin, Muffin-Age, and Muffin-Sites are posted. If we expand the paired model into 3 or 4, we can still identify a series of models. It is obvious that even though we increase the number of paired models, the parameters will explode, but the reward is maintained at the same level. This is a good illustration of the tradeoff between unfairness, accuracy, and parameters.
V Conclusion
In this work, we have put multiple unfair attributes in a holistic optimization loop and proposed a framework, namely Muffin. On top of it, a model fusing structure is built to unit the existing models and then it will be trained with a proxy dataset in order to improve fairness. As such, Muffin can form a much better Pareto frontier on accuracy and multi-dimensional fairness compared with the existing models. As a whole, an automatic tool is developed to generate an AI system that can achieve multi-dimensional fairness. Extensive experiments are carried out to evaluate Muffin, where it is consistent for all models’ optimization and the maximum fairness improvement can reach 26.32% for age and 20.37% for the site.
Acknowledgement
We gratefully acknowledge the support of National Institutes of Health (NIH) (Award No. 1R01EB033387-01).
References
- [1] J. Yang, Y. Sheng, Y. Zhang, W. Jiang, and L. Yang, “On-device unsupervised image segmentation,” arXiv preprint arXiv:2303.12753, 2023.
- [2] Y. Sheng, J. Yang, Y. Wu, K. Mao, Y. Shi, J. Hu, W. Jiang, and L. Yang, “The larger the fairer? small neural networks can achieve fairness for edge devices,” in Proceedings of the 59th ACM/IEEE Design Automation Conference, pp. 163–168, 2022.
- [3] H. Peng, D. Gurevin, S. Huang, T. Geng, W. Jiang, O. Khan, and C. Ding, “Towards sparsification of graph neural networks,” in 2022 IEEE 40th International Conference on Computer Design (ICCD), pp. 272–279, IEEE, 2022.
- [4] J. Liao, Y. Liu, G. Xing, H. Wei, J. Chen, and S. Xu, “Shadow detection via predicting the confidence maps of shadow detection methods,” in Proceedings of the 29th ACM International Conference on Multimedia, pp. 704–712, 2021.
- [5] C. Yang, P. Zhao, Y. Li, W. Niu, J. Guan, H. Tang, M. Qin, B. Ren, X. Lin, and Y. Wang, “Pruning parameterization with bi-level optimization for efficient semantic segmentation on the edge,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15402–15412, 2023.
- [6] Z. Yan, Y. Qin, X. S. Hu, and Y. Shi, “On the viability of using llms for sw/hw co-design: An example in designing cim dnn accelerators,” arXiv preprint arXiv:2306.06923, 2023.
- [7] H. Peng, S. Huang, T. Zhou, Y. Luo, C. Wang, Z. Wang, J. Zhao, X. Xie, A. Li, T. Geng, et al., “Autorep: Automatic relu replacement for fast private network inference,” arXiv preprint arXiv:2308.10134, 2023.
- [8] Y. Huang, F. Zhao, X. Gui, and H. Jin, “Path-enhanced explainable recommendation with knowledge graphs,” World Wide Web, vol. 24, no. 5, pp. 1769–1789, 2021.
- [9] Dogan and et al., “Autonomous driving: A comparison of machine learning techniques by means of the prediction of lane change behavior,” in 2011 IEEE International Conference on Robotics and Biomimetics, pp. 1837–1843, IEEE, 2011.
- [10] Corston-Oliver and et al., “A machine learning approach to the automatic evaluation of machine translation,” in Proceedings of the 39th Annual Meeting of the Association for Computational Linguistics, pp. 148–155, 2001.
- [11] Rambhajani and et al., “A survey on implementation of machine learning techniques for dermatology diseases classification,” International Journal of Advances in Engineering & Technology, vol. 8, no. 2, p. 194, 2015.
- [12] Chen and et al., “A mobile phone-based wearable vital signs monitoring system,” in The Fifth International Conference on Computer and Information Technology (CIT’05), pp. 950–955, IEEE, 2005.
- [13] “tesla.” https://www.npr.org/2022/06/15/1105252793/nearly-400-car-crashes-in-11-months-involved-automated-tech-companies-tell-regul/.
- [14] G. Núñez and et al., “Translating for linguistic minorities in northern ireland: a look at translation policy in the judiciary, healthcare, and local government,” Current Issues in Language Planning, vol. 14, no. 3-4, pp. 474–489, 2013.
- [15] Suite and et al., “Beyond misdiagnosis, misunderstanding and mistrust: relevance of the historical perspective in the medical and mental health treatment of people of color.,” Journal of the National Medical Association, vol. 99, no. 8, p. 879, 2007.
- [16] Holstein and et al., “Improving fairness in machine learning systems: What do industry practitioners need?,” in Proceedings of the 2019 CHI conference on human factors in computing systems, pp. 1–16, 2019.
- [17] Xu and et al., “To be robust or to be fair: Towards fairness in adversarial training,” in International Conference on Machine Learning, pp. 11492–11501, PMLR, 2021.
- [18] Wang and et al., “Towards fairness in visual recognition: Effective strategies for bias mitigation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 8919–8928, 2020.
- [19] Liu and et al., “Fair loss: Margin-aware reinforcement learning for deep face recognition,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 10052–10061, 2019.
- [20] Sattigeri and et al., “Fairness gan: Generating datasets with fairness properties using a generative adversarial network,” IBM Journal of Research and Development, vol. 63, no. 4/5, pp. 3–1, 2019.
- [21] “Skin lesion analysis.” https://challenge2019.isic-archive.com/.
- [22] Sandler and et al., “Mobilenetv2: Inverted residuals and linear bottlenecks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4510–4520, 2018.
- [23] Targ and et al., “Resnet in resnet: Generalizing residual architectures,” arXiv preprint arXiv:1603.08029, 2016.
- [24] Huang and et al., “Densely connected convolutional networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 4700–4708, 2017.
- [25] Ma and et al., “Shufflenet v2: Practical guidelines for efficient cnn architecture design,” in Proceedings of the European conference on computer vision (ECCV), pp. 116–131, 2018.
- [26] Sharma and et al., “Data augmentation for discrimination prevention and bias disambiguation,” in Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, pp. 358–364, 2020.
- [27] Vannur and et al., “Data augmentation for fairness in personal knowledge base population,” in Pacific-Asia Conference on Knowledge Discovery and Data Mining, pp. 143–152, Springer, 2021.
- [28] Iosifidis and et al., “Dealing with bias via data augmentation in supervised learning scenarios,” Jo Bates Paul D. Clough Robert Jäschke, vol. 24, 2018.
- [29] Du and et al., “Fairness-aware agnostic federated learning,” in Proceedings of the 2021 SIAM International Conference on Data Mining (SDM), pp. 181–189, SIAM, 2021.
- [30] Madras and et al., “Predict responsibly: improving fairness and accuracy by learning to defer,” Advances in Neural Information Processing Systems, vol. 31, 2018.
- [31] Yan and et al., “Forml: Learning to reweight data for fairness,” arXiv preprint arXiv:2202.01719, 2022.
- [32] Williams and et al., “Simple statistical gradient-following algorithms for connectionist reinforcement learning,” Machine learning, vol. 8, no. 3-4, pp. 229–256, 1992.
- [33] Weiss and et al., “Cost-sensitive learning vs. sampling: Which is best for handling unbalanced classes with unequal error costs?,” Dmin, vol. 7, no. 35-41, p. 24, 2007.
- [34] Jozani and et al., “On estimation with weighted balanced-type loss function,” Statistics & Probability Letters, vol. 76, no. 8, pp. 773–780, 2006.
- [35] Groh and et al., “Evaluating deep neural networks trained on clinical images in dermatology with the fitzpatrick 17k dataset,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1820–1828, 2021.
- [36] Howard and et al., “Searching for mobilenetv3,” in Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), October 2019.
- [37] Zhang and et al., “Shufflenet: An extremely efficient convolutional neural network for mobile devices,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 6848–6856, 2018.
- [38] “Fitzpatrick scale.” https://suttonderm.com/whats-your-type/.