Fair Attribute Completion on Graph with Missing Attributes
Abstract
Tackling unfairness in graph learning models is a challenging task, as the unfairness issues on graphs involve both attributes and topological structures. Existing work on fair graph learning simply assumes that attributes of all nodes are available for model training and then makes fair predictions. In practice, however, the attributes of some nodes might not be accessible due to missing data or privacy concerns, which makes fair graph learning even more challenging. In this paper, we propose FairAC, a fair attribute completion method, to complement missing information and learn fair node embeddings for graphs with missing attributes. FairAC adopts an attention mechanism to deal with the attribute missing problem and meanwhile, it mitigates two types of unfairness, i.e., feature unfairness from attributes and topological unfairness due to attribute completion. FairAC can work on various types of homogeneous graphs and generate fair embeddings for them and thus can be applied to most downstream tasks to improve their fairness performance. To our best knowledge, FairAC is the first method that jointly addresses the graph attribution completion and graph unfairness problems. Experimental results on benchmark datasets show that our method achieves better fairness performance with less sacrifice in accuracy, compared with the state-of-the-art methods of fair graph learning. Code is available at: https://github.com/donglgcn/FairAC.
1 Introduction
Graphs, such as social networks, biomedical networks, and traffic networks, are commonly observed in many real-world applications. A lot of graph-based machine learning methods have been proposed in the past decades, and they have shown promising performance in tasks like node similarity measurement, node classification, graph regression, and community detection. In recent years, graph neural networks (GNNs) have been actively studied (Scarselli et al., 2008; Wu et al., 2020; Jiang et al., 2019; 2020; Zhu et al., 2021c; b; a; Hua et al., 2020; Chu et al., 2021), which can model graphs with high-dimensional attributes in the non-Euclidean space and have achieved great success in many areas such as recommender systems (Sheu et al., 2021). However, it has been observed that many graphs are biased, and thus GNNs trained on the biased graphs may be unfair with respect to certain sensitive attributes such as demographic groups. For example, in a social network, if the users with the same gender have more active connections, the GNNs tend to pay more attention to such gender information and lead to gender bias by recommending more friends to a user with the same gender identity while ignoring other attributes like interests. And from the data privacy perspective, it is possible to infer one’s sensitive information from the results given by GNNs (Sun et al., 2018). In a time when GNNs are widely deployed in the real world, this severe unfairness is unacceptable. Thus, fairness in graph learning emerges and becomes notable very recently.
Existing work on fair graph learning mainly focuses on the pre-processing, in-processing, and post-processing steps in the graph learning pipeline in order to mitigate the unfairness issues. The pre-processing approaches modify the original data to conceal sensitive attributes. Fairwalk (Rahman et al., 2019) is a representative pre-processing method, which enforces each group of neighboring nodes an equal chance to be chosen in the sampling process. In many in-processing methods, the most popular way is to add a sensitive discriminator as a constraint, in order to filter out sensitive information from original data. For example, FairGNN (Dai & Wang, 2021) adopts a sensitive classifier to filter node embeddings. CFC (Bose & Hamilton, 2019) directly adds a filter layer to deal with unfairness issues. The post-processing methods directly force the final prediction to satisfy fairness constraints, such as (Hardt et al., 2016).
When the graphs have complete node attributes, existing fair graph learning methods could obtain promising performance on both fairness and accuracy. However, in practice, graphs may contain nodes whose attributes are entirely missing due to various reasons (e.g., newly added nodes, and data privacy concerns). Taking social networks as an example, a newly registered user may have incomplete profiles. Given such incomplete graphs, existing fair graph learning methods would fail, as they assume all the nodes have attributes for model training. Although FairGNN (Dai & Wang, 2021) also involves the missing attribute problem, it only assumes that a part of the sensitive attributes are missing. To the best of our knowledge, addressing the unfairness issue on graphs with some nodes whose attributes are entirely missing has not been investigated before. Another relevant topic is graph attribute completion (Jin et al., 2021; Chen et al., 2020). It mainly focuses on completing a precise graph but ignores the unfairness issues. In this work, we aim to jointly complete a graph with missing attributes and mitigate unfairness at both feature and topology levels.
In this paper, we study the new problem of learning fair embeddings for graphs with missing attributes. Specifically, we aim to address two major challenges: (1) how to obtain meaningful node embeddings for graphs with missing attributes, and (2) how to enhance fairness of node embeddings with respect to sensitive attributes. To address these two challenges, we propose a Fair Attribute Completion (FairAC) framework. For the first challenge, we adopt an autoencoder to obtain feature embeddings for nodes with attributes and meanwhile we adopt an attention mechanism to aggregate feature information of nodes with missing attributes from their direct neighbors. Then, we address the second challenge by mitigating two types of unfairness, i.e., feature unfairness and topological unfairness. We adopt a sensitive discriminator to regulate embeddings and create a bias-free graph.
The main contributions of this paper are as follows: (1) We present a new problem of achieving fairness on a graph with missing attributes. Different from the existing work, we assume that the attributes of some nodes are entirely missing. (2) We propose a new framework, FairAC, for fair graph attribute completion, which jointly addresses unfairness issues from the feature and topology perspectives. (3) FairAC is a generic approach to complete fair graph attributes, and thus can be used in many graph-based downstream tasks. (4) Extensive experiments on benchmark datasets demonstrate the effectiveness of FairAC in eliminating unfairness and maintaining comparable accuracy.
2 Related Work
2.1 Fairness in Graph Learning
Recent work promotes fairness in graph-based machine learning (Bose & Hamilton, 2019; Rahman et al., 2019; Dai & Wang, 2021; Wang et al., 2022). They can be roughly divided into three categories, i.e., the pre-processing methods, in-processing methods, and post-processing methods.
The pre-processing methods are applied before training downstream tasks by modifying training data. For instance, Fairwalk (Rahman et al., 2019) improves the sampling procedure of node2vec (Grover & Leskovec, 2016). Our FairAC framework can be viewed as a pre-processing method, as it seeks to complete node attributes and use them as input of graph neural networks. However, our problem is much harder than existing problems, because the attributes of some nodes in the graph are entirely missing, including both the sensitive ones and non-sensitive ones. Given an input graph with missing attributes, FairAC generates fair and complete feature embeddings and thus can be applied to many downstream tasks, such as node classification, link prediction (Liben-Nowell & Kleinberg, 2007; Taskar et al., 2003), PageRank (Haveliwala, 2003), etc. Graph learning models trained on the refined feature embeddings would make fair predictions in downstream tasks.
There are plenty of fair graph learning methods as in-processing solutions. Some work focus on dealing with unfairness issues on graphs with complete features. For example, GEAR (Ma et al., 2022) mitigates graph unfairness by counterfactual graph augmentation and an adversarial learning method to learn sensitive-invariant embeddings. However, in order to generate counterfactual subgraphs, they need precise and entire features for every node. In other words, it cannot work well if it encounters a graph with full missing nodes since it cannot generate counterfactual subgraph based on a blank node. But we can deal with the situation. The most related work is FairGNN (Dai & Wang, 2021). Different from the majority of problem settings on graph fairness. It learns fair GNNs for node classification in a graph where only a limited number of nodes are provided with sensitive attributes. FairGNN adopts a sensitive classifier to predict the missing sensitive labels. After that, it employs a classic adversarial model to mitigate unfairness.Specifically, a sensitive discriminator aims to predict the known or estimated sensitive attributes, while a GNN model tries to fool the sensitive discriminator and meanwhile predicts node labels. However, it cannot predict sensitive information if a node misses all features in the first place and thus will fail to achieve its final goal. Our FairAC can get rid of the problem because we recover the node embeddings from their neighbors. FairAC learns attention between neighbors according to existing full attribute nodes, so we can recover the node embeddings for missing nodes from their neighbors by aggregating the embeddings of neighbors. With the help of the adversarial learning method, it can also remove sensitive information. In addition to attribute completion, we have also designed novel de-biasing strategies to mitigate feature unfairness and topological unfairness.
2.2 Attribution Completion on Graphs
The problem of missing attributes is ubiquitous in reality. Several methods (Liao et al., 2016; You et al., 2020; Chen et al., 2020; He et al., 2022; Jin et al., 2021; 2022; Tu et al., 2022; Taguchi et al., 2021) have been proposed to address this problem. GRAPE (You et al., 2020) tackles the problem of missing attributes in tabular data using a graph-based approach. SAT (Chen et al., 2020) assumes that the topology representation and attributes share a common latent space, and thus the missing attributes can be recovered by aligning the paired latent space. He et al. (2022) and Jin et al. (2021) extend such problem settings to heterogeneous graphs. HGNN-AC (Jin et al., 2021) is an end-to-end model, which does not recover the original attributes but generates attribute representations that have sufficient information for the final prediction task. It is worth noting that existing methods on graph attribute completion only focus on the attribute completion accuracy or performance of downstream tasks, but none of them takes fairness into consideration. Instead, our work pays attention to the unfairness issue in graph learning, and we aim to generate fair feature embeddings for each node by attribute completion, which contain the majority of information inherited from original attributes but disentangle the sensitive information.

3 Methodology
3.1 Problem Definition
Let denote an undirected graph, where is the set of nodes, is the set of undirected edges in the graph, is the node attribute matrix, and is the dimension of attributes. is the adjacency matrix of the graph , where if nodes and are connected; otherwise, . In addition, denotes a set of sensitive attributes (e.g., age or gender) of nodes, and denotes the node labels. The goal of fair graph learning is to make fair predictions of node labels with respect to the sensitive attribute, which is usually measured by certain fairness notations like statistical parity (Dwork et al., 2012) and equal opportunity (Hardt et al., 2016). Statistical Parity and Equal Opportunity are two group fairness definitions. Their detailed formulations are presented below. The label denotes the ground-truth node label, and the sensitive attribute indicates one’s sensitive group. For example, for binary node classification task, y only has two labels. Here we consider two sensitive groups, i.e. .
-
•
Statistical Parity (Dwork et al., 2012). It refers to the equal acceptance rate, which can be formulated as:
(1) where denotes the probability that occurs.
-
•
Equal Opportunity (Hardt et al., 2016). It means the probability of a node in a positive class being classified as a positive outcome should be equal for both sensitive group nodes. It mathematically requires an equal true positive rate for each subgroup.
(2)
In this work, we mainly focus on addressing unfairness issues on graphs with missing attributes, i.e., attributes of some nodes are totally missing. Let denote the set of nodes whose attributes are available, and denote the set of nodes whose attributes are missing, . If , both and are unavailable during model training.
With the notations given below, the fair attribute completion problem is formally defined as:
Problem 1. Given a graph , where node set with the corresponding attributes available and the corresponding sensitive attributes in , learn a fair attribute completion model to generate fair feature embeddings for each node in , i.e.,
(3) |
where is the function we aim to learn. should exclude any sensitive information while preserve non-sensitive information.
3.2 Fair Attribute Completion (FairAC) Framework
Input: ,
Output: Autoencoder , Sensitive classifier , Attribute completion
We propose a fair attribute completion (FairAC) framework to address Problem 1. Existing fair graph learning methods tackle unfairness issues by training fair graph neural networks in an end-to-end fashion, but they cannot effectively handle graphs that are severely biased due to missing attributes. Our FairAC framework, as a data-centric approach, deals with the unfairness issue from a new perspective, by explicitly debiasing the graph with feature unfairness mitigation and fairness-aware attribute completion. Eventually, FairAC generates fair embeddings for all nodes including the ones without any attributes. The training algorithms are shown in Algorithm 1.
To train the graph attribute completion model, we follow the setting in (Jin et al., 2021) and divide the nodes with attributes (i.e., ) into two sets: and . For nodes in , we keep their attributes, while for nodes in , we temporally drop their attributes and try to recover them using our attribute completion model. Although the nodes are randomly assigned to and , the proportion of is consistent with the attribute missing rate of graph , i.e., .
Different from existing work on fair graph learning, we consider unfairness from two sources. The first one is from node features. For example, we can roughly infer one’s sensitive information, like gender, from some non-sensitive attributes like hobbies. It means that non-sensitive attributes may imply sensitive attributes and thus lead to unfairness in model prediction. We adopt a sensitive discriminator to mitigate feature unfairness. The other source is topological unfairness introduced by graph topological embeddings and node attribute completion. To deal with the topological unfairness, we force the estimated feature embeddings to fool the sensitive discriminator, by updating attention parameters during the attribute completion process.
As illustrated in Figure 1, our FairAC framework first mitigates feature unfairness for nodes with attributes (i.e., ) by removing sensitive information implicitly contained in non-sensitive attributes with an auto-encoder and sensitive classifier (Section 3.2.1). For nodes without features (i.e., ), FairAC performs attribute completion with an attention mechanism (Section 3.2.2) and meanwhile mitigates the topological unfairness (Section 3.2.3). Finally, the FairAC model trained on and can be used to infer fair embeddings for nodes in .
The overall loss function of FairAC is formulated as:
(4) |
where represents the loss for mitigating feature unfairness, is the loss for attribute completion, and is the loss for mitigating topological unfairness. is a trade-off hyperparameter.
3.2.1 Mitigating feature unfairness
The nodes in have full attributes , while some attributes may implicitly encode information about sensitive attributes and thus lead to unfair predictions. To address this issue, FairAC aims to encode the attributes of node into a fair feature embedding . Specifically, we use a simple autoencoder framework together with a sensitive classifier. The autoencoder maps into embedding , and meanwhile the sensitive classifier is trained in an adversarial way, such that the embeddings are invariant to sensitive attributes.
Autoencoder.
The autoencoder contains an encoder and a decoder . encodes the original attributes to feature embeddings , i.e., , and reconstructs attributes from the latent embeddings, i.e., , where the reconstructed attributes should be close to as possible. The loss function of the autoencoder is written as:
(5) |
Sensitive classifier
The sensitive classifier is a simple multilayer perceptron (MLP) model. It takes the feature embedding as input and predicts the sensitive attribute , i.e., . When the sensitive attributes are binary, we can use the binary cross entropy loss to optimize :
(6) |
With the sensitive classifier , we could leverage it to adversarially train the autoencoder, such that is able to generate fair feature embeddings that can fool . The loss is written as: .
3.2.2 Completing node embeddings via attention mechanism
For nodes without attributes (), FairAC makes use of topological embeddings and completes the node embeddings with an attention mechanism.
Topological embeddings.
Recent studies reveal that the topology of graphs has similar semantic information as the attributes (Chen et al., 2020; McPherson et al., 2001; Pei et al., 2020; Zhu et al., 2020). Inspired by this observation, we assume that the nodes’ topological information can reflect the relationship between nodes’ attributes and the attributes of their neighbors. There are a lot of off-the-shelf node topological embedding methods, such as DeepWalk (Perozzi et al., 2014) and node2vec (Grover & Leskovec, 2016). For simplicity, we adopt the DeepWalk method to extract topological embeddings for nodes in .
Attention mechanism.
For graphs with missing attributes, a commonly used strategy is to use average attributes of the one-hop neighbors. This strategy works in some cases, however, simply averaging information from neighbors might be biased, as the results might be dominated by some high-degree nodes. In fact, different neighbors should have varying contributions to the aggregation process in the context of fairness. To this end, FairAC adopts an attention mechanism (Vaswani et al., 2017) to learn the influence of different neighbors or edges with the awareness of fairness, and then aggregates attributes information for nodes in .
Given a pair of nodes which are neighbors, the contribution of node is the attention , which is defined as: , where are the topological embeddings of nodes and , respectively. Specifically, we only focus on the neighbor pairs and ignore those node pairs that are not directly connected. denotes the attention between two topological embeddings, i.e., , where is the learnable parametric matrix, and is an activation function. After we get all the attention scores between one node and its neighbors, we can get the coefficient of each pair by applying the softmax function:
(7) |
where is the coefficient of node pair , and is the set of neighbors of node . For node , FairAC calculates its feature embedding by the weighted aggregation with multi-head attention:
(8) |
where is the number of attention heads. The loss for attribute completion with topological embedding and attention mechanism is formulated as:
(9) |
3.2.3 Mitigating topological unfairness
The attribute completion procedure may introduce topological unfairness since we assume that topology information is similar to attributes relation. It is possible that the completed feature embeddings of would be unfair with respect to sensitive attributes . To address this issue, FairAC leverages sensitive classifier to help mitigate topological unfairness by further updating the attention parameter matrix and thus obtaining fair feature embeddings . Inspired by (Gong et al., 2020), we expect that the feature embeddings can fool the sensitive classifier to predict the probability distribution close to the uniform distribution over the sensitive category, by minimizing the loss:
(10) |
3.3 FairAC for Node Classification
The proposed FairAC framework could be viewed as a generic data debiasing approach, which achieves fairness-aware attribute completion and node embedding for graphs with missing attributes. It can be easily integrated with many existing graph neural networks (e.g., GCN (Kipf & Welling, 2016), GAT (Veličković et al., 2018), and GraphSAGE (Hamilton et al., 2017)) for tasks like node classification. In this work, we choose the basic GCN model for node classification and assess how FairAC enhances model performance in terms of accuracy and fairness.
4 Experiments
Dataset | Method | M | Acc | AUC | + | ||
---|---|---|---|---|---|---|---|
GCN | ✓ | 70.660.24 | 74.230.63 | 3.432.44 | 2.740.67 | 6.163.1 | |
ALFR | 64.31.3 | 71.50.3 | 2.30.9 | 3.21.5 | 5.52.4 | ||
ALFR-e | 66.00.4 | 72.91.0 | 4.71.8 | 4.71.7 | 9.43.4 | ||
NBA | Debias | 63.11.1 | 71.30.7 | 2.51.5 | 3.11.9 | 5.63.4 | |
Debias-e | 65.62.4 | 72.91.2 | 5.30.9 | 3.11.3 | 8.42.2 | ||
FCGE | 66.01.5 | 73.61.5 | 2.91.0 | 3.01.2 | 5.92.2 | ||
FairGNN | ✓ | 70.730.44 | 76.770.1 | 0.950.7 | 1.630.67 | 2.581.37 | |
FairAC (Ours) | ✓ | 70.660.73 | 74.440.67 | 0.280.25 | 0.630.34 | 0.910.59 | |
GCN | ✓ | 67.40.88 | 72.041.78 | 1.960.64 | 4.170.54 | 6.121.18 | |
ALFR | 65.40.3 | 71.30.3 | 2.80.5 | 1.10.4 | 3.90.9 | ||
ALFR-e | 68.00.6 | 74.00.7 | 5.80.4 | 2.80.8 | 8.61.2 | ||
Pokec-z | Debias | 65.20.7 | 71.40.6 | 1.90.6 | 1.90.4 | 3.81.0 | |
Debias-e | 67.50.7 | 74.20.7 | 4.71.0 | 3.01.4 | 7.72.4 | ||
FCGE | 65.90.2 | 71.00.2 | 3.10.5 | 1.70.6 | 4.81.1 | ||
FairGNN | ✓ | 66.540.45 | 70.100.07 | 0.950.70 | 1.630.67 | 2.580.57 | |
FairAC (Ours) | ✓ | 66.940.14 | 72.870.13 | 0.190.07 | 0.120.07 | 0.310.14 | |
GCN | ✓ | 66.120.88 | 71.50.1 | 0.460.1 | 1.410.14 | 1.870.24 | |
ALFR | 63.10.6 | 67.70.5 | 3.050.5 | 3.90.6 | 3.951.1 | ||
ALFR-e | 66.20.4 | 71.91.0 | 4.11.8 | 4.61.7 | 8.73.5 | ||
Pokec-n | Debias | 62.61.1 | 67.90.7 | 2.41.5 | 2.61.9 | 5.03.4 | |
Debias-e | 65.62.4 | 71.71.2 | 3.60.9 | 4.41.3 | 8.02.2 | ||
FCGE | 64.81.5 | 69.51.5 | 4.11.0 | 5.51.2 | 9.62.2 | ||
FairGNN | ✓ | 68.540.45 | 70.100.07 | 0.760.15 | 0.480.09 | 1.240.24 | |
FairAC (Ours) | ✓ | 66.350.24 | 72.320.08 | 0.270.12 | 0.140.11 | 0.410.23 |
In this section, we evaluate the performance of the proposed FairAC framework on three benchmark datasets in terms of node classification accuracy and fairness w.r.t. sensitive attributes. We compare FairAC with other baseline methods in settings with various sensitive attributes or different attribute missing rates. Ablation studies are also provided and discussed.
4.1 Datasets and Settings
Datasets. In the experiments, we use three public graph datasets, NBA, Pokec-z, and Pokec-n. A detailed description is shown in supplementary materials.
Baselines. We compare our FairAC method with the following baseline methods: GCN (Kipf & Welling, 2016), ALFR (Edwards & Storkey, 2015), ALFR-e, Debias (Zhang et al., 2018), Debias-e, FCGE (Bose & Hamilton, 2019), and FairGNN (Dai & Wang, 2021). ALFR-e concatenates the feature embeddings produced by ALFR with topological embeddings learned by DeepWalk (Perozzi et al., 2014). Debias-e also concatenates the topological embeddings learned by DeepWalk with feature embeddings learned by Debias. FairGNN is an end-to-end debias method which aims to mitigate unfairness in label prediction task. GCN and FairGNN uses the average attribute completion method, while other baselines use original complete attributes.
Evaluation Metrics. We evaluate the proposed framework with respect to two aspects: classification performance and fairness performance. For classification, we use accuracy and AUC scores.As for fairness, we adopt and as evaluation metrics, which can be defined as:
(11) |
(12) |
The smaller and are, the more fair the model is. In addition, we use + as an overall indicator of a model’s performance on fairness.
Method | Acc (%) | AUC (%) | (%) | (%) | + | |
---|---|---|---|---|---|---|
GCN | 66.10 | 69.14 | 0.88 | 0.20 | 1.08 | |
0.1 | FairGNN | 69.37 | 76.93 | 0.17 | 0.29 | 0.46 |
BaseAC (Ours) | 66.37 | 69.34 | 0.07 | 1.46 | 1.53 | |
FairAC (Ours) | 66.33 | 69.35 | 0.14 | 0.37 | 0.51 | |
GCN | 67.40 | 72.04 | 1.96 | 4.17 | 6.12 | |
0.3 | FairGNN | 66.54 | 70.10 | 0.95 | 1.63 | 2.58 |
BaseAC (Ours) | 66.10 | 69.67 | 0.78 | 1.57 | 2.35 | |
FairAC (Ours) | 66.94 | 72.87 | 0.19 | 0.12 | 0.31 | |
GCN | 66.41 | 70.14 | 0.07 | 2.11 | 2.18 | |
0.5 | FairGNN | 66.25 | 70.13 | 0.47 | 1.71 | 2.18 |
BaseAC (Ours) | 66.13 | 69.93 | 0.38 | 1.60 | 1.98 | |
FairAC (Ours) | 66.45 | 72.95 | 0.14 | 1.06 | 1.20 | |
GCN | 66.99 | 73.13 | 0.57 | 1.74 | 2.31 | |
0.8 | FairGNN | 66.60 | 71.35 | 0.94 | 0.09 | 1.03 |
BaseAC (Ours) | 66.06 | 71.21 | 0.08 | 2.20 | 2.28 | |
FairAC (Ours) | 66.10 | 71.66 | 0.01 | 0.69 | 0.70 |
4.2 Results and Analysis
4.2.1 Unfairness issues in Graph Neural Networks
According to the results showed in Table 1, they reveal several unfairness issues in Graph Neural Networks. We divided them into two categories.
-
•
Feature unfairness Feature unfairness is that some non-sensitive attributes could infer sensitive information. Hence, some Graph Neural Networks may learn this relation and make unfair prediction. In most cases, ALFR and Debias and FCGE have better fairness performance than GCN method. It is as expected because the non-sensitive features may contain proxy variables of sensitive attributes which would lead to biased prediction. Thus, ALFR and Debias methods that try to break up these connections are able to mitigate feature unfairness and obtain better fairness performance. These results further prove the existence of feature unfairness.
-
•
Topological unfairness Topological unfairness is sourced from graph structure. In other words, edges in graph, i.e. the misrepresentation due to the connection(Mehrabi et al., 2021) can bring topological unfairness. From the experiments, ALFR-e and Debias-e have worse fairness performance than ALFR and Debias, respectively. It shows that although graph structure can improve the classification performance, it will bring topological unfairness consequently. The worse performance on fairness verifies that topological unfairness exists in GNNs and graph topological information could magnify the discrimination.
4.2.2 Effectiveness of FairAC on mitigating feature and topological unfairness
The results of our FairAC method and baselines in terms of the node classification accuracy and fairness metrics on three datasets are shown in Table 1. The best results are shown in bold. Generally speaking, we have the following observations. (1). The proposed method FairAC shows comparable classification performance with these baselines, GCN and FairGNN. This suggests that our attribute completion method is able to preserve useful information contained in the original attributes. (2). FairAC outperforms all baselines regarding fairness metrics, especially in +. FairAC outperform baselines that focus on mitigate feature fairness, like ALFR, which proves that FairAC also mitigate topological unfairness. Besides, it is better than those who take topological fairness into consideration, like FCGE, which also validates the effectiveness of FairAC. FairGNN also has good performance on fairness, because it adopts a discriminator to deal with the unfairness issue. Our method performs better than FairGNN in most cases. For example, our FairAC method can significantly improve the performance in terms of the fairness metric , i.e., 65, 87, and 67 improvement over FairGNN on the NBA, pokec-z, pokec-n datasets, respectively. Overall, the results in Table 1 validate the effectiveness of FairAC in mitigating unfairness issues.
4.3 Ablation Studies
Attribute missing rate
In our proposed framework, the attribute missing rate indicates the integrity of node attribute matrix, which has a great impact on model performance. Here we investigate the performance of our FairAC method and baselines on dealing with graphs with varying degrees of missing attributes. In particular, we set the attribute missing rate to 0.1, 0.3, 0.5 and 0.8, and evaluate FairAC and baselines on the pokec-z dataset. The detailed results are presented in Table 2. From the table, we have the following observation that with varying values of , FairAC is able to maintain its high fairness performance. Especially when reaches 0.8, FairAC can greatly outperform other methods. It proves that FairAC is effective even if the attributes are largely missing.
The effectiveness of adversarial learning
A key module in FairAC is adversarial learning, which is used to mitigate feature unfairness and topological unfairness. To investigate the contribution of adversarial learning in FairAC, we implement a BaseAC model, which only has the attention-based attribute completion module, but does not contain the adversarial learning loss terms. Comparing BaseAC with FairAC in Table 2, we can find that the fairness performance drops desperately when the adversarial training loss is removed. Since BaseAC does not have an adversarial discriminator to regulate feature encoder as well as attribute completion parameters, it is unable to mitigate unfairness. Overall, the results confirm the effectiveness of the adversarial learning module.

Parameter analysis
We investigate how the hyperparameters affect the performance of FairAC. The most important hyperparameter in FairAC is , which adjusts the trade-off between fairness and attribute completion. We report the results with different hyperparameter values. We set to 0.2, 0.4, 0.7, 0.8 and 0 that is equivalent to the BaseAC. We also fix other hyperparameters by setting to 0.3. As shown in Figure 2, we can find that, as increases, the fairness performance improves while the accuracy of node classification slightly declined. Therefore, it validates our assumption that there is a trade-off between fairness and attribute completion, and our FairAC is able to enhance fairness without compromising too much on accuracy.
5 Conclusions
In this paper, we presented a novel problem, i.e., fair attribute completion on graphs with missing attributes. To address this problem, we proposed the FairAC framework, which jointly completes the missing features and mitigates unfairness. FairAC leverages the attention mechanism to complete missing attributes and adopts a sensitive classifier to mitigate implicit feature unfairness as well as topological unfairness on graphs. Experimental results on three real-world datasets demonstrate the superiority of the proposed FairAC framework over baselines in terms of both node classification performance and fairness performance. As a generic fair graph attributes completion approach, FairAC can also be used in other graph-based downstream tasks, such as link prediction, graph regression, pagerank, and clustering.
Acknowledgement
This research is supported by the Cisco Faculty Award and Adobe Data Science Research Award.
References
- Bose & Hamilton (2019) Avishek Bose and William Hamilton. Compositional fairness constraints for graph embeddings. In International Conference on Machine Learning, pp. 715–724. PMLR, 2019.
- Chen et al. (2020) Xu Chen, Siheng Chen, Jiangchao Yao, Huangjie Zheng, Ya Zhang, and Ivor W Tsang. Learning on attribute-missing graphs. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
- Chu et al. (2021) Zhixuan Chu, Stephen L Rathbun, and Sheng Li. Graph infomax adversarial learning for treatment effect estimation with networked observational data. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 176–184, 2021.
- Dai & Wang (2021) Enyan Dai and Suhang Wang. Say no to the discrimination: Learning fair graph neural networks with limited sensitive attribute information. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, pp. 680–688, 2021.
- Dwork et al. (2012) Cynthia Dwork, Moritz Hardt, Toniann Pitassi, Omer Reingold, and Richard Zemel. Fairness through awareness. In Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, pp. 214–226, 2012.
- Edwards & Storkey (2015) Harrison Edwards and Amos Storkey. Censoring representations with an adversary, 2015. URL https://arxiv.org/abs/1511.05897.
- Gong et al. (2020) Sixue Gong, Xiaoming Liu, and Anil K Jain. Jointly de-biasing face recognition and demographic attribute estimation. In European Conference on Computer Vision, pp. 330–347. Springer, 2020.
- Grover & Leskovec (2016) Aditya Grover and Jure Leskovec. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 855–864, 2016.
- Hamilton et al. (2017) Will Hamilton, Zhitao Ying, and Jure Leskovec. Inductive representation learning on large graphs. Advances in Neural Information Processing Systems, 30, 2017.
- Hardt et al. (2016) Moritz Hardt, Eric Price, and Nati Srebro. Equality of opportunity in supervised learning. Advances in Neural Information Processing Systems, 29, 2016.
- Haveliwala (2003) Taher H Haveliwala. Topic-sensitive pagerank: A context-sensitive ranking algorithm for web search. IEEE Transactions on Knowledge and Data Engineering, 15(4):784–796, 2003.
- He et al. (2022) Dongxiao He, Chundong Liang, Cuiying Huo, Zhiyong Feng, Di Jin, Liang Yang, and Weixiong Zhang. Analyzing heterogeneous networks with missing attributes by unsupervised contrastive learning. IEEE Transactions on Neural Networks and Learning Systems, 2022.
- Hua et al. (2020) Yuncheng Hua, Yuan-Fang Li, Guilin Qi, Wei Wu, Jingyao Zhang, and Daiqing Qi. Less is more: Data-efficient complex question answering over knowledge bases. Journal of Web Semantics, 65:100612, 2020.
- Jiang et al. (2019) Xiaodong Jiang, Pengsheng Ji, and Sheng Li. Censnet: convolution with edge-node switching in graph neural networks. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, pp. 2656–2662, 2019.
- Jiang et al. (2020) Xiaodong Jiang, Ronghang Zhu, Pengsheng Ji, and Sheng Li. Co-embedding of nodes and edges with graph neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2020.
- Jin et al. (2021) Di Jin, Cuiying Huo, Chundong Liang, and Liang Yang. Heterogeneous graph neural network via attribute completion. In Proceedings of the Web Conference 2021, pp. 391–400, 2021.
- Jin et al. (2022) Di Jin, Rui Wang, Tao Wang, Dongxiao He, Weiping Ding, Yuxiao Huang, Longbiao Wang, and Witold Pedrycz. Amer: A new attribute-missing network embedding approach. IEEE Transactions on Cybernetics, pp. 1–14, 2022. doi: 10.1109/TCYB.2022.3166539.
- Kingma & Ba (2014) Diederik P Kingma and Jimmy Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
- Kipf & Welling (2016) Thomas N Kipf and Max Welling. Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907, 2016.
- Liao et al. (2016) Lizi Liao, Qirong Ho, Jing Jiang, and Ee-Peng Lim. Slr: A scalable latent role model for attribute completion and tie prediction in social networks. In 2016 IEEE 32nd International Conference on Data Engineering (ICDE), pp. 1062–1073. IEEE, 2016.
- Liben-Nowell & Kleinberg (2007) David Liben-Nowell and Jon Kleinberg. The link-prediction problem for social networks. Journal of the American society for information science and technology, 58(7):1019–1031, 2007.
- Ma et al. (2022) Jing Ma, Ruocheng Guo, Mengting Wan, Longqi Yang, Aidong Zhang, and Jundong Li. Learning fair node representations with graph counterfactual fairness. In Proceedings of the Fifteenth ACM International Conference on Web Search and Data Mining, pp. 695–703, 2022.
- McPherson et al. (2001) Miller McPherson, Lynn Smith-Lovin, and James M Cook. Birds of a feather: Homophily in social networks. Annual review of sociology, 27(1):415–444, 2001.
- Mehrabi et al. (2021) Ninareh Mehrabi, Fred Morstatter, Nripsuta Saxena, Kristina Lerman, and Aram Galstyan. A survey on bias and fairness in machine learning. ACM Computing Surveys (CSUR), 54(6):1–35, 2021.
- Pei et al. (2020) Hongbin Pei, Bingzhe Wei, Kevin Chen-Chuan Chang, Yu Lei, and Bo Yang. Geom-gcn: Geometric graph convolutional networks. arXiv preprint arXiv:2002.05287, 2020.
- Perozzi et al. (2014) Bryan Perozzi, Rami Al-Rfou, and Steven Skiena. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, pp. 701–710, 2014.
- Rahman et al. (2019) Tahleen Rahman, Bartlomiej Surma, Michael Backes, and Yang Zhang. Fairwalk: Towards fair graph embedding. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, pp. 3289–3295. International Joint Conferences on Artificial Intelligence Organization, 7 2019. doi: 10.24963/ijcai.2019/456. URL https://doi.org/10.24963/ijcai.2019/456.
- Rozemberczki et al. (2020) Benedek Rozemberczki, Oliver Kiss, and Rik Sarkar. Karate Club: An API Oriented Open-source Python Framework for Unsupervised Learning on Graphs. In Proceedings of the 29th ACM International Conference on Information and Knowledge Management (CIKM ’20), pp. 3125–3132. ACM, 2020.
- Scarselli et al. (2008) Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini. The graph neural network model. IEEE Transactions on Neural Networks and Learning Systems, 20(1):61–80, 2008.
- Sheu et al. (2021) Heng-Shiou Sheu, Zhixuan Chu, Daiqing Qi, and Sheng Li. Knowledge-guided article embedding refinement for session-based news recommendation. IEEE Transactions on Neural Networks and Learning Systems, 33(12):7921–7927, 2021.
- Sun et al. (2018) Lichao Sun, Yingtong Dou, Carl Yang, Ji Wang, Philip S Yu, Lifang He, and Bo Li. Adversarial attack and defense on graph data: A survey. arXiv preprint arXiv:1812.10528, 2018.
- Taguchi et al. (2021) Hibiki Taguchi, Xin Liu, and Tsuyoshi Murata. Graph convolutional networks for graphs containing missing features. Future Generation Computer Systems, 117:155–168, 2021.
- Takac & Zabovsky (2012) Lubos Takac and Michal Zabovsky. Data analysis in public social networks. In International scientific conference and international workshop present day trends of innovations, volume 1. Present Day Trends of Innovations Lamza Poland, 2012.
- Taskar et al. (2003) Ben Taskar, Ming-Fai Wong, Pieter Abbeel, and Daphne Koller. Link prediction in relational data. Advances in Neural Information Processing Systems, 16, 2003.
- Tu et al. (2022) Wenxuan Tu, Sihang Zhou, Xinwang Liu, Yue Liu, Zhiping Cai, En Zhu, Changwang Zhang, and Jieren Cheng. Initializing then refining: A simple graph attribute imputation network. In Lud De Raedt (ed.), Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22, pp. 3494–3500. International Joint Conferences on Artificial Intelligence Organization, 7 2022. doi: 10.24963/ijcai.2022/485. URL https://doi.org/10.24963/ijcai.2022/485. Main Track.
- Vaswani et al. (2017) Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in Neural Information Processing Systems, 30, 2017.
- Veličković et al. (2018) Petar Veličković, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua Bengio. Graph attention networks. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=rJXMpikCZ.
- Wang et al. (2022) Nan Wang, Lu Lin, Jundong Li, and Hongning Wang. Unbiased graph embedding with biased graph observations. In Proceedings of the ACM Web Conference 2022, pp. 1423–1433, 2022.
- Wu et al. (2020) Zonghan Wu, Shirui Pan, Fengwen Chen, Guodong Long, Chengqi Zhang, and S Yu Philip. A comprehensive survey on graph neural networks. IEEE Transactions on Neural Networks and Learning Systems, 32(1):4–24, 2020.
- You et al. (2020) Jiaxuan You, Xiaobai Ma, Yi Ding, Mykel J Kochenderfer, and Jure Leskovec. Handling missing data with graph representation learning. Advances in Neural Information Processing Systems, 33:19075–19087, 2020.
- Zhang et al. (2018) Brian Hu Zhang, Blake Lemoine, and Margaret Mitchell. Mitigating unwanted biases with adversarial learning. In Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society, pp. 335–340, 2018.
- Zhu et al. (2020) Jiong Zhu, Yujun Yan, Lingxiao Zhao, Mark Heimann, Leman Akoglu, and Danai Koutra. Beyond homophily in graph neural networks: Current limitations and effective designs. Advances in Neural Information Processing Systems, 33:7793–7804, 2020.
- Zhu et al. (2021a) Ronghang Zhu, Xiaodong Jiang, Jiasen Lu, and Sheng Li. Cross-domain graph convolutions for adversarial unsupervised domain adaptation. IEEE Transactions on Neural Networks and Learning Systems, 2021a.
- Zhu et al. (2021b) Ronghang Zhu, Xiaodong Jiang, Jiasen Lu, and Sheng Li. Transferable feature learning on graphs across visual domains. In 2021 IEEE International Conference on Multimedia and Expo (ICME), pp. 1–6. IEEE, 2021b.
- Zhu et al. (2021c) Ronghang Zhu, Zhiqiang Tao, Yaliang Li, and Sheng Li. Automated graph learning via population based self-tuning gcn. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 2096–2100, 2021c.
Appendix A Appendix
A.1 Datasets and Settings
Datasets.
In the experiments, we use three public graph datasets, NBA, Pokec-z, and Pokec-n. The detailed explanation is shown in supplementary materials. The NBA dataset (Dai & Wang, 2021) is extended from a Kaggle dataset containing around 400 NBA basketball players. It provides the performance statistics of those players in the 2016-2017 season and their personal profiles, e.g., nationality, age, and salary. Their relationships are obtained from Twitter. We use their nationality, whether one is U.S. player or oversea player, as the sensitive attribute. The node label is binary, indicating whether the salary of the player is over median or not. Pokec (Takac & Zabovsky, 2012) is an online social network in Slovakia, which contains millions of anonymized data of users. It has a variety of attributes, such as gender, age, education, region, etc. Based on the region where users belong to, (Dai & Wang, 2021) sampled two datasets named as: Pokec-z and Pokec-n. In our experiments, we consider the region or gender as sensitive attribute, and working field as label for node classification. The statistics of three datasets are summarized in supplementary materials. The statistics of three datasets are summarized in Table 3.
Dataset | NBA | Pokec-z | Pokec-n |
---|---|---|---|
# of nodes | 403 | 67,797 | 66,569 |
# of edges | 16,570 | 882,765 | 729,129 |
Density | 0.10228 | 0.00019 | 0.00016 |
Baselines.
We compare our FairAC method with the following baseline methods:
-
•
GCN (Kipf & Welling, 2016) with average attribute completion. GCN is a classical graph neural network model, which has obtained very promising performance in numerous applications. The standard GCN cannot handle graphs with missing attributes. In the experiments, we use the average attribute completion strategy to preprocess the feature matrix, by using the averaged attributes of one’s neighbors to approximate the missing attributes. After average attribute completion, GCN takes the graph with completed feature matrix as inputs to learn node embeddings and predict node labels.
-
•
ALFR (Edwards & Storkey, 2015) with full attributes. This is a pre-processing method. It utilize a discriminator to remove the sensitive feature information in feature embeddings produced by an Autoencoder. Since this method need full sensitive attributes and full features, we give them complete information. In other words, the missing rate is set to 0.
-
•
ALFR-e with full attributes. Based on ALFR, ALFR-e utilize the topological information. It concatenates the feature embeddings produced by ALFR with topological embeddings learned by DeepWalk (Perozzi et al., 2014). It also relys on complete information.
-
•
Debias (Zhang et al., 2018) with full attributes. This is an in-processing method. It applies a discriminator on node classifier in order to make the probability distribution be the same w.r.t. sensitive attribute. Since the discriminator needs the full sensitive attributes, we provide full node features.
-
•
Debias-e with full attributes. Similar to ALFR-e. It also concatenates the topological embeddings learned by DeepWalk (Perozzi et al., 2014) with feature embeddings learned by Debias.
-
•
FCGE (Bose & Hamilton, 2019) with full attributes. It learns fair node embeddings in graph without node features through edge prediction only. An discriminator is also applied to mitigate sensitive information in topological perspective.
-
•
FairGNN (Dai & Wang, 2021) with average attribute completion. Although FairGNN trains a sensitive attribute discriminator as an adversarial regularizer to enhance the fairness of GNNs, it still cannot deal with graphs with missing attributes. Thus, we use the average attribute completion method to complete the feature matrix, and then train a FairGNN model for node classification.
Implementation Details.
Each dataset is randomly split into 75%/25% training/test set as (Dai & Wang, 2021). Besides, we randomly drop node attributes based on the attribute missing rate, , which means the attributes of nodes will be unavailable. For each datasets, we choose a specific attribute as the sensitive attribute. In particular, region, and nation are selected as the sensitive attribute for the pokec, and nba datasets, respectively. Unless otherwise specified, we generate 128-dimension node embeddings and set the attribute missing rate to 0.3, and set the hyperparameters of FairAC as: for pokec-z and nba datasets, and for pokec-n dataset. We adopt Adam (Kingma & Ba, 2014) with the learning rate of and weight decay as . We adopt the DeepWalk (Perozzi et al., 2014) method to generate topological embedding for each node. Specifically, we use the DeepWalk implementation provided by the Karate Club library (Rozemberczki et al., 2020). We set walk length as 100, embedding dimension as 64, window size as 5, and epochs as 10. To evaluate fairness of compared methods, we follow the widely used evaluation protocol in fair graph learning and set a threshold for accuracy, because there is a trade-off between accuracy and fairness. Since we mainly focus on the fairness metric, we set the accuracy threshold that all methods can satisfy. we evaluated our models three times and calculated the mean and standard deviation(std). We estimate the std of by adding std of and , because for some methods, we use the reported data from (Dai & Wang, 2021) which does not provide the metric.
Method | Acc (%) | AUC (%) | (%) | (%) | + | |
---|---|---|---|---|---|---|
GAT | 67.77 | 73.57 | 1.02 | 3.38 | 3.40 | |
0.3 | FairGNN | 66.55 | 68.64 | 0.45 | 0.99 | 1.44 |
FairAC (Ours) | 67.25 | 72.96 | 0.23 | 0.10 | 0.33 | |
GAT | 68.59 | 73.97 | 0.30 | 1.96 | 2.26 | |
0.5 | FairGNN | 68.09 | 72.22 | 0.81 | 1.55 | 2.36 |
FairAC (Ours) | 66.36 | 70.66 | 0.09 | 0.32 | 0.41 | |
GAT | 67.59 | 71.62 | 3.69 | 7.15 | 10.84 | |
0.7 | FairGNN | 62.36 | 67.99 | 2.95 | 3.55 | 6.50 |
FairAC (Ours) | 66.64 | 69.59 | 0.18 | 4.19 | 4.37 |
A.2 Additional Experiments
Evaluations on GAT (Veličković et al., 2018) model.
As discussed in the main paper, the proposed FairAC method can be easily integrated with existing graph neural networks. Extensive results in Section 4 of the main paper demonstrate that the combination of FairAC and GCN performs very well. In this section, we integrate FairAC with another representative graph neural network model, GAT (Veličković et al., 2018). The results of our method and two main baselines in terms of the node classification accuracy and fairness metrics are shown in Table 4. In these experiments, FairAC generates fair and complete node features, and then GAT is trained for node classification. We also investigate the performance of our FairAC method and baselines on dealing with graphs with varying degrees of missing attributes. We set the attribute missing rate to 0.1, 0.3, 0.5 and 0.7, and evaluate FairAC and baselines on the Pokec-n dataset. In addition, we set to 1.0. The best results are shown in bold. Generally speaking, we have the following observations. (1). The proposed method FairAC shows comparable classification performance with two baselines, GAT and FairGNN. This suggests that our attribute completion method is able to work well under different downstream models. It further demonstrates that FairAC can preserve useful information implied in the original attributes. (2). FairAC has comparable results with two baselines regarding fairness metrics. Especially when is greater than 0.3, FairAC can greatly outperform other methods, which proves that FairAC is effective even if the attributes are largely missing. Overall, the results in Table 4 validate the effectiveness of FairAC in mitigating unfairness issues and show the compatibility with varying downstream models.