This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

FLEE-GNN: A Federated Learning System for Edge-Enhanced Graph Neural Network in Analyzing Geospatial Resilience of Multicommodity Food Flows

Yuxiao Qu1,2, Jinmeng Rao2, Song Gao2, Qianheng Zhang2, Wei-Lun Chao3, Yu Su3, Michelle Miller4, Alfonso Morales5, Patrick R. Huber6 1Department of Computer Science, Carnegie Mellon University
2Geospatial Data Science lab, University of Wisconsin, Madison
3Department of Computer Science and Engineering, Ohio State University
4Center for Integrated Agricultural Systems, University of Wisconsin, Madison
5Department of Planning and Landscape Architecture, University of Wisconsin, Madison
6 Institute of the Environment, University of California, Davis
Email: [email protected]
(2023)
Abstract.

Understanding and measuring the resilience of food supply networks is a global imperative to tackle increasing food insecurity. However, the complexity of these networks, with their multidimensional interactions and decisions, presents significant challenges. This paper proposes FLEE-GNN, a novel Federated Learning System for Edge-Enhanced Graph Neural Network, designed to overcome these challenges and enhance the analysis of geospatial resilience of multicommodity food flow network, which is one type of spatial networks. FLEE-GNN addresses the limitations of current methodologies, such as entropy-based methods, in terms of generalizability, scalability, and data privacy. It combines the robustness and adaptability of graph neural networks with the privacy-conscious and decentralized aspects of federated learning on food supply network resilience analysis across geographical regions. This paper also discusses FLEE-GNN’s innovative data generation techniques, experimental designs, and future directions for improvement. The results show the advancements of this approach to quantifying the resilience of multicommodity food flow networks, contributing to efforts towards ensuring global food security using AI methods. The developed FLEE-GNN has the potential to be applied in other spatial networks with spatially heterogeneous sub-network distributions.

federated learning, smart foodsheds, food supply networks, GeoAI
journalyear: 2023copyright: acmlicensedconference: 6th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery; November 13, 2023; Hamburg, Germanybooktitle: 6th ACM SIGSPATIAL International Workshop on AI for Geographic Knowledge Discovery (GeoAI ’23), November 13, 2023, Hamburg, Germanyprice: 15.00doi: 10.1145/3615886.3627742isbn: 979-8-4007-0348-5/23/11ccs: Networks Network reliabilityccs: Information systems Spatial-temporal systems

1. Introduction

According to the State of Food Security and Nutrition in the World 2023 report jointly prepared by FAO, IFAD, UNICEF, WFP, and WHO (FAO and WHO, 2023), the world is moving backwards in its efforts to end hunger, food insecurity, and malnutrition in all its forms. The number of people experiencing acute food insecurity and requiring urgent food, nutrition, and livelihood assistance increased for the fourth consecutive year in 2022. The need to transform agrifood systems for increased resilience is urgent, as it can help provide nutritious foods at lower costs and ensure affordable, healthy diets for everyone in a sustainable and inclusive manner. Within the agrifood systems, food supply networks are pivotal in upholding global food security and facilitating the transit, dissemination, and sale of food. It’s imperative that these networks demonstrate resilience and sturdiness (Suweis et al., 2015; Miller, 2021; Karakoc and Konar, 2021).

However, the complexity inherent in them, arising from diverse food needs, shipment timeframes and costs, promotional strategies, cultural and environmental considerations, among others, complicates the assessment of their durability and adaptability (Chaturvedi et al., 2014; Yadav et al., 2022). Given the intricate nature of food supply networks, the concept of resilience is often interpreted in diverse ways by different individuals and groups (Fair et al., 2017; Puma, 2019; Tu et al., 2019; Karakoc and Konar, 2021). The term ”resilience” in this study predominantly pertains to the capacity of the food flow networks to sustain essential food supplies across geographical regions despite potential disruptions. In general terms, a node within the food flow network is deemed less resilient if its dependence is heavily skewed towards a sole supplier, a far-off supplier, a singular consumer, or one specific commodity (Rao et al., 2022; Inman and Blumenfeld, 2014).

Currently, resilience measurement methods such as the network topological characteristics and motifs (Dey et al., 2019; Gao et al., 2016; Karakoc and Konar, 2021), geospatial knowledge graph (GeoKG) and entropy-based metrics (Rao et al., 2022) have been employed to clarify the semantics of spatial networks (Barthélemy, 2011) (e.g., a multicommodity flow network between regions) and to evaluate its resilience. However, these mostly centralized methods encounter some limitations:

  • Robustness: the existing model is vulnerable to missing data or poorly curated data.

  • Requirement for centralized data access: the conventional entropy-based method necessitates access to the entire dataset, which is problematic given the recent emphasis on privacy concerns.

As a decentralized approach, Federated Learning (FL) has been adopted in various agricultural applications and supply chain systems. Existing efforts include utilizing federated learning to facilitate agricultural data sharing (Durrant et al., 2022), combat food fraud (Gavai et al., 2023), and predict supply chain risks (Zheng et al., 2023), yielding promising results. Hence, to address the issues exhibited in centralized methods, we propose a Federated Learning System for Edge-enhanced Graph Neural Network (FLEE-GNN). This innovative approach aims to improve the robustness and decentralize the measurements related to the geospatial resilience of multicommodity food flows, which refers to the capacity of a geographical region to withstand and recover from a variety of environmental, social, and economic challenges and disruptions on food supplies. By fusing the generalization capabilities of Graph Neural Networks (GNNs) (Scarselli et al., 2008; Gilmer et al., 2017; Zhang et al., 2021) with the decentralized features of federated learning (Chen and Chao, 2020), the FLEE-GNN offers a more resilient, flexible, and privacy-aware solution than centralized approaches for analyzing multicommodity flow networks and beyond.

This paper makes the following key contributions:

  • We designed an edge-enhanced graph neural network and a federated learning framework to measure network resilience across geographical regions in a decentralized manner.

  • We introduced an adjustable data generator to overcome data scarcity in measuring food supply resilience with deep learning methods.

  • We conducted a comprehensive analysis of the performance of both centralized and decentralized machine learning methods, comparing them to an entropy-based approach using the U.S. food flow data.

The rest of the paper is structured as follows:

  • Section 2 presents the elements of FLEE-GNN, encompassing GNN architecture, the federated learning framework, and a data generator that aids in addressing data shortages in the realm of multicommodity food flows.

  • Section 3 outlines the datasets and models we assessed in the experiments.

  • Section 4 delves into the results of prediction error, rank evaluation, and robustness across various models and datasets.

  • Section 5 outlines the directions of our future research, spotlighting possibilities like unsupervised learning and enhanced GeoKG embedding.

  • Section 6 provides a conclusion drawn from this study.

2. Methods

2.1. System Design

In this section, we introduce the components of FLEE-GNN, which include a message-passing-based GNN and a specially designed federated learning system to handle cases where there is no freight transportation or information exchange between geographical regions (“hypothesized disruptions”) in the U.S.

2.1.1. Edge-enhanced Graph Neural Network

Refer to caption
Figure 1. The architecture for the edge-enhanced graph neural network

As shown in Figure 1, We have adapted a message-passing model to develop the edge-enhanced GNN. This model accepts a graph, GG, as its input. Here, G=(V,E)G=(V,E) where VV represents the set of all nodes in a graph. For the purpose of this work, we consider each U.S. state as a node in the spatial network. EE is the set of all commodity food transported from a source state, ss, to a destination state, dd. Each transport of a commodity type between a pair of origin and destination is characterized by its monetary value vv, weight tonnage tt, and average transportation distance in mile aa. It denotes as E={(s,d,vs,ts,as)i|in}E=\{(s,d,v_{s},t_{s},a_{s})_{i}|i\in n\}, where nn is the number of edges |E||E|.

In this study, we use the Standard Classification of Transported Goods (SCTG) Codes from the U.S. Commodity Flow Survey (CFS) to categorize each food commodity type:

  • 01: Animals and fish (live)

  • 02: Cereal grains (includes seed)

  • 03: Agricultural products (excluding animal feed, cereal grains, and forage products)

  • 04: Animal feed, eggs, honey, and other products of animal origin

  • 05: Meat, poultry, fish, seafood, and their preparations

  • 06: Milled grain products, preparations, and bakery products

  • 07: Other prepared foodstuffs, and fats and oils (CFS10)

  • 08: Alcoholic beverages and denatured alcohol (CFS20)

For each node, we relay its neighboring node features, which include the location information (i.e., latitude and longitude), as well as the edge features, which encompasses the value, tonnage, and average transportation miles of all commodity food types transported from a source state to a destination state. The features of the destination state itself are captured by an edge linking from itself. If there is no transportation within the state, this results in zero edge feature values. Conversely, if transportation has occurred within the state, it results in non-zero edge feature values.

For illustration, consider the value of agricultural products (SCTG 03) from Alabama (AL) to Georgia (GA), which stands at 145 dollars per ton. The value for other prepared foodstuffs (SCTG 07) from AL to GA is 1497 dollars per ton. In 2012, 197 tons of agricultural products were transported from AL to GA, with an average transportation distance of 249 miles. Conversely, 613 tons of other prepared foodstuffs were transported from AL to GA, averaging a distance of 152 miles. These are the only two food types transported from AL to GA. Assuming AL is the ii’th neighbor for GA, then the edge features viv_{i} during computation are represented as a vector with 24 dimensions, with specific dimensions as {V3:145,T3:197,A3:249,V7:1497,T7:613,A7:152}\{V_{3}:145,T_{3}:197,A_{3}:249,V_{7}:1497,T_{7}:613,A_{7}:152\}. The remaining dimensions are zero:

V3\textbf{V}_{3} T3\textbf{T}_{3} A3\textbf{A}_{3} V7\textbf{V}_{7} T7\textbf{T}_{7} A7\textbf{A}_{7}
0 145 197 249 0 1497 613 152 0

Every message, denoted as [ni,vi][n_{i},v_{i}] (a combination of edge and node features), passes through a fully connected layer in the neural network (see Figure 1). It then merges into a latent vector, uiu_{i}, encapsulating the information within the transitions from the source to destination states. This condensation into a lower dimension bolsters computational efficiency. After procuring all latent vectors for the destination state’s neighbors, every message undergoes aggregation via an summation function. This is then transmuted into another latent vector, UU, signifying all data relevant to import resilience for the destination state. This vector is then projected into a scalar value, which undergoes another fully connected layer, followed by a sigmoid activation function, ensuring an output range between 0 and 1, which represents the predicted resilience score.

This design ensures that both node and edge information is leveraged at each processing stage, aiming to provide a richer representation for various graph-based tasks.

2.1.2. Federated Learning

Refer to caption
Figure 2. The federated learning architecture for the edge-enhanced graph neural networks

This study introduces a cutting-edge federated learning framework to assess the geospatial resilience of multicommodity food flow as shown in Figure 2.

The federated learning approach embodies the concept of decentralized training (Konecný et al., 2016). Here’s a detailed overview of the process:

  1. (1)

    Initialization: The central server initializes a global model. This model is then dispatched to all distributed region nodes (or ’silos’) participating in the federated learning network. In this research, each silo refers to each geographical region in the U.S. and there are four regions (i.e., West, South, Midwest, and Northeast) in our experiments.

  2. (2)

    Local Training: Each node computes its model locally using its own dataset. Importantly, no raw data is exchanged or shared between nodes or with the central server. Training is performed over a pre-defined number of epochs (e.g., 100 epochs) to adjust and refine the model using local data.

  3. (3)

    Local Model Update: At the end of local training, each node computes a summary of its model updates. These are essentially changes or differences compared to the global model.

  4. (4)

    Aggregation at Central Server: Nodes send their model updates to the central server periodically (e.g., after every 10 epochs). The central server then aggregates these updates using methods like weighted averaging.

  5. (5)

    Global Model Refinement: After aggregation, the central server refines the global model parameters, which benefit from the knowledge gathered from all distributed nodes.

  6. (6)

    Dispatching Updated Model: The improved global model is then sent back to all regional nodes. This cyclic process of local training, aggregation, and global update continues until the model converges or another pre-defined criterion to stop is met.

2.2. Data Generation

The U.S. Bureau of Transportation (BTS) and the U.S. Census Bureau publish the detailed commodity food flow survey (CFS) 111https://www.bts.gov/product/commodity-flow-survey once every five years, which presents a data scarcity challenge for the idea of measuring food supply resilience using deep learning methods. To overcome this challenge, we introduce an adjustable data generator (Algorithm 1), in conjunction with the previous entropy-based method (Rao et al., 2022), to create nearly authentic food flow data. This data generator requires a real commodity flow dataset as input, accompanied by a noise ratio. The noise ratio is defined as the proportion of data we wish to modify based on the actual dataset. We have also delineated three methods to adjust the data:

Algorithm 1 GRAPH_GENERATOR
1:GG is the true commodity food flow graph with nn edges, where G=(V,E)G=(V,E) and n=|E|n=|E|, rr is the noise ratio
2:GG is the generated commodity food flow graph
3:nr×n3n^{\prime}\leftarrow\left\lfloor\frac{r\times n}{3}\right\rfloor \triangleright each operation [REMOVE, CHANGE, ADD] contribute a third of the total noise
4:for i=1i=1 to nn^{\prime} do
5:     GREMOVE(G)G\leftarrow\text{REMOVE}(G)
6:     GCHANGE(G)G\leftarrow\text{CHANGE}(G)
7:     GADD(G)G\leftarrow\text{ADD}(G)
8:end for
  1. (1)

    ADD (Algorithm 2): Introduce a transportation event for food type cc^{\prime} from source ss^{\prime} to destination dd^{\prime} with a random value vv^{\prime}, tonnage tt^{\prime}, and average transportation miles aa^{\prime} sampled from its corresponding distribution in the original dataset. If the sampled tuple (s,d,c)(s^{\prime},d^{\prime},c^{\prime}) is already present, we sample another one.

  2. (2)

    REMOVE (Algorithm 3): Randomly select an edge and expunge it from the current food flow.

  3. (3)

    CHANGE (Algorithm 4): Randomly select an edge and resample value vv^{\prime}, tonnage tt^{\prime}, and average transportation miles aa^{\prime} from its corresponding distribution in the original dataset.

Algorithm 2 ADD
1:GG is a directed graph with nn edges, where n=|E|n=|E|, G=(V,E)G=(V,E), and E={(si,di,ci,vi,ti,ai)|i[1,n]}E=\{(s_{i},d_{i},c_{i},v_{i},t_{i},a_{i})|i\in[1,n]\}. CC is a set of commodity food
2:GG^{\prime} is a directed graph with (n+1)(n+1) edges
3:MAXVmaxi=1nviMAX_{V}\leftarrow\max_{i=1}^{n}v_{i}, MINVmini=1nviMIN_{V}\leftarrow\min_{i=1}^{n}v_{i}
4:MAXTmaxi=1ntiMAX_{T}\leftarrow\max_{i=1}^{n}t_{i}, MINTmini=1ntiMIN_{T}\leftarrow\min_{i=1}^{n}t_{i}
5:MAXAmaxi=1naiMAX_{A}\leftarrow\max_{i=1}^{n}a_{i}, MINAmini=1naiMIN_{A}\leftarrow\min_{i=1}^{n}a_{i}
6:repeat
7:     ssample a source node from Vs^{\prime}\leftarrow\text{sample a source node from }V
8:     dsample a destination node from Vd^{\prime}\leftarrow\text{sample a destination node from }V
9:     csample a commodity food from Cc^{\prime}\leftarrow\text{sample a commodity food from }C
10:until there is no edge in GG that goes from ss^{\prime} to dd^{\prime} with c=cc=c^{\prime}
11:vsample a real number from [MINV,MAXV]v^{\prime}\leftarrow\text{sample a real number from }[MIN_{V},MAX_{V}]
12:tsample a real number from [MINT,MAXT]t^{\prime}\leftarrow\text{sample a real number from }[MIN_{T},MAX_{T}]
13:asample a real number from [MINA,MAXA]a^{\prime}\leftarrow\text{sample a real number from }[MIN_{A},MAX_{A}]
14:EE{(s,d,c,v,t,a)}E^{\prime}\leftarrow E\cup\{(s^{\prime},d^{\prime},c^{\prime},v^{\prime},t^{\prime},a^{\prime})\}
15:G(V,E)G^{\prime}\leftarrow(V,E^{\prime})
Algorithm 3 REMOVE
1:GG is a directed graph with nn edges, where n=|E|n=|E|, and G=(V,E)G=(V,E)
2:GG^{\prime} is a directed graph with (n1)(n-1) edges
3:isample an integer from 1 to ni\leftarrow\text{sample an integer from }1\text{ to }n
4:EE{(si,di,ci,vi,ti,ai)}E^{\prime}\leftarrow E\setminus\{(s_{i},d_{i},c_{i},v_{i},t_{i},a_{i})\}
5:G(V,E)G^{\prime}\leftarrow(V,E^{\prime})
Algorithm 4 CHANGE
1:GG is a directed graph with nn edges, where n=|E|n=|E|, and G=(V,E)G=(V,E)
2:GG^{\prime} is a directed graph with nn edges
3:MAXVmaxi=1nviMAX_{V}\leftarrow\max_{i=1}^{n}v_{i}, MINVmini=1nviMIN_{V}\leftarrow\min_{i=1}^{n}v_{i}
4:MAXTmaxi=1ntiMAX_{T}\leftarrow\max_{i=1}^{n}t_{i}, MINTmini=1ntiMIN_{T}\leftarrow\min_{i=1}^{n}t_{i}
5:MAXAmaxi=1naiMAX_{A}\leftarrow\max_{i=1}^{n}a_{i}, MINAmini=1naiMIN_{A}\leftarrow\min_{i=1}^{n}a_{i}
6:Isample an integer from 1 to nI\leftarrow\text{sample an integer from }1\text{ to }n
7:(s,d,c,v,t,a)(si,di,ci,vi,ti,ai)(s^{\prime},d^{\prime},c^{\prime},v^{\prime},t^{\prime},a^{\prime})\leftarrow(s_{i},d_{i},c_{i},v_{i},t_{i},a_{i})
8:vsample a real number from [MINV,MAXV]v^{\prime}\leftarrow\text{sample a real number from }[MIN_{V},MAX_{V}]
9:tsample a real number from [MINT,MAXT]t^{\prime}\leftarrow\text{sample a real number from }[MIN_{T},MAX_{T}]
10:asample a real number from [MINA,MAXA]a^{\prime}\leftarrow\text{sample a real number from }[MIN_{A},MAX_{A}]
11:EE{(si,di,ci,vi,ti,ai)}E^{\prime}\leftarrow E\setminus\{(s_{i},d_{i},c_{i},v_{i},t_{i},a_{i})\}
12:EE{(s,d,c,v,t,a)}E^{\prime}\leftarrow E^{\prime}\cup\{(s^{\prime},d^{\prime},c^{\prime},v^{\prime},t^{\prime},a^{\prime})\}
13:G(V,E)G^{\prime}\leftarrow(V,E^{\prime})

Given a noise ratio, we evenly apply the three aforementioned operations to generate a new food flow dataset for data augmentation purpose. For instance, if the original datasets contain 100 edges and we have a noise ratio of 0.3, we will add 10 edges, remove 10 edges, and change 10 edges. This results in a dataset with 100 edges, albeit different from the original set.

3. Experiments

3.1. Datasets

In our experiments, we utilize the above-mentioned U.S. CFS data. The CFS provides comprehensive information of domestic freight shipments including commodity type, value, weight, distance shipped, origin and destination, etc. from national-level to state-level. As an example, we focus on the agricultural multi-commodity flows in 2012 and 2017 at differential geographical regions by extracting the data with SCTG code from 01 to 08. We further process the raw data into two primary datasets:

  1. (1)

    Silo Dataset: Assumes that each silo (i.e., geographic region) can only access the links and transitions within its domain.

  2. (2)

    Centralized Dataset: Encompasses all links and transitions, both within and across silos, thus serving as a comprehensive graph of transitions.

The basic graph statistics of the whole food supply network and different sub-networks by regions are illustrated in Table 1 and Table 2. The sub-networks of food flows in different geographical regions have distinctive graph statistics (e.g., degree, centrality, connectivity), which reflects spatial heterogeneity under different spatial configurations and poses challenges on geospatial resilience modeling and evaluation.

Area Whole Whole w/o cross-silo edges West Silo Midwest Silo South Silo Northeast Silo
average degree 63.7255 27.7647 16.0000 37.1667 32.7059 22.8889
average weighted degrees (VAL) 25797.3333 20902.3529 16193.8462 28875.5000 18695.4118 21241.3333
average degree centrality 1.2745 0.5553 1.3333 3.3788 2.0441 2.8611
average closeness centrality 0.6153 0.1681 0.5533 0.8211 0.6620 0.7492
average betweenness centrality 0.0134 0.0025 0.0594 0.0250 0.0355 0.0536
average node connectivity 13.6184 1.2894 2.5897 6.9697 6.2169 3.7917
edge connectivity 1 0 0 3 1 1
Table 1. Graph Statistics for 2012 CFS data
Area Whole Whole w/o cross-silo edges West Silo Midwest Silo South Silo Northeast Silo
average degree 49.7647 21.0980 13.3846 28.8333 25.7647 13.1111
average weighted degrees (VAL) 16365.5294 12778.3137 9137.0769 17025.3333 13682.1176 10668.0000
average degree centrality 0.9953 0.4220 1.1154 2.6212 1.6103 1.6389
average closeness centrality 0.5515 0.1477 0.4289 0.7759 0.6264 0.5110
average betweenness centrality 0.0152 0.0027 0.0606 0.0303 0.0336 0.1012
average node connectivity 11.2208 1.1031 1.9295 6.5985 5.5662 1.7639
edge connectivity 0 0 0 4 0 0
Table 2. Graph Statistics for 2017 CFS data

3.2. Models

The baseline model, derived from our previous work (Rao et al., 2022), employs an entropy-based method using the centralized food flow dataset. Given its access to all links between geographical regions, we designate this model as our ground truth.

A node-level resilience RiR_{i} that comprehensively measures the single-supplier/customer dependence, single-commodity-type dependence, transport distance, and geographic adjacency of the node ii is computed as:

(1) Ri=1DiAAggV(i,A)ViR_{i}=1-D_{i}\frac{\sum_{A\in Agg}V^{\prime}_{(i,A)}}{V^{\prime}_{i}}

where DiD_{i} is the overall single-commodity-type dependence of node ii (measured by the Shannon information entropy); AggAgg is a set of aggregated commodity types; V(i,A)V^{\prime}_{(i,A)} is the value of aggregated commodity AA reflecting single-commodity-type dependence; Vi=AAggcAjV(ij,c)V^{\prime}_{i}=\sum_{A\in Agg}\sum_{c\in A}\sum_{j}V^{\prime}_{(i\rightarrow j,c)} denotes the total commodity value of node ii combined with average transport miles and geography adjacency information. The higher the RiR_{i}, the less dependent node ii is on single or geographically distant/non-adjacent supplier/customer or single commodity type, thus the higher the resilience.

We set the developed FLEE-GNN model against the entropy-based method and a centralized GNN (CT-GNN). We analyzed how each model performed under various noise levels by perturbing the 2012 food flow data at noise ratios of 0.1, 0.3, and 0.5, resulting in three datasets of 500 graphs each. Both CT-GNN and FLEE-GNN models underwent training for 100 epochs. Evaluations were then carried out on real food flow data from 2012 and 2017, respectively.

4. Results

4.1. Relative Resilience Difference

We define the relative resilience difference as the difference between each model output and the entropy-based computation result (ground truth). As seen in Figure 3, it is evident that the resilience values generated by the entropy-based method within sub-networks only take into account data within a silo, causing the model to underestimate its global resilience in the food supply network. This can be attributed to the omission of transportation between different geographical silos. By neglecting these transportation movements, we effectively remove edges from the original spatial network. As a result, nodes, especially those that lose transportation connections, become more vulnerable to potential disruptive changes imposed on the food systems.

In centralized training (Figure 4), the full graph is used both for training and evaluation. Consequently, there isn’t any significant overestimation or underestimation observable from the graph. In the case of FLEE-GNN (Figure 5), the silo sub-networks and ground truth label from the full graph are used for training. This simulates the scenario where we only have access to silo data in the real world due to privacy constraints. Yet, from that silo data, we still aim to recover the original true resilience of the full graph. During the training, our model tends to overestimate its resilience. However, this overestimation is offset by the lack of information in the actual global graph during evaluation. Through such a strategy, we also achieve balanced predictions when using FLEE-GNN without significant overestimation or underestimation.

Refer to caption
Refer to caption
Figure 3. The difference in resilience between the entropy-based method using silo data and the entropy-based method using global data for 2012 (L) and 2017 (R)
Refer to caption
Refer to caption
Figure 4. The prediction error on 2012 (L) / 2017 (R) data by using centralized model trained on data generated by adjusting real 2012 data with noise = 0.30.3
Refer to caption
Refer to caption
Figure 5. The prediction error on 2012 (L) / 2017 (R) data by using federated learning model trained on silo data generated by adjusting real 2012 data noise = 0.30.3

4.2. Absolute Prediction Error

While the relative resilience difference captures the proportionality of errors, highlighting under- or over-estimations, we also examine the absolute prediction error, which quantifies the absolute deviations. For each model evaluated on data from both 2012 (Table 3) and 2017 (Table 4), we calculated the mean, standard deviation, minimum, 25th percentile, median (50th percentile), 75th percentile, and maximum values. The results indicate that both centralized training CT-GNN and federated training FLEE-GNN (with smaller prediction errors), trained on datasets generated with varying noise ratios, outperformed the entropy-based method across all metrics.

Another trend we observed from the results is that the performance tends to degrade slightly as we increase the noise ratio. This can be attributed to the fact that by injecting more noise, the generated food flow data becomes less representative of real data. Modern machine learning techniques often encounter this out-of-distribution issue (Yang et al., 2021). Furthermore, when comparing results between 2012 and 2017, the model maintains its performance even when evaluated on the 2017 food flow data. This is notable since the training data comes solely from the 2012 food flow dataset. Despite this, the model still performs admirably in predicting resilience for the food flow network in a different time period, demonstrating its versatility and effectiveness.

Noise 0 0.1 0.3 0.5
Dataset SILO WHOLE SILO WHOLE SILO WHOLE SILO
Method Entropy CT FL CT FL CT FL CT FL CT FL CT FL
mean 0.1216 0.0815 0.0628 0.2250 0.0640 0.0515 0.0685 0.1359 0.0635 0.0569 0.0746 0.1456 0.0620
std 0.1475 0.0504 0.0677 0.0745 0.0548 0.0575 0.0689 0.0688 0.0556 0.0615 0.0709 0.0746 0.0580
min 0.0028 0.0004 0.0004 0.0311 0.0019 0.0054 0.0036 0.0120 0.0006 0.0024 0.0117 0.0049 0.0036
25% 0.0246 0.0366 0.0360 0.1733 0.0249 0.0205 0.0378 0.0874 0.0209 0.0236 0.0446 0.1002 0.0204
50% 0.0797 0.0838 0.0458 0.2346 0.0545 0.0347 0.0533 0.1397 0.0477 0.0378 0.0564 0.1432 0.0469
75% 0.1574 0.1100 0.0590 0.2627 0.0785 0.0596 0.0663 0.1916 0.0921 0.0682 0.0726 0.1993 0.0935
max 0.6881 0.2008 0.4482 0.3823 0.2361 0.3694 0.4612 0.3074 0.2626 0.3775 0.4769 0.3063 0.3093
Table 3. Absolute Prediction Error for 2012 Data (CT: Centralized Training; FL: Federating Learning)
Noise 0 0.1 0.3 0.5
Dataset SILO WHOLE SILO WHOLE SILO WHOLE SILO
Method Entropy CT FL CT FL CT FL CT FL CT FL CT FL
mean 0.1193 0.0570 0.0737 0.1976 0.0835 0.0599 0.0751 0.1988 0.0808 0.0612 0.0749 0.2229 0.0852
std 0.1480 0.0500 0.0484 0.1097 0.0925 0.0496 0.0505 0.1120 0.0836 0.0556 0.0515 0.1108 0.0787
min 0.0022 0.0008 0.0108 0.0250 0.0018 0.0042 0.0308 0.0011 0.0022 0.0042 0.0297 0.0088 0.0067
25% 0.0305 0.0251 0.0453 0.1305 0.0227 0.0290 0.0456 0.1163 0.0228 0.0266 0.0457 0.1327 0.0284
50% 0.0652 0.0334 0.0628 0.1744 0.0419 0.0396 0.0612 0.1747 0.0446 0.0394 0.0612 0.2104 0.0538
75% 0.1292 0.0621 0.0919 0.2601 0.0975 0.0665 0.0921 0.2760 0.1139 0.0642 0.0885 0.3200 0.1093
max 0.7940 0.1848 0.3058 0.5014 0.3832 0.2094 0.3309 0.4845 0.3424 0.2327 0.3390 0.4601 0.3352
Table 4. Absolute Prediction Error for 2017 Data (CT: Centralized Training; FL: Federating Learning)

4.3. Rank Evaluation

Rather than focusing solely on the specific resilience provided by the model, we are also interested in how the rankings of node resilience in food supply networks are maintained. Specifically, we aim to determine whether a state with high or low resilience will consistently show high or low resilience values in comparison to other states in the FLEE-GNN model’s output. To assess the consistency between predicted and actual state rankings based on their resilience values, we sorted the states accordingly and computed the following concordance metrics:

  1. (1)

    The coincidence rate between the top n% ranked nodes of actual data and the top n% ranked nodes of predicted data.

  2. (2)

    Pearson’s correlation coefficient.

  3. (3)

    Spearman’s rank correlation coefficient.

From the results (Table 5 and Table 6), it is evident that the machine learning-based method also boasts a superior coincidence rate at different noise levels of data, signifying that these models offer higher consistency and alignment with the actual resilience values in the original dataset. Additionally, a higher Spearman’s rho indicates that the rankings from the predictions of CT-GNN and FLEE-GNN closely match the actual rankings. Likewise, a larger Pearson’s R indicates a robust linear relationship between predicted and actual resilience values. This means that as the actual resilience values rise (or fall), the model’s predictions will also increase (or decrease) in a similar fashion.

Noise 0 0.1 0.3 0.5
Dataset SILO WHOLE SILO WHOLE SILO WHOLE SILO
Method Entropy CT FL CT FL CT FL CT FL CT FL CT FL
Coincidence / Recall (Top 10%) 0.0000 0.5000 0.3333 0.1667 0.5000 0.5000 0.5000 0.3333 0.5000 0.5000 0.5000 0.3333 0.3333
Coincidence / Recall (Top 30%) 0.4375 0.5625 0.7500 0.5625 0.5000 0.7500 0.7500 0.5625 0.4375 0.7500 0.7500 0.5625 0.4375
Coincidence / Recall (Top 50%) 0.6154 0.8462 0.8462 0.7692 0.8077 0.8462 0.8462 0.7308 0.7308 0.8077 0.7692 0.7308 0.6923
Pearson R 0.4805 0.8098 0.7066 0.5251 0.7093 0.7272 0.6979 0.5680 0.6619 0.7068 0.6673 0.5620 0.6380
Spearman rho 0.3135 0.8110 0.8642 0.5738 0.6624 0.8251 0.8338 0.5932 0.6118 0.8195 0.8121 0.5600 0.5542
Table 5. Rank Evaluation Metrics for 2012 CFS Data (CT: Centralized Training; FL: Federating Learning)
Noise 0 0.1 0.3 0.5
Dataset SILO WHOLE SILO WHOLE SILO WHOLE SILO
Method Entropy CT FL CT FL CT FL CT FL CT FL CT FL
Coincidence / Recall (Top 10%) 0.2000 0.4000 0.4000 0.1667 0.5000 0.6000 0.4000 0.3333 0.5000 0.4000 0.4000 0.3333 0.3333
Coincidence / Recall (Top 30%) 0.5333 0.5333 0.6667 0.5625 0.5000 0.5333 0.6667 0.5625 0.4375 0.6667 0.6667 0.5625 0.4375
Coincidence / Recall (Top 50%) 0.6000 0.8800 0.8000 0.7692 0.8077 0.8800 0.8400 0.7308 0.7308 0.8400 0.8400 0.7308 0.6923
Pearson R 0.7704 0.7710 0.7670 0.5251 0.7093 0.7886 0.7641 0.5680 0.6619 0.7811 0.7581 0.5620 0.6380
Spearman rho 0.4736 0.8008 0.8210 0.5738 0.6624 0.8215 0.8317 0.5932 0.6118 0.8372 0.8206 0.5600 0.5542
Table 6. Rank Evaluation Metrics for 2017 CFS Data (CT: Centralized Training; FL: Federating Learning)

4.4. Robustness with Missing Data

Noise 0.3
Columns VAT VT VA TA V T A NONE
Method CT FL CT FL CT FL CT FL CT FL CT FL CT FL CT FL
mean 0.0599 0.0808 0.0600 0.0840 0.0574 0.0786 0.0602 0.0794 0.0601 0.0834 0.0792 0.0806 0.0583 0.0872 0.0793 0.0978
std 0.0496 0.0836 0.0549 0.0910 0.0540 0.0823 0.0648 0.0829 0.0595 0.0935 0.0864 0.0873 0.0672 0.1011 0.0864 0.0991
min 0.0042 0.0023 0.0053 0.0001 0.0004 0.0021 0.0013 0.0042 0.0011 0.0004 0.0006 0.0019 0.0011 0.0004 0.0001 0.0001
25% 0.0290 0.0228 0.0257 0.0188 0.0227 0.0218 0.0162 0.0233 0.0226 0.0169 0.0156 0.0160 0.0142 0.0194 0.0194 0.0236
50% 0.0396 0.0446 0.0402 0.0485 0.0371 0.0418 0.0289 0.0394 0.0403 0.0398 0.0321 0.0439 0.0259 0.0451 0.0421 0.0557
75% 0.0665 0.1139 0.0625 0.1130 0.0743 0.1090 0.0913 0.1160 0.0611 0.1239 0.1287 0.1055 0.0760 0.1118 0.1314 0.1335
max 0.2094 0.3424 0.2344 0.3655 0.2202 0.3332 0.2421 0.3289 0.2452 0.3677 0.2938 0.3413 0.2571 0.3685 0.3131 0.3724
Table 7. Absolute Prediction Error of the ablation study for 2017 CFS Data (CT: Centralized Training; FL: Federating Learning)

Not only are we interested in the performance of FLEE-GNN at different time periods (2012/2017) and in different spaces (global transportation, silo transportation), but we are also keen on its performance across different missing data scenarios. In these experiments, we compare the absolute prediction error for CT-GNN trained with data generated from the 2012 food flow network and evaluated on the 2017 global network data, with that of FLEE-GNN trained with the same data and evaluated on 2017 sub-network silo data. However, during training, we consider different data combinations to simulate scenarios where certain edge features missing from a dataset.

  • V: Represents that the value of commodity food flow exists in edge features.

  • A: Represents that the average transportation mile of commodity food flow exists in edge features.

  • T: Represents that the tonnage mile of commodity food flow exists in edge features.

For example, ”AT” means the graph structure of the food supply network remains the same, but instead of having access to all the shipment value, average transportation mile, and tonnage, we lose the shipment value information.

As seen in Table 7, interestingly, both CT-GNN and FLEE-GNN maintain stability even with some missing edge features. We do notice that some performance decreases when the data constrain the type of information the model can access. The maximum difference in terms of mean absolute prediction error for CT-GNN is 0.07930.0599=0.01950.0793-0.0599=0.0195 and for FLEE-GNN is 0.09780.0808=0.0170.0978-0.0808=0.017.

To some extent, this research proves that the most relevant information supporting current resilience measurement is embedded in the graph structure. In the extreme NONE case (in Table 7), all we retain is the basic graph structure (topological relations between nodes) without any edge features. Still, both CT-GNN and FLEE-GNN provide a satisfactory estimation of the resilience value. However, the original entropy-based method, which relies heavily on the edge attribution, fails to generate a meaningful resilience score in this case.

4.5. Summary

The silo-data-based resilience metrics using the entropy method revealed its constraints, particularly in its limited access to comprehensive linkages among geographical regions. While both the CT-GNN and FLEE-GNN demonstrated performance robustness against noise, and FLEE-GNN outperformed consistently, especially in higher noise scenarios. The Spearman rank correlation of the FLEE-GNN indicates a strong agreement with the ground-truth data, emphasizing the model’s utility in accurately ranking the resilience of U.S. states in food supply. Given that the eventual aim is to bolster food security, a model like FLEE-GNN offers valuable insights into spatial network vulnerabilities and opportunities to improve supply network resilience. This could guide food systems policy-making, resource allocation and negotiation in critical scenarios in advance of cascading system failures.

In summary, FLEE-GNN not only holds promise as an innovative tool for assessing food network resilience but also sets the stage for leveraging federated learning in a myriad of complex, geospatial AI applications. Future directions might delve deeper into optimizing FLEE-GNN and expanding its applicability to other domains.

5. Future Work

5.1. Unsupervised Learning

Our initial work utilized labels generated by the entropy-based method as the ground truth. However, this limits us to the upper bound provided by the entropy-based method. Unsupervised learning has demonstrated greate success in the field of image classification (Schmarje et al., 2021), feature selection (Khanum et al., 2015), association rule detection (Reshef et al., 2011), etc. Developing unsupervised learning techniques may allow us to exceed this bound and improve our system’s predictive power.

5.2. GeoKG Embedding and Semantic Extraction

We currently input different attributes into the FLEE-GNN as scalar values. The model thus lacks an understanding of the differences between values, tons, average miles, temperature control, and other attributes in food supply transport. Moreover, it cannot extract comprehensive relationships between these attributes. To improve our model’s predictive ability and interpretability, we need to devise a better explainable AI method for embedding GeoKG information directly into our system.

5.3. Other Domains

Moreover, the emphasis on federated learning within FLEE-GNN opens doors for its application in scenarios where data privacy and decentralization are paramount. This consideration will undoubtedly grow more significant in an era marked by heightened data protection regulations and the need for more inclusive, globally-spread research collaborations.

In conclusion, while the immediate implications of our work pertain to food supply networks, the FLEE-GNN framework offers a versatile blueprint with applications across a multitude of sectors. We eagerly anticipate the breakthroughs and innovations that can arise as we venture into other types of spatial networks and application domains.

6. Conclusion

In the face of increasing global food insecurity, understanding and fortifying the resilience of food supply networks is of paramount importance. Our study introduced the FLEE-GNN, a pioneering Federated Learning System for Edge-enhanced Graph Neural Network, adeptly tailored to navigate the complexities of spatial networks. Our comprehensive experimental design and evaluations elucidated the method’s superiority over traditional entropy-based systems, especially in terms of robustness, scalability, and data privacy considerations. The novel integration of graph neural networks with the decentralized architecture of federated learning showcased how data privacy and model robustness can be harmoniously balanced when dealing with spatially heterogeneous sub-network datasets. The FLEE-GNN’s performance, particularly in scenarios with noise, demonstrates its potential as a valuable tool for stakeholders across the global food supply network. Not only does it offer insights into the current state of the food system, but its predictive prowess can be instrumental in proactive decision-making and vulnerability mitigation.

Furthermore, the limitations of entropy methods, particularly with constrained data access, were evident. This highlights the pressing need for innovative, decentralized models like FLEE-GNN in our evolving digital age, where data is often fragmented and siloed. As we contemplate a future with increasing challenges to food security, AI tools like the FLEE-GNN will be indispensable. They promise a more informed, agile, and resilient approach to safeguarding our most crucial supply networks. While this study illuminates a promising path forward, it also beckons further exploration and refinement in the realm of federated learning and neural network integration for complex network analysis especially for public sector data. We are optimistic that this fusion of geospatial AI technologies can serve as a linchpin in the global endeavor to ensure food security for all.

Data and code availability statement

The data and codes that support the findings of this study are available at the following link on GitHub: https://github.com/GeoDS/FLEE-GNN/.

Acknowledgements.
We acknowledge the funding support from the National Science Foundation funded AI institute [Grant No. 2112606] for Intelligent Cyberinfrastructure with Computational Learning in the Environment (ICICLE). Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the funder(s).

References

  • (1)
  • Barthélemy (2011) Marc Barthélemy. 2011. Spatial networks. Physics reports 499, 1-3 (2011), 1–101.
  • Chaturvedi et al. (2014) Alok Chaturvedi, Brian Armstrong, and Rashmi Chaturvedi. 2014. Securing the food supply chain: understanding complex interdependence through agent-based simulation. Health and Technology 4, 2 (2014), 159–169.
  • Chen and Chao (2020) Hong-You Chen and Wei-Lun Chao. 2020. FedBE: Making Bayesian Model Ensemble Applicable to Federated Learning. In International Conference on Learning Representations.
  • Dey et al. (2019) Asim K Dey, Yulia R Gel, and H Vincent Poor. 2019. What network motifs tell us about resilience and reliability of complex networks. Proceedings of the National Academy of Sciences 116, 39 (2019), 19368–19373.
  • Durrant et al. (2022) Aiden Durrant, Milan Markovic, David Matthews, David May, Jessica Enright, and Georgios Leontidis. 2022. The role of cross-silo federated learning in facilitating data sharing in the agri-food sector. Computers and Electronics in Agriculture 193 (2022), 106648.
  • Fair et al. (2017) Kathyrn R Fair, Chris T Bauch, and Madhur Anand. 2017. Dynamics of the global wheat trade network and resilience to shocks. Scientific reports 7, 1 (2017), 1–14.
  • FAO and WHO (2023) UNICEF WFP FAO, IFAD and WHO. 2023. The State of Food Security and Nutrition in the World 2023. (2023).
  • Gao et al. (2016) Jianxi Gao, Baruch Barzel, and Albert-László Barabási. 2016. Universal resilience patterns in complex networks. Nature 530, 7590 (2016), 307–312.
  • Gavai et al. (2023) Anand Gavai, Yamine Bouzembrak, Wenjuan Mu, Frank Martin, Rajaram Kaliyaperumal, Johan van Soest, Ananya Choudhury, Jaap Heringa, Andre Dekker, and Hans JP Marvin. 2023. Applying federated learning to combat food fraud in food supply chains. npj Science of Food 7, 1 (2023), 46.
  • Gilmer et al. (2017) Justin Gilmer, Samuel S. Schoenholz, Patrick F. Riley, Oriol Vinyals, and George E. Dahl. 2017. Neural Message Passing for Quantum Chemistry. In Proceedings of the 34th International Conference on Machine Learning (Proceedings of Machine Learning Research), Doina Precup and Yee Whye Teh (Eds.), Vol. 70. PMLR, 1263–1272. https://proceedings.mlr.press/v70/gilmer17a.html
  • Inman and Blumenfeld (2014) Robert R Inman and Dennis E Blumenfeld. 2014. Product complexity and supply chain design. International Journal of Production Research 52, 7 (2014), 1956–1969.
  • Karakoc and Konar (2021) Deniz Berfin Karakoc and Megan Konar. 2021. A complex network framework for the efficiency and resilience trade-off in global food trade. Environmental Research Letters 16, 10 (2021), 105003.
  • Khanum et al. (2015) Memoona Khanum, Tahira Mahboob, Warda Imtiaz, Humaraia Abdul Ghafoor, and Rabeea Sehar. 2015. A survey on unsupervised machine learning algorithms for automation, classification and maintenance. International Journal of Computer Applications 119, 13 (2015).
  • Konecný et al. (2016) Jakub Konecný, H. B. McMahan, Daniel Ramage, and Peter Richtárik. 2016. Federated Optimization: Distributed Machine Learning for On-Device Intelligence. ArXiv abs/1610.02527 (2016). https://api.semanticscholar.org/CorpusID:2549272
  • Miller (2021) Michelle Miller. 2021. Identifying Critical Thresholds for Resilient Regional Food Flows: A Case Study From the US Upper Midwest. Frontiers in Sustainable Food Systems 5 (2021), 684159.
  • Puma (2019) Michael J Puma. 2019. Resilience of the global food system. Nature Sustainability 2, 4 (2019), 260–261.
  • Rao et al. (2022) Jinmeng Rao, Song Gao, Michelle Miller, and Alfonso Morales. 2022. Measuring network resilience via geospatial knowledge graph: a case study of the us multi-commodity flow network. In Proceedings of the 1st ACM SIGSPATIAL International Workshop on Geospatial Knowledge Graphs. 17–25.
  • Reshef et al. (2011) David N Reshef, Yakir A Reshef, Hilary K Finucane, Sharon R Grossman, Gilean McVean, Peter J Turnbaugh, Eric S Lander, Michael Mitzenmacher, and Pardis C Sabeti. 2011. Detecting novel associations in large data sets. science 334, 6062 (2011), 1518–1524.
  • Scarselli et al. (2008) Franco Scarselli, Marco Gori, Ah Chung Tsoi, Markus Hagenbuchner, and Gabriele Monfardini. 2008. The graph neural network model. IEEE transactions on neural networks 20, 1 (2008), 61–80.
  • Schmarje et al. (2021) Lars Schmarje, Monty Santarossa, Simon-Martin Schröder, and Reinhard Koch. 2021. A survey on semi-, self-and unsupervised learning for image classification. IEEE Access 9 (2021), 82146–82168.
  • Suweis et al. (2015) Samir Suweis, Joel A Carr, Amos Maritan, Andrea Rinaldo, and Paolo D’Odorico. 2015. Resilience and reactivity of global food security. Proceedings of the National Academy of Sciences 112, 22 (2015), 6902–6907.
  • Tu et al. (2019) Chengyi Tu, Samir Suweis, and Paolo D’Odorico. 2019. Impact of globalization on the resilience and sustainability of natural resources. Nature Sustainability 2, 4 (2019), 283–289.
  • Yadav et al. (2022) Vinay Surendra Yadav, AR Singh, Angappa Gunasekaran, Rakesh D Raut, and Balkrishna E Narkhede. 2022. A systematic literature review of the agro-food supply chain: Challenges, network design, and performance measurement perspectives. Sustainable Production and Consumption 29 (2022), 685–704.
  • Yang et al. (2021) Jingkang Yang, Kaiyang Zhou, Yixuan Li, and Ziwei Liu. 2021. Generalized out-of-distribution detection: A survey. arXiv preprint arXiv:2110.11334 (2021).
  • Zhang et al. (2021) Xitong Zhang, Yixuan He, Nathan Brugnone, Michael Perlmutter, and Matthew Hirn. 2021. Magnet: A neural network for directed graphs. Advances in neural information processing systems 34 (2021), 27003–27015.
  • Zheng et al. (2023) Ge Zheng, Lingxuan Kong, and Alexandra Brintrup. 2023. Federated machine learning for privacy preserving, collective supply chain risk prediction. International Journal of Production Research (2023), 1–18.