A generalised novel loss function for computational fluid dynamics
Abstract
Computational fluid dynamics (CFD) simulations are crucial in automotive, aerospace, maritime and medical applications, but are limited by the complexity, cost and computational requirements of directly calculating the flow, often taking days of compute time. Machine-learning architectures, such as controlled generative adversarial networks (cGANs) hold significant potential in enhancing or replacing CFD investigations, due to cGANs ability to approximate the underlying data distribution of a dataset. Unlike traditional cGAN applications, where the entire image carries information, CFD data contains small regions of highly variant data, immersed in a large context of low variance that is of minimal importance. This renders most existing deep learning techniques that give equal importance to every portion of the data during training, inefficient. To mitigate this, a novel loss function is proposed called Gradient Mean Squared Error (GMSE) which automatically and dynamically identifies the regions of importance on a field-by-field basis, assigning appropriate weights according to the local variance. To assess the effectiveness of the proposed solution, three identical networks were trained; optimised with Mean Squared Error (MSE) loss, proposed GMSE loss and a dynamic variant of GMSE (DGMSE). The novel loss function resulted in faster loss convergence, correlating to reduced training time, whilst also displaying an 83.6% reduction in structural similarity error between the generated field and ground truth simulations, a 76.6% higher maximum rate of loss and an increased ability to fool a discriminator network. It is hoped that this loss function will enable accelerated machine learning within computational fluid dynamics.
keywords:
Loss function , Neural networks , Generational adversarial networks , Computational fluid dynamics.[label1]organization=Flinders University, Centre for Defence Engineering, Research and Training,city=Adelaide, postcode=5042, state=South Australia, country=Australia
1 Introduction
Computational fluid dynamics (CFD) data developed from automotive, aerospace, maritime and medical applications diverges markedly from traditional images. This is due to the fact that CFD studies the movements of fluids around or through objects of interest. CFD techniques then result in the generation of gradient containing fields that offer insights into the physical changes to the velocity, density and pressure of the fluid. These produced fields, both in 2D and 3D forms, do not conform to traditional visual patterns seen in images [1], such as faces [2], text and numbers [3] or more general images [4], upon which the focus of machine learning often sits. The gradients in CFD flow fields are frequently irregular and influenced by various fluid specific effects [5]. This results in CFD data containing critical components that manifest at a small scale within a much larger flow field, where the majority of the field carries minimal information. Effectively adapting and translating generative machine learning architectures for CFD requires a more nuanced approach [6] to isolate and evaluate predictions in these intricate regions. Historically, this has seen the development of highly specific, tailored loss functions [6, 7, 8, 9]. Where, in machine learning, a loss function serves to map a change in performance to a numerical value, often representing a performance metric. It is employed to assess the learned behaviour of a network, architecture, algorithm, or system of algorithms. The resulting value from the loss function is then used for optimization, where the objective is to decrease or increase the value of the function, based on the specific task.
There are two common approaches to loss function development for a deep learning architecture; generalised or specific. A generalised loss function is not tailored to a specific problem in isolation, but assesses overall qualities of a generated or predicted instance, such as the difference between pixels in a generated image and those in a ground truth [10, 11]. Contrary to this is the development of a specific or custom loss function. This is an involved process requiring knowledge of the important aspects of the optimisation being undertaken and the features of the data to focus on [12]. Tailored loss functions are often problem-dependent, relying on a deep understanding of the domain and specific task at hand. This is challenging when applied to the generation of fluid flow fields, as these specific loss functions [9], make comparisons between related works difficult, as the learnt performance of two identical architectures may be different. Additionally, any loss functions utilized for deep learning must be robust enough to measure deviations or convergences concerning a goal, whilst also exhibiting stability in evaluating performance. Loss function instability can significantly impact the effective training of machine learning architectures by not allowing a network to quickly or effectively capture, learn and subsequently represent the underlying data distribution [13]. Selecting a suitable loss function is a non-trivial decision, often requiring tailoring it to the specific nuances of the problem at hand.
This paper proposes, develops and tests a dynamic loss function that identifies, extracts and uses the gradient containing regions of an image to produce higher quality flow field predictions at faster rates throughout the training process by automatically and dynamically assigning different weights to different regions of the field, based on estimating the local velocity gradient. Additionally, a further variation of the loss function is tested where a priori knowledge is used to further accelerate training speed, loss convergence and image quality by varying known optimal loss function parameters during training. The loss function was implemented in a way that allows for successful operation regardless of the dataset in question, where the weighting is determined on a case by case basis without adaptation or modification for any specific features of a given dataset. The approach is based on the commonly used Mean Square Error (MSE) loss [10, 11], and is modified to automatically identify and dynamically produce a weighting relating to the regions of greatest pixel disparity according to their gradient intensities. The motivation for the development is discussed (Section 2), with the method detailed and equations provided in full (Section 3). The experimental design and algorithm (Section 4) are provided, with the results (Section 4) explained, after the proposed method was comprehensively evaluated. Conclusions are drawn in Section 7.
2 Related Work
Deep learning architectures, particularly in generative tasks, rely heavily on loss functions to guide the learning process. These functions quantify the discrepancy between generated data and ground truth, enabling the model to adjust its parameters and improve its output. While generic loss functions like Mean Squared Error (MSE), Cross-Entropy (CE), and Binary Cross-Entropy (BCE) are widely used, their inherent limitations necessitate exploring tailored alternatives, especially for complex domains like Computational Fluid Dynamics (CFD). MSE loss, a common choice for image generation [14], operates by calculating the average squared difference between corresponding pixels in the generated image and the ground truth [15]. Its effectiveness stems from its straightforward pixel-by-pixel comparison, assuming equal importance of all pixels [16]. However, this assumption can be problematic in applications where specific regions or features hold greater significance.
CE and BCE loss, on the other hand, are frequently employed in classification tasks involving probability distributions [17]. CE loss measures the difference between the predicted probability distribution and the actual distribution, making it suitable for multi-class scenarios. BCE loss, a special case of CE, is specifically designed for binary classification, evaluating the difference between the predicted probability of the positive class and the true binary label [17]. Both CE and BCE are extensively utilized in training deep learning architectures such as neural networks (NNs), convolutional neural networks (CNNs) and generative adversarial networks (GANs).
GANs have emerged as a powerful tool in fluid dynamics, particularly because of their ability to learn from limited datasets [18] – a crucial advantage given the cost and difficulty of generating large CFD datasets [19]. These networks consist of a generator, which synthesizes new data instances, and a discriminator, which attempts to differentiate between generated and real samples. Loss functions play a critical role in training both networks, providing feedback on the generator’s performance and the discriminator’s ability to distinguish real from fake data [20]. GANs, in their original and conditional (cGAN) forms, have been successfully applied to various fluid flow problems, including laminar vortex shedding prediction [21], airfoil flow field reconstruction [22], and supercritical airfoil analysis [23]. These networks have been trained with a range of loss functions, from basic implementations [24] to highly specialized variants [25].
Despite their successes, generic loss functions like MSE, CE, and BCE loss suffer from a key limitation: they treat all errors equally, irrespective of their relative importance within the specific application [26]. This equal weighting scheme fails to capture the intricacies of complex data distributions and relationships [16], potentially leading to generated instances that exhibit superficial similarity to the ground truth while missing crucial underlying features [27]. While modifications to these generic functions have been proposed to address this issue, they often result in highly specialized solutions tailored to a single problem or dataset [12] [28] [29]. These specialized functions lack generalizability and require substantial re-engineering for adaptation to different datasets, making cross-study comparisons challenging.
Consequently, a pressing need exists for a loss function that is general in nature, but can identify, weight and score the regions of importance in fluid field generation. Such a loss function needs to be implemented such that it does not require complex modification or manual intervention by researchers [30] when applied to different contexts or datasets. To achieve this, a novel adaptation to the MSE loss function is proposed. This modification seeks to automatically and dynamically identify the regions where the field changes quickly. This gradient region is then used to weight the loss between the generated field and a ground truth CFD dataset instance on a field by field basis. The determination of gradient weighting is designed such that it will work on any given velocity, pressure and density field, to ensure broad applicability. The novel loss function is further explained in Section 3.
3 Method
The most widely applicable loss function for general image reconstruction, the Mean Squared Error (MSE) loss [10, 11], is used to determine a scalar (0D) mean error value between a ground truth instance () and a secondary, or generated instance () by subtraction and then summation of the error in each direction () over instances. The MSE loss is denoted by Eq. (1). The proposed novel method differs from Eq. (1) by seeking to dynamically identify the regions of importance in the flow fields where strong gradients of interest are present and then weight these regions accordingly in the loss calculation. Computational fluid dynamics data is well suited to this, as strong gradients hold importance in the evaluation of the underlying data. To achieve this custom loss, the MSE loss, Eq. (1), has been modified to include a 2D dynamic weighting () for each generated image (), where is produced from the disparity of the known CFD ground truth instance () on an field-by-field basis. The detailed process to determine this weighting is provided in Section 3.3. The modified form of the MSE loss equation is given below, Eq. (2), and it is called Gradient Mean Squared Error loss (GMSE).
(1) |
(2) |
The following sections detail the conceptualisation, development and testing of the dynamic weighting function. The assessment of it, the test architecture, implementation and algorithm structure are provided, as are the experimental specifics.
3.1 Computational fluid dynamics data
CFD data is intrinsically highly dynamic, gradient driven data. CFD data is produced in the assessment of fluid flow around, or through, an object of interest [31]. When the fluid moves its velocity, pressure and density change in difficult to predict ways, due to the turbulent nature of fluid flow. Subsequently, the results of such analyses contain small regions of interest, with highly varying gradients. The cross sections of these regions are often represented in planar, 2D formats, similar to traditional 2D images [31], making them an ideal candidate for modified generative ML frameworks. An example cross section of such a CFD simulation is provided in Fig. 1, showing a central submarine body, with a fluid flow field (velocity magnitude) around it on a 2D plane. This image shows a highly complex separated flow structure, that is of interest to fluid dynamics researchers. The structure, driven by the submarines propeller, exhibits a large degree of separation from the body and erratic flow structure behaviour due to turbulence. The variant regions, of this wake structure, whilst small in size, are important in the subsequent analysis of the flow field, as they can impact the vessel’s drag and lift characteristics. Optimising the generation and accurate approximation of flow fields such as this and their structures, through deep learning is a primary focus of current research [32]. The next sections will detail the process to identify and dynamically weight these regions using the proposed GMSE loss function.

3.2 Network Architecture
A neural network is required to test the effectiveness of the proposed GMSE loss function. A controlled generational adversarial network (CGAN) was selected to serve this purpose. The CGAN architecture was comprised of two neural networks that worked against each other. A schematic diagram of this architecture is provided in Fig. 2. The first neural network, the Generator (G), was responsible for generating 512x512 grey-scale images of the flow fields, when given the speed and angle of the vessel. A discriminator (D) was used to view both the generated field and a ground truth field from the CFD dataset and determine if the generated field was a true dataset instance or not. The entire architecture was controlled (C) so that the generator was constrained to generate a field based on the speed and angle data it was passed. The architecture of the CGAN was be held constant across testing, with only the loss function used varied to assess the image quality with different modifications to the loss function. Further detail is provided in Section 4.2.

The explicit architecture of the generator and discriminator are provided in Appendix A. The generator architecture took a controlled 1-dimensional latent space in the form of a vector with shape (1, 128) that specified the parameters (flow speed, heading and direction) that the generated image should contain and reshaped it into a 4D tensor, for input into the first convolutional transpose layer. A series of linearly increasing transposed convolutions were used to gradually increase the spatial dimensions whilst learning features [33]. The spatial resolution was increased and up-sampling undertaken by a factor of 2, allowing for a controlled increase in each layer [34]. This occurred in layers 1 through 7. A mixture of ReLU and Leaky ReLU activation functions were used to mitigate the vanishing gradient problem in the deeper layers of the architecture [35]. Batch normalisation was used to normalise the later weights to reduce internal co-variate shift [36]. The generator network was optimised using either the MSE loss or the GMSE loss value, which was then back-propagated through the architecture for optimisation.
The discriminator architecture based on a CNN structure, utilizing 4 convolutional layers were used to take the single channel generated image, extract its features, and then feed it into a fully connected linear layer after flattening. Convolutional layers were used to allow the discriminator to learn from features at different spatial scales whilst the dimensionality of the data was reduced [37]. Leaky ReLU was again used to address vanishing gradients in the deeper layers and batch normalisation used to stabilise discriminator training [35], [36]. The fully connected linear layer then took the feature vector and mapped it to a single output value, to which the Sigmoid activation function was applied, resulting in a classification between 0 (fake) and 1 (real) [38]. This output represented the probability that the image was real (belonging to the CFD dataset, and not the generator). The discriminator was then optimised using cross entropy loss.
3.3 Gradient Mean Squared Error (GMSE) Loss
The Gradient Mean Squared Error (GMSE) is designed to operate using the change in the field to generate more accurate predictions. The operation of the GMSE loss function is sufficiently generalised to allow for velocity, pressure or density fields to be used to determine the weighting used for the loss calculation. To calculate the Gradient Mean Squared Error (GMSE) loss, a series of modifications were made to the MSE loss equation. A visual depiction of this process is provided in Fig. 3, showing the process undertaken to determine the 2D weighting array (Fig. 3d) for the GMSE loss function operation using the ground truth flow field (Fig. 3a), disparity array (Fig. 3b) and blurred array (Fig. 3c).

First, element-wise subtraction was conducted with respect to the x (Eq. 3) and y axis (Eq. 4) of the ground truth field. This determined the gradient regions by the relative difference between adjacent cells/pixels, similar to a unidirectional spatial highpass filter. denotes the x-axis pixel disparity. This is comprised of the amplitude of each pixel , subtracted by the amplitude of the previous pixel . This was then also conducted for the y axis, using and . was then calculated as the magnitude of the resulting vector addition of and (Eq. 5).
(3) |
(4) |
(5) |
This operation (5) is non-linear. The disparity calculation was undertaken to ensure that small scale disparities were included in the disparity array. Traditional difference of Gaussians (DOG) filters with linear addition, unlike Eq. (5), can produce reduced disparity where inverse signs exist. The modification for non-linear addition, Eq. (5), was found to result in increased disparity capture, particularly in high frequency but low amplitude regions of interest. The difference between this operation and a normal DOG filter, followed by rectification, can be seen in Fig. 4. Provided in Fig. 4(a) is the result of a typical DOG filter. The non-linear disparity operation is provided in Fig. 4(b).


A Gaussian blur was then applied to the disparity array to blur and increase the size of the areas of importance, where the disparity occurred. The Gaussian function is given by Eq. (6), where is the distance from the origin in the lateral direction and is the distance from the origin in the vertical direction. In Eq. (6) the value, denotes the standard deviation of the Gaussian distribution, which can be manipulated to shift the values during the blur process, depending on the desired extent of the salience expansion from the point of variation.
(6) |
Next, the blurred array (), was raised to the power of a fixed value, denoted by gamma (), to strengthen or weaken the gradient of the blur in the resulting array (). This modification resulted in different weighting values and was be used to affect the final weighting and performance of the loss function. The formula for this is given in Eq. (7).
(7) |
The modified array () was then normalised between its maximum and minimum values in accordance with Eq. (8). Thus all values, for the normalised weight, , fell in the range [0,1]. A lower bound of 0 can have a negative impact on image generation, where it would indicate that these regions of the image are unimportant. This is not the case, as the region away from the highest disparity areas (the freestream in this case) still retains a limited relative importance. To ensure this is accounted for is adjusted by an offset to achieve a non-zero lower bound.
(8) |
(9) |
The offset was applied, resulting in the range [, 1]. Larger values of increased the relative weight of the surrounding gradient region compared to the strongest gradient structures. The final weighting, developed from the ground truth image, was then determined and is given by Eq. (9). This weighting then applied within the GMSE loss function as given by Eq. (2) to yield the GMSE weighting for each set of image. This loss is used to score the quality of the generated instances produced at each epoch of training.
3.4 Algorithm
Algorithm 1 shows the implementation of the equations required to determine the gradient mean squared error loss between a set of generated images (F) and a set of real images (R). To begin, F and R were both moved to the GPU, in the interest of compute speed. Next, a disparity tensor (D) was constructed. After which on an image-by-image basis using R, the disparity (x, y) was determined, negated and normalised, before being added to D. Then a Gaussian blur was applied to all images within D, in accordance with the selected sigma value. Then a gamma operation was performed, before the blurred arrays were normalised between 0 and 1, and offset used to include a relative importance for all regions of the image. After which the dynamic weighted squared error was computed and then the mean of these values taken. This GMSE value across all images was then returned and used for loss propagation through the generator network.
4 Experiments
This section provides a detailed overview of the dataset utilised for the experiments and the testing methodology. The evaluation metrics employed to assess the impact of parameter variation on the proposed loss metric are outlined, along with the settings and results derived from the conducted experiments.
4.1 Dataset
A dataset of submarine flow was used. This dataset contained 1200 distinct instances. In each instance, a different flow speed and flow angle was used, and the movement of fluid around a submarine was determined. Each instance is a fully resolved computational fluid dynamics (CFD) assessment, using the k-epislon turbulence model in conjunction with the Reynolds Averaged Navier Stokes (RANS) equations as outlined in an existing work [39]. Each CFD assessment in the dataset was then interpolated onto a uniform 2D cross sectional plane (as in Fig. 1) containing the velocity magnitude of the flow. The speed of the flow hitting the submarine was varied from 0.1 m/s to 5.0 m/s and the angle of impact varied from 0 degrees to 60 degrees. This dataset was observed to contains small regions of interest where strong gradients were present, making it ideal to assess the custom GMSE loss function. Each instance in the dataset was developed using ANSYS Fluent, a commercial computational fluid dynamics software. Further details on the generation of this dataset can be found in [39].
4.2 Comparison with related methods
The performance of the novel loss function was assessed via a comparison to the existing Mean Squared Error (MSE) loss, as well as two baseline variants of the GMSE loss function. The first variant (GMSE) had a fixed set of parameters () and the second (DGMSE) had them dynamically changed during training. The GMSE loss function had the parameters fixed such that and . DGMSE was used to assess the effect of beginning with a high blur strength (high ) and a focus on the gradient containing region specifically (low ), before switching to a more equal treatment of the image. The DGMSE loss function had (30, 0.20, 0.1) for the first epoch, then (20, 0.40, 0.1) for epochs 1 to 5, before (20, 0.40, 0.2) was adopted until 20 epochs and then (25, 0.40, 0.2) fixed until the end of training. These settings were selected based on analysis of the results from an exploration of the effect of different parameters on network training performance.
The architecture used to provide the assessment of the loss functions was fixed. The generator of the controlled generational adversarial network (cGAN) from Fig. 2 was optimised using the different loss functions one at a time. Each loss function was used to optimise the generator of the CGAN architecture for a duration of 300 epochs, until convergence was reached. The hyper-parameters used in this training loop were held constant across all tests. During training, the generator loss value, discriminator loss value and the structural similarity index measure (SSIM) were recorded. The justification and explanation of these metrics are provided in Section 4.3.
After the baselines had been tested, variations of the GMSE loss equations were used where the sigma (), gamma () and offset () values were incremented to assess the effect of non-linear shifts, blur strength and relative weighting respectively. The gamma value was incremented between 0.20 and 2.00 in 0.20 increments. Sigma was varied and tested at and . The offset values were varied between 0.1 and 0.9 in 0.1 increments. To ensure the testing was comparable, fixed seeds were used to set the random NumPy variables used in testing and the Torch seed. This ensured that all generative activities, optimisation and back-propagation would begin from the same initial start-point.
4.3 Evaluation metrics
The accuracy and convergence of the proposed method must be evaluated with respect to a known ground truth. Additionally, the evaluation of the accuracy needs to be sufficiently complex to capture the spatial and flow driven relationships present in the data. To ensure this, the effectiveness of the generative network and the loss functions themselves were monitored in several ways.
To make an analysis of the loss convergence during training, the loss was recorded at both the generator and discriminator for both baseline optimised networks as well as all GMSE variants. During training of the generator, the generator loss was recorded using the validation set, for each batch, within each epoch. From this, the batch averaged generator loss at each epoch was produced. This approach was replicated to record the discriminator loss for each batch, within each epoch, resulting in the batch averaged discriminator loss at each epoch. These loss curves were then normalised to a maximum value of 1. The normalised rate of loss per epoch () is also recorded, to provide comparison with regard to the rate of convergence.
To assess the quality of the generated fields as the network learnt, the network was progressively saved over epochs. Both the generator architecture and its associated weights were saved at epochs 1, 5, 10, 20, 50, 100 and 300. Whilst generated fields can have similar MSE values, but different qualitative performance [40], the structural similarity index measure (SSIM) was also determined at 1, 5, 20, 100 and 300 epochs. The SSIM gives a better indication of the structural similarity between the generated flow field and the ground truth RANS flow field. The SSIM determines the image degradation where there are changes, or perceived changes to the structural information of a field [41] occur. The structural similarity index was computed by application of Eq. 10. Where the two fields (generated and RANS) are used, as denoted by (x, y). The pixel mean is given by and the variance , specific to each field and the covariance by . Two constants () were used to stabilise the division operations [42]. The two constants are given by Eq. 11 and 12 respectively, where denotes the dynamic range of the pixel-values, and are fixed constants of 0.01 and 0.03 respectively, as per [43].
(10) |
(11) |
(12) |
In addition to the quantitative metrics, a set of fields were produced for qualitative comparison between the MSE and GMSE baseline optimised networks during training for a number of different epochs.
4.4 Experimental setting
Experimental testing was conducting using a PC with an AMD Threadripper Pro 5955WX, 256GB of RAM and 2 NVIDIA A6000 GPUs. The testing was conducted using Python 3.11.6 and the Pytorch (2.0.0 + cu117), Numpy (1.22.4), Pandas (2.1.4) and Kornia (0.7.1) libraries. To assess the loss functions, the networks were optimised for 100 epochs, after which GMSE convergence was seen. The learning rate of the generator architecture was fixed to and for the discriminator. A batch size () of 20 images was used with the Adam optimizer used for both generator and discriminator.
5 Results
Table 1 presents the structural similarity index measure (SSIM) at epochs 1, 5, 20, 100 and 300. The maximum normalised rate of loss is also indicated. The standard deviation () is also provided to give an indication of the performance variability with respect to modifying the gamma (), sigma () or offset () parameters of the baseline GMSE loss function. The MSE loss optimised network performance and that of the dynamic gradient mean squared error loss function (DGMSE) are also provided. Indicated in bold is the GMSE baseline for the gamma (), sigma () and offset () variations.
The GMSE and DGMSE loss functions were seen to outperform the MSE loss function in both SSIM and maximum loss rate across all epochs of training. At early epochs, when the SSIM is low, the GMSE (0.068) and DGMSE (0.075) still outperformed MSE optimisation (0.032). The disparity between the approaches widened further through training, where the GMSE and DGMSE loss functions proved beneficial, reporting final SSIM’s of 0.988 and 0.989 respectively, compared to 0.933 for the MSE loss function. Marking a reduction in structural dissimilarity of 82.1% and 83.6% in comparison to the MSE loss. An increased rate of maximum loss was also observed for the GMSE and DGMSE variants of -0.143 (33.6% higher) and -0.189 (76.6% higher) respectively, compared to -0.107 for the MSE loss function. This stronger rate of normalised loss convergence indicates that the network can better optimise in response to the initial loss between the generated instances and the real dataset instances.
Variations to the gamma () variable were seen to influence both the SSIM and the rate of loss. The standard deviation of gamma variation was observed to be larger (0.082, 0.108 and 0.146) during early training (1, 5 and 20 epochs) before reducing to smaller values (0.033, 0.018) towards the completion of training (100 and 300 epochs). This correlates to the close SSIM observed at later training stages, with most structural similarities being clustered around 0.983 to 0.989, with a maximum recorded SSIM of 0.991. The maximum SSIM outperformed the MSE baseline by 86.6%. There is more variance reported in the rate of loss recorded for the gamma variations, with a standard deviation of 0.017, a maximum of -0.177 () and a minimum of -0.115 (). are seen to correspond to reduced rates of loss, and variations where correspond to an increased rate of loss.
Variations to sigma () were also observed to effect the SSIM and the rate of loss, with different performance depending on the strength of the blur. Higher blur strengths () were seen to correspond the highest SSIM at epochs 1 (0.285), 5 (0.654), 20 (0.926) and 100 (0.982). Blur strengths between 10 and 30 also corresponded to the highest rates of loss (-0.143, -0.141 and -0.133), outperforming the MSE baseline by 33.6%, 31.7% and 24.3% respectively. Again, the variation of SSIM was seen to decrease as training progressed, with a standard deviation of 0.022 reported at the completion of training.
Variations to the offset () had the largest SSIM deviation at the completion of training (0.028) indicating that the offset weighting can still affect the image quality further into the training cycle. This is reflected by the higher SSIM values recorded for offsets of 0.1, 0.2 and 0.3, which highly weight the gradient regions. These offset values corresponded to the highest SSIM values recorded for the offsets (0.986, 0.988, 0.976 respectively). The opposite is true of which were observed to record SSIMs between 0.923 and 0.926, thus under performing the MSE loss by 10.5% to 14.9%. It should be noted that all variations to the baseline and offset (() still reported higher rates of loss (-0.110 to -0.158) than the MSE baseline by a minimum of 2.8% to a maximum of 47.7%.
Variant | # | SSIM at epoch | |||||
1 | 5 | 20 | 100 | 300 | |||
Baselines | MSE | 0.032 | 0.239 | 0.456 | 0.829 | 0.933 | -0.107 |
GMSE | 0.068 | 0.617 | 0.833 | 0.972 | 0.988 | -0.143 | |
DGMSE | 0.075 | 0.674 | 0.935 | 0.978 | 0.989 | -0.189 | |
0.20 | 0.282 | 0.474 | 0.918 | 0.978 | 0.991 | -0.124 | |
0.40 | 0.167 | 0.632 | 0.940 | 0.983 | 0.990 | -0.142 | |
0.60 | 0.185 | 0.557 | 0.916 | 0.980 | 0.989 | -0.177 | |
0.80 | 0.103 | 0.593 | 0.847 | 0.974 | 0.984 | -0.130 | |
1.00 | 0.068 | 0.617 | 0.833 | 0.972 | 0.988 | -0.143 | |
1.20 | 0.053 | 0.561 | 0.722 | 0.961 | 0.983 | -0.120 | |
1.40 | 0.038 | 0.487 | 0.715 | 0.949 | 0.983 | -0.121 | |
1.60 | 0.026 | 0.441 | 0.642 | 0.936 | 0.976 | -0.130 | |
1.80 | 0.036 | 0.332 | 0.545 | 0.905 | 0.954 | -0.121 | |
2.00 | 0.028 | 0.307 | 0.523 | 0.881 | 0.935 | -0.115 | |
0.082 | 0.108 | 0.146 | 0.033 | 0.018 | 0.017 | ||
2 | 0.021 | 0.281 | 0.408 | 0.863 | 0.929 | -0.118 | |
5 | 0.040 | 0.460 | 0.664 | 0.943 | 0.981 | -0.126 | |
10 | 0.068 | 0.617 | 0.833 | 0.972 | 0.988 | -0.143 | |
20 | 0.213 | 0.654 | 0.926 | 0.982 | 0.989 | -0.141 | |
30 | 0.285 | 0.630 | 0.922 | 0.980 | 0.991 | -0.133 | |
40 | 0.196 | 0.623 | 0.915 | 0.977 | 0.990 | -0.126 | |
0.099 | 0.134 | 0.189 | 0.042 | 0.022 | 0.009 | ||
0.1 | 0.103 | 0.538 | 0.923 | 0.979 | 0.986 | -0.158 | |
0.2 | 0.068 | 0.617 | 0.833 | 0.972 | 0.988 | -0.143 | |
0.3 | 0.036 | 0.429 | 0.524 | 0.949 | 0.976 | -0.126 | |
0.4 | 0.020 | 0.274 | 0.361 | 0.859 | 0.924 | -0.134 | |
0.5 | 0.019 | 0.261 | 0.447 | 0.874 | 0.924 | -0.133 | |
0.6 | 0.017 | 0.203 | 0.372 | 0.875 | 0.926 | -0.132 | |
0.7 | 0.025 | 0.250 | 0.380 | 0.855 | 0.926 | -0.115 | |
0.8 | 0.024 | 0.247 | 0.421 | 0.841 | 0.923 | -0.118 | |
0.9 | 0.027 | 0.255 | 0.444 | 0.860 | 0.925 | -0.110 | |
0.028 | 0.136 | 0.189 | 0.050 | 0.028 | 0.013 |






Three plots detail the generator loss during training. The Adam optimiser has been used to reduce this loss during training for each individual test. The maximum loss values have been normalised to 1 for all data shown. Fig. 5 provides the gamma () generator loss, Fig. 6 provides the sigma () generator loss and Fig. 7 provides the offset loss () generator loss. The x axis details the epochs of training, and the y axis provides the normalised loss of each test. Indicated in red is the MSE baseline, black details the GMSE baseline and blue the DGMSE loss. Three additional figures are provided, covering the discriminator loss during training. A separate CNN network was used to assess both the real and generated images during training. The goal of the generator was to produce images that would be awarded a value of 1, indicating the generated fake image belonged to the real dataset when scored by the discriminator. Fig. 8 provides the gamma variant discriminator performance, Fig. 9 provides the sigma variant and Fig. 10 the offset variant performance. The y axis shows the score from the discriminator where 1 is the target and the x axis denotes the epoch of training. Again, indicated in red is the MSE baseline, black is the GMSE baseline and blue the DGMSE loss.
The generator loss of the gamma variations provided in Fig. 5 were seen to converge faster than the MSE loss optimised network. Both the GMSE and DGMSE loss functions provided beneficial loss characteristics in comparison to the gamma variants themselves, which were seen to converge at different rates, but to a slower extent than the GMSE baseline (black). Initial rates of loss in the gamma 0.20, 0.40 and 0.60 were seen to outperform the GMSE baseline to a limited extent, but with ultimately lower maximum rates, as reflected by Table 1. Gamma variations of 0.20 to 0.80 were seen to converge to a higher final loss than (Fig. 5), indicating that the loss function may benefit from dynamic changes to during training. The dynamic variant, DGMSE, was seen to outperform all gamma variants and both the MSE and GMSE baselines and resulted in the lowest final loss. The ability of the gamma variants to fool a discriminator network, Fig. 8, is seen to have limited variation and relatively robust performance compared to the sigma discriminator loss (Fig. 9) and the offset discriminator loss (Fig. 10). All gamma variants are observed to fool the discriminator network, at earlier epochs, compared to the MSE baseline. variations are seen to have better performance at early epochs (), but worse performance later in training (). The GMSE baseline, with no shift was observed to perform robustly, with moderate performance in early training but a better ability to overcome the discriminator at later stages. The DGMSE loss was observed to perform well early in training and until completion, where it recorded the highest final loss. The second highest final loss was the GMSE loss function, followed by the variants, then the MSE loss and finally the variants. Thus, further supporting the adoption of a dynamic approach (DGMSE) or .
Sigma was seen to also effect the generator loss during training (Fig. 6), but to differing extents depending on the strength of the blur. In the initial stages of training ( epochs), strong blurs () corresponded to better loss rates. As training progressed however ( epochs) smaller blurs became more performant and resulted in lower final loss values at the completion of training. The MSE baseline was found to outperform the sigma 2 and 5 variants, only at the completion of training. Both the GMSE and DGMSE outperformed the baselines and all sigma variants, with the DGMSE displaying beneficial initial loss, where it outperformed all variants with the highest recorded normalised loss rate (-0.189). The ability of the sigma variants to fool the discriminator (Fig. 9) is also dependent on blur strength. Low strength blurs () are observed to perform similarly to the MSE baseline, whilst stronger blurs offer a discriminator loss advantage with markedly better early perform loss (epochs ), after which the strength of the blur is less impactful and closer loss performance is observed. Again, the DGMSE is the most performant, followed by the sigma variants in descending order of strength (40 to 20), then the GMSE variant, sigma 5 variant, the MSE baseline and finally the sigma 2 variant.
The offset weighting of the gradient free region was also seen to have an effect on generator loss convergence, as provided in Fig. 7. Lower values indicate where the gradient region is highly important and higher values prioritise the gradient free region. Initial convergence of all offset variants outperformed the MSE baseline, with this continuing for the entirety of training, with a notable exception to this, as it resulted in higher loss from 130 epochs onwards. The variant was also found to result in higher loss than the MSE baseline at the completion of training. These findings indicate that both the gradient and gradient free regions retain importance during training and the weighting should account for both. The higher performance of the GMSE and DGMSE baselines indicate the benefits of either a balanced or dynamic approach. For the offset discriminator loss performance (Fig. 10), offsets equal to or above 0.5 are seen to perform close to the MSE baseline in early training between 0 and 100 epochs, before converging to lower loss values (indicating decreased performance). Conversely, offsets less than 0.5, where the gradient regions are weighted heavily, correspond to better initial training (0 to 100 epochs) as well as better final convergence. The most beneficial final loss is achieved by the DGMSE variant, followed by , then the GMSE baseline (with ), then , the MSE baseline and then .
Qualitative field assessment also offers insight into the behaviour of the baseline MSE and GMSE loss optimised networks. Fig. 11 shows the generation abilities of the two baseline networks as training progressed. Both networks have similar performance after a single epoch, before deviation at the displayed 5, 10, 20, 50 and 300 epochs. It was seen that a circular spatial reconstruction error was present where the payload bay of the submarine was located in the fluid flows of [39]. This error does not appear in the GMSE loss optimised network, where these region had a higher weighting for accurate generation. The strong gradient regions around the submarine developed earlier in the training process for the GMSE loss optimised network and were refined at later stages of training.

6 Discussion
The results of these tests demonstrate the potential of gradient capture driven GMSE and DGMSE loss functions for accelerating training speeds of generative architectures and ensuring higher quality generations for fluidic applications. The dynamic extraction and weighting of regions of importance was demonstrated to be a robust means of accelerating loss convergence at both the generator and discriminator network of a cGAN architecture, resulting in several advantages when compared to MSE loss. These advantages are evident in several aspects. The first is the structural similarity index measure (SSIM) performance which was observed to consistently surpass that of the MSE baseline across all variants of the GMSE loss function. The higher recorded SSIM indicates a more accurate reconstruction of the complex gradient structures of the flow field, as evidenced by both Fig. 11 and the data of Table 1. The 82.1% and 83.6% reduction in relative SSIM error achieved by the GMSE and DGMSE loss function highly the potential of leveraging gradient information to improve generation fidelity.
Notably, the DGMSE and GMSE variants resulted in a faster rate of convergence during training, evidenced by the higher maximum loss rates in comparison to the MSE baseline. This accelerated learning can be attributed to the GMSE function’s ability to provide more informative gradients to the generator, allowing for faster and more effective optimisation. This finding carries significant practical implications, as faster training directly translates to reduced computational cost and faster development cycles for deep learning. Furthermore, the DGMSE and GMSE loss functions success in consistently and rapidly fooling the discriminator network highlights its capacity to guide the generator towards producing more realistic and challenging flow fields for the discriminator to score. This superior performance, particularly at early epochs (Fig. 8, Fig. 9, Fig. 10) suggests that the GMSE loss facilities a more efficient exploration of the data and results in a more effective approximation of the underlying data distribution of the RANS dataset.
Exploration of the GMSE function’s parameters reveals their influence on its efficacy. The blur strength, controlled by , highlights a crucial trade-off. Weaker blurs () retain precise gradient localization but risk overfitting to small-scale features, while stronger blurs () provide a more generalized weighting, potentially diluting the focus on crucial regions, but are still seen to outperform MSE loss. The optimal blur strength is likely dataset-dependent, but DGMSE substantiates the benefits of starting with a strong blur and decreasing the strength during training. This technique is seen to result in beneficial early and late stage SSIM, as well as a higher maximum loss rate (Fig. 6).
The gamma parameter, , governs the contrast enhancement of the weighting mask. While its overall impact on SSIM appears limited within the tested range, it significantly affects the convergence behaviour (Table 1). Higher gamma values (), emphasizing high-gradient regions, lead to faster initial convergence but potentially at the cost of a higher final loss. This suggests a dynamic gamma adjustment strategy could further optimize training by initially prioritizing rapid error reduction and later shifting focus to fine-grained details.
The offset parameter, , dictates the relative weighting of gradient-rich versus gradient-free regions. Assigning high weights to gradient-free regions () consistently degraded performance, highlighting the importance of prioritizing areas containing crucial flow structures. Conversely, excessively low offset values () might overshadow the contribution of the surrounding context. The optimal offset likely depends on the specific dataset and the relative significance of high-gradient features versus their surroundings.
The DGMSE and GMSE loss functions are also seen to achieve these advantages without requiring domain specific knowledge or complex modifications to the training data, to account for highly customised loss functions [44]. While these customised loss functions for CFD data do exist, they introduce significant complexity [29] and can lead to unstable or incomparable training [29] – limiting their generalizability. The DGMSE and GMSE loss function, in contrast, relies on the simpler method of prioritizing the regions of high gradient disparity – making it easily adaptable to other datasets exhibiting similar characteristics – such as temperature fields [45], hydrodynamic [46] and aerodynamic data [47] and fields with strong edge structures [1].
Generational architectures are susceptible to a range of different failure modes or issues that render training challenging. Most often discussed are mode collapse and convergence failure [48]. Mode collapse occurs where the generator struggles to produce a variety of data and produces one output for a larger range of inputs [49]. This is prevalent when working with small datasets and model sizes, where either the model or dataset is not sufficiently large, and struggles to approximate the underlying data distribution. This often leads to generative errors, in the case of GAN architectures. In a limited capacity, this can be seen in Fig. 11 where the centre of the submarine in the MSE images (top) can be seen to feature a consistently low field generation, despite this not being the desired output. This can be contrast against the more correct GMSE generation (bottom). The effective prevention of mode collapse is challenging but may be alleviated by the function of the GMSE loss function. Giving an increased weighting to the images regions of importance in the GMSE loss function results in the generator being proportionally punished for generating poor outputs, in areas such as the submarine itself, leading to a more robust generative ability. Using the natural weighting behaviour of the loss function to combat mode collapse offers an easier pathway to prevention than the implementation of other, more complex methods of prevention, like those explored in literature [50]. This is because the generator is incentivised to produce higher quality images by way of the weighted loss term.
7 Conclusions
This research introduces the GMSE loss function as a powerful and versatile tool for enhancing the training of generative deep learning architectures, with a focus on fluid fields and gradient-drive data. The ability of both the DGMSE and GMSE loss function to dynamically prioritise gradient containing regions, whilst remaining conceptually simple and adaptable underscores its potential for broader applicability. Future research should investigate the performance of the GMSE loss function with other gradient-containing datasets, explore its application to higher dimension data (such as 3D volumes) and delve into its potential for mitigating mode collapse in generative models. This work represents a significant step towards efficient, effective and generally applicable training techniques for generative architectures across a range of fields.
CRediT authorship contribution statement
Zachary Cooper-Baldock: Investigation, Conceptualization, Methodology, Formal analysis, Writing - original draft, Writing - review and editing. Paulo E. Santos, Russell S.A. Brinkworth, Karl Sammut: Supervision, Conceptualization, Methodology, Writing - review and editing.
Declaration of competing interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Data availability
Data will be made available upon reasonable request.
Acknowledgments
This work was supported by an NCI HPC-AI Talent Program 2023 Scholarship, with computational resources provided by NCI Australia, an NCRIS enabled capability supported by the Australian Government.
Appendix A Architectures
Contained here are the architecture schematics of the controlled generative adversarial network used in the investigation of the GMSE and DGMSE loss functions. Fig. 12 displays the discriminator architecture and Fig. 13 displays the generator architecture.


References
- [1] J. R. Clough, N. Byrne, I. Oksuz, V. A. Zimmer, J. A. Schnabel, A. P. King, A topological loss function for deep-learning based image segmentation using persistent homology, IEEE Transactions on Pattern Analysis and Machine Intelligence 44 (12) (2022) 8766–8778. doi:10.1109/TPAMI.2020.3013679.
-
[2]
A. Kammoun, R. Slama, H. Tabia, T. Ouni, M. Abid, Generative adversarial networks for face generation: A survey, ACM Comput. Surv. 55 (5) (dec 2022).
doi:10.1145/3527850.
URL https://doi.org/10.1145/3527850 - [3] N. Elaraby, S. Barakat, A. Rezk, A conditional gan-based approach for enhancing transfer learning performance in few-shot hcr tasks, Nature Scientific Reports 12 (2022) 16271. doi:10.1038/s41598-022-20654-1.
- [4] I. H. Rather, S. Kumar, Generative adversarial network based synthetic data training model for lightweight convolutional neural networks, Multimedia Tools and Applications 83 (2024) 6429–6271. doi:10.1007/s11042-023-15747-6.
- [5] C. Duru, H. Alemdar, O. U. Baran, A deep learning approach for the transonic flow field predictions around airfoils, Computers & Fluids 236 (2022) 105312. doi:https://doi.org/10.1016/j.compfluid.2022.105312.
- [6] Z. Pang, K. Liu, H. Xiao, T. Jin, K. Luo, J. Fan, A deep-learning super-resolution reconstruction model of turbulent reacting flow, Computers & Fluids 275 (2024) 106249. doi:https://doi.org/10.1016/j.compfluid.2024.106249.
- [7] F. Jaffar, T. Farid, M. Sajid, Y. Ayaz, M. Jawad, Prediction of drag force on vehicles in a platoon configuration using machine learning, IEEE Access 8 (11 2020). doi:10.1109/ACCESS.2020.3035318.
- [8] G. Calzolari, W. Liu, Deep learning to replace, improve, or aid cfd analysis in built environment applications: A review, Building and Environment 206 (2021) 108315. doi:https://doi.org/10.1016/j.buildenv.2021.108315.
- [9] M. Z. Yousif, L. Yu, S. Hoyas, R. Vinuesa, H. Lim, A deep-learning approach for reconstructing 3d turbulent flows from 2d observation data, Nature Scientific Reports 13 (2023) 2359. doi:https://doi.org/10.1038/s41598-023-29525-9.
- [10] H.-S. Choi, M. Jeon, K. Song, M. Kang, Semantic fire segmentation model based on convolutional neural network for outdoor image, Fire Technology 57 (2021) 3005–3019. doi:10.1007/s10694-020-01080-z.
- [11] J. F. Ciprián-Sánchez, G. Ochoa-Ruiz, L. Rossi, F. Morandini, Assessing the impact of the loss function, architecture and image type for deep learning-based wildfire segmentation, Applied Sciences 11 (15) (2021). doi:10.3390/app11157046.
- [12] X. Ling, C. Li, P. Yang, Y. Huang, F. Qin, Tu2net-gan: A temporal precipitation nowcasting model with multiple decoding modules, Pattern Recognition Letters 178 (2024) 98–105. doi:10.1016/j.patrec.2023.12.025.
- [13] X. Zhang, W. Lu, Y. Pan, H. Wu, R. Wang, R. Yu, Empirical study on tangent loss function for classification with deep neural networks, Computers & Electrical Engineering 90 (2021) 107000. doi:10.1016/j.compeleceng.2021.107000.
- [14] Y. Chen, L. Li, W. Li, Q. Guo, Z. Du, Z. Xu, AI Computing Systems, Chapter 2 - Fundamentals of neural networks, 1st Edition, Morgan Kaufmann, Burlington, MA, 2024.
- [15] A. Lucas, A. K. Katsaggelos, S. Lopez-Tapuia, R. Molina, Generative adversarial networks and perceptual losses for video super-resolution, in: 2018 25th IEEE International Conference on Image Processing (ICIP), 2018, pp. 51–55. doi:10.1109/ICIP.2018.8451714.
- [16] H. Zhao, O. Gallo, I. Frosio, J. Kautz, Loss functions for image restoration with neural networks, IEEE Transactions on Computational Imaging 3 (1) (2017) 47–57. doi:10.1109/TCI.2016.2644865.
- [17] Z. Zhang, M. R. Sabuncu, Generalized cross entropy loss for training deep neural networks with noisy labels, in: 32nd Conference on Neural Information Processing Systems, 2018, pp. 1–11. doi:10.48550/arXiv.1805.07836.
-
[18]
S. Gurumurthy, R. Sarvadevabhatla, R. Babu, Deligan: Generative adversarial networks for diverse and limited data, in: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE Computer Society, Los Alamitos, CA, USA, 2017, pp. 4941–4949.
doi:10.1109/CVPR.2017.525.
URL https://doi.ieeecomputersociety.org/10.1109/CVPR.2017.525 - [19] Z. Wang, X. Liu, J. Yu, H. Wu, H. Lyu, A general deep transfer learning framework for predicting the flow field of airfoils with small data, Computers & Fluids 251 (2023) 105738. doi:https://doi.org/10.1016/j.compfluid.2022.105738.
- [20] J. Xu, X. Ren, J. Lin, X. Sun, Diversity-promoting GAN: A cross-entropy based generative adversarial network for diversified text generation, in: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, 2018, pp. 3940–3949. doi:10.18653/v1/D18-1428.
- [21] S. Lee, D. You, Data-driven prediction of unsteady flow over a circular cylinder using deep learning, Journal of Fluid Mechanics 879 (2019) 217–254.
- [22] D. Chen, X. Gao, C. Xu, S. Chen, J. Fang, Z. Wang, Z. Wang, Flowgan: A conditional generative adversarial network for flow prediction in various conditions, in: 2020 IEEE 32nd International Conference on Tools with Artificial Intelligence (ICTAI), 2020, pp. 315–322. doi:10.1109/ICTAI50040.2020.00057.
- [23] H. Wu, X. Liu, W. An, S. Chen, H. Lyu, A deep learning approach for efficiently and accurately evaluating the flow field of supercritical airfoils, Computers & Fluids 198 (2020) 104393. doi:https://doi.org/10.1016/j.compfluid.2019.104393.
- [24] L.-W. Chen, N. Thuerey, Towards high-accuracy deep learning inference of compressible flows over aerofoils, Computers & Fluids 250 (2023) 105707. doi:https://doi.org/10.1016/j.compfluid.2022.105707.
- [25] P. Heger, D. Hilger, M. Full, N. Hosters, Investigation of physics-informed deep learning for the prediction of parametric, three-dimensional flow based on boundary data, Computers & Fluids 278 (2024) 106302. doi:https://doi.org/10.1016/j.compfluid.2024.106302.
- [26] J. Andrés-Ferrer, D. Ortiz-Martínez, I. García-Varea, F. Casacuberta, On the use of different loss functions in statistical pattern recognition applied to machine translation, Pattern Recognition Letters 29 (8) (2008) 1072–1081, pattern Recognition in Interdisciplinary Perception and Intelligence. doi:10.1016/j.patrec.2007.06.015.
- [27] Y. Tian, D. Su, S. Lauria, X. Liu, Recent advances on loss functions in deep learning for computer vision, Neurocomputing 497 (2022) 129–158. doi:10.1016/j.neucom.2022.04.127.
- [28] Y. Kim, Y. Lee, M. Jeon, Imbalanced image classification with complement cross entropy, Pattern Recognition Letters 151 (2021) 33–40. doi:10.1016/j.patrec.2021.07.017.
- [29] K. R. Lennon, G. H. McKinley, J. W. Swan, Scientific machine learning for modeling and simulating complex fluids, Proceedings of the National Academy of Sciences of the United States of America 120 (2023) 1–9. doi:10.1073/pnas.2304669120.
- [30] M. Abdar, F. Pourpanah, S. Hussain, D. Rezazadegan, L. Liu, M. Ghavamzadeh, P. Fieguth, X. Cao, A. Khosravi, U. R. Acharya, V. Makarenkov, S. Nahavandi, A review of uncertainty quantification in deep learning: Techniques, applications and challenges, Information Fusion 76 (2021) 243–297. doi:https://doi.org/10.1016/j.inffus.2021.05.008.
- [31] J. Alexandersen, C. S. Andreasen, A review of topology optimisation for fluid-based problems, Fluids 5 (1) (2020). doi:10.3390/fluids5010029.
- [32] S. Masoumi-Verki, F. Haghighat, N. Bouguila, U. Eicker, The use of GANs and transfer learning in model-order reduction of turbulent wake of an isolated high-rise building, Building and Environment 246 (2023) 110948. doi:10.1016/j.buildenv.2023.110948.
-
[33]
T. Karras, T. Aila, S. Laine, J. Lehtinen, Progressive growing of GANs for improved quality, stability, and variation, in: International Conference on Learning Representations, 2018, pp. 1–26.
URL https://openreview.net/forum?id=Hk99zCeAb - [34] V. Dumoulin, F. Visin, A guide to convolution arithmetic for deep learning (2018). arXiv:1603.07285.
- [35] O. A. Montesinos López, A. Montesinos López, J. Crossa, Fundamentals of Artificial Neural Networks and Deep Learning, Springer International Publishing, Cham, 2022, Ch. 10, pp. 379–425.
- [36] S. Santurkar, D. Tsipras, A. Ilyas, A. Madry, How does batch normalization help optimization?, in: 32nd Conference on Neural Information Processing Systems (NeurIPS), 2018, pp. 1–11.
- [37] R. Ayachi, M. Afif, Y. Said, M. Atri, Strided convolution instead of max pooling for memory efficiency of convolutional neural networks, in: M. S. Bouhlel, S. Rovetta (Eds.), Proceedings of the 8th International Conference on Sciences of Electronics, Technologies of Information and Telecommunications (SETIT’18), Vol.1, Springer International Publishing, Cham, 2020, pp. 234–243.
- [38] N. Chintalapudi, G. Battineni, M. A. Hossain, F. Amenta, Cascaded deep learning frameworks in contribution to the detection of parkinson’s disease, Bioengineering (Basel) 9 (2022) 116–128. doi:10.3390/bioengineering9030116.
- [39] Z. Cooper-Baldock, P. E. Santos, R. S. A. Brinkworth, K. Sammut, Hydrodynamic analysis of payload bay berthing for underwater vehicles, IEEE Journal of Oceanic Engineering 49 (2024) 727–748. doi:10.1109/JOE.2024.3352714.
- [40] H. Yeganeh, Cross dynamic range and cross resolution objective image quality assessment with applications, Ph.D. thesis, University of Waterloo (July 2014). doi:10.13140/RG.2.1.1900.2320.
- [41] Z. Wang, A. Bovik, H. Sheikh, E. Simoncelli, Image quality assessment: from error visibility to structural similarity, IEEE Transactions on Image Processing 13 (4) (2004) 600–612. doi:10.1109/TIP.2003.819861.
- [42] Z. Wang, E. Simoncelli, A. Bovik, Multiscale structural similarity for image quality assessment, in: The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003, Vol. 2, 2003, pp. 1398–1402 Vol.2. doi:10.1109/ACSSC.2003.1292216.
- [43] Z. Wang, E. Simoncelli, A. Bovik, Multiscale structural similarity for image quality assessment, in: The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003, Vol. 2, 2003, pp. 1398–1402 Vol.2. doi:10.1109/ACSSC.2003.1292216.
- [44] L. Hu, W. Wang, Y. Xiang, J. Zhang, Flow field reconstructions with gans based on radial basis functions, IEEE Transactions on Aerospace and Electronic Systems 58 (4) (2022) 3460–3476. doi:10.1109/TAES.2022.3152706.
- [45] M. Kuzay, A. Dogan, S. Yilmaz, O. Herkiloglu, A. S. Atalay, A. Cemberci, C. Yilmaz, E. Demirel, Numerical and experimental dataset for a retrofitted data center, Data in Brief 45 (2022) 108587. doi:https://doi.org/10.1016/j.dib.2022.108587.
- [46] T. Nijssen, J. Padding, M. Ottens, Data companion to: Hydrodynamics of expanded bed adsorption studied through cfd-dem (2023). doi:10.4121/ebaf0dc5-d8d2-4eb7-8f28-21f7f9b2335c.v1.
- [47] R. McConkey, E. Yee, F.-S. Lien, A curated dataset for data-driven turbulence modelling, Nature Scientific Data 8 (255) (2021). doi:10.1038/s41597-021-01034-2.
- [48] Z. Zhang, M. Li, J. Yu, On the convergence and mode collapse of gan, in: SIGGRAPH Asia 2018 Technical Briefs, SA ’18, Association for Computing Machinery, New York, NY, USA, 2018, pp. 1–4. doi:10.1145/3283254.3283282.
- [49] Bhagyashree, V. Kushwaha, G. C. Nandi, Study of prevention of mode collapse in generative adversarial network (gan), in: 2020 IEEE 4th Conference on Information & Communication Technology (CICT), 2020, pp. 1–6. doi:10.1109/CICT51604.2020.9312049.
- [50] S. Pei, R. Y. D. Xu, S. Xiang, G. Meng, Alleviating mode collapse in GAN via diversity penalty module, CoRR abs/2108.02353 (2021). arXiv:2108.02353.