This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Machine-Learned HASDM Model with Uncertainty Quantification

Abstract

The first thermospheric neutral mass density model with robust and reliable uncertainty estimates is developed based on the SET HASDM density database. This database, created by Space Environment Technologies (SET), contains 20 years of outputs from the U.S. Space Force’s High Accuracy Satellite Drag Model (HASDM), which represents the state-of-the-art for density and drag modeling. We utilize principal component analysis (PCA) for dimensionality reduction, creating the coefficients upon which nonlinear machine-learned (ML) regression models are trained. These models use three unique loss functions: mean square error (MSE), negative logarithm of predictive density (NLPD), and continuous ranked probability score (CRPS). Three input sets are also tested, showing improved performance when introducing time histories for geomagnetic indices. These models leverage Monte Carlo (MC) dropout to provide uncertainty estimates, and the use of the NLPD loss function results in well-calibrated uncertainty estimates without sacrificing model accuracy (¡10% mean absolute error). By comparing the best HASDM-ML model to the HASDM database along satellite orbits, we found that the model provides robust and reliable uncertainties in the density space over all space weather conditions. A storm-time comparison shows that HASDM-ML also supplies meaningful uncertainty measurements during extreme events.

\journalname

Space Weather

Department of Mechanical and Aerospace Engineering, West Virginia University, Morgantown, West Virginia, USA. Space Environment Technologies, Pacific Palisades, California, USA. School of Mathematical and Data Sciences, West Virginia University, Morgantown, West Virginia, USA.

\correspondingauthor

Richard J. [email protected]

{keypoints}

First thermosphere model with robust and reliable uncertainty estimates is developed.

HASDM-ML provides predictions with approximately 10% error relative to HASDM database across 20 years of available data.

Heteroskedastic loss function provides model with well-calibrated uncertainty estimates across all conditions.

Plain Language Summary

The first upper-atmospheric density model with robust and reliable uncertainty estimates is developed based on the SET HASDM density database. This database contains 20 years of outputs from the High Accuracy Satellite Drag Model (HASDM), which represents the state-of-the-art for density and drag modeling. We use a decomposition tool called principal component analysis (PCA) to reduce the dimensionality of the dataset. Three loss functions, mean square error (MSE), negative logarithm of predictive density (NLPD), and continuous ranked probability score (CRPS), are tested with three input sets to identify the best-performing model. We optimize nine models (all three loss functions and input sets) and compare the prediction accuracy and the reliability of its uncertainty estimates. The models leverage Monte Carlo dropout to generate probabilistic outputs from which we extract model uncertainty. We find that using an input set containing a time series for the geomagnetic indices results in the most accurate models. In addition, the model using these inputs with the NLPD loss function has sufficient performance (\sim10% absolute error) and the most calibrated/reliable uncertainty estimates on independent data. We test this model’s uncertainty capabilities in the density space along satellite orbits from 2002-2010, showing the model’s reliability across all conditions.

1 Introduction

Space Situational Awareness (SSA) is a major focus of space agencies and private companies worldwide. With the number of objects in low-Earth orbit (LEO) continuously growing, knowledge of future satellite/debris positions is becoming increasingly important (Radtke et al. 2017). While there are numerous perturbations affecting the trajectories of objects, atmospheric drag is the largest source of uncertainty in the LEO regime. Our current understanding of the thermosphere is incomplete, resulting in imperfect modeling of neutral mass density. Over the past several decades, researchers have developed increasingly accurate models and made improvements to existing ones. This has come from a combination of the incorporation of new measurements and refinements of the underlying physics (Emmert 2015).

The thermosphere is the neutral region of the upper atmosphere ranging from \sim90 km to 500-1000 km depending on space weather conditions. The thermosphere is impacted by external forces (e.g. space weather events) and internal dynamics (e.g. species transport). Extreme Ultraviolet (EUV) emissions from the Sun heat up the thermosphere forcing expansion and increasing density at a given location. Effects from solar emissions can be represented by different solar indices and proxies (e.g. F10.7) which serve as reliable model drivers (Tobiska et al. 2008). The Sun continuously emits particles in the form of solar wind which interact with and disrupt the Earth’s magnetosphere. During eruptive events, such as coronal mass ejections, there are strong interactions between the particles and the magnetosphere which result in energy being deposited into the thermosphere through Joule heating and particle precipitation at high latitudes (Knipp et al. 2004). This energy enhancement causes large increases in density and is difficult to model (Oliveira et al. 2017; Bruinsma et al. 2021). The planetary amplitude index (ap) has 28 discrete values that represent the level of global geomagnetic activity. During geomagnetic storms, the storm-time disturbance index (Dst) can be used to improve density modeling (Bowman et al. 2008). While space weather indices and proxies are useful as model drivers, their relationship with density is imperfect, as is our ability to forecast these drivers (Licata et al. 2020b).

These modeling limitations put a stress on Space Domain Awareness (SDA) which shifts the focus from catalog maintenance to threat assessment. Uncertainties in density modeling can result in large discrepancies between expected and observed satellite positions (Bussy-Virat et al. 2018; Licata et al. 2020a, 2021a). A major improvement in density modeling capabilities came with the introduction of real-time data assimilation. HASDM (Storz et al. 2005) is an assimilative model that leverages an empirical model (JB2008) and Dynamic Calibration of the Atmosphere (DCA) to correct the density nowcast with satellite observations. The updated nowcast can then be propagated forward in time. This model/technique represents the current state-of-the-art in our ability to accurately predict atmospheric density. HASDM outputs had not been publicly available until the recent release of the SET HASDM density database (Tobiska et al. 2021).

The continued commercial satellite launches are creating a necessity for managing the increasing traffic in orbit. A model that balances prediction accuracy and robustness of uncertainty estimates will be a critical tool for the rising importance of Space Traffic Management (STM) (Muelhaupt et al. 2019). In this work, we investigate different modeling techniques using machine learning (ML) to create a surrogate for the database. This allows us to extrapolate beyond the limits of the existing database. A straightforward way to accomplish this is by creating a nonlinear regression model on the full density grid. However, this creates limitations in the context of uncertainty quantification (UQ). To combat this, we focus on the development of nonlinear reduced order models (ROMs). For dimensionality reduction, a linear technique called principal component analysis (PCA) is utilized. The predictive models are standard feed-forward neural networks, or artificial neural networks (ANNs), that employ nonlinear activation functions. We explore various custom loss functions in training in order to obtain reliable or calibrated uncertainty estimates, crucial for collision avoidance operations.

The paper is organized as follows. We first explain the data and the various techniques used for model development. Then, we show the performance and UQ capabilities using a standard mean square error (MSE) loss function. We improve the UQ capabilities with the introduction of the negative logarithm of predictive density (NLPD) and the continuous ranked probability score (CRPS) loss functions. The best HASDM-ML model, in terms of accuracy and reliability, is then compared to the HASDM database in the density space.

2 Methodology

2.1 Data

HASDM is an assimilative framework using the Jacchia-Bowman 2008 Empirical Thermospheric Density Model (JB2008) as a background density model. HASDM improves upon the density correction work of Marcos et al. (1998) and Nazerenko et al. (1998) to modify 13 global temperature correction coefficients with its DCA algorithm. HASDM uses observations of more than 70 carefully chosen calibration satellites to estimate local density values. The satellite orbits span an altitude range of 190-900 km although a majority are between 300 and 600 km (Bowman and Storz 2003). The HASDM algorithm uses a prediction filter that employs wavelet and Fourier analysis for the correction coefficients (Storz et al. 2005). Another highlight of HASDM’s novel framework is its segmented solution for the ballistic coefficient (SSB). This allows the ballistic coefficient estimate to deviate over the fitting period for the satellite trajectory estimation.

SET validates the HASDM output each week and archives the results. The archived values from 2000-2020 make up the SET HASDM density database, upon which this work is based. The database contains density predictions with 15 longitude, 10 latitude, and 25 km altitude increments spanning from 175 - 825 km. This results in 12,312 grid points for every three hours from the start of 2000 to the end of 2019. For further details on HASDM, the reader is referred to Storz et al. (2005), and for details on SET’s validation process and on the database, the reader is referred to Tobiska et al. (2021).

JB2008, and subsequently HASDM, use solar and geomagnetic indices/proxies to drive them. F10.7 (referred to as F10 in this work) is a legacy solar proxy used widely in thermospheric density models. It represents 10.7 cm solar radio flux which does not directly interact with the thermosphere, but it is a reliable proxy for solar EUV heating. The S10 index is a measure of the integrated 26-34 nm solar EUV emission which is absorbed by atomic oxygen in the middle thermosphere. M10 is a proxy that represents the Mg II core-to-wing ratio and is a surrogate for far ultraviolet photospheric 160-nm Schumann-Runge Continuum emissions that cause molecular oxygen dissociation in the lower thermosphere. The last solar index used by JB2008 is Y10 which is a hybrid measure of solar coronal X-ray emissions and Lyman-alpha. S10, M10, and Y10 have no relation to the 10.7 cm wavelength but are converted to solar flux units (sfu) through linear regression with F10. The 81-day solar centered averages of these four solar drivers are used as inputs to JB2008 and are denoted by the ”81c” subscript.

For temperature equations corresponding to geomagnetic activity, JB2008 utilizes ap and Dst. The 3-hour ap is indicative of overall geomagnetic activity and is only measured at low to mid-latitudes. It has 28 discrete values which are capped at 400 2nT leading to further limitations. To combat this, JB2008 uses Dst during geomagnetic storms if the minimum Dst ¡ -75 nT, and it is the primary parameter for energy being deposited into the thermosphere during storms. The data (drivers and density) is split into training, validation, and test data using 60%, 20%, and 20%, respectively as shown in Figure 1. Table 1 shows the number of time steps in the SET HASDM density database across various space weather conditions. The cutoff values for F10 and ap are obtained from Licata et al. (2020b) where they were used to bin space weather driver forecasts.

Refer to caption
Figure 1: F10 and ap at available data points shaded to show the training, validation, and test splits.
Table 1: Number of time steps for different space weather conditions across the SET HASDM density database.
𝐅𝟏𝟎𝟕𝟓\mathbf{F_{10}\leq 75} 𝟕𝟓<𝐅𝟏𝟎𝟏𝟓𝟎\mathbf{75<F_{10}\leq 150} 𝟏𝟓𝟎<𝐅𝟏𝟎𝟏𝟗𝟎\mathbf{150<F_{10}\leq 190} 𝐅𝟏𝟎>𝟏𝟗𝟎\mathbf{F_{10}>190} All 𝐅𝟏𝟎\mathbf{F_{10}}
𝐚𝐩𝟏𝟎\mathbf{a_{p}\leq 10} 13,839 22,034 4,126 2,088 42,087
𝟏𝟎<𝐚𝐩𝟓𝟎\mathbf{10<a_{p}\leq 50} 3,003 9,226 1,982 1,091 15,302
𝐚𝐩>𝟓𝟎\mathbf{a_{p}>50} 54 652 196 149 1,051
All 𝐚𝐩\mathbf{a_{p}} 16,896 31,912 6,304 3,328 58,440

In Table 1, there is clear under-representation of geomagnetic storms in this vast dataset. This can cause limitations in model development, because over 98% of the dataset corresponds to ap \leq 50. Hierarchical modeling could be used for data of this nature, but we proceed with the development of a single comprehensive model.

2.2 Principal Component Analysis

Principal Component Analysis is an eigendecomposition technique that determines uncorrelated linear combinations of the data that maximize variance (F.R.S. 1901; Hotelling 1933). PCA has been previously used to analyze atmospheric density models and accelerometer-derived density sets. Researchers applied PCA to CHAllenging Minisatellite Payload (CHAMP) and Gravity Recovery and Climate Experiment (GRACE) accelerometer-derived density data in order to analyze the dominant modes of variation and identify phenomona encountered by the satellites (Matsuo and Forbes 2010; Lei et al. 2012; Calabia and Jin 2016). PCA has also been leveraged for dimensionality reduction to create linear dynamic reduced order models (Mehta and Linares 2017; Mehta et al. 2018; Mehta and Linares 2020; Gondelach and Linares 2020). Licata et al. (2021b) recently applied PCA to the SET HASDM density database for investigation.

The SET HASDM dataset is a prime candidate for PCA with the goal of predictive modeling due to its high dimensionality. The three spatial dimensions are first flattened to make the dataset two-dimensional (time ×\times space). Then a common logarithm (log10) of the density values is taken in order to reduce the variance of the dataset. Next, we subtract the temporal mean for each cell to center the data and save it for later use. Finally, we perform PCA using the svds function in MATLAB to acquire the UU, Σ\Sigma, and VV matrices (shown in Equation 2). PCA decomposes the data and separates spatial and temporal variations as in Equation 1,

𝐱(𝐬,t)=𝐱¯(𝐬)+𝐱~(𝐬,t)and𝐱~(𝐬,t)=i=1rαi(t)Ui(s)\begin{split}\mathbf{x}\left(\mathbf{s},t\right)=\mathbf{\bar{x}}\left(\mathbf{s}\right)+\mathbf{\widetilde{x}}\left(\mathbf{s},t\right)\;\;\;\textrm{and}\;\;\;\mathbf{\widetilde{x}}\left(\mathbf{s},t\right)=\sum^{r}_{i=1}\alpha_{i}\left(t\right)U_{i}\left(s\right)\end{split} (1)

where 𝐱n\mathbf{x}\in\mathbb{R}^{n} is the log-transformed model output state (HASDM density on its full 3D grid), 𝐱¯\mathbf{\bar{x}} is the mean, 𝐱~\mathbf{\widetilde{x}} is the deviation about the mean, 𝐬\mathbf{s} represents the spatial dimension, rr is the choice of order truncation (here rr=10), αi(t)\alpha_{i}(t) are temporal coefficients, and UiU_{i} are orthogonal modes, or basis functions. Equation 2 shows how to derive these modes.

𝐗=[𝐱~𝟏𝐱~𝟐𝐱~𝟑𝐱~𝐦]and𝐗=UΣVT\begin{split}\mathbf{X}=\left[\begin{array}[]{ccccc}\vrule&\vrule&\vrule&&\vrule\\ \mathbf{\widetilde{x}_{1}}&\mathbf{\widetilde{x}_{2}}&\mathbf{\widetilde{x}_{3}}&\ldots&\mathbf{\widetilde{x}_{m}}\\ \vrule&\vrule&\vrule&&\vrule\end{array}\right]\;\;\;\textrm{and}\;\;\;\mathbf{X}=U\Sigma V^{T}\end{split} (2)

In Equation 2, mm represents the ensemble size (two solar cycles with HASDM) and 𝐗\mathbf{X} represents the log-transformed, mean-centered data . The left unitary matrix, UU, is made of orthogonal vectors that represent the modes of variation, and Σ\Sigma is a diagonal matrix consisting of the squares of the eigenvalues that correspond to the vectors in UU. We encode the data (𝐗\mathbf{X}) into temporal coefficients (αi\alpha_{i}) by performing matrix multiplication between Σ\Sigma and VTV^{T}. To decode back to the density, the coefficients are multiplied by UU with the temporal mean added at each cell, and taking the antilogarithm (10y10^{y}) of the resulting value completes the back-transformation. Figure 2 shows the first 10 PCA coefficients generated using the methods described above. The first five are discussed in detail by Licata et al. (2021b).

Refer to caption
Figure 2: First 10 PCA coefficients (αi\alpha_{i}) for the HASDM database along with F10 and ap.

2.3 Hyperparameter Tuning

In machine learning, developing an accurate model previously required a manual architecture selection process that did not guarantee optimal performance for a given application. Recent developments in hyperparameter tuning make this process much quicker and more thorough. In this work, we use KerasTuner which allows the user to define the ranges and choices of any hyperparameter and to choose a search algorithm (O’Malley et al. 2019). We train all models with 100 trials (or architectures) with the first 25 being randomly sampled from our hyperparameter space while the last 75 trials take advantage of the tuner’s Bayesian optimization scheme.

The tuner trained three identical models for each trial and returned the lowest validation loss after 100 training iterations, or epochs. The Bayesian optimization scheme chooses future architectures based on the validation losses resulting from previous trials. Once completed, the tuner returns the best 10 models which we can evaluate on all three subsets of our data to find the most accurate and generalized model.

2.4 Regression Modeling

A traditional approach to regression modeling with UQ is to develop Gaussian process (GP) models (Rasmussen 2004; Chandorkar et al. 2017). In the early stages of model development, we attempted this approach – training GP regression models on each of the PCA coefficients. However, we could only use 10 years of data for training, limited by computational resources, which resulted in higher predictive error. In addition, the ten resulting models were 6.83 GB each making subsequent work cumbersome. Therefore, the results only pertain to the feedforward neural networks with MC dropout. There are three methods we explore in developing the best HASDM-ML model. First, we create a ROM using PCA for dimensionality reduction and a nonlinear ANN for prediction. We then modify this technique using a custom loss function in an attempt to obtain a model with calibrated uncertainty estimates. We test two loss functions to achieve this: NLPD and CRPS; details of which are given in the following section.

ANNs have two key components: features (or inputs) and labels (or outputs). We try using three separate input sets for our regression models, the first two are explained in Table 2. The first set is the JB2008 input set, referred to as JB. Licata et al. (2021b) showed that the SET HASDM density database contained evidence of post-storm cooling mechanisms which cannot be captured solely by geomagnetic indices at epoch. Therefore, we introduce a second set that is similar to the first but with a time history of the geomagnetic indices. We refer to this as JBH (Historical JB2008). Unlike the actual JB2008 inputs, all of our input sets contain sinusoidal transformations to the day of year (doy) and universal time (UT) inputs. The four resulting temporal inputs (shown in Equation 3) represent annual (t1 and t2) and diurnal (t3 and t4) variations. The use of cyclical functions of time as inputs has been previously demonstrated (Weimer et al. 2020; Licata et al. 2021c).

t1=sin(2πdoy365.25),t2=cos(2πdoy365.25),t3=sin(2πUT24),t4=cos(2πUT24).\begin{split}t_{1}=sin\left(\frac{2\pi doy}{365.25}\right),\;\;\;\;t_{2}=cos\left(\frac{2\pi doy}{365.25}\right),\;\;\;\;t_{3}=sin\left(\frac{2\pi UT}{24}\right),\;\;\;\;t_{4}=cos\left(\frac{2\pi UT}{24}\right).\end{split} (3)
Table 2: List of inputs in the two sets used for model development.
JB2008 Inputs Historical JB2008 Inputs
Solar Geomagnetic Temporal Solar Geomagnetic Temporal
F10F_{10}, S10S_{10}, apa_{p}, Dst t1t_{1}, t2t_{2}, F10F_{10}, S10S_{10}, apAa_{pA}, apa_{p}, ap3a_{p3}, t1t_{1}, t2t_{2},
M10M_{10}, Y10Y_{10}, t3t_{3}, t4t_{4} M10M_{10}, Y10Y_{10}, ap6a_{p6}, ap9a_{p9}, ap1233a_{p12-33}, t3t_{3}, t4t_{4}
F81cF_{81c}, S81cS_{81c}, F81cF_{81c}, S81cS_{81c}, ap3657a_{p36-57}, DstA, Dst,
M81cM_{81c}, Y81cY_{81c} M81cM_{81c}, Y81cY_{81c} Dst3, Dst6, Dst9,
Dst12, Dst15, Dst18, Dst21

In the JBH set, the geomagnetic indices are extensive in an effort to improve storm-time and post-storm modeling. The ”A” subscript refers to the daily average, and the numerical subscripts refer to the value of the index that many hours prior to the epoch. The combination of two numbers references the number of previous hours the index is being averaged over (e.g. ap12-33 refers to the average ap value between 12 and 33 hours prior to the prediction epoch). The authors found that using different time histories of ap and Dst (shown in Table 2) resulted in more optimal performance. For completeness, the results will also be shown using an input set that adopts the same time history for Dst as the ap time history in Table 2. This input set will be referred to as JBH0.

As the geomagnetic inputs in the JBH and JBH0 sets are in a time series format, we tested the time-delay neural network (TDNN) approach. TDNNs use convolutional layers to absorb the information of the time history of a given input and make the model time invariant (Waibel et al. 1989). We ran tuners for the two historical input sets using the MSE loss function and did not see error reductions worthy of further investigation. Additionally, the architecture required for a TDNN adds significant complexity when implementing the NLPD and CRPS loss functions explained in Section 2.5.

2.5 Uncertainty Quantification

Dropout is a regularization tool often used in machine learning (ML) to prevent the model from overfitting to the training data (Srivastava et al. 2014). In standard feed-forward neural networks, each layer sends outputs from all nodes to those in the subsequent layer where they are introduced to weights and biases. Deep neural networks can have millions of parameters and thus are prone to overfitting. This causes undesired performance when interpolating or extrapolating.

Dropout layers use Bernoulli distributions, one for each node, with probability P. This makes the model probabilistic since the distributions are sampled each time that a set of inputs are given to the model. If a sample is ”true”, the node’s output is nullified, and the output of the layer is scaled based on the number of nullified outputs. Dropout is believed to make each node independently sufficient and not reliant on the outputs of other nodes in the layer (Alake 2020).

In traditional use, dropout layers are only activated during training to uphold the deterministic nature of the model. However, measures can be taken in order for this feature to remain activated during prediction making the model probabilistic. When passing the same input set to the model a significant number of times (e.g. 1,000), there is a distribution of model outputs for each unique input. This process is referred to as Monte Carlo (MC) dropout. Essentially, every time the model is presented with a set of inputs, random nodes are dropped out providing a different functional representation of the model. Gal and Ghahramani (2016) show that MC dropout is a Bayesian approximation of a Gaussian process.

In this case, the model uses the input set to predict the ten (10) PCA coefficients. Through the bootstrapping method, we estimate the mean and variance for each of the coefficients (Efron 1979). As the variance is different for each output/coefficient, this is a heteroskedastic problem. Each unique input can be passed to the model k times and there will be k ×\times 10 outputs. The mean and variance are computed from those outputs with respect to the repeated axis, k. The two loss functions used to improve uncertainty estimation (in addition to MSE) are NLPD and CRPS. NLPD is based on the logarithm of the probability density function (pdf) of the Gaussian distribution, and is shown in Equation 4 (Matheson and Winkler 1976; Camporeale and Care 2021).

NLPD(y,μ,σ)=(yμ)22σ2+log(σ2)2+log(2π)2\begin{split}NLPD(y,\mu,\sigma)=\frac{\left(y-\mu\right)^{2}}{2\sigma^{2}}+\frac{log(\sigma^{2})}{2}+\frac{log(2\pi)}{2}\end{split} (4)

In Equation 4, y is the ground truth, μ\mu is the mean prediction, and σ\sigma is the predictive standard deviation, each being computed for all 10 outputs making it a heteroskedastic loss function. For clarity, the log used in the NLPD loss function is the natural logarithm. The second loss function for uncertainty estimation is CRPS which is shown analytically in Equation 5 (Gneiting et al. 2005). The main difference between NLPD and CRPS is that CRPS is also includes the cumulative distribution function (cdf) of the Gaussian distribution as opposed to only the pdf,

CRPS(y,μ,σ)=σ[yμσ erf(yμ2σ)+2π exp((yμ)22σ2)1π].\begin{split}CRPS(y,\mu,\sigma)=\sigma\left[\frac{y-\mu}{\sigma}\textrm{ erf}\left(\frac{y-\mu}{\sqrt{2}\sigma}\right)+\sqrt{\frac{2}{\pi}}\textrm{ exp}\left(-\frac{\left(y-\mu\right)^{2}}{2\sigma^{2}}\right)-\frac{1}{\sqrt{\pi}}\right].\end{split} (5)

As with Equation 4, μ\mu and σ\sigma are computed using bootstrap methods, and this loss function is also heteroskedastic. An important aspect of using the loss functions described in Equations 4 and 5 is the preparation of the training data. The data is traditionally in the following form. The n features are set up as the number of samples, with nI denoting the number of inputs, resulting in the input shape dimension as (n ×\times nI). The n labels are the number of samples with no the number of outputs, resulting in the output shape dimension as (n ×\times no). To implement these loss functions, we stack each input and output by the number of MC samples, k. This is a repeated axis, meaning all samples along this axis are identical. The resulting shapes of the features and labels are (n ×\times k ×\times nI) and (n ×\times k ×\times no), respectively. This allows us to determine the mean and standard deviation for each unique input within the loss function.

2.5.1 Latent Space UQ

Since there are multiple models and loss functions to compare, we have to implement a metric to judge each model’s ability to provide reliable uncertainty estimates. To do so, we modified a calibration error equation from Anderson et al. (2020), shown as

Calibration Error=i=1ri=jm|p(zi,j)p(z^i,j)|.\begin{split}\textrm{Calibration Error}=\sum^{r}_{i=1}\sum^{m}_{i=j}\Big{|}p(z_{i,j})-p(\hat{z}_{i,j})\Big{|}.\end{split} (6)

In Equation 6, m is the number of prediction intervals investigated, r is the choice order of truncation for PCA (the number of model outputs), p(z) is the expected cumulative probability, and p(z^\hat{z}) is the observed cumulative probability. The prediction intervals of interest in this work span from 5% to 95% with 5% increments appended with 99%. p(z^\hat{z}) is computed empirically shown in Equation 7,

p(z^)=1ni=1n𝕀(z^il<zi<z^iu)\begin{split}p\left(\hat{z}\right)=\frac{1}{n}\sum^{n}_{i=1}\mathbb{I}\left(\hat{z}_{i}^{l}<z_{i}<\hat{z}_{i}^{u}\right)\end{split} (7)

where n is the number of samples, 𝕀\mathbb{I} is the indicator function, z^il\hat{z}_{i}^{l}, is the lower bound of the prediction interval, z^iu\hat{z}_{i}^{u} is the upper bound of the prediction interval, and ziz_{i} is the sample. To compute the bounds, we use the pair of equations given in Equation 8,

z^il=μ𝐳σandz^iu=μ+𝐳σ,\begin{split}\hat{z}_{i}^{l}=\mu-\mathbf{z}\sigma\;\;\;\textrm{and}\;\;\;\hat{z}_{i}^{u}=\mu+\mathbf{z}\sigma,\end{split} (8)

where z is the critical value used for the prediction interval. This is calculated using the Gaussian CDF

𝐳=2 erf1(PI)\begin{split}\mathbf{z}=\sqrt{2}\textrm{ erf}^{-1}\left(PI\right)\end{split} (9)

where PI is the prediction interval of interest (e.g. 95% or 0.95).

2.5.2 Density UQ

While latent space calibration is important because the model is trained on the PCA coefficients, determining the reliability of the model’s predictions on the resulting density is the ultimate goal. To examine this, we look at the orbits of CHAMP and GRACE (Luhr et al. 2002; Bettadpur 2012). These satellites had onboard accelerometers from which mass density has been estimated (Bruinsma and Biancale 2003; Liu et al. 2005; Sutton 2008; Doornbos 2012; Calabia and Jin 2016; Mehta et al. 2017). Licata et al. (2021b) recently compared the SET HASDM density database to both CHAMP and GRACE-A density estimates over the lifetime of their measurements.

We evaluate the model 1,000 times every 3 hours across the entire availability of CHAMP and GRACE measurements from Mehta et al (2017) and interpolate them to the satellite locations. This allows us to compute the observed cumulative probability of the HASDM-ML model relative to the HASDM database. Due to the redundancy and computational expense, we only interpolate the model and database density every 500 samples (5,000 and 2,500 seconds for CHAMP and GRACE-A, respectively). This provides a wide view of the model’s UQ capabilities. The CHAMP comparison uses 23,795 model prediction epochs (40.7% of the total available data) with the density being interpolated to at least two satellite positions per prediction due to the cadence of this study. The GRACE comparison uses 24,602 model prediction epochs (42.1% of the total available data) with the density being interpolated to at least four satellite positions per prediction.

Geomagnetic storms are particularly difficult conditions to model accurately. Therefore, we look at four (4) geomagnetic storms from 2002-2004 where we can evaluate HASDM-ML’s reliability across unique events. Two of the events take place in 2002, which is outside the training dataset, while the other two are from 2003 and 2004 and are seen in training. To conduct this assessment, we employ the methodology from the Licata et al. (2021b) storm-time case study. The model is evaluated over a six (6) day period encompassing a storm then interpolated to the CHAMP locations (10 second cadence). During each three-hour prediction period, the density grids remain constant. All 1,000 HASDM-ML density variations are then averaged across each orbit using the mean altitude to determine the orbital period. The mean and 95% predition interval bounds are computed to compare to the corresponding HASDM densities and shown in the figure in an orbit-averaged form. In total, the six days amounts to 48 model prediction epochs which results in 51,840 interpolated densities (1,000 MC runs) from which we compute the observed cumulative probabilities.

3 Results

Upon running each input set with all three loss functions through individual tuners, the best ten models from each tuner are evaluated on the entire training, validation and test sets 1,000 times. The mean absolute error for each subset is computed based on the mean prediction transformed back to the density space. The calibration error score (Equation 6) is computed using the mean and standard deviation from the 1,000 runs in the latent space. The mean absolute error results from the best of the ten models for each input set/loss function are shown in Table 3.

Table 3: Mean absolute for the best model from each technique across training, validation, and test data.
Training Validation Test
Technique JB JBH JBH0 JB JBH JBH0 JB JBH JBH0
MSE 10.38% 8.73% 8.47% 12.00% 10.48% 9.91% 11.95% 10.71% 10.51%
NLPD 10.07% 9.07% 8.81% 11.93% 10.69% 9.87% 11.41% 10.69% 10.05%
CRPS 9.67% 8.64% 8.26% 11.56% 10.55% 9.69% 11.76% 10.43% 10.69%

The addition of historical geomagnetic indices clearly improves the model performance with error reductions upwards of 2%. As mentioned in Section 2.4, the motivation for using the time series geomagnetic indices was to improve storm-time and post-storm performance. However, Table 1 shows that these conditions account for a small subset of the data meaning the notable performance improvement with the JBH and JBH0 input sets show that it likely improves the predictions across all conditions. In general, the CRPS models have the lowest error, and the JBH0 models have the lowest error in terms of the input sets. Table 4 shows the calibration error score for the same models.

Table 4: Calibration error score for the best model from each technique across training, validation, and test data.
Training Validation Test
Technique JB JBH JBH0 JB JBH JBH0 JB JBH JBH0
MSE 3.871 3.744 3.758 3.962 3.898 3.901 4.004 3.979 3.972
NLPD 0.3404 0.3061 0.2794 0.3084 0.2506 0.2836 0.2212 0.1757 0.2790
CRPS 0.3292 0.7933 0.4464 0.2270 0.4634 0.2735 0.2399 0.2387 0.2950

The incorporation of the custom loss functions reduce the calibration error score by an order of magnitude relative to models trained with MSE, which tend to underestimate the uncertainty. The best performing loss function, in regards to calibration, is NLPD. To choose the best overall model, we focus on the test performance as that data is completely independent from the training process. Furthermore, we value the most calibrated model as reliable uncertainty estimates are the most valuable asset for a thermospheric density model. The JBH + NLPD model is within 1% of the error of all better-performing models, and it has the lowest test calibration error score with satisfactory values for the training and validation data. As the calibration error score is a composite of the scores from each coefficient, we show the calibration curves of all coefficients on the independent test set for the best JBH + NLPD model, in panel (b), alongside the best JBH + MSE model, in panel (a), for comparison in Figure 3.

Refer to caption
Figure 3: Expected vs observed cumulative probability (Pr.) of all 10 PCA coefficients for HASDM-ML on the test set using JBH + MSE (a) and JBH + NLPD (b).

The calibration curve in panel (b) for all PCA coefficients roughly follows the perfectly-calibrated 45 line with α5\alpha_{5} being the only coefficient that prominently underestimates uncertainty. However, there is minimal contribution to the full-state (density) after the first few coefficients, so this should not greatly impact the resulting density. In sharp contrast, panel (a) shows the model trained with the MSE loss function is not nearly calibrated, as is evident in Table 4. There is a strong underestimation of the uncertainty at all cumulative probability levels, because the model is not trained with any terms for its variance.

The JBH + NLPD model shown in Figure 3 will be the focus of all subsequent analyses and will be referred to as HASDM-ML. To investigate the model’s reliability on density in an operational nature, we look at the orbits of CHAMP and GRACE-A each over eight year periods with a cumulative altitude range of 300 - 530 km. HASDM-ML was evaluated in three-hour increments from 2002-2011, and was interpolated to the satellite positions. The results for the CHAMP orbit are displayed in Figure 4.

Refer to caption
Figure 4: (a) shows the HASDM and mean HASDM-ML density interpolated to CHAMP positions, (b) shows the expected vs observed calibration curve, (c) shows F10 and ap for the period corresponding to (a), and (d) shows the difference between expected and observed cumulative probability corresponding to (b). Discontinuities in (a) and (c) represent data gaps.

Panel (a) shows a clear similarity in the HASDM and HASDM-ML mean densities. The ML model’s density formulation does not appear to contain any major discrepancies. The surrogate ML model is imperfect in its mean prediction, as seen in Table 3, but panels (b) and (d) show that the density uncertainty is reliable. The calibration curve is exceptional with the observed cumulative probability being within 1% of the expected value for all 20 cumulative probability levels tested. Figure 5 shows the same analysis along the GRACE-A orbit.

Refer to caption
Figure 5: (a) shows the HASDM and mean HASDM-ML density interpolated to GRACE-A positions, (b) shows the expected vs observed calibration curve, (c) shows F10 and ap for the period corresponding to (a), and (d) shows the difference between expected and observed cumulative probability corresponding to (b). Discontinuities in (a) and (c) represent data gaps.

Again, panel (a) shows that HASDM and HASDM-ML mean densities are quite similar. Panels (b) and (d) also demonstrate that although the densities are not identical, HASDM-ML provides uncertainty estimates that are reliable. Panel (d) reveals that at the higher GRACE altitudes, there is slightly less agreement with the expected and observed cumulative probabilities with the largest discrepancy being just over 1%. To perform the simulations displayed in Figures 4 and 5, the model had to be evaluated 23,795,000 and 24,602,000 times for CHAMP and GRACE, respectively. These numbers come from the number of HASDM prediction epochs (Section 2.5.2) and the number of MC runs (1,000). HASDM-ML can perform these predictions in 17.27 seconds for CHAMP and 17.54 seconds for GRACE on a laptop with a NVIDIA GeForce GTX 1070 Mobile graphics card. Using CPU, the model takes 143 seconds for CHAMP and 152 seconds for GRACE. Figure 6 shows HASDM and HASDM-ML orbit-averaged densities during four geomagnetic storms with prediction intervals and the associated calibration curves.

Refer to caption
Figure 6: Panels (a), (b), (e), and (f) show HASDM and HASDM-ML mean orbit-averaged density for CHAMP’s orbit across various geomagnetic storms. The shaded region represents the 95% prediction interval for HASDM-ML, and -Dst is shown on the right axis in each panel. Panels (c), (d), (g), and (h) show the calibration curves corresponding to panels (a), (b), (e), and (f), respectively.

Across all of the storms investigated, the mean prediction of HASDM-ML follows the trend of HASDM density. Even when the model deviates, HASDM densities are mostly captured by the uncertainty bounds. Panels (a) and (b) represent storms not contained in the training dataset which show that HASDM-ML is well-generalized, even during these highly nonlinear events. The calibration curves corresponding to each event show the robust nature of HASDM-ML’s uncertainty estimates. While the observed cumulative probability values deviated from the expected values, these are highly nonlinear periods where density models tend to be unreliable.

3.1 HASDM-ML Performance Metrics

To assess the conditions in which HASDM-ML can improve, the global mean absolute errors are computed as a function of space weather conditions. The results are shown in Table 5 and the number of samples in each bin can be found in Table 1.

Table 5: Mean absolute error across global grid for HASDM-ML as a function of space weather conditions.
𝐅𝟏𝟎𝟕𝟓\mathbf{F_{10}\leq 75} 𝟕𝟓<𝐅𝟏𝟎𝟏𝟓𝟎\mathbf{75<F_{10}\leq 150} 𝟏𝟓𝟎<𝐅𝟏𝟎𝟏𝟗𝟎\mathbf{150<F_{10}\leq 190} 𝐅𝟏𝟎>𝟏𝟗𝟎\mathbf{F_{10}>190} All 𝐅𝟏𝟎\mathbf{F_{10}}
𝐚𝐩𝟏𝟎\mathbf{a_{p}\leq 10} 8.96% 9.78% 9.97% 9.14% 9.50%
𝟏𝟎<𝐚𝐩𝟓𝟎\mathbf{10<a_{p}\leq 50} 9.76% 10.05% 10.87% 9.90% 10.09%
𝐚𝐩>𝟓𝟎\mathbf{a_{p}>50} 15.35% 12.86% 13.23% 12.55% 13.01%
All 𝐚𝐩\mathbf{a_{p}} 9.12% 9.92% 10.36% 9.55% 9.71%

The results from Table 5 show that HASDM-ML is robust to different F10 and ap ranges when ap \leq 50. The only conditions in which the mean absolute error exceeds 11% is when ap ¿ 50, which only accounts for 1.80% of the samples. This shows that more samples are required for this specific condition in both the training and evaluation phases. The last row contains the errors only as a function of F10 which shows that across all four solar activity levels, the error deviates by only 1.24%. The bottom-right cell shows that the error across all 20 years of available data is only 9.71%.

4 Summary

In this work, we developed a surrogate ML model for the SET HASDM density database with robust and reliable uncertainty estimation capabilities. Principal component analysis was selected for dimensionality reduction due to its versatile nature. A Bayesian search algorithm was leveraged to identify an optimal architecture for each input set and loss function tested. We found that of the nine input-loss function combinations explored, the combination of a JB2008 input set with historical geomagnetic drivers and the heteroskedastic NLPD loss function resulted in the most comprehensive model. This model, HASDM-ML, has 9.07% error across the 12 year training set and an average 10.69% error over the combined 8 year validation/test sets. It also had the lowest calibration error score on the test set, meaning it was the most calibrated to independent data. We also compared its calibration curves for each output across the test set to that of the MSE model with the same inputs. This showed that the MSE model considerably underestimated the uncertainty while the NLPD model was well-calibrated across the 10 outputs.

Upon selecting HASDM-ML, we evaluated its uncertainty capabilities across the entire orbits of CHAMP and GRACE-A, both using almost half of the time span of the dataset. This assessment showed that the mean prediction at the satellite locations closely matched that of the HASDM dataset. Across all 20 prediction intervals tested over this period, the model provided an observed cumulative probability that never deviated more than 1% of the expected value for CHAMP’s orbit and never deviated more than 1.15% for GRACE-A’s orbit. A separate storm-time evaluation unveiled that across four storms, HASDM-ML provides similar density to HASDM and its uncertainty estimates remained robust and reliable. The results from the density calibration tests are significant, because no thermospheric density model to date provides uncertainty estimates. Additionally, uncertainty estimates themselves are not meaningful unless they are well-calibrated, and HASDM-ML is able to provide that.

5 Limitations and Future Work

To further improve HASDM-ML, additional data can be introduced, particularly in the coming solar maximum. As seen in Section 3.1, periods of ap ¿ 50 were the source of the highest overall error and the introduction of additional storms could reduce that error. Hierarchical modeling is another approach to potentially combat this limitation. A nonlinear dimensionality reduction method could also improve performance, especially in modeling the highly nonlinear storm response. The developed model can be used in research (e.g. to study historical storms where HASDM outputs are unavailable) as well as in an operational setting. Another area of future work could be to develop a nonlinear dynamic ROM based on the SET HASDM density database. While the background model is static, the assimilative framework represents the dynamic thermosphere, and a dynamic ROM could better model certain phenomena and provide a framework for further data assimilation (Mehta and Linares 2018).

6 Data Statement

Requests can be submitted for access to the SET HASDM density database at https://spacewx.com/hasdm/ and all reasonable requests for scientific research will be accepted as explained in the rules of road document on the website. The historical space weather indices used in this study can also be found at https://spacewx.com/jb2008/. Original CHAMP and GRACE density estimates from Mehta et al. (2017) can be found at http://tinyurl.com/densitysets.

7 Acknowledgment

This work was made possible by NASA Established Program to Stimulate Competitive Research, Grant #\#80NSSC19M0054. SET and WVU gratefully acknowledge support from the NASA SBIR contract #\#80NSSC20C0292 for Machine learning Enabled Thermosphere Advanced by HASDM (META-HASDM). The authors would like to thank the anonymous reviewers for all of their time and effort.

References

  • Alake (2020) Alake, R. (2020), Understanding and Implementing Dropout in TensorFlow and Keras, https://towardsdatascience.com/understanding-and-implementing-dropout-in-tensorflow-and-keras-a8a3a02c1bfa.
  • Anderson et al. (2020) Anderson, G. J., J. A. Gaffney, B. K. Spears, P.-T. Bremer, R. Anirudh, and J. J. Thiagarajan (2020), Meaningful uncertainties from deep neural network surrogates of large-scale numerical simulations.
  • Bettadpur (2012) Bettadpur, S. (2012), Gravity Recovery and Climate Experiment: Product Specification Document, GRACE 327-720, CSR-GR-03-02, cent. for Space Res., The Univ. of Texas, Austin, TX.
  • Bowman and Storz (2003) Bowman, B., and M. Storz (2003), High Accuracy Satellite Drag Model (HASDM) Review, in AIAA/AAS Astrodynamics Specialist Conference, AAS 03-625.
  • Bowman et al. (2008) Bowman, B., W. K. Tobiska, F. Marcos, C. Huang, C. Lin, and W. Burke (2008), A New Empirical Thermospheric Density Model JB2008 Using New Solar and Geomagnetic Indices, in AIAA/AAS Astrodynamics Specialist Conference, AIAA 2008-6438.
  • Bruinsma and Biancale (2003) Bruinsma, S., and R. Biancale (2003), Total Densities Derived from Accelerometer Data, Journal of Spacecraft and Rockets, 40(2), 230–236, 10.2514/2.3937.
  • Bruinsma et al. (2021) Bruinsma, S., C. Boniface, E. K. Sutton, and M. Fedrizzi (2021), Thermosphere modeling capabilities assessment: geomagnetic storms, J. Space Weather Space Clim., 11, 12, 10.1051/swsc/2021002.
  • Bussy-Virat et al. (2018) Bussy-Virat, C. D., A. J. Ridley, and J. W. Getchius (2018), Effects of Uncertainties in the Atmospheric Density on the Probability of Collision Between Space Objects, Space Weather, 16(5), 519–537, 10.1029/2017SW001705.
  • Calabia and Jin (2016) Calabia, A., and S. Jin (2016), New modes and mechanisms of thermospheric mass density variations from GRACE accelerometers, Journal of Geophysical Research: Space Physics, 121(11), 11,191–11,212, https://doi.org/10.1002/2016JA022594.
  • Camporeale and Care (2021) Camporeale, E., and A. Care (2021), ACCRUE: Accurate and Reliable Uncertainty Estimate in Deterministic Models, International Journal for Uncertainty Quantification, 11(4), 81–94.
  • Chandorkar et al. (2017) Chandorkar, M., E. Camporeale, and S. Wing (2017), Probabilistic forecasting of the disturbance storm time index: An autoregressive Gaussian process approach, Space Weather, 15(8), 1004–1019, https://doi.org/10.1002/2017SW001627.
  • Doornbos (2012) Doornbos, E. (2012), Producing Density and Crosswind Data from Satellite Dynamics Observations, pp. 91–126, Springer Berlin Heidelberg, Berlin, Heidelberg, 10.1007/978-3-642-25129-0_4.
  • Efron (1979) Efron, B. (1979), Bootstrap Methods: Another Look at the Jackknife, The Annals of Statistics, 7(1), 1–26.
  • Emmert (2015) Emmert, J. (2015), Thermospheric mass density: A review, Advances in Space Research, 56, 10.1016/j.asr.2015.05.038.
  • F.R.S. (1901) F.R.S., K. P. (1901), LIII. On lines and planes of closest fit to systems of points in space, The London, Edinburgh, and Dublin Philosophical Magazine and Journal of Science, 2(11), 559–572, 10.1080/14786440109462720.
  • Gal and Ghahramani (2016) Gal, Y., and Z. Ghahramani (2016), Dropout as a bayesian approximation: Representing model uncertainty in deep learning.
  • Gneiting et al. (2005) Gneiting, T., A. E. Raftery, A. H. Westveld, and T. Goldman (2005), Calibrated Probabilistic Forecasting Using Ensemble Model Output Statistics and Minimum CRPS Estimation, Monthly Weather Review, 133(5), 1098 – 1118, 10.1175/MWR2904.1.
  • Gondelach and Linares (2020) Gondelach, D. J., and R. Linares (2020), Real-Time Thermospheric Density Estimation via Two-Line Element Data Assimilation, Space Weather, 18(2), e2019SW002,356, https://doi.org/10.1029/2019SW002356, e2019SW002356 10.1029/2019SW002356.
  • Hotelling (1933) Hotelling, H. (1933), Analysis of a complex of statistical variables into principal components, Journal of Educational Psychology, 24(6), 417–441, 10.1037/h0071325.
  • Knipp et al. (2004) Knipp, D., W. K. Tobiska, and B. Emery (2004), Direct and Indirect Thermospheric Heating Sources for Solar Cycles 21-23, Solar Physics - SOL PHYS, 224, 495–505, 10.1007/s11207-005-6393-4.
  • Lei et al. (2012) Lei, J., T. Matsuo, X. Dou, E. Sutton, and X. Luan (2012), Annual and semiannual variations of thermospheric density: EOF analysis of CHAMP and GRACE data, Journal of Geophysical Research: Space Physics, 117(A1), https://doi.org/10.1029/2011JA017324.
  • Licata et al. (2020a) Licata, R., P. Mehta, and W. K. Tobiska (2020a), Impact of Space Weather Driver Forecast Uncertainty on Drag and Orbit Prediction, in Proceedings of the 2020 AAS/AIAA Astrodynamics Specialist Conference.
  • Licata et al. (2021a) Licata, R., P. Mehta, and W. K. Tobiska (2021a), Impact of Driver and Model Uncertainty on Drag and Orbit Prediction, in Proceedings of the 31st AAS/AIAA Space Flight Mechanics Meeting.
  • Licata et al. (2020b) Licata, R. J., W. K. Tobiska, and P. M. Mehta (2020b), Benchmarking Forecasting Models for Space Weather Drivers, Space Weather, 18(10), e2020SW002,496, https://doi.org/10.1029/2020SW002496.
  • Licata et al. (2021b) Licata, R. J., P. M. Mehta, W. K. Tobiska, B. R. Bowman, and M. D. Pilinski (2021b), Qualitative and Quantitative Assessment of the SET HASDM Database, Space Weather, 19(8), e2021SW002,798, https://doi.org/10.1029/2021SW002798.
  • Licata et al. (2021c) Licata, R. J., P. M. Mehta, D. R. Weimer, and W. K. Tobiska (2021c), Improved Neutral Density Predictions through Machine Learning Enabled Exospheric Temperature Model, Earth and Space Science Open Archive, p. 16, 10.1002/essoar.10507687.1.
  • Liu et al. (2005) Liu, H., H. Lühr, V. Henize, and W. Köhler (2005), Global distribution of the thermospheric total mass density derived from CHAMP, Journal of Geophysical Research: Space Physics, 110(A4), https://doi.org/10.1029/2004JA010741.
  • Luhr et al. (2002) Luhr, H., L. Grunwaldt, and C. Forste (2002), CHAMP Reference Systems, Transformations and Standards, Tech. rep., CH-GFZ-RS-002, GFZ-Potsdam, Postdam, Germany.
  • Marcos et al. (1998) Marcos, F., M. Kendra, J. Griffin, J. Bass, D. Larson, and J. J. Liu (1998), Precision Low Earth Orbit Determination Using Atmospheric Density Calibration, Journal of The Astronautical Sciences, 46, 395–409.
  • Matheson and Winkler (1976) Matheson, J. E., and R. L. Winkler (1976), Scoring Rules for Continuous Probability Distributions, Management Science, 22(10), 1087–1096.
  • Matsuo and Forbes (2010) Matsuo, T., and J. M. Forbes (2010), Principal modes of thermospheric density variability: Empirical orthogonal function analysis of CHAMP 2001–2008 data, Journal of Geophysical Research: Space Physics, 115(A7), https://doi.org/10.1029/2009JA015109.
  • Mehta and Linares (2017) Mehta, P. M., and R. Linares (2017), A methodology for reduced order modeling and calibration of the upper atmosphere, Space Weather, 15(10), 1270–1287, 10.1002/2017SW001642.
  • Mehta and Linares (2018) Mehta, P. M., and R. Linares (2018), A new transformative framework for data assimilation and calibration of physical ionosphere-thermosphere models, Space Weather, 16(8), 1086–1100, https://doi.org/10.1029/2018SW001875.
  • Mehta and Linares (2020) Mehta, P. M., and R. Linares (2020), Real-Time Thermospheric Density Estimation from Satellite Position Measurements, Journal of Guidance, Control, and Dynamics, 43(9), 1656–1670, 10.2514/1.G004793.
  • Mehta et al. (2017) Mehta, P. M., A. C. Walker, E. K. Sutton, and H. C. Godinez (2017), New density estimates derived using accelerometers on board the CHAMP and GRACE satellites, Space Weather, 15(4), 558–576, https://doi.org/10.1002/2016SW001562.
  • Mehta et al. (2018) Mehta, P. M., R. Linares, and E. K. Sutton (2018), A Quasi-Physical Dynamic Reduced Order Model for Thermospheric Mass Density via Hermitian Space-Dynamic Mode Decomposition, Space Weather, 16(5), 569–588, 10.1029/2018SW001840.
  • Muelhaupt et al. (2019) Muelhaupt, T. J., M. E. Sorge, J. Morin, and R. S. Wilson (2019), Space traffic management in the new space era, Journal of Space Safety Engineering, 6(2), 80–87, https://doi.org/10.1016/j.jsse.2019.05.007.
  • Nazarenko et al. (1998) Nazarenko, A., P. Cefola, and V. Yurasov (1998), Estimating atmospheric density variations to improve LEO orbit prediction accuracy, in AIAA/AAS Space Flight Mechanics Meeting, AAS 98-190.
  • Oliveira et al. (2017) Oliveira, D. M., E. Zesta, P. W. Schuck, and E. K. Sutton (2017), Thermosphere Global Time Response to Geomagnetic Storms Caused by Coronal Mass Ejections, Journal of Geophysical Research: Space Physics, 122(10), 10,762–10,782, https://doi.org/10.1002/2017JA024006.
  • O’Malley et al. (2019) O’Malley, T., E. Bursztein, J. Long, F. Chollet, H. Jin, L. Invernizzi, et al. (2019), Keras Tuner, https://github.com/keras-team/keras-tuner.
  • Radtke et al. (2017) Radtke, J., C. Kebschull, and E. Stoll (2017), Interactions of the space debris environment with mega constellations—Using the example of the OneWeb constellation, Acta Astronautica, 131, 55–68, https://doi.org/10.1016/j.actaastro.2016.11.021.
  • Rasmussen (2004) Rasmussen, C. E. (2004), Gaussian Processes in Machine Learning, pp. 63–71, Springer Berlin Heidelberg, Berlin, Heidelberg, 10.1007/978-3-540-28650-9_4.
  • Srivastava et al. (2014) Srivastava, N., G. Hinton, A. Krizhevsky, I. Sutskever, and R. Salakhutdinov (2014), Dropout: A simple way to prevent neural networks from overfitting, Journal of Machine Learning Research, 15, 1929–1958.
  • Storz et al. (2005) Storz, M., B. Bowman, and J. Branson (2005), High Accuracy Satellite Drag Model (HASDM), in AIAA/AAS Astrodynamics Specialist Conference and Exhibit, 10.2514/6.2002-4886.
  • Sutton (2008) Sutton, E. K. (2008), Effects of solar disturbances on the thermosphere densities and winds from CHAMP and GRACE satellite accelerometer data, Ph.D. thesis, University of Colorado at Boulder.
  • Tobiska et al. (2008) Tobiska, W. K., S. D. Bouwer, and B. R. Bowman (2008), The development of new solar indices for use in thermospheric density modeling, Journal of Atmospheric and Solar-Terrestrial Physics, 70(5), 803–819, https://doi.org/10.1016/j.jastp.2007.11.001.
  • Tobiska et al. (2021) Tobiska, W. K., B. R. Bowman, D. Bouwer, A. Cruz, K. Wahl, M. Pilinski, P. M. Mehta, and R. J. Licata (2021), The SET HASDM density database, Space Weather, p. e2020SW002682, https://doi.org/10.1029/2020SW002682.
  • Waibel et al. (1989) Waibel, A., T. Hanazawa, G. Hinton, K. Shikano, and K. Lang (1989), Phoneme recognition using time-delay neural networks, IEEE Transactions on Acoustics, Speech, and Signal Processing, 37(3), 328–339, 10.1109/29.21701.
  • Weimer et al. (2020) Weimer, D. R., P. M. Mehta, W. K. Tobiska, E. Doornbos, M. G. Mlynczak, D. P. Drob, and J. T. Emmert (2020), Improving Neutral Density Predictions Using Exospheric Temperatures Calculated on a Geodesic, Polyhedral Grid, Space Weather, 18(1), e2019SW002,355, https://doi.org/10.1029/2019SW002355.