ChronoGAN: Supervised and Embedded Generative Adversarial Networks for Time Series Generation
Abstract
Generating time series data using Generative Adversarial Networks (GANs) presents several prevalent challenges, such as slow convergence, information loss in embedding spaces, instability, and performance variability depending on the series length. To tackle these obstacles, we introduce a robust framework aimed at addressing and mitigating these issues effectively. This advanced framework integrates the benefits of an Autoencoder-generated embedding space with the adversarial training dynamics of GANs. This framework benefits from a time series-based loss function and oversight from a supervisory network, both of which capture the stepwise conditional distributions of the data effectively. The generator functions within the latent space, while the discriminator offers essential feedback based on the feature space. Moreover, we introduce an early generation algorithm and an improved neural network architecture to enhance stability and ensure effective generalization across both short and long time series. Through joint training, our framework consistently outperforms existing benchmarks, generating high-quality time series data across a range of real and synthetic datasets with diverse characteristics.
Index Terms:
Time Series Generation, Generative Adversarial Networks, Autoencoders, Data AugmentationI Introduction
Fields such as biomedical signal processing [1] and solar flare prediction [2, 3] often face data shortages due to complex and noisy data environments, scarcity of events, and privacy concerns [4], all of which complicate accurate model training and evaluation. Developing methods that leverage Generative Adversarial Networks (GANs) [5] to produce realistic synthetic data can foster scientific progress. By creating balanced datasets and mitigating data shortages, GANs can improve the performance of machine learning tasks [6].
Generative modeling of time series data poses unique challenges due to the temporal nature of the data. These models must not only capture the distribution of features at individual time points but also unravel the complex dynamics between these points over time. For instance, when managing multivariate sequential data represented as , an effective model should accurately determine the conditional distribution , which dictates the temporal transitions. Without this capability, the generated data fails to capture the characteristics of the real dataset [7]. This leads to misleading and inaccurate evaluations when used alongside real data for downstream machine learning tasks [8].
In the field of time series generation, a substantial body of research has focused on enhancing the temporal dynamics of autoregressive models for sequence forecasting. The primary aim is to reduce the propagation of sampling errors through various training-time adjustments, leading to more precise conditional distribution modeling [9, 10, 11]. Autoregressive models decompose the sequence distribution into a chain of conditionals, , which proves useful for forecasting due to their deterministic nature. However, they lack true generative capabilities, as generating new sequences from them does not require external input. In contrast, research applying GANs to sequential data often employs sequence-to-sequence neural network layers for both the generator and discriminator. This approach pursues a direct adversarial objective [12, 13, 14] to learn the probability distribution of the data and generate new samples by feeding random noise into the model. While straightforward, this adversarial goal focuses on modeling the joint distribution [15] without considering the autoregressive structure. This may be inadequate, as aggregating standard GAN losses across vectors does not necessarily ensure the capture of stepwise dependencies in time series samples.
In this paper, we introduce a novel framework that significantly enhances stability, accuracy, and generalizability. Our approach, termed ChronoGAN, effectively integrates the two research streams into a robust and precise generative model specifically designed to preserve temporal dynamics through supervised GAN training. Additionally, it leverages latent space during training, ensuring more reliable convergence. Therefore, ChronoGAN offers a comprehensive method for generating realistic time-series data applicable across various fields. The key contributions of our study are:
-
1.
Generating data within the latent space using a generator, while utilizing a discriminator that operates in the feature space, offers significant advantages. This method not only provides more precise adversarial feedback to the generator but also delivers crucial adversarial feedback to the autoencoder, enhancing the overall performance of the model.
-
2.
The development of a novel time series-based loss function for the generator network, combined with a supervised loss, enhances the quality of the generated data by more effectively learning the temporal dynamics. Additionally, a new loss function is designed for the autoencoder to improve its reconstruction capabilities.
-
3.
The implementation of an early generation algorithm to stabilize the framework and ensure optimal results after each training session.
-
4.
The implementation of a novel GRU-LSTM architecture across the framework’s five neural networks to enhance the generation of high-quality data for sequences of varying lengths, both short and long.
We demonstrate the advantages of ChronoGAN by conducting a series of experiments on a variety of real-world and synthetic datasets. Our findings indicate that ChronoGAN consistently outperforms existing benchmarks, including TimeGAN [16], in generating realistic time-series data.
II Related Work
Autoregressive recurrent networks trained using maximum likelihood methods are susceptible to significant prediction errors during multi-step sampling [17]. This issue arises from the difference between closed-loop training (conditioned on actual data) and open-loop inference (based on prior predictions). Further, inspired by adversarial domain adaptation [18], Professor Forcing trains an additional discriminator to differentiate between autonomous and teacher-driven hidden states [19], helping to align training and sampling dynamics. However, although these methods share our aim of modeling stepwise transitions, they are deterministic and do not explicitly involve sampling from a learned distribution, which is crucial for our objective of synthetic data generation.
The foundational paper on GANs [5] introduced a novel framework for generating synthetic data. The model consists of two neural networks (the generator and the discriminator) that are trained simultaneously in a zero-sum game setup. However, despite being capable of generating data by sampling from a learned distribution, they struggle to capture the stepwise dependencies inherent in time series data. The adversarial feedback from the discriminator alone is insufficient for the generator to effectively learn the patterns of sequences.
Several studies have adopted the GAN framework for use in time series analysis. The earliest, C-RNN-GAN [12], applied the GAN directly to sequential data with LSTM networks serving as both generator and discriminator. It generates data recurrently, starting with a noise vector and the data from the previous time step. RCGAN [13] modified this by removing the reliance on previous outputs and incorporating additional inputs for conditioning [20]. However, these models depend solely on binary adversarial feedback for learning, which may not capture the temporal dynamics of time series data.
TimeGAN [16] presented a sophisticated method for generating time-series data, combining the versatility of unsupervised learning with the accuracy of supervised training. By optimizing an embedding space through both supervised and adversarial objectives, it aimed to closely mirror the dynamics of time series data. Despite its novel approach, TimeGAN encounters challenges with the quality of the generated data, primarily due to its reliance on adversarial training within the embedding space rather than the feature space. Furthermore, TimeGAN suffers from stability issues, yielding inconsistent outcomes across identical iteration counts and hyperparameter settings. It also faces difficulties in generating both short and long time series sequences.
The ChronoGAN framework is developed to enhance the efficacy and robustness of time series generation by accomplishing several critical objectives. First, it optimizes performance across both short and long sequences. Second, it enhances data reconstruction by the decoder and data generation by the generator through providing more accurate adversarial feedback to both the autoencoder and generator. Third, it facilitates the convergence of both the generator and autoencoder networks through the implementation of novel loss functions. Finally, it incorporates an early generation algorithm to achieve consistent optimal results under the same hyperparameters. Fig. 1 illustrates the implementation of ChronoGAN.
III Proposed Model: ChronoGAN

Based on Fig. 1, the framework includes five networks: an autoencoder (encoder and decoder), a generator, a supervisor, and a discriminator. The autoencoder’s role is to facilitate training by generating compressed representations in the latent space, thereby reducing the likelihood of non-convergence within the GAN framework. The generator produces data in this lower-dimensional latent space, as opposed to the feature space. The supervisor network, integrated with a supervised loss function, is specifically designed to learn the temporal dynamics of the time series data. This is crucial, as sole reliance on the discriminator’s adversarial feedback may not sufficiently prompt the generator to capture the data’s stepwise conditional distributions. The discriminator network differentiates between fake and real data in the feature space, providing more accurate feedback to both the generator and autoencoder.
In Fig. 1, represents the encoding of the input data into a latent space using the encoder function . The reconstructed data is obtained by decoding using the recovery function , aiming to replicate the original input data as closely as possible. The generator function transforms a random noise vector into synthetic latent data , which is then reconstructed into synthetic data . The supervisor network processes to produce a supervised latent representation , from which the final synthetic data is reconstructed. The discriminator evaluates the authenticity of the synthetic and real data by outputting for synthetic data and for real data.
III-A Adversarial Training
In a joint training scheme involving a GAN network and an autoencoder, relying solely on reconstruction loss for the autoencoder results in noisy outputs, where the autoencoder’s output fails to fully retain the input’s characteristics [21]. Additionally, adversarial training within an embedding space leads to the generation of noisy data after decoding the generator’s output. The issue arises when the encoder’s output () is regarded as real data and the generator’s output () as synthetic during the adversarial training process. This practice reduces the discriminator’s ability to accurately differentiate between the attributes of real and synthetic data. A significant limitation is that the discriminator does not account for the error rate and data loss inherent in the autoencoder’s performance. This oversight may compromise the efficacy of the discriminator, resulting in suboptimal performance in distinguishing between real and generated data attributes. Consequently, this leads to less precise feedback being provided to the generator network, potentially affecting the overall quality of the synthetic data. To address this, as shown in Fig. 1, discriminating in the feature space allows for defining real data as the dataset () and fake data as the decoding of the generator’s output (). This facilitates more accurate training for the discriminator, thus yielding improved feedback for the generator. Additionally, discrimination in the feature space provides valuable adversarial feedback to the autoencoder, enhancing its reconstruction capabilities in conjunction with conventional reconstruction loss. In the context of time series data, the feature space denotes the original dimensions, such as individual time points and their observed values. The latent or embedding space, achieved through an encoding process, represents the data in a lower-dimensional form, capturing its essential patterns and structures in a more compact and informative manner [22].
Through a joint learning scheme, the autoencoder is initially trained using a combination of reconstruction loss and binary feedback from the discriminator, where real data is the dataset () and fake data is its reconstruction (). This approach enhances the autoencoder’s precision in reconstructing outputs. In the subsequent phase, only the supervisor network is trained. The supervisor utilizes real data embeddings from the previous two time steps generated by the embedding network to create the subsequent latent vector . Finally, all five networks are trained jointly. During this final phase, the same discriminator distinguishes between real data, denoted as the dataset (), and the dataset reconstructions (), where the fake data comprises the generator’s decoded outputs () and the supervisor’s decoded outputs (). The generator undergoes training through this adversarial feedback , in addition to other feedback mechanisms including , , and . This phase involves a shift in the characterization of fake and real data compared to the initial phase.
III-B Novel Loss Functions
Based on the feedback from the discriminator, we introduce a new loss function for the autoencoder (), which comprises both reconstruction loss () and adversarial loss (). The proportion of reconstruction loss to adversarial loss decreases in the third phase of training compared to the first phase, where the primary purpose of the discriminator is to provide feedback for the generator rather than the autoencoder.
(1) |
Where denotes an individual time step, and represents the total number of time steps within the series. In addition, represents the real data at timestamp , and denotes the output of the autoencoder corresponding to the real data at the same timestamp.
(2) |
(3) |
Here, indicates the probability distribution of real data, and represents the probability distribution of synthetic data. Moreover, the discriminator generates the output when evaluating the autoencoder’s output and produces the output when assessing the real samples
The sole reliance on the discriminator’s binary adversarial feedback might not sufficiently drive the generator to capture the data’s stepwise conditional distributions. To address this, ChronoGAN introduces an additional component, the supervisor, along with a novel loss mechanism denoted by . ChronoGAN employs a closed-loop training mode, where the supervisor utilizes actual data embeddings from the previous two time steps produced by the embedding network to generate the subsequent latent vector . This looped training involves the generator’s loss , which encompasses the adversarial loss , the stepwise transition loss , the distribution loss , and our innovative time series loss . This structure ensures the generation of realistic sequences with accurate temporal transitions. The distribution loss leverages the mean absolute error (MAE) of the mean and variance between the real data and the generated data . This approach effectively assists the generator in learning the real data distribution, enabling it to produce data across the entire distribution, which also serves as a key metric for evaluating GAN techniques.
(4) |
Where is the MAE of the mean between a batch of real and generated samples, and is the MAE of the variance between the same batch of real and generated data.
(5) |
Where each sample is labeled by and the batch is represented as .
(6) |
Where indicates the mean of , and represents the mean of for a batch of data.
(7) |
Where is the supervisor network, is the output of the generator at timestamp , and is the output of the generator at timestamp . This technique is more efficient than predicting timestamp using timestamp .
In the third phase of training, referred to as joint training, represents the output of the discriminator for synthetic samples and , while denotes the output of for real samples and .
(8) |
Furthermore, we introduce a novel loss function for the generator called the time series loss, , which not only facilitates convergence but also enhances the quality of the generated data. This loss function is defined as the mean squared error (MSE) of the mean and standard deviation (std) of four key time series characteristics, including slope, skewness, weighted average, and median, between real and synthetic data. The aim is to boost the generator’s convergence and its ability to learn the real data characteristics and distribution, as relying solely on the adversarial loss is insufficient for learning the characteristics of real time series data. The time series loss is a novel contribution, comprising the slope loss (), weighted average loss (), skewness loss (), and median loss ().
(9) |
The slope loss includes the MSE of the mean () and the MSE of the std () between the slopes of real and generated samples.
(10) |
The slope is calculated using the provided formula,
(11) |
In these equations, is the slope of real samples, and is the slope of generated samples.
(12) |
(13) |
Other components of , such as skewness, weighted average, and median, are calculated similarly to (10), (12), and (III-B). The only difference is that instead of using the formula for slope, the formulas for skewness (skew), weighted average (wAvg), and median are applied.
(14) |
(15) |
Where represents the std of , and denotes the weight assigned to the value at timestamp .
III-C GRU-LSTM Network Architecture
Leveraging the strengths of different neural network architectures by combining them has long been a powerful and effective approach. In auditory attention detection (AAD), combining GRU and CNN architectures has been particularly effective. CNNs, while good at extracting spatial features from EEG data, struggle to capture long-term dependencies. To address this, the AAD-GCQL model [1] integrates GRU with CNN to capture both spatial and temporal dynamics in EEG signals, enhancing the detection of auditory attention.
The GRU used in this combination belongs to a broader family of recurrent neural networks (RNNs), which are tailored for sequence modeling tasks. Among these, LSTM and GRU are the two most prominent architectures, frequently applied in domains such as natural language processing [23] and time series forecasting. LSTMs are equipped with memory cells and three distinct gates (input, output, and forget), which help manage the flow of information and address the vanishing gradient problem seen in traditional RNNs [24]. This architecture makes LSTMs particularly well-suited for longer sequence data, where maintaining information over extended intervals is critical. On the other hand, GRUs simplify the structure by merging the input and forget gates into a single update gate, complemented by a reset gate that determines the extent of past information retention [25]. GRUs tend to be more efficient and quicker to train, making them ideal for tasks with shorter sequences or when computational resources are limited. The decision between using LSTM and GRU often hinges on the specific sequence length and complexity of the task, with LSTMs generally preferred for longer sequences and GRUs for shorter ones [26].
A time series generation framework should be capable of handling both short and long sequences and, more importantly, be accurate on both. The exclusive use of either LSTM or GRU as the network architecture can lead to weaknesses in handling either long or short sequences. As shown in Fig. 2, by implementing both network architectures and merging the results via a multilayer perceptron, the network becomes more generalized, making it more powerful in learning both long and short sequences. We employ multiple layers of GRU and LSTM separately to produce output, and then merge them using a multilayer perceptron network to obtain the final output. We utilize the same architecture and number of layers for all five networks within the ChronoGAN framework.
III-D Early Generation
Another prevalent issue with GANs is stability. To enhance the stability of the network, we employ an early generation algorithm since the optimal results may be achieved after a random, rather than a specific, number of iterations. Accordingly, as per Algorithm 1, after half the number of epochs, we generate synthetic data and calculate the discriminative score and predictive score between real and synthetic data at intervals of every 500 epochs. Additionally, we compute the MSE of the mean and MSE of the std of real and synthetic data to verify whether the synthetic data matches the distribution of the real data. By integrating the results of the discriminative score, predictive score, and MSE of the mean and std, we determine whether to save the current model and generated data. Upon the completion of training, we ensure that the framework has produced the optimal results, consistently delivering reliable and precise outcomes after each training session. It is crucial to determine the appropriate weights for these metrics in order to integrate them and compare them with the previously saved model. The proportion of the discriminative score, predictive score, and MSE of the mean and std can vary depending on the characteristics of the dataset. Therefore, it is inappropriate to establish fixed hyperparameters to combine these three metrics. To address this issue, we initially calculate the hyperparameters and during the first assessment of these metrics. Once established, these hyperparameters are consistently applied in all subsequent epochs.

IV Experimnets
The codebase for the ChronoGAN framework, along with a detailed tutorial on its usage, implementation, and hyperparameter settings, is publicly available for review and application 111The codebase of ChronoGAN is available here: https://github.com/samresume/ChronoGAN. The framework is designed to be straightforward, allowing users to simply call a Python function and provide the necessary data and hyperparameters.
IV-A Datasets
We evaluate ChronoGAN’s effectiveness on time-series datasets with varying attributes such as periodicity, discreteness, noise levels, length, and feature correlation over time. We choose the datasets based on different combinations of these characteristics:
-
1.
Stocks: Stock price sequences are continuous but aperiodic and features are correlated. We use daily historical data from Google stocks spanning 2004 to 2019, which includes features such as volume, high, low, opening, closing, and adjusted closing prices.
-
2.
Sines: We generate multivariate sinusoidal sequences with varying frequencies and phases , providing continuous, periodic, and multivariate data with each feature being independent.
-
3.
ECG: The ECG5000 dataset from Physionet, which covers a 20-hour long ECG recording with 140 timestamps, is a univariate time series that is continuous and periodic. The data is classified as a long time series.
-
4.
SWAN-SF: The Space Weather Analytics for Solar Flares (SWAN-SF) [27] dataset consists of multivariate time series of photospheric magnetic field parameters for solar flare prediction tasks [28]. The SWAN-SF dataset is recognized as challenging due to its complex temporal dynamics and the numerous data preprocessing issues it presents. In [29], the authors thoroughly addressed these challenges by implementing an innovative preprocessing pipeline [30]. This effort resulted in the creation of an enhanced version of the SWAN-SF dataset [31], which was subsequently utilized in our evaluation in place of the original, unprocessed dataset.
IV-B Baseline Techniques and Evaluation Metrics
We conduct a comparison between ChronoGAN, TimeGAN [16], Teacher Forcing (T-Forcing) [19], Professor Forcing (P-Forcing) [18] and Standard GAN [13], which represent the five best-performing techniques in various fields of time series generation, including GAN-based and Autoregressive approaches. To ensure unbiased results, we maintain identical hyperparameters across all five models. To evaluate the quality of the generated data, we focus on three key criteria:
- 1.
-
2.
Discriminative Score: For a quantitative measure of similarity, each sequence from the original dataset is labeled as ‘real‘, while each from the generated set is labeled as ‘synthetic‘. An LSTM classifier is then trained to differentiate these two categories in a standard supervised learning task. The classification error on a reserved test set provides a quantitative measure of this score. We then subtract the result from 0.5, making the optimal result 0 instead of 0.5 for easier comparison.
-
3.
Predictive Score: To evaluate the quality of the generated data in capturing step-wise conditional distributions, we utilize the synthetic dataset to train an LSTM for sequence prediction. This involves forecasting the next-step temporal vectors for each input sequence. The model’s accuracy is subsequently tested on the original dataset, with performance assessed using the MAE.
For each discriminative or predictive score experiment, we replicated the experiments eight times to avoid incidental results. We present the mean and std of each experiment in Tables I and II.
Stocks | Sines | ECG | SWAN-SF | |
ChronoGAN | 0.204 \stackon[-2pt] 0.03 | 0.190 \stackon[-2pt] 0.08 | 0.213 \stackon[-2pt] 0.04 | 0.304 \stackon[-2pt] 0.06 |
TimeGAN | 0.326 \stackon[-2pt] 0.03 | 0.283 \stackon[-2pt] 0.13 | 0.271 \stackon[-2pt] 0.08 | 0.374 \stackon[-2pt] 0.10 |
GAN | 0.499 \stackon[-2pt] 0.01 | 0.320 \stackon[-2pt] 0.22 | 0.486 \stackon[-2pt] 0.01 | 0.5 \stackon[-2pt] 0.00 |
T-Forcing | 0.476 \stackon[-2pt] 0.01 | 0.348 \stackon[-2pt] 0.13 | 0.351 \stackon[-2pt] 0.10 | 0.5 \stackon[-2pt] 0.00 |
P-Forcing | 0.5 \stackon[-2pt] 0.00 | 0.5 \stackon[-2pt] 0.00 | 0.329 \stackon[-2pt] 0.10 | 0.5 \stackon[-2pt] 0.00 |
Stocks | Sines | ECG | SWAN-SF | |
ChronoGAN | 0.045 \stackon[-2pt] 0.00 | 0.225 \stackon[-2pt] 0.01 | 0.129 \stackon[-2pt] 0.00 | 0.055 \stackon[-2pt] 0.00 |
TimeGAN | 0.046 \stackon[-2pt] 0.00 | 0.245 \stackon[-2pt] 0.01 | 0.129 \stackon[-2pt] 0.01 | 0.082 \stackon[-2pt] 0.00 |
GAN | 0.186 \stackon[-2pt] 0.01 | 0.233 \stackon[-2pt] 0.01 | 0.191 \stackon[-2pt] 0.00 | 0.219 \stackon[-2pt] 0.01 |
T-Forcing | 0.050 \stackon[-2pt] 0.01 | 0.275 \stackon[-2pt] 0.01 | 0.130 \stackon[-2pt] 0.01 | 0.066 \stackon[-2pt] 0.01 |
P-Forcing | 0.147 \stackon[-2pt] 0.02 | 0.224 \stackon[-2pt] 0.01 | 0.194 \stackon[-2pt] 0.01 | 0.241 \stackon[-2pt] 0.01 |


IV-C Results and Discussion
Based on the results presented in Tables I and II, the ChronoGAN framework consistently outperforms state-of-the-art models, including TimeGAN, Teacher Forcing, Professor Forcing, and Standard GAN. In terms of the discriminative score, ChronoGAN achieves an average reduction of approximately 27.60% across the four datasets compared to TimeGAN. This substantial improvement indicates that ChronoGAN generates more realistic temporal data than other techniques. Furthermore, this improvement in the discriminative score can be attributed to the early generation algorithm, which enhances stability and ensures the best data is preserved during training. The improvement is also evident across all four datasets, each with different lengths, demonstrating the effectiveness of the GRU-LSTM layers within our framework. Additionally, according to discriminative score evaluations, ChronoGAN and TimeGAN emerge as superior compared to Teacher Forcing and Standard GAN. This underscores the importance of developing GAN-based techniques specifically tailored for time series data.
In terms of predictive score, ChronoGAN reduces the error by approximately 10.82% across the four datasets compared to TimeGAN. This underscores the effectiveness of our novel time series-based () and supervised () loss functions, which significantly improve the generator’s ability to capture the temporal dynamics of the data more accurately. As demonstrated in Figs. 3 and 4, we present several examples of synthetic samples generated by ChronoGAN for both the Sines and ECG datasets. These examples highlight ChronoGAN’s ability to effectively learn the temporal distributions of the real data and generate high-quality synthetic data that accurately reflect those patterns.
Based on Figs. 5 and 6, ChronoGAN demonstrates a superior ability to learn the probability distribution of real datasets more efficiently than all other baseline techniques. This is crucial, as a GAN-based model must generate data that accurately covers the entire distribution of the real dataset. The PCA and t-SNE results for the Stocks dataset show highly accurate outcomes. This achievement is primarily due to the loss, which enables the network to effectively capture the mean and variance of each batch of real data.








































V Conclusion and Future Work
In this study, we present ChronoGAN, an innovative model designed for generating time series data. ChronoGAN consists of five networks: an autoencoder (comprising an encoder and decoder), a generator, a supervisor, and a discriminator. These networks are trained together to learn the probability distribution and stepwise temporal dynamics of time series data. The model employs adversarial training in the feature space while generating data in the latent space, which significantly enhances the performance of both the autoencoder and generator networks. Additionally, ChronoGAN introduces novel loss functions for the autoencoder, generator, and supervisor networks, along with a new neural network architecture and an early generation mechanism. This framework consistently outperforms leading methods in generating realistic time series data, both qualitatively and quantitatively. In future research, we aim to integrate these concepts into adversarial autoencoders to develop an advanced framework for producing high-quality time series data.
VI Acknowledgment
Support for this work has been provided by the Division of Atmospheric and Geospace Sciences within the Directorate for Geosciences through NSF awards #2301397, #2204363, and #2240022, as well as by the Office of Advanced Cyberinfrastructure within the Directorate for Computer and Information Science and Engineering under NSF award #2305781.
References
- [1] M. EskandariNasab, Z. Raeisi, R. A. Lashaki, and H. Najafi, “A GRU–CNN model for auditory attention detection using microstate and recurrence quantification analysis,” Scientific Reports, vol. 14, no. 1, p. 8861, Apr. 2024, doi: 10.1038/s41598-024-58886-y.
- [2] S. M. Hamdi, D. Kempton, R. Ma, S. F. Boubrahimi, and R. A. Angryk, “A time series classification-based approach for solar flare prediction,” in 2017 IEEE International Conference on Big Data (Big Data), Boston, MA, USA, 2017, pp. 2543-2551, doi: 10.1109/BigData.2017.8258213.
- [3] Y. Velanki, P. Hosseinzadeh, S. F. Boubrahimi, and S. M. Hamdi, “Time-series feature selection for solar flare forecasting,” Universe, vol. 10, no. 9, Art. no. 373, 2024, doi: 10.3390/universe10090373.
- [4] A. Behfar, H. Atashpanjeh, and M. N. Al-Ameen, “Can password meter be more effective towards user attention, engagement, and attachment? A study of metaphor-based designs,” in Companion Publication of the 2023 Conference on Computer Supported Cooperative Work and Social Computing (CSCW ’23 Companion), Minneapolis, MN, USA, 2023, pp. 164-171, doi: 10.1145/3584931.3606983.
- [5] I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio, “Generative adversarial networks,” arXiv preprint arXiv:1406.2661, 2014.
- [6] A. Ahmadzadeh, B. Aydin, M. K. Georgoulis, D. J. Kempton, S. S. Mahajan, and R. A. Angryk, “How to train your flare prediction model: Revisiting robust sampling of rare events,” The Astrophysical Journal Supplement Series, vol. 254, no. 2, p. 23, 2021, doi: 10.3847/1538-4365/abec88.
- [7] O. Bahri, P. Li, S. F. Boubrahimi, and S. M. Hamdi, “Multiloss-based optimization for time series data augmentation,” in 2023 IEEE International Conference on Big Data (BigData), 2023, pp. 325–330, doi: 10.1109/BigData59044.2023.10386614.
- [8] K. Saini, K. Alshammari, S. M. Hamdi, and S. Filali Boubrahimi, “Classification of major solar flares from extremely imbalanced multivariate time series data using minimally random convolutional kernel transform,” Universe, vol. 10, p. 234, 2024, doi: 10.3390/universe10060234.
- [9] S. Bengio, O. Vinyals, N. Jaitly, and N. Shazeer, “Scheduled sampling for sequence prediction with recurrent neural networks,” in Advances in Neural Information Processing Systems, vol. 28, C. Cortes, N. Lawrence, D. Lee, M. Sugiyama, and R. Garnett, Eds., Curran Associates, Inc., 2015.
- [10] A. Lamb, A. Goyal, Y. Zhang, S. Zhang, A. Courville, and Y. Bengio, “Professor forcing: A new algorithm for training recurrent networks,” arXiv preprint arXiv:1610.09038, 2016.
- [11] D. Bahdanau, P. Brakel, K. Xu, A. Goyal, R. Lowe, J. Pineau, A. Courville, and Y. Bengio, “An actor-critic algorithm for sequence prediction,” arXiv preprint arXiv:1607.07086, 2017.
- [12] O. Mogren, “C-RNN-GAN: Continuous recurrent neural networks with adversarial training,” arXiv preprint arXiv:1611.09904, 2016.
- [13] C. Esteban, S. L. Hyland, and G. Rätsch, “Real-valued (medical) time series generation with recurrent conditional GANs,” arXiv preprint arXiv:1706.02633, 2017.
- [14] G. Ramponi, P. Protopapas, M. Brambilla, and R. Janssen, “T-CGAN: Conditional generative adversarial network for data augmentation in noisy time series with irregular sampling,” arXiv preprint arXiv:1811.08295, 2019.
- [15] P. Li, P. Hosseinzadeh, O. Bahri, S. F. Boubrahimi, and S. M. Hamdi, “Adversarial attack driven data augmentation for time series classification,” in 2023 International Conference on Machine Learning and Applications (ICMLA), 2023, pp. 653–658, doi: 10.1109/ICMLA58977.2023.00096.
- [16] J. Yoon, D. Jarrett, and M. van der Schaar, “Time-series generative adversarial networks,” in Advances in Neural Information Processing Systems, vol. 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. Fox, and R. Garnett, Eds., Curran Associates, Inc., 2019.
- [17] R. J. Williams and D. Zipser, “A learning algorithm for continually running fully recurrent neural networks,” Neural Computation, vol. 1, no. 2, pp. 270-280, Jun. 1989, doi: 10.1162/neco.1989.1.2.270.
- [18] Y. Ganin, E. Ustinova, H. Ajakan, P. Germain, H. Larochelle, F. Laviolette, M. Marchand, and V. Lempitsky, “Domain-adversarial training of neural networks,” Journal of Machine Learning Research, vol. 17, no. 1, pp. 2096–2030, 2016.
- [19] A. Lamb, A. Goyal, Y. Zhang, S. Zhang, A. Courville, and Y. Bengio, “Professor forcing: A new algorithm for training recurrent networks,” arXiv preprint arXiv:1610.09038 [stat.ML], 2016.
- [20] M. Mirza and S. Osindero, “Conditional generative adversarial nets,” arXiv preprint arXiv:1411.1784, 2014.
- [21] D. P. Kingma and M. Welling, “An introduction to variational autoencoders,” Foundations and Trends in Machine Learning, vol. 12, no. 4, pp. 307–392, 2019, doi: 10.1561/2200000056.
- [22] J. Beck and S. Chakraborty, “Fully embedded time series generative adversarial networks,” Neural Computation and Applications, vol. 36, pp. 14885–14894, 2024, doi: 10.1007/s00521-024-09825-5.
- [23] C. J. Cascalheira, S. Chapagain, R. E. Flinn, D. Klooster, D. Laprade, Y. Zhao, E. M. Lund, A. Gonzalez, K. Corro, R. Wheatley, A. Gutierrez, O. Garcia Villanueva, K. Saha, M. De Choudhury, J. R. Scheer, and S. M. Hamdi, “The LGBTQ+ minority stress on social media (MiSSoM) dataset: A labeled dataset for natural language processing and machine learning,” Proceedings of the International AAAI Conference on Web and Social Media, vol. 18, no. 1, pp. 1888–1899, May 2024, doi: 10.1609/icwsm.v18i1.31433.
- [24] S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Computation, vol. 9, no. 8, pp. 1735–1780, 1997.
- [25] K. Cho, B. van Merriënboer, C. Gulcehre, D. Bahdanau, F. Bougares, H. Schwenk, and Y. Bengio, “Learning phrase representations using RNN encoder-decoder for statistical machine translation,” in Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), A. Moschitti, B. Pang, and W. Daelemans, Eds., Doha, Qatar, Association for Computational Linguistics, Oct. 2014, pp. 1724-1734. [Online]. Available: https://aclanthology.org/D14-1179
- [26] R. Cahuantzi, X. Chen, and S. Güttel, “A comparison of LSTM and GRU networks for learning symbolic sequences,” in Intelligent Computing, Springer Nature Switzerland, 2023, pp. 771–785, doi: 10.1007/978-3-031-37963-5_53.
- [27] R. A. Angryk, P. C. Martens, B. Aydin, D. Kempton, S. S. Mahajan, S. Basodi, A. Azim, X. Cai, S. Filali Boubrahimi, S. Soukaina, S. M. Hamdi, M. Muhammad, M. A. Schuh, and M. K. Georgoulis, “Multivariate time series dataset for space weather data analytics,” Scientific Data, vol. 7, no. 1, p. 227, 2020, doi: 10.1038/s41597-020-0548-x.
- [28] K. Alshammari, K. Saini, S. M. Hamdi, and S. F. Boubrahimi, “End-to-end attention/transformer model for solar flare prediction from multivariate time series data,” in 2023 International Conference on Machine Learning and Applications (ICMLA), 2023, pp. 558–565, doi: 10.1109/ICMLA58977.2023.00083.
- [29] M. EskandariNasab, S. M. Hamdi, and S. F. Boubrahimi, “Impacts of data preprocessing and sampling techniques on solar flare prediction from multivariate time series data of photospheric magnetic field parameters,” Astrophysical Journal Supplement Series, in press, doi: 10.3847/1538-4365/ad7c4a.
- [30] M. EskandariNasab, S. M. Hamdi, and S. F. Boubrahimi, “SWAN-SF Data Preprocessing and Sampling Notebooks”. Zenodo, Jun. 11, 2024. doi: 10.5281/zenodo.11564789.
- [31] M. EskandariNasab, S. M. Hamdi, and S. F. Boubrahimi, “Cleaned SWANSF Dataset”. Zenodo, Jun. 11, 2024. doi: 10.5281/zenodo.11566472.
- [32] L. van der Maaten and G. Hinton, “Visualizing data using t-SNE,” Journal of Machine Learning Research, vol. 9, no. 86, pp. 2579–2605, 2008.
- [33] F. B. Bryant and P. R. Yarnold, “Principal-components analysis and exploratory and confirmatory factor analysis,” in Reading and Understanding Multivariate Statistics, L. G. Grimm and P. R. Yarnold, Eds., American Psychological Association, 1995, pp. 99–136.