This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Addressing Concept Shift in Online Time Series Forecasting: Detect-then-Adapt

YiFan Zhang, Weiqi Chen, Zhaoyang Zhu, Dalin Qin, Liang Sun, Xue Wang, Qingsong Wen, Zhang Zhang, Liang Wang , Rong Jin Yi-Fan Zhang, Zhang Zhang, and Liang Wang are with the State Key Laboratory of Multimodal Artificial Intelligence Systems (MAIS), Center for Research on Intelligent Perception and Computing (CRIPAC), Institute of Automation, Chinese Academy of Sciences (CASIA), Beijing 100190, China, and also with the School of Artificial Intelligence, University of Chinese Academy of Sciences (UCAS), Beijing 100049, China (e-mail: [email protected]).Weiqi Chen, Zhaoyang Zhu, Dalin Qin, Liang Sun, and Xue Wang are with the Alibaba Group.Qingsong Wen is with the Squirrel AI Group.Rong Jin is with the Meta AI Group.
Abstract

Online updating of time series forecasting models aims to tackle the challenge of concept drifting by adjusting forecasting models based on streaming data. While numerous algorithms have been developed, most of them focus on model design and updating. In practice, many of these methods struggle with continuous performance regression in the face of accumulated concept drifts over time. To address this limitation, we present a novel approach, Concept Drift Detection anD Adaptation (D3A), that first detects drifting conception and then aggressively adapts the current model to the drifted concepts after the detection for rapid adaption. To best harness the utility of historical data for model adaptation, we propose a data augmentation strategy introducing Gaussian noise into existing training instances. It helps mitigate the data distribution gap, a critical factor contributing to train-test performance inconsistency. The significance of our data augmentation process is verified by our theoretical analysis. Our empirical studies across six datasets demonstrate the effectiveness of D3A in improving model adaptation capability. Notably, compared to a simple Temporal Convolutional Network (TCN) baseline, D3A reduces the average Mean Squared Error (MSE) by 43.9%43.9\%. For the state-of-the-art (SOTA) model, the MSE is reduced by 33.3%33.3\%.

Index Terms:
Time series forecasting, concept drift, online learning, concept drift detection

I Introduction

In recent years, there has been a notable upswing in research efforts dedicated to applying deep learning techniques to time series forecasting [17, 31]. Deep models have demonstrated exceptional proficiency, excelling not only in forecasting tasks but also in representation learning. However, predominant studies have focused on the batch learning setting, assuming the availability of the entire training dataset in advance. These approaches operate under the assumption that the relationship between input and output variables remains constant throughout the learning process. This paradigm, however, falls short in real-world scenarios where concepts are dynamic and subject to change, a phenomenon known as concept drift [28]. Concept drift implies that future data may exhibit patterns different from those observed in the past. Although re-training the model from scratch may help address the concept drift problem, it is very time-consuming. Hence, there is a growing need for the development of online training methodologies for deep forecasters, enabling incremental updates to the forecasting model with new samples. These adaptive approaches have been proved to be essential for capturing the evolving dynamics of the environment.

Refer to caption
Figure 1: Model performance comparison on various tasks.

Engaging in online time series forecasting within real-world scenarios introduces formidable challenges marked by heightened noisy gradients, distinct from the conventional offline mini-batch training methodologies [3]. The continuous shifts in data distribution further compound the intricacies, diminishing the efficacy of models trained on historical data for contemporary predictions. Despite commendable endeavors to address these challenges through the design of sophisticated updating structures or learning objectives [22, 34], current methodologies fall short when confronted with substantial shifts in the underlying concept. This deficiency manifests in a notable degradation of mean-average-error (MAE), as depicted in Figure 2. To delve deeper into this challenge, it is imperative to consider the limitations of individual-example-based online learning algorithms in achieving rapid adaptation. These algorithms grapple with the intricate task of balancing long-term performance and swift adaptation. A large learning rate can facilitate rapid adaptation but risks compromising long-term generalization performance by making the model overly sensitive to noise in time series data. On the contrary, a small learning rate ensures sustained long-term performance but falters in adapting quickly to distribution changes (Figure 2b).

Refer to caption
(a)
Refer to caption
(b)
Figure 2: Accumulated MAE curves of different online learning (OL) methods on ECL dataset, the default model for (b) is FSNet.

These observations underscore the necessity of adapting the model aggressively upon the occurrence of a significant concept change, as a proactive approach for adaptation becomes crucial for mitigating the inherent trade-off between rapid adjustment and sustained performance in the dynamic landscape of online time series forecasting. As a result, we introduce the framework of detect and then adapt, with one component for drift detection and one for model adaption. Unlike typical out-of-distribution detection methods that make the detection decision independently for individual data instances, our detector evaluates whether the online testing distribution significantly deviates from the historical one or surpasses the model’s capacity. We make drift decision by comparing error rates. It is grounded in the hypothesis that a statistically significant increase in the error rate serves as a trigger for a drift alarm. Our adaption module is carefully designed to address critical issues for significant concept drift. To avoid the catastrophic forgetting problem that often arises in quick adaption, our adaption module leverages both recently acquired data and the wealth of extensive historical data for model training. Unlike existing replay-buffer based methods that directly mix old and new data for training and often overlook the inherent disparities between two distributions, our method makes special efforts to mitigate the performance regression caused by the distribution gap. In particular, we incorporate data dependent Gaussian noise into the history data before they are used for model adaption, which is theoretically justified to bridge the distribution gap in the case of linear regression models and is verified in our empirical study (see Figure 2b).

Our approach, referred to as Concept Drift Detection and Adaptation (D3A), encompasses a novel online learning framework and training strategy. This methodology significantly alleviates the issue of sharp performance decline encountered by models in the face of substantial concept drift, as depicted in Figure 2. Our method stands orthogonal to the majority of existing models and when applying our updating strategy, these models all demonstrate faster adaptation to new concepts, leading to a substantial enhancement in performance (the average MSEs of different tasks are shown in Figure 1). The contributions of this paper are:

1. Introduce a Concept Detection Framework: Our framework monitors loss distribution drift, aiming to predict the occurrence of concept drift. This detector provides instructions for our model updating, enhancing model robustness and AI safety, particularly in high-risk tasks.

2. Innovative Model Adaptation Strategy: Our theoretical analysis commences with a linear case, uncovering the pivotal role played by the training-test distribution gap in the resultant model. In our pursuit of harnessing a substantial volume of historical data for enhanced model updating, we introduce a novel data augmentation strategy designed to bridge and mitigate the existing distribution gap effectively.

3. More realistic Evaluation setting: We observe that previous benchmarks [22] often presume a substantial overlap in the forecasting target during testing. In this paper, we advocate for the evaluation of online time series forecasting models with delayed feedback, demonstrating a more realistic and challenging assessment.

4. Empirical Studies: With six datasets, D3A improves model adaptation capability across various methods. For example, compared to a simple TCN baseline, D3A reduces the average MSE by 43.9%43.9\% and MAE by 26.9%26.9\%. For the previous SOTA model FSNet [22], the reductions in MSE and MAE are 33.3%33.3\% and 16.7%16.7\%, respectively. In a more challenging real-world experimental setting, D3A consistently outperforms existing methods. For TCN, FSNet, and OneNet [36], the MSE reductions are 32%32\%, 33.1%33.1\%, and 22.2%22.2\%, respectively.

II Preliminary and Related Work

Concept drift. Concepts in the real world are often dynamic and can change over time, which is especially true for scenarios like weather prediction and customer preferences. Because of unknown changes in the underlying data distribution, models learned from historical data may become inconsistent with new data, thus requiring regular updates to maintain accuracy. This phenomenon, known as concept drift [28], adds complexity to the process of learning a model from data. In this paper, we focus on online learning for time series forecasting. Unlike most existing studies for online time series forecasting [15, 23, 22] that only focus on how to online update their models, this work goes beyond parameter updating by each online instance and introduces to detect and adapt concept drift initiatively.

Time Series Modeling: Time series models, pivotal in various domains, have undergone decades of development. Early data-driven approaches, such as autoregressive models like ARIMA [6], paved the way but grapple with challenges posed by nonlinearity and non-stationarity. Recurrent Neural Networks (RNNs) were introduced to manage sequential data, with LSTM [11] and GRU [10] incorporating gated structures to mitigate gradient-related issues. Attention-based RNNs [24], leveraging temporal attention for capturing long-range dependencies, face limitations in parallelizability and struggle with extended dependencies. Temporal Convolutional Networks [26] offer efficiency but exhibit restricted reception fields, making them less effective for handling prolonged dependencies. In recent times, transformer-based models [29, 31] have been revitalized and applied to time series forecasting. Despite extensive efforts to enhance the efficiency and potency of Transformer models [38, 37, 21], we uniquely contribute by evaluating the robustness of advanced forecasting models under concept drift. This novel approach enhances their adaptability to new distributions.

Out-of-distribution detection and anomaly detection. Out-of-distribution detection methods primarily encompass (1) classification-based methods, which derive Out-of-Distribution (OOD) scores based on classifiers trained on In-Distribution (ID) data [13]. Examples include using the maximum softmax probability [13] and energy [19] as OOD scores. (2) Density-based methods model in-distribution using probabilistic models, classifying data samples in low-density regions as OOD [40, 1, 32]. (3) Distance-based methods operate under the assumption that OOD samples should be relatively distant from representations of in-distribution classes [14, 30, 35]. Detecting anomalies in unsupervised time series presents a significant challenge in practical scenarios. The model must acquire informative representations of intricate temporal dynamics through unsupervised tasks and establish a distinctive criterion capable of identifying rare anomalies among numerous normal time points. Various classic methods for anomaly detection provide unsupervised paradigms, such as density-estimation methods introduced in the local outlier factor (LOF) [7], clustering-based approaches exemplified by one-class SVM (OC-SVM) [25], and SVDD [27]. These traditional methods overlook temporal information and encounter difficulties when generalizing to unforeseen real-world scenarios. Leveraging the representation learning capabilities of neural networks, recent deep models [16, 33] have demonstrated superior performance. However, in this paper, our focus is on detecting concept shifts rather than OOD instances that differ from the ID distribution. We employ a simple sliding-window detection method to emphasize the importance of recognizing when the concept drift occurs and adapting to new environments proactively. More complex considerations involving existing OOD detection metrics can be explored in future work.

Online time series forecasting: streaming data. Traditional time series forecasting tasks have a collection of multivariate time series with a look-back window LL: (𝐱i)i=1L(\mathbf{x}_{i})_{i=1}^{L}, where each 𝐱i\mathbf{x}_{i} is a MM-channel vector 𝐱i=(xij)j=1M\mathbf{x}_{i}=(x_{i}^{j})_{j=1}^{M}. Given a forecast horizon HH, the target is to forecast HH future values (𝐱i)i=L+1L+H(\mathbf{x}_{i})_{i=L+1}^{L+H}. In real-world applications, the model builds on the historical data needs to forecast the future data, that is, given time offset K>LK^{\prime}>L, and (𝐱i)i=KL+1K(\mathbf{x}_{i})_{i=K^{\prime}-L+1}^{K^{\prime}}, the model needs to forecast (𝐱)i=K+1K+H(\mathbf{x})_{i=K^{\prime}+1}^{K^{\prime}+H}. Online time series forecasting [5, 18, 22] is a widely used technique in real-world due to the sequential nature of the data and the frequent drift of concepts. In this approach, the learning process takes place over a sequence of rounds, where the model receives a look-back window and predicts the forecast window. The true values are then revealed to improve the model’s performance in the next rounds. When we perform online adaptation, the model is retrained using the online data stream with the MSE loss over each channel: =1Mj=1Mx^K+1:K+HjxK+1:K+Hj\mathcal{L}=\frac{1}{M}\sum_{j=1}^{M}\parallel\hat{x}_{K^{\prime}+1:K^{\prime}+H}^{j}-{x}_{K^{\prime}+1:K^{\prime}+H}^{j}\parallel.

III Concept Drift Detection and Adaptation

Refer to caption
(a)
Refer to caption
(b)
Figure 3: (a) Accumulated MAE curves of ETTh2 and (b) Distribution of Loss Over Time, where lw=100l_{w}=100.

The current research landscape in online time series forecasting primarily revolves around model design and training strategies. However, when confronted with a substantial shift in the underlying concept, these methods grapple with mitigating the significant performance decline of the model under the new concept (see Figure 2 and Figure 3a). One primary limitation with most of these online forecasting algorithms is that they gradually update their model based on one data point at each time, making it inefficient to generate enough changes in parameters when encountering significant concept drift. A recent approach, SIESTA [12], suggests the need for online learning but requires a long period of time (sleep stages) for offline memory consolidation, thus not applicable to our scenario. In addition, this method requires a full fine tuning of the model, thus introducing significant overhead in computation.

An intuitive idea to address this challenge is to perform aggressive updates of the model with new data when a substantial concept change occurs, allowing for rapid adaptation to the new concept and avoiding unnecessary full fine-tuning. To make this detect-then-adapt approach work in practice, we need to address three critical questions:

  • How to determine if a significant change occurs in concept ?,

  • What data to use for effective model update?, and

  • How to reduce the computation overhead in detect-then-adapt ?.

In this section, we will delve into addressing these questions.

III-A Concept drift detection

Unlike traditional out-of-distribution (OOD) detection or anomaly detection methods, which aim to determine whether a given data instance deviates significantly from the training or normal distribution, our focus is on assessing whether the current online testing distribution substantially differs from the training distribution or exceeds the model’s capability. Our detector operates on the hypothesis that a statistically significant increase in the error rate triggers a drift alarm.

Depicted in Figure 3b, our online monitoring strategy involves tracking testing losses, assuming a normal distribution 𝒩(μ,σ)\mathcal{N}(\mu,\sigma), and subsequently subjecting them to a statistical test. Specifically, employing a test window size lwl_{w}, we designate the latest lwl_{w} losses as the test target. The mean value of losses within the test window, denoted as μ~\tilde{\mu}, is computed, along with the corresponding z-score μ~μσ/|B|\frac{\tilde{\mu}-\mu}{\sigma/\sqrt{|B|}}. We then assess the confidence level at which the score of new data instances deviates from the overall distribution of test losses. Notably, in the context of a normal distribution, both significantly smaller and larger scores should trigger the drift alarm. However, in our scenario, a smaller loss signifies commendable model performance under new data instances. Consequently, the alarm is activated exclusively when μ~>μ\tilde{\mu}>\mu. In addition, we implement a detector reset every mtm_{t} steps, where mtlwm_{t}\geq l_{w}. This design offers two distinct advantages: firstly, it prevents a scenario where outliers with high losses persist, leading to an inflated μ\mu that does not accurately reflect the true loss distribution; secondly, this flexible approach empowers users to determine the restart time based on their preferences or the model’s idleness. The corresponding pseudocode is presented in Algorithm 1.

Algorithm 1 Concept Drift Detection and Adaptation
1:  Input: Online multivariate time series 𝐱M×L\mathbf{x}\in\mathbb{R}^{M\times L}, target 𝐲M×H\mathbf{y}\in\mathbb{R}^{M\times H}, trained forecaster fθf_{\theta}
2:  Parameters: Window size lwl_{w}, a predefined confidence level α\alpha, and a restart step mtlwm_{t}\geq l_{w}.  
3:  Initialize a memory bank \mathcal{M} with FIFO updating strategy, ||=lw|\mathcal{M}|=l_{w}, and loss list B=[]B=[]
4:  while i=1,,Ti=1,\cdots,T do
5:     𝐲~fθ(𝐱)\tilde{\mathbf{y}}\leftarrow f_{\theta}(\mathbf{x})
6:     fθAdam(fθ,(𝐲~,𝐲))f_{\theta}\leftarrow\text{Adam}(f_{\theta},\mathcal{L}(\tilde{\mathbf{y}},\mathbf{y})) //Update forecaster
7:     (𝐱,𝐲)\mathcal{M}\leftarrow\mathcal{M}\cup(\mathbf{x},\mathbf{y})
8:     Bappend(B,(𝐲~,𝐲))B\leftarrow\text{append}(B,\mathcal{L}(\tilde{\mathbf{y}},\mathbf{y})) //Store loss and data
9:      // Loss statistics.
10:     μmean(B)\mu\leftarrow\text{mean}(B), σstd(B)\sigma\leftarrow\text{std}(B), μ~mean(B[lw:])\tilde{\mu}\leftarrow\text{mean}(B[-l_{w}:])
11:     if (|B|>w|B|>w and μ~μσ/|B|>α\frac{\tilde{\mu}-\mu}{\sigma/\sqrt{|B|}}>\alpha) or i%mt==0i\%m_{t}==0 then
12:        Fully tune the model using \mathcal{M};  Reset B=[]B=[]
13:     end if
14:  end while

III-B Better Adaptation by Reduced Distribution Gap

Simply fully fine-tuning the model fθf_{\theta} on new data points in lwl_{w} may lead to limited performance. This is because (a) it could result in catastrophic forgetting of knowledge learned from historical data if we let the optimization procedure reach its optimum, and (b) the limited number of training examples in lwl_{w} makes it impossible to learn an optimal model. It then prompts a crucial question: Can we leverage historical data to enhance our model adaptation? Given historical examples stem from different distributions compared to test instances, directly using historical data for adaption could result in sub-optimal performance. We propose a data augmentation method aimed at alleviating the gap in feature distribution. It augments existing historical data with synthetically generated examples. These synthetic examples are crafted by introducing new instances-dependent Gaussian noise to historical instances. Below, we justify this approach by an analysis of a linear model, with all proofs available in appendix VI.

III-B1 An Analysis based on Linear Model

Given the substantial size of our model training dataset, our analysis conveniently navigates around the intricacies introduced by finite data sampling. We employ the simplifying assumption of having direct access to the oracle of data distribution. Let dd denote the dimension of the input data. Our approach entails the segmentation of the overall input data into two distributions: 𝒫A\mathcal{P}_{A} representing historical/training data and 𝒫B\mathcal{P}_{B} representing newly observed instances. The true data distribution 𝒫\mathcal{P} emerges as a linear combination of 𝒫A\mathcal{P}_{A} and 𝒫B\mathcal{P}_{B}, articulated as 𝒫=(1γ)𝒫A+γ𝒫B\mathcal{P}=(1-\gamma)\mathcal{P}_{A}+\gamma\mathcal{P}_{B}, where γ(0,1)\gamma\in(0,1) is a fixed yet unknown weight employed to linearly blend the two distributions.

Consider y(x):d[0,1]y(x):\mathbb{R}^{d}\mapsto[0,1], the deterministic output function 111For simplicity, we disregard the randomness in the output function. Nonetheless, it’s relatively straightforward to extend our analysis to accommodate stochastic outputs. Our assumption rests on the belief that 𝒫A\mathcal{P}_{A} exhibits significantly higher quality and a more substantial volume than 𝒫B\mathcal{P}_{B}, forming the cornerstone of our analysis.

Ex𝒫A[y(x)]Ex𝒫B[y(x)]\mathrm{E}_{x\sim\mathcal{P}_{A}}[y(x)]\gg\mathrm{E}_{x\sim\mathcal{P}_{B}}[y(x)]

In other words, we can almost treat Ex𝒫B[y(x)]0\mathrm{E}_{x\sim\mathcal{P}_{B}}[y(x)]\approx 0222This does not imply that it is exactly zero, but considered negligibly small relative to 𝒫A\mathcal{P}_{A}. This assumption is a simplification aimed at emphasizing that, the historical data distribution dominates the model training.. If we have access to the true data distribution 𝒫\mathcal{P}, in case of a linear predictor, we need to solve the following optimization problem minwd(w)=12Ex𝒫[|y(x)x,w|2]\min\limits_{w\in\mathbb{R}^{d}}\;\mathcal{L}(w)=\frac{1}{2}\mathrm{E}_{x\sim\mathcal{P}}\left[|y(x)-\langle x,w\rangle|^{2}\right]. The optimal ww is then given by

w=Σ1Ex𝒫[y(x)x]w=\Sigma^{-1}\mathrm{E}_{x\sim\mathcal{P}}\left[y(x)x\right]
Ex𝒫[y(x)x]=Ex𝒫A[y(x)x]+Ex𝒫B[y(x)x]\mathrm{E}_{x\sim\mathcal{P}}[y(x)x]=\mathrm{E}_{x\sim\mathcal{P}_{A}}[y(x)x]+\mathrm{E}_{x\sim\mathcal{P}_{B}}[y(x)x]

where Σ=Ex𝒫[xx]\Sigma=\mathrm{E}_{x\sim\mathcal{P}}[xx^{\top}]. Since we assume 𝒫A\mathcal{P}_{A} has much higher influence on the model, we have

Ex𝒫[y(x)x]Ex𝒫A[y(x)x]:=zA\mathrm{E}_{x\sim\mathcal{P}}[y(x)x]\approx\underbrace{\mathrm{E}_{x\sim\mathcal{P}_{A}}[y(x)x]}_{:=z_{A}}

Hence, we will only consider the following approximately optimal solution in comparison, i.e.

w=Σ1zA\displaystyle w_{*}=\Sigma^{-1}z_{A} (1)

Now, we consider the case when we only train the model with input data from distribution 𝒫A\mathcal{P}_{A}. The resulting solution is given by

wA=ΣA1zA\displaystyle w_{A}=\Sigma_{A}^{-1}z_{A} (2)

where ΣA=Ex𝒫A[xx]\Sigma_{A}=\mathrm{E}_{x\sim\mathcal{P}_{A}}[xx^{\top}]. When comparing wAw_{A} to ww_{*}, a discernible distinction becomes apparent in the feature correlation matrices ΣA\Sigma_{A} and Σ\Sigma. Consequently, for the linear prediction model, we observe that the discrepancy between the learned model wAw_{A} and the underlying optimal model ww_{*} can be attributed to the distribution gap specifically present in the feature correlation matrices.

III-B2 How to Fill out the Gap?

We will first bound how the difference between ww_{*} and wAw_{A} affects the overall prediction performance in terms of the gap in feature correlations. To this end, we first define the gap matrix Δ(ΣA)\Delta(\Sigma_{A}) as

Δ(ΣA)=Σ1/2(ΣΣA)Σ1/2=γΣ1/2(ΣBΣA)Σ1/2\Delta(\Sigma_{A})=\Sigma^{-1/2}(\Sigma-\Sigma_{A})\Sigma^{-1/2}=\gamma\Sigma^{-1/2}\left(\Sigma_{B}-\Sigma_{A}\right)\Sigma^{-1/2}

The following theorem allows us to bound the performance degradation of the linear predictor in terms of |Δ(ΣA)|2|\Delta(\Sigma_{A})|_{2}, i.e., the spectral norm of the gap matrix.

Theorem 1.

If |Δ(ΣA)|21/2|\Delta(\Sigma_{A})|_{2}\leq 1/2, we have Ex𝒫[|(wwA)x|2]40|Δ(ΣA)|22\mathrm{E}_{x\sim\mathcal{P}}\left[\left|(w_{*}-w_{A})^{\top}x\right|^{2}\right]\leq 4\mathcal{L}_{0}|\Delta(\Sigma_{A})|_{2}^{2}, where 0=zAΣ1zA\mathcal{L}_{0}=z_{A}^{\top}\Sigma^{-1}z_{A}.

As indicated by Theorem 1, to reduce the prediction gap with respect to the true optimal solution ww_{*}, we need to come up with a different estimation of ΣA\Sigma_{A} such that |Δ(ΣA)|2|\Delta(\Sigma_{A})|_{2} can be reduced significantly. To motivate our approach, we consider a special structure for ΣA\Sigma_{A} and ΣB\Sigma_{B}. We assume that for the input data from distribution 𝒫B\mathcal{P}_{B}, most features are completely uncorrelated, and as a result, we approximate ΣB\Sigma_{B} by a diagonal matrix, i.e.

ΣB=αI\displaystyle\Sigma_{B}=\alpha I (3)

where α>0\alpha>0. For input distribution 𝒫A\mathcal{P}_{A}, we assume that it has k<dk<d groups of features being strongly correlated. As a result, we assume that ΣA\Sigma_{A} can be well approximated by a combination of a diagonal matrix and a low-rank matrix, i.e.

ΣA=βI+Udiag(ν)U\displaystyle\Sigma_{A}=\beta I+U\mbox{diag}(\nu)U^{\top} (4)

where β>0\beta>0, Ud×kU\in\mathbb{R}^{d\times k} is an orthonormal matrix and ν+k\nu\in\mathbb{R}_{+}^{k}. The following proposition computes |Δ(ΣA)|2|\Delta(\Sigma_{A})|_{2}

Proposition 1.

Assume αβ\alpha\geq\beta and maxi[k]νi[αβ,2(αβ)]\max_{i\in[k]}\nu_{i}\in[\alpha-\beta,2(\alpha-\beta)]. We have |Δ(ΣA)|2=γ(αβ)τ|\Delta(\Sigma_{A})|_{2}=\frac{\gamma(\alpha-\beta)}{\tau}.

Now, we use a different ΣA\Sigma_{A} for computing wAw_{A}. Define

ΣA=ΣA+γI\displaystyle\Sigma_{A^{\prime}}=\Sigma_{A}+\gamma I (5)

The resulting new solution wAw_{A^{\prime}} becomes

wA=ΣA1zA\displaystyle w_{A^{\prime}}=\Sigma_{A^{\prime}}^{-1}z_{A} (6)

According to Theorem 1, we have

Ex𝒫[|(wwA)x|2]40|Δ(ΣA)|22\mathrm{E}_{x\sim\mathcal{P}}\left[\left|(w_{*}-w_{A^{\prime}})^{\top}x\right|^{2}\right]\leq 4\mathcal{L}_{0}|\Delta(\Sigma_{A^{\prime}})|^{2}_{2}

where Δ(ΣA)=γΣ1/2(ΣBΣA)Σ1/2\Delta(\Sigma_{A^{\prime}})=\gamma\Sigma^{-1/2}(\Sigma_{B}-\Sigma_{A^{\prime}})\Sigma^{-1/2}. The theorem below bounds |Δ(ΣA|2|\Delta(\Sigma_{A^{\prime}}|_{2}.

Theorem 2.

Assume α2γ3γβ\alpha\geq\frac{2-\gamma}{3-\gamma}\beta, and

αβ|ν|min(2(αβ),τ(3γ(αβ)β)τγ(αβ))\alpha-\beta\leq|\nu|_{\infty}\leq\min\left(2(\alpha-\beta),\frac{\tau\left(3\gamma(\alpha-\beta)-\beta\right)}{\tau-\gamma(\alpha-\beta)}\right)

we have |Δ(ΣA)|2|Δ(ΣA)|2|\Delta(\Sigma_{A^{\prime}})|_{2}\leq|\Delta(\Sigma_{A})|_{2}

How to implement it? As Theorem 2 shows, if we can replace ΣA\Sigma_{A} with ΣA=ΣA+τI\Sigma_{A}^{\prime}=\Sigma_{A}+\tau I, we will be able to reduce the gap between historical and new data. Now, the question is how to implement it in practice. First, adding τI\tau I to ΣA\Sigma_{A} is equivalent to adding Gaussian noise to the input data. More specifically, for each xx sampled from 𝒫A\mathcal{P}_{A}, we add a random vector u𝒩(0,τI)u\sim\mathcal{N}(0,\tau I) to form a new sample x=x+ux^{\prime}=x+u. It is easy to verify that Ex[x(x)]=ΣA+γI=ΣA\mathrm{E}_{x^{\prime}}\left[x^{\prime}(x^{\prime})^{\top}\right]=\Sigma_{A}+\gamma I=\Sigma_{A^{\prime}} and zA=Ex[y(x)x]=zAz_{A}^{\prime}=\mathrm{E}_{x^{\prime}}[y(x)x^{\prime}]=z_{A}. And our solution ww learned from the perturbed data becomes

w~\displaystyle\widetilde{w} =(Ex[x(x)])1Ex[y(x)x]\displaystyle=\left(\mathrm{E}_{x^{\prime}}[x^{\prime}(x^{\prime})^{\top}]\right)^{-1}\mathrm{E}_{x^{\prime}}[y(x)x^{\prime}]
=(ΣA+γI)1zA=ΣA1zA\displaystyle=\left(\Sigma_{A}+\gamma I\right)^{-1}z_{A}^{\prime}=\Sigma^{-1}_{A^{\prime}}z_{A}

which is the same as the solution given in (6).

Second, we notice that τI\tau I is close to diag(Σ)\mbox{diag}(\Sigma). Hence, we can approximate τI\tau I by using diag(Σ)\mbox{diag}(\Sigma), which requires computing the variance of individual features over distribution 𝒫\mathcal{P}. Hence, in practice, we will first compute variance vector s=diag(Σ)+ds=\mbox{diag}(\Sigma)\in\mathbb{R}_{+}^{d}, with each component of ss corresponding to the variance of a different feature. We then construct the synthesis dataset. Let 𝒟={(xi,yi),i=1,,n}\mathcal{D}=\{(x_{i},y_{i}),i=1,\ldots,n\} be the original training dataset. For each (xi,yi)𝒟(x_{i},y_{i})\in\mathcal{D} from the original dataset 𝒟\mathcal{D}, we sample a noise vector ui𝒩(0,diag(s))u_{i}\sim\mathcal{N}(0,\mbox{diag}(s)) and create a synthesized example (xi=xi+ui,yi)(x_{i}^{\prime}=x_{i}+u_{i},y_{i}). Let 𝒟\mathcal{D}^{\prime} be the synthesized dataset, i.e.

𝒟={(xi+ui,yi),ui𝒩(0,diag(s)),i=1,,n}\mathcal{D}^{\prime}=\left\{(x_{i}+u_{i},y_{i}),u_{i}\sim\mathcal{N}(0,\mbox{diag}(s)),i=1,\ldots,n\right\} (7)

In our experimental setup, employing the original training dataset as 𝒫A\mathcal{P}_{A} proves to be less intuitive due to the potential significant deviation of the concept at a later time from the training set. To address this issue, we introduce an auxiliary memory, denoted as prev\mathcal{M}_{prev}, which encompasses a substantial number of data instances from the period preceding the concept drift. Specifically, \mathcal{M} may include instances (xi)i=tt+lw(x_{i})_{i=t}^{t+l_{w}}, while prev\mathcal{M}_{prev} consists of instances (xi)t1lvt1(x_{i})_{t-1-l_{v}}^{t-1}, where lvlwl_{v}\gg l_{w}. We then construct the synthesized dataset as outlined in 7. Furthermore, we exclusively utilize data samples from 𝒟\mathcal{D}^{\prime} as a regularization strategy to compel the model to enhance its performance on the drifted distribution. In each iteration, we randomly sample pairs (𝐱,𝐲)(\mathbf{x},\mathbf{y})\in\mathcal{M} and (𝐱~,𝐲~)𝒟(\tilde{\mathbf{x}},\tilde{\mathbf{y}})\in\mathcal{D}^{\prime}, and subsequently compute the loss (fθ(𝐱),𝐲)+λ(fθ(𝐱~),𝐲~)\mathcal{L}(f_{\theta}(\mathbf{x}),\mathbf{y})+\lambda\mathcal{L}(f_{\theta}(\tilde{\mathbf{x}}),\tilde{\mathbf{y}}).

TABLE I: MSE (top) and MAE (bottom) of various adaptation methods. H: forecast horizon. Avg-C: the average of metric values under challenging benchmarks including ETTH2, Weather, and ECL.
ETTH2 ETTm1 WTH Traffic Weather ECL
MSE 1 24 48 1 24 48 1 24 48 1 24 48 1 24 48 1 24 48 Avg Avg-C
Informer 7.571 4.629 5.692 0.456 0.478 0.388 0.426 0.380 0.367 - - - - - - - - - 2.265 5.964
OnlineTCN 0.502 0.830 1.183 0.214 0.258 0.283 0.206 0.308 0.302 0.329 0.463 0.498 0.375 0.506 0.522 3.309 11.339 11.534 1.831 2.616
TFCL 0.557 0.846 1.208 0.087 0.211 0.236 0.177 0.301 0.323 - - - - - - 2.732 12.094 12.110 2.574 4.925
ER 0.508 0.808 1.136 0.086 0.202 0.220 0.180 0.293 0.297 0.286 0.383 0.394 0.312 0.441 0.442 2.579 9.327 9.685 1.532 2.192
MIR 0.486 0.812 1.103 0.085 0.192 0.210 0.179 0.291 0.297 - - - - - - 2.575 9.265 9.411 2.076 3.942
DER++ 0.508 0.828 1.157 0.083 0.196 0.208 0.174 0.287 0.294 0.280 0.374 0.384 0.308 0.434 0.435 2.657 8.996 9.009 1.478 2.114
FSNet 0.466 0.687 0.846 0.085 0.115 0.127 0.162 0.188 0.223 0.288 0.358 0.379 0.309 0.357 0.399 3.143 6.051 7.034 1.179 1.693
OneNet 0.372 0.559 0.603 0.087 0.098 0.133 0.154 0.181 0.197 0.223 0.348 0.381 0.296 0.314 0.428 2.481 2.747 3.279 0.735 1.059
D3A-TCN 0.407 0.753 0.845 0.091 0.193 0.201 0.203 0.329 0.326 0.304 0.386 0.405 0.348 0.456 0.477 2.347 5.033 5.507 1.034 1.439
D3A-FSNet 0.381 0.643 0.767 0.084 0.111 0.121 0.160 0.185 0.219 0.276 0.336 0.354 0.308 0.332 0.377 2.542 3.162 3.781 0.786 1.105
D3A-OneNet 0.362 0.543 0.584 0.082 0.098 0.129 0.150 0.173 0.191 0.219 0.337 0.352 0.296 0.329 0.371 2.188 2.122 2.406 0.608 0.842
ETTH2 ETTm1 WTH Traffic Weather ECL
MAE 1 24 48 1 24 48 1 24 48 1 24 48 1 24 48 1 24 48 Avg Avg-C
Informer 0.850 0.668 0.752 0.512 0.525 0.460 0.458 0.417 0.419 - - - - - - - - - 0.562 0.757
OnlineTCN 0.436 0.547 0.589 0.214 0.381 0.403 0.276 0.367 0.362 0.281 0.361 0.382 0.235 0.312 0.326 0.635 1.196 1.235 0.467 0.545
TFCL 0.472 0.548 0.592 0.198 0.341 0.363 0.240 0.363 0.382 - - - - - - 0.524 1.256 1.303 0.549 0.783
ER 0.376 0.543 0.571 0.197 0.333 0.351 0.244 0.356 0.363 0.247 0.299 0.307 0.16 0.264 0.281 0.506 1.057 1.074 0.418 0.474
MIR 0.410 0.541 0.565 0.197 0.325 0.342 0.244 0.355 0.361 - - - - - - 0.504 1.066 1.079 0.499 0.694
DER++ 0.375 0.540 0.577 0.192 0.326 0.340 0.235 0.351 0.359 0.241 0.289 0.295 0.154 0.256 0.273 0.421 1.035 1.048 0.406 0.459
FSNet 0.368 0.467 0.515 0.191 0.249 0.263 0.216 0.276 0.301 0.249 0.284 0.298 0.161 0.244 0.292 0.472 0.997 1.061 0.384 0.451
OneNet 0.352 0.412 0.435 0.197 0.225 0.267 0.201 0.261 0.278 0.202 0.269 0.299 0.203 0.269 0.395 0.341 0.578 0.609 0.322 0.364
D3A-TCN 0.381 0.495 0.516 0.206 0.322 0.331 0.266 0.358 0.370 0.251 0.279 0.296 0.199 0.273 0.289 0.343 0.484 0.481 0.341 0.357
D3A-FSNet 0.358 0.457 0.509 0.190 0.247 0.258 0.213 0.269 0.289 0.250 0.279 0.290 0.173 0.225 0.268 0.382 0.528 0.56 0.319 0.357
D3A-OneNet 0.349 0.401 0.431 0.187 0.225 0.261 0.196 0.253 0.268 0.202 0.266 0.276 0.141 0.200 0.242 0.291 0.365 0.388 0.275 0.296

III-C Importance of D3A

Concept drift detection and adaptation play a crucial role in addressing real-world needs, especially in the context of on-device continual learning and inference.

Real-World Relevance: The paradigm of concept drift detection aligns with the practical demands of on-device continual learning and inference, as highlighted in the work by [12]. For instance, VR/AR headsets can leverage continual learning to define play boundaries and recognize locations in the physical world for virtual overlays. Similarly, home robots, smart appliances, and smartphones need to adapt and learn about their environments and user preferences. In these scenarios, online learning is essential, but there are also significant periods of time during which offline memory consolidation is feasible, such as when a mobile device is being charged or when its owner is asleep.

Adaptation Amidst Device Busyness: Even during busy device operation, the need for full-model updates becomes imperative when a substantial concept drift occurs. As depicted in Figure 3, our model’s performance under a new concept might deteriorate significantly. In such cases, predictions become highly unreliable and may lead to adverse consequences, especially in high-risk tasks. Therefore, prioritizing model adjustments before deployment becomes essential, ensuring that the model is aligned with the current concept and avoiding potential negative impacts. This not only enhances the reliability of predictions but also ensures the safety and performance of intelligent systems across various applications.

IV Experiments

In this section, we present compelling evidence showcasing the efficacy of D3A. Firstly, we demonstrate that (1) the proposed model achieves superior forecasting performance, reducing MSE by over 30%30\% compared to the previous SOTA model. This underscores the model’s adeptness in detecting and adapting to concept drift. (2) we conduct comprehensive ablation studies, along with visualization and analysis, to elucidate the significance of each design choice. (3) we immerse ourselves in a more realistic and challenging evaluation setting, revealing D3A’s consistent improvements over all baselines. Lastly, we introduce a variant of D3A, denoted as D3A , which proves more efficient than D3A while still outperforming the previous SOTA model by a substantial margin.

IV-A Experimental setting

Datasets We investigate a diverse set of datasets for time series forecasting. ETT [37]333https://github.com/zhouhaoyi/ETDataset logs the target variable of the ”oil temperature” and six features of the power load over a two-year period. We also analyze the hourly recorded observations of ETTh2 and the 15-minute intervals of ETTm1 benchmarks. Additionally, we study ECL444https://archive.ics.uci.edu/ml/datasets/ElectricityLoadDiagrams20112014 (Electricity Consuming Load), which gathers electricity consumption data from 321 clients between 2012 and 2014. The WTH555https://www.ncei.noaa.gov/data/local-climatological-data/ dataset contains hourly records of 11 climate features from almost 1,600 locations across the United States. The Weather666https://www.bgc-jena.mpg.de/wetter/ is recorded every 10 minutes for 2020 whole year, which contains 21 meteorological indicators, such as air temperature, humidity. The Traffic777http://pems.dot.ca.gov/ is a collection of hourly data from California Department of Transportation, which describes the road occupancy rates measured by different sensors on San Francisco Bay area freeways.

Implementation Details For all benchmarks, we set the look-back window length at 60 and vary the forecast horizon from H=1,24,48H={1,24,48}. We split the data into two phases: warm-up and online training, with a ratio of 25:75. We follow the optimization details outlined in [37] and utilize the AdamW optimizer [20] to minimize the mean squared error (MSE) loss. To ensure a fair comparison, we set the epoch and batch sizes to one, which is consistent with the online learning setting. We make sure that all baseline models based on the TCN backbone use the same total memory budget as FSNet, which includes three times the network sizes: one working model and two exponential moving averages (EMAs) of its gradient. For ER, MIR, and DER++, we allocate an episodic memory to store previous samples to meet this budget. For transformer backbones, we find that a large number of parameters do not benefit the generalization results and always select the hyperparameters such that the number of parameters for transformer baselines is fewer than that for FSNet. In the warm-up phase, we calculate the mean and standard deviation to normalize the online training samples and perform hyperparameter cross-validation. For different structures, we use the optimal hyperparameters that are reported in the corresponding paper.

Baselines of adaptation methods We evaluate several baselines for our experiments, including methods for continual learning, time series forecasting, and online learning. Our first baseline is OnlineTCN [39], which continuously trains the model without any specific strategy. The second baseline is Experience Replay (ER) [9], where previous data is stored in a buffer and interleaved with newer samples during learning. Additionally, we consider three advanced variants of ER: TFCL [4], which uses a task-boundary detection mechanism and a knowledge consolidation strategy; MIR [3], which selects samples that cause the most forgetting; and DER++ [8], which incorporates a knowledge distillation strategy. It is worth noting that ER and its variants are strong baselines in the online setting, as we leverage mini-batches during training to reduce noise from single samples and achieve faster and better convergence. Finally, we compare our method to FSNet [22] and OneNet [36], which are the previous state-of-the-art online adaptation model and strategy respectively. For OneNet, we use the average version for clarity, namely the combination strategy of two branches is just averaging.

Metrics Because learning occurs over a sequence of rounds. At each round, the model receives a look-back window and predicts the forecast window. All models are commonly evaluated by their accumulated mean-squared errors (MSE) and mean-absolute errors (MAE); namely, the model is evaluated based on its accumulated errors over the entire learning.

License. All the assets (i.e., datasets and the codes for baselines) we use include an MIT license containing a copyright notice and this permission notice shall be included in all copies or substantial portions of the software.

Environment. We conduct all the experiments on a machine with an Intel R Xeon (R) Platinum 8163 CPU @ 2.50GHZ, 32G RAM, and four Tesla-V100 (32G) instances. All experiments are repeated 33 times with different seeds.

IV-B Baseline details

We present a brief overview of the baselines employed in our experiments.

First, OnlineTCN adopts a conventional TCN backbone [39] consisting of ten hidden layers, each layer containing two stacks of residual convolution filters.

Secondly, ER [9] expands on the OnlineTCN baseline by adding an episodic memory that stores previous samples and interleaves them during the learning process with newer ones.

Third, MIR [3] replaces the random sampling technique of ER with its MIR sampling approach, which selects the samples in the memory that cause the highest forgetting and applies ER to them.

Fourthly, DER++ [8] enhances the standard ER method by incorporating a knowledge distillation loss on the previous logits.

Finally, TFCL [4] is a task-free, online continual learning method that starts with an ER process and includes a task-free MAS-styled [2] regularization.

All the ER-based techniques utilize a reservoir sampling buffer, which is identical to that used in [22].

Hyper-parameters For the hyper-parameters of FSNet and the baselines mentioned in Section IV-B, we follow the setting in [22]. Besides, we cross-validate the hyper-parameters on the ETTh2 dataset and use them for the remaining ones. In particular, we use the following configuration:

  • Learning rate 3e33e-3 on Traffic and ECL and 1e31e-3 for other datasets. Learning rate 1e21e-2 for the EGD algorithm and 1e31e-3 for the offline reinforcement learning block, where the selection scope is {1e3,3e3,1e2,3e2}\{1e-3,3e-3,1e-2,3e-2\}.

  • Number of hidden layers 1010 for both cross-time and cross-variable branches, where the selection scope is {6,8,10,12}\{6,8,10,12\}.

  • Adapter’s EMA coefficient 0.90.9, Gradient EMA for triggering the memory interaction 0.30.3, where the selection scope is {0.1,0.2,,1.0}\{0.1,0.2,\dots,1.0\}.

  • Memory triggering threshold 0.750.75, where the selection scope is {0.6,0.65,0.7,0.9}\{0.6,0.65,0.7\dots,0.9\}.

  • Episodic memory size: 5000 (for ER, MIR, and DER++), 50 (for TFCL).

IV-C Online forecasting results

Cumulative Performance: This section explores the cumulative performance of various baselines, as illustrated in Table I, focusing on mean-squared errors (MSE) and mean-absolute errors (MAE). Notably, the D3A-OneNet, a novel approach that deploys OneNet with our D3A strategy, outperforms the majority of competing baselines across diverse forecasting horizons. Moreover, the flexibility of D3A enables seamless integration with advanced online forecasting methods and representation learning structures, enhancing the model’s overall robustness. Interestingly, in datasets with higher forecasting error rates, D3A demonstrates superior performance. For instance, when applied to the challenging ECL dataset with a forecasting horizon of 48, D3A achieves a remarkable improvement of 52.2%52.2\% over TCN and a substantial improvement of 46.2%46.2\% over FSNet. The average MSE and MAE values of D3A consistently outshine those obtained using traditional online learning strategies in isolation. This highlights the pivotal role played by D3A in incorporating concept drift detection and bias correlation during adaptation, underscoring its significance in achieving superior predictive accuracy.

TABLE II: MSE (top) and MAE (bottom)of various adaptation methods with delayed feedback. H: forecast horizon. Avg-C: the average of metric values under challenging benchmarks including ETTH2, Weather, and ECL.
ETTH2 ETTm1 WTH Traffic Weather ECL
MSE 1 24 48 1 24 48 1 24 48 1 24 48 1 24 48 1 24 48 Avg Avg-C
OnlineTCN 0.502 5.871 11.074 0.214 0.410 0.535 0.206 0.429 0.504 0.331 0.629 0.804 0.371 0.999 1.565 2.900 11.101 24.159 3.478 6.505
ER 0.508 5.461 17.329 0.086 0.367 0.498 0.180 0.373 0.435 0.286 0.471 0.567 0.312 0.896 1.389 2.388 8.205 19.528 3.293 6.224
DER++ 0.508 5.387 17.334 0.083 0.347 0.465 0.174 0.369 0.431 0.281 0.469 0.564 0.307 0.836 1.281 2.657 7.878 17.692 3.170 5.987
D3A-TCN 0.401 4.752 9.300 0.091 0.319 0.381 0.209 0.383 0.493 0.308 0.549 0.800 0.370 0.826 1.038 2.341 6.681 13.582 2.379 4.366
FSNet 0.466 6.511 12.624 0.085 0.383 0.502 0.162 0.382 0.489 0.290 0.478 0.576 0.309 0.822 1.286 3.143 9.094 30.523 3.785 7.198
D3A-FSNet 0.402 5.712 8.300 0.085 0.336 0.398 0.163 0.326 0.406 0.278 0.433 0.532 0.324 0.795 1.087 2.456 7.168 16.390 2.533 4.737
OneNet 0.372 2.807 5.884 0.087 0.322 0.368 0.154 0.323 0.391 0.223 0.426 0.514 0.296 0.722 1.193 2.472 5.495 11.363 1.856 3.400
D3A-OneNet 0.377 2.458 4.323 0.085 0.285 0.333 0.152 0.319 0.390 0.213 0.391 0.474 0.286 0.711 0.964 2.179 4.865 7.192 1.444 2.595
ETTH2 ETTm1 WTH Traffic Weather ECL
MAE 1 24 48 1 24 48 1 24 48 1 24 48 1 24 48 1 24 48 Avg Avg-C
OnlineTCN 0.436 1.109 1.348 0.240 0.511 0.548 0.276 0.465 0.508 0.282 0.414 0.495 0.229 0.575 0.787 0.611 1.141 1.076 0.614 0.812
ER 0.376 0.976 1.651 0.197 0.456 0.525 0.244 0.421 0.459 0.248 0.311 0.363 0.161 0.505 0.710 0.506 0.606 0.790 0.528 0.698
DER++ 0.375 0.967 1.644 0.192 0.443 0.508 0.235 0.415 0.456 0.241 0.309 0.359 0.154 0.475 0.671 0.421 0.591 0.758 0.512 0.673
D3A-TCN 0.377 0.925 1.171 0.205 0.411 0.448 0.271 0.421 0.493 0.259 0.369 0.467 0.190 0.459 0.552 0.335 0.569 0.603 0.474 0.576
FSNet 0.368 0.989 1.472 0.191 0.468 0.512 0.216 0.467 0.518 0.252 0.320 0.371 0.160 0.449 0.665 0.472 0.923 1.391 0.493 0.765
D3A-FSNet 0.366 1.000 1.173 0.192 0.428 0.461 0.212 0.385 0.446 0.254 0.304 0.356 0.160 0.429 0.569 0.371 0.876 1.112 0.444 0.673
OneNet 0.352 0.743 0.993 0.196 0.415 0.445 0.201 0.381 0.436 0.202 0.294 0.342 0.141 0.396 0.621 0.341 0.539 0.865 0.406 0.555
D3A-OneNet 0.349 0.720 0.852 0.193 0.388 0.417 0.182 0.377 0.433 0.198 0.277 0.323 0.137 0.418 0.512 0.282 0.515 0.579 0.379 0.485

Explore more efficient adaptation strategies. As discussed in subsection III-C, D3A plays a crucial role in addressing significant concept drift even when the model is occupied. However, we aim to enhance the efficiency of this process. Initially, we propose that, during the adaptation phase, only the linear regressor undergoes full fine-tuning until convergence, reducing backward computation time and avoiding the need for gradient storage from the resource-intensive encoder. However, a simple retraining approach proves ineffective, leading to prolonged and challenging convergence. To expedite the learning process, we employ a large learning rate (1e31e-3 by default) for warm-up, decaying it by a factor of 33 when the average training losses in one epoch do not decrease. This strategy, termed D3A , achieves a favorable balance between effectiveness and efficiency. As depicted in Figure 4, on two representative datasets, the inference cost of D3A is significantly lower than that of D3A and comparable to the baseline models. Simultaneously, the performance of D3A does not compromise much compared to D3A.

Refer to caption
(a) MSE
Refer to caption
(b) Inference Time
Figure 4: Comparison of MSE and Inference Time among the naive method, augmented with D3A, and enhanced D3A

Forecasting results visualization In contrast to baseline models that encounter challenges in adapting to new concepts and yield subpar forecasting outcomes, D3A adeptly captures intricate patterns within time series data. As illustrated in Figure 5, the baseline model struggles to dynamically adjust to evolving data patterns, leading to prolonged overfitting to the existing trend. For instance, when FSNet encounters a sequence of data instances depicting an increasing trend, it tends to persistently retain this trend for an extended duration. Conversely, D3A actively monitors forecasting losses, prompting timely updates to the model parameters whenever they are deemed inadequate for the current data pattern. This dynamic adaptation capability enables D3A to navigate shifting trends more effectively, resulting in improved forecasting accuracy.

Refer to caption
(a) Channel 0.
Refer to caption
(b) Channel 1.
Figure 5: Visualizing the model’s prediction on the ECL dataset.

IV-D Online forecasting results with delayed feedback

This paper adheres to the experimental setup introduced in FSNet [22] and OneNet [36], where the true values of each time step are disclosed to enhance the model’s performance in subsequent rounds. Specifically, at time step tt, with a look-back window LL: (𝐱i)i=tL+t(\mathbf{x}_{i})_{i=t}^{L+t} and forecasting window HH: (𝐱i)i=L+1+tL+H+t(\mathbf{x}_{i})_{i=L+1+t}^{L+H+t}. Post-evaluation, the true values of (𝐱i)i=L+1+tL+H+t(\mathbf{x}_{i})_{i=L+1+t}^{L+H+t} are observed and utilized for training the model. Subsequently, at time step t+1t+1, the input becomes (𝐱i)i=t+1L+t+1(\mathbf{x}_{i})_{i=t+1}^{L+t+1} and the task is to forecast (𝐱i)i=L+2+tL+H+t+1(\mathbf{x}_{i})_{i=L+2+t}^{L+H+t+1}. Notably, there is a significant overlap in the forecasting target, with only one new point added. It is crucial to highlight that in real-world scenarios, the true values for the forecast horizon HH may not be accessible until HH rounds later, introducing the concept of online forecasting with delayed feedback. This scenario presents increased challenges and efficiency constraints, as the model cannot undergo retraining at each round, and training is restricted to every HH rounds. Tables II presents cumulative performance metrics, considering MSE and MAE. As expected, all methods exhibit reduced performance under delayed feedback compared to the traditional online forecasting setting. Notably, the SOTA method FSNet shows sensitivity to delayed feedback, particularly when H=48H=48, falling behind a simple TCN baseline on specific datasets. Our proposed method, D3A, proves advantageous across all baselines, particularly on the challenging datasets like ETTH2, Weather, and ECL. Specifically, concerning the average MSE on the challenging subset, D3A reduces TCN’s MSE by 𝟑𝟑%\mathbf{33\%}, while achieving performance gains of 34.2%\mathbf{34.2\%} and 23.7%\mathbf{23.7\%} for FSNet and OneNet, respectively. Additionally, D3A outperforms existing online learning baselines by a significant margin.

IV-E Ablation studies and analysis

Refer to caption
(a) Loss weight λ\lambda.
Refer to caption
(b) Memory size lwl_{w}.
Refer to caption
(c) Confidence level αt\alpha_{t}.
Figure 6: Ablation studies of loss weight and memory size.

Ablation studies of different hyper-parameters. We present ablation studies investigating the impact of different hyperparameters, specifically the loss weight λ\lambda during adaptation and the memory size lwl_{w}. As depicted in Figure 6a, various datasets exhibit preferences for distinct choices. Overall, datasets with a greater number of variables tend to benefit more from the proposed augmentation strategy. This is intuitively explained by the challenging nature of learning the ECL dataset, and our augmentation strategy effectively utilizes debiased historical data to expedite the learning process. For the ETTh2 dataset, a smaller weight of 0.10.1 yields optimal performance. This suggests that only a modest regularization is necessary, and training the model solely on the drifted instances allows for quick adaptation to new concepts in these datasets. Consequently, for datasets with a substantial number of variables, we default to λ=2.0\lambda=2.0, and 0.10.1 otherwise. Regarding memory size, we observe a tradeoff, as a small memory size proves sensitive to concept drift and noise, while a large memory may not effectively signal when concept drift occurs. For simplicity, we choose lw=16l_{w}=16 as the default memory size for all datasets.

The confidence level is represented by the threshold αt\alpha_{t} rather than the direct use of α\alpha. Given a specific value for αt\alpha_{t}, we calculate the corresponding threshold α\alpha such that the area under the normal curve beyond this threshold aligns with the specified significance level (αt\alpha_{t}). Utilizing the factor α=1αt/2\alpha=1-\alpha_{t}/2 is essential for a two-tailed test. In our ablation studies, it becomes evident that as the confidence level decreases, we impose a tighter constraint on drift detection. Only when the shift in the loss distribution is significant enough does the detector trigger an alarm. Consequently, the occurrence of drift alarms and the total re-training instances are reduced. Regarding model performance, our observations indicate that more frequent retraining does not consistently enhance performance. In contrast, optimal model performance is achieved when retraining occurs judiciously based on proper data and timing considerations.

Detecting Concept Drift Events: When Does the Concept Detector Trigger an Alarm? We explore the scenarios in which our concept detector initiates a concept drift alarm by visualizing data instances and losses. The model is trained on the ECL dataset with HH set to 4848. As illustrated in Figure 7, in cases 1 and 2, discernible patterns emerge simultaneously in both data distribution and losses, prompting the detector to raise an alarm—an intuitive response. Intriguingly, in case 3, data patterns may not be as straightforward for human observers, yet our detector can discern a significant increase on the forecasting error, prompting the model to adapt to the new concept.

Refer to caption
(a) Case 1. Data instance.
Refer to caption
(b) Case 1. Error.
Refer to caption
(c) Case 2. Data instance.
Refer to caption
(d) Case 2. Error.
Refer to caption
(e) Case 3. Data instance.
Refer to caption
(f) Case 3. Error.
Figure 7: Visualizing instances and losses where the detector triggers a concept drift alarm in several cases, where lw=64l_{w}=64.

V Conclusion and Limitations

This paper introduces a novel online learning framework, Concept Drift Detection and Adaptation (D3A), designed to tackle the challenges posed by significant concept drifts in time series forecasting. Our approach integrates a concept drift detector, which assesses the divergence between online testing and training distributions, with a data augmentation strategy aimed at enhancing adaptation using historical data. The proposed methodology showcases a substantial improvement in model adaptation capability, effectively reducing forecasting errors across a range of datasets and models. While acknowledging the introduction of additional computational burdens, we thoroughly discuss the necessity of D3A, particularly in high-risk tasks. Additionally, we propose a trivial method that strikes a better balance between efficiency and effectiveness.

Despite the promising outcomes, it is imperative to recognize certain limitations. The effectiveness of the proposed method may be influenced by the choice of hyperparameters and the specific characteristics of the datasets. Furthermore, the performance of the data augmentation strategy might exhibit variability in scenarios with limited historical data or under specific noise conditions. These considerations are crucial when applying D3A to different real-world applications. Looking ahead, exploring the adaptability of D3A across diverse application domains and datasets with distinct characteristics will contribute to a more comprehensive understanding of its generalizability.

References

  • [1] Davide Abati, Angelo Porrello, Simone Calderara, and Rita Cucchiara. Latent space autoregression for novelty detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 481–490, 2019.
  • [2] Rahaf Aljundi, Francesca Babiloni, Mohamed Elhoseiny, Marcus Rohrbach, and Tinne Tuytelaars. Memory aware synapses: Learning what (not) to forget. In Proceedings of the European conference on computer vision (ECCV), pages 139–154, 2018.
  • [3] Rahaf Aljundi, Eugene Belilovsky, Tinne Tuytelaars, Laurent Charlin, Massimo Caccia, Min Lin, and Lucas Page-Caccia. Online continual learning with maximal interfered retrieval. In Advances in Neural Information Processing Systems 32, pages 11849–11860. 2019.
  • [4] Rahaf Aljundi, Klaas Kelchtermans, and Tinne Tuytelaars. Task-free continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11254–11263, 2019.
  • [5] Oren Anava, Elad Hazan, Shie Mannor, and Ohad Shamir. Online learning for time series prediction. In Conference on learning theory, pages 172–184. PMLR, 2013.
  • [6] George EP Box and David A Pierce. Distribution of residual autocorrelations in autoregressive-integrated moving average time series models. Journal of the American statistical Association, 65(332):1509–1526, 1970.
  • [7] Markus M Breunig, Hans-Peter Kriegel, Raymond T Ng, and Jörg Sander. Lof: identifying density-based local outliers. In Proceedings of the 2000 ACM SIGMOD international conference on Management of data, pages 93–104, 2000.
  • [8] Pietro Buzzega, Matteo Boschini, Angelo Porrello, Davide Abati, and Simone Calderara. Dark experience for general continual learning: a strong, simple baseline. Advances in neural information processing systems, 33:15920–15930, 2020.
  • [9] Arslan Chaudhry, Marcus Rohrbach, Mohamed Elhoseiny, Thalaiyasingam Ajanthan, Puneet K Dokania, Philip HS Torr, and Marc’Aurelio Ranzato. On tiny episodic memories in continual learning. arXiv preprint arXiv:1902.10486, 2019.
  • [10] Junyoung Chung, Caglar Gulcehre, KyungHyun Cho, and Yoshua Bengio. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555, 2014.
  • [11] Alex Graves and Alex Graves. Long short-term memory. Supervised sequence labelling with recurrent neural networks, pages 37–45, 2012.
  • [12] Md Yousuf Harun, Jhair Gallardo, Tyler L Hayes, Ronald Kemker, and Christopher Kanan. Siesta: Efficient online continual learning with sleep. arXiv preprint arXiv:2303.10725, 2023.
  • [13] Dan Hendrycks and Kevin Gimpel. A baseline for detecting misclassified and out-of-distribution examples in neural networks. arXiv preprint arXiv:1610.02136, 2016.
  • [14] Kimin Lee, Kibok Lee, Honglak Lee, and Jinwoo Shin. A simple unified framework for detecting out-of-distribution samples and adversarial attacks. Advances in neural information processing systems, 31, 2018.
  • [15] Wendi Li, Xiao Yang, Weiqing Liu, Yingce Xia, and Jiang Bian. Ddg-da: Data distribution generation for predictable concept drift adaptation. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 36, pages 4092–4100, 2022.
  • [16] Zhihan Li, Youjian Zhao, Jiaqi Han, Ya Su, Rui Jiao, Xidao Wen, and Dan Pei. Multivariate time series anomaly detection and interpretation using hierarchical inter-metric and temporal embedding. In Proceedings of the 27th ACM SIGKDD conference on knowledge discovery & data mining, pages 3220–3230, 2021.
  • [17] Bryan Lim and Stefan Zohren. Time-series forecasting with deep learning: a survey. Philosophical Transactions of the Royal Society A, 379(2194):20200209, 2021.
  • [18] Chenghao Liu, Steven CH Hoi, Peilin Zhao, and Jianling Sun. Online arima algorithms for time series prediction. In Proceedings of the AAAI conference on artificial intelligence, volume 30, 2016.
  • [19] Weitang Liu, Xiaoyun Wang, John Owens, and Yixuan Li. Energy-based out-of-distribution detection. Advances in Neural Information Processing Systems, 33:21464–21475, 2020.
  • [20] Ilya Loshchilov and Frank Hutter. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101, 2017.
  • [21] Yuqi Nie, Nam H. Nguyen, Phanwadee Sinthong, and Jayant Kalagnanam. A time series is worth 64 words: Long-term forecasting with transformers. ICLR, 2023.
  • [22] Quang Pham, Chenghao Liu, Doyen Sahoo, and Steven CH Hoi. Learning fast and slow for online time series forecasting. ICLR, 2023.
  • [23] Tiexin Qin, Shiqi Wang, and Haoliang Li. Generalizing to evolving domains with latent structure-aware sequential autoencoder. In International Conference on Machine Learning, pages 18062–18082. PMLR, 2022.
  • [24] Yao Qin, Dongjin Song, Haifeng Chen, Wei Cheng, Guofei Jiang, and Garrison Cottrell. A dual-stage attention-based recurrent neural network for time series prediction. arXiv preprint arXiv:1704.02971, 2017.
  • [25] Bernhard Schölkopf, John C Platt, John Shawe-Taylor, Alex J Smola, and Robert C Williamson. Estimating the support of a high-dimensional distribution. Neural computation, 13(7):1443–1471, 2001.
  • [26] Rajat Sen, Hsiang-Fu Yu, and Inderjit S Dhillon. Think globally, act locally: A deep neural network approach to high-dimensional time series forecasting. Advances in neural information processing systems, 32, 2019.
  • [27] David MJ Tax and Robert PW Duin. Support vector data description. Machine learning, 54:45–66, 2004.
  • [28] Alexey Tsymbal. The problem of concept drift: definitions and related work. Computer Science Department, Trinity College Dublin, 106(2):58, 2004.
  • [29] Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  • [30] Haoqi Wang, Zhizhong Li, Litong Feng, and Wayne Zhang. Vim: Out-of-distribution with virtual-logit matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4921–4930, 2022.
  • [31] Qingsong Wen, Tian Zhou, Chaoli Zhang, Weiqi Chen, Ziqing Ma, Junchi Yan, and Liang Sun. Transformers in time series: A survey. arXiv preprint arXiv:2202.07125, 2022.
  • [32] Zhisheng Xiao, Qing Yan, and Yali Amit. Likelihood regret: An out-of-distribution detection score for variational auto-encoder. Advances in neural information processing systems, 33:20685–20696, 2020.
  • [33] Jiehui Xu, Haixu Wu, Jianmin Wang, and Mingsheng Long. Anomaly transformer: Time series anomaly detection with association discrepancy. arXiv preprint arXiv:2110.02642, 2021.
  • [34] Xiaoyu You, Mi Zhang, Daizong Ding, Fuli Feng, and Yuanmin Huang. Learning to learn the future: Modeling concept drifts in time series prediction. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, pages 2434–2443, 2021.
  • [35] YiFan Zhang, Xue Wang, Tian Zhou, Kun Yuan, Zhang Zhang, Liang Wang, Rong Jin, and Tieniu Tan. Model-free test time adaptation for out-of-distribution detection. arXiv preprint arXiv:2311.16420, 2023.
  • [36] YiFan Zhang, Qingsong Wen, Xue Wang, Weiqi Chen, Liang Sun, Zhang Zhang, Liang Wang, Rong Jin, and Tieniu Tan. Onenet: Enhancing time series forecasting models under concept drift by online ensembling. In Thirty-seventh Conference on Neural Information Processing Systems (NeurIPS), 2023.
  • [37] Haoyi Zhou, Shanghang Zhang, Jieqi Peng, Shuai Zhang, Jianxin Li, Hui Xiong, and Wancai Zhang. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the AAAI conference on artificial intelligence, volume 35, pages 11106–11115, 2021.
  • [38] Tian Zhou, Ziqing Ma, Qingsong Wen, Xue Wang, Liang Sun, and Rong Jin. Fedformer: Frequency enhanced decomposed transformer for long-term series forecasting. In International Conference on Machine Learning, pages 27268–27286. PMLR, 2022.
  • [39] Martin Zinkevich. Online convex programming and generalized infinitesimal gradient ascent. In Proceedings of the 20th international conference on machine learning (icml-03), pages 928–936, 2003.
  • [40] Bo Zong, Qi Song, Martin Renqiang Min, Wei Cheng, Cristian Lumezanu, Daeki Cho, and Haifeng Chen. Deep autoencoding gaussian mixture model for unsupervised anomaly detection. In International conference on learning representations, 2018.

VI Proofs of Theoretical Statements

Theorem 1. If |Δ(ΣA)|21/2|\Delta(\Sigma_{A})|_{2}\leq 1/2, we have

Ex𝒫[|(wwA)x|2]40|Δ(ΣA)|22\mathrm{E}_{x\sim\mathcal{P}}\left[\left|(w_{*}-w_{A})^{\top}x\right|^{2}\right]\leq 4\mathcal{L}_{0}|\Delta(\Sigma_{A})|_{2}^{2}

where 0=zAΣ1zA\mathcal{L}_{0}=z_{A}^{\top}\Sigma^{-1}z_{A}.

Proof.

We have

Ex𝒫[|(wwA)x|2]\displaystyle\mathrm{E}_{x\sim\mathcal{P}}\left[\left|(w_{*}-w_{A})^{\top}x\right|^{2}\right]
=\displaystyle= tr((wwA)(wwA)Σ)\displaystyle\mbox{tr}\left((w_{*}-w_{A})(w_{*}-w_{A})^{\top}\Sigma\right)
=\displaystyle= zA[Σ1ΣA1]Σ[Σ1ΣA1]zA\displaystyle z_{A}^{\top}\left[\Sigma^{-1}-\Sigma_{A}^{-1}\right]\Sigma\left[\Sigma^{-1}-\Sigma_{A}^{-1}\right]z_{A}
=\displaystyle= zAΣ1/2[IΣ1/2ΣA1Σ1/2]2Σ2zA\displaystyle z_{A}^{\top}\Sigma^{-1/2}\left[I-\Sigma^{1/2}\Sigma_{A}^{-1}\Sigma^{1/2}\right]^{2}\Sigma^{-2}z_{A}
=\displaystyle= zAΣ1/2[I(IΔ(ΣA))1]2Σ2zA\displaystyle z_{A}^{\top}\Sigma^{-1/2}\left[I-(I-\Delta(\Sigma_{A}))^{-1}\right]^{2}\Sigma^{-2}z_{A}

When |Δ(ΣA)|2<1/2|\Delta(\Sigma_{A})|_{2}<1/2, we

minu(I+2Δ(ΣA))(IΔ(ΣA))u|u|2=mini[d]1+λi(Δ(ΣA))2λi2(Δ(ΣA))1\min\frac{u^{\top}(I+2\Delta(\Sigma_{A}))(I-\Delta(\Sigma_{A}))u}{|u|^{2}}=\min\limits_{i\in[d]}1+\lambda_{i}(\Delta(\Sigma_{A}))-2\lambda^{2}_{i}(\Delta(\Sigma_{A}))\geq 1

and therefore (I+2Δ(ΣA))(IΔ(ΣA))I(I+2\Delta(\Sigma_{A}))(I-\Delta(\Sigma_{A}))\succeq I or

(IΔ(ΣA))1I+2Δ(ΣA))(I-\Delta(\Sigma_{A}))^{-1}\preceq I+2\Delta(\Sigma_{A}))

Since (IΔ(ΣA))1I(I-\Delta(\Sigma_{A}))^{-1}\succeq I, we have

Ex𝒫[|(wwA)x|2]4|Δ(ΣA)|22zAΣ1zA:=0\mathrm{E}_{x\sim\mathcal{P}}\left[|(w_{*}-w_{A})^{\top}x|^{2}\right]\leq 4|\Delta(\Sigma_{A})|^{2}_{2}\underbrace{z_{A}^{\top}\Sigma^{-1}z_{A}}_{:=\mathcal{L}_{0}}

Proposition 1. Assume αβ\alpha\geq\beta and maxi[k]νi[αβ,2(αβ)]\max_{i\in[k]}\nu_{i}\in[\alpha-\beta,2(\alpha-\beta)]. We have

|Δ(ΣA)|2=γ(αβ)τ|\Delta(\Sigma_{A})|_{2}=\frac{\gamma(\alpha-\beta)}{\tau}
Proof.

Using the expressions for ΣB\Sigma_{B} and ΣA\Sigma_{A} in (3) and (4), we can rewrite Δ\Delta as

Δ(ΣA)\displaystyle\Delta(\Sigma_{A}) =\displaystyle= γ(τI+γUdiag(ν)U)1/2((αβ)IUdiag(ν)U)(τI+γUdiag(α)U)1/2\displaystyle\gamma\left(\tau I+\gamma U\mbox{diag}(\nu)U^{\top}\right)^{-1/2}\left((\alpha-\beta)I-U\mbox{diag}(\nu)U^{\top}\right)\left(\tau I+\gamma U\mbox{diag}(\alpha)U^{\top}\right)^{-1/2}
=\displaystyle= I+γ(αβ+τγ)(τI+γUdiag(ν)U)1\displaystyle-I+\gamma\left(\alpha-\beta+\frac{\tau}{\gamma}\right)\left(\tau I+\gamma U\mbox{diag}(\nu)U^{\top}\right)^{-1}

where τ=(1γ)β+γα\tau=(1-\gamma)\beta+\gamma\alpha. Since α>β\alpha>\beta, we have

|Δ(ΣA)|2=max(γ(αβ)τ,1γ(αβ)+ττ+γ|ν|)|\Delta(\Sigma_{A})|_{2}=\max\left(\frac{\gamma(\alpha-\beta)}{\tau},1-\frac{\gamma(\alpha-\beta)+\tau}{\tau+\gamma|\nu|_{\infty}}\right)

where |ν||\nu|_{\infty} is infinity norm of vector ν\nu. When |ν|[αβ,2(αβ)]|\nu|_{\infty}\in[\alpha-\beta,2(\alpha-\beta)], it is easy to verify that

1γ(αβ)+ττ+γ|ν|=γ(|ν|[αβ])τ+γ|ν|γ(αβ)τ+γ|ν|<γ(αβ)τ1-\frac{\gamma(\alpha-\beta)+\tau}{\tau+\gamma|\nu|_{\infty}}=\frac{\gamma\left(|\nu|_{\infty}-[\alpha-\beta]\right)}{\tau+\gamma|\nu|_{\infty}}\leq\frac{\gamma(\alpha-\beta)}{\tau+\gamma|\nu|_{\infty}}<\frac{\gamma(\alpha-\beta)}{\tau}

and therefore

|Δ(ΣA)|2=γ(αβ)τ|\Delta(\Sigma_{A})|_{2}=\frac{\gamma(\alpha-\beta)}{\tau}

Theorem 2. Assume

α2γ3γβ,αβ|ν|min(2(αβ),τ(3γ(αβ)β)τγ(αβ))\alpha\geq\frac{2-\gamma}{3-\gamma}\beta,\quad\alpha-\beta\leq|\nu|_{\infty}\leq\min\left(2(\alpha-\beta),\frac{\tau\left(3\gamma(\alpha-\beta)-\beta\right)}{\tau-\gamma(\alpha-\beta)}\right)

we have

|Δ(ΣA)|2|Δ(ΣA)|2|\Delta(\Sigma_{A^{\prime}})|_{2}\leq|\Delta(\Sigma_{A})|_{2}
Proof.

Plugging the expression for ΣA\Sigma_{A^{\prime}} (i.e. 5) into the expression for Δ(ΣA\Delta(\Sigma_{A^{\prime}}, we have

Δ(ΣA)\displaystyle\Delta(\Sigma_{A^{\prime}}) =\displaystyle= (τI+γUdiag(ν)U)1/2(γ(αβ)I[τI+Udiag(ν)U])(τI+γUdiag(α)U)1/2\displaystyle\left(\tau I+\gamma U\mbox{diag}(\nu)U^{\top}\right)^{-1/2}\left(\gamma(\alpha-\beta)I-\left[\tau I+U\mbox{diag}(\nu)U^{\top}\right]\right)\left(\tau I+\gamma U\mbox{diag}(\alpha)U^{\top}\right)^{-1/2}
=\displaystyle= (γ(αβ)+(1γ)τ)(τI+γUdiag(ν)U)11\displaystyle\left(\gamma(\alpha-\beta)+(1-\gamma)\tau\right)\left(\tau I+\gamma U\mbox{diag}(\nu)U^{\top}\right)^{-1}-1

Since |ν|>αβ|\nu|_{\infty}>\alpha-\beta, we have

|Δ(ΣA)|2=1γ(αβ)+(1γ)ττ+γ|ν||\Delta(\Sigma_{A^{\prime}})|_{2}=1-\frac{\gamma(\alpha-\beta)+(1-\gamma)\tau}{\tau+\gamma|\nu|_{\infty}}

Since

|ν|τ(2[αβ]τ)τγ(αβ)|\nu|_{\infty}\leq\frac{\tau(2[\alpha-\beta]-\tau)}{\tau-\gamma(\alpha-\beta)}

we have

|Δ(ΣA)|2<γ(αβ)τ=|Δ(ΣA)|2|\Delta(\Sigma_{A^{\prime}})|_{2}<\frac{\gamma(\alpha-\beta)}{\tau}=|\Delta(\Sigma_{A})|_{2}

leading to less discrepancy in prediction between wAw_{A^{\prime}} and ww_{*}. ∎