This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Adaptive State-Space Multitaper Spectral Estimation

Andrew H. Song,  Seong-Eun Kim,  and Emery N. Brown These authors contributed equally to this work.This work was partially supported by NRF grants (2020R1C1C1011857, 2020M3C1B8081320, 2019S1A5A2A03053308), and by NIH grant P01-GM118629 and the JPB Foundation (Corresponding author: Seong-Eun Kim)A. H. Song is with Electrical Engineering and Computer Science, Massachusetts Institute of Technology, Cambridge, MA 02139, USA (e-mail:[email protected]).S.-E. Kim is with the Department of Applied Artificial Intelligence, Seoul National University of Science and Technology, Nowon-gu, Seoul 01811, South Korea (e-mail:[email protected]).E. N. Brown is with the Picower Institute for Learning and Memory, Massachusetts Institute of Technology, Cambridge, MA 02139 and the Department of Anesthesia, Critical Care, and Pain Medicine, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA (e-mail:[email protected]).
Abstract

Short-time Fourier transform (STFT) is the most common window-based approach for analyzing the spectrotemporal dynamics of time series. To mitigate the effects of high variance on the spectral estimates due to finite-length, independent STFT windows, state-space multitaper (SSMT) method used a state-space framework to introduce dependency among the spectral estimates. However, the assumed time-invariance of the state-space parameters makes the spectral dynamics difficult to capture when the time series is highly nonstationary. We propose an adaptive SSMT (ASSMT) method as a time-varying extension of SSMT. ASSMT tracks highly nonstationary dynamics by adaptively updating the state parameters and Kalman gains using a heuristic, computationally efficient exponential smoothing technique. In analyses of simulated data and real human electroencephalogram (EEG) recordings, ASSMT showed improved denoising and smoothing properties relative to standard multitaper and SSMT approaches.

Index Terms:
State-space model, spectral estimation, multitaper method, adaptive estimation, Kalman filter

This manuscript is an extended version of the IEEE Signal Processing Letters paper (doi:10.1109/LSP.2022.3142670), with the supplementary material as the appendix.

I Introduction

Nonstationary time series with time-varying probability structures are ubiquitous. Some examples include speech and image recordings [1, 2], oceanographic and seismic signals [3], neural spike trains [4], and electroencephalogram (EEG) [5]. We are interested in analyzing the nonstationary data through the lens of the time-varying spectral dynamics, which yields valuable information on the underlying system. The traditional approach has been to segment the data into independent overlapping or non-overlapping windows, assuming local stationarity [6] within each window, and to apply Fourier or wavelet transform [7, 8].

Despite the popularity, the windowing approach suffers from spectral estimates of high variance within each window, due to finite window-length [9] and the restrictive independence assumption for different windows. The state-space multitaper (SSMT) framework [10] and its extensions [11, 12, 13, 14] have been proposed as the solutions, by positing a time-invariant latent state-space model in the time-frequency domain, with each state representing the Fourier coefficients in each window. Since the states are linked by stochastic continuity constraint across windows, the spectral estimates are not independent.

However, the use of time-invariant parameters limits the capacity of SSMT to track strong nonstationarity that commonly occurs in systems neuroscience and manifests as strong spectral fluctuations due to fluctuations in system properties. Although time-varying state-space model offers an obvious solution, traditional parameter estimation approaches based on expectation-maximization (EM) algorithm [15] makes it less practical for real-time applications - The parameters must be re-estimated with every incoming batch of data, and thus incurs high computational costs.

We propose a time-varying extension of SSMT with an efficient and adaptive parameter estimation scheme, termed an adaptive SSMT (ASSMT) method. Based on a novel nonstationarity metric derived from spectral fluctuation, ASSMT adaptively adjusts the state parameters and consequently Kalman gain. This allows ASSMT to switch between removing background noise and tracking spectral fluctuation in a data-driven manner. The estimation for time-varying parameters requires only single pass through the data, making it well suited for real-time applications. The paper is organized as follows. In Section II, we review the SSMT framework. In Section III, we formulate the ASSMT framework. In Section IV and V, the results and conclusion are presented.

II Review of State-Space Multitaper Method

We first review the SSMT algorithm in [10]. Consider a nonstationary time series yty_{t} sampled at frequency fsf_{s} as

yt=xt+εt,t=1,,T\displaystyle y_{t}=x_{t}+\varepsilon_{t},\quad t=1,\ldots,T (1)

where xtx_{t} is a locally stationary latent Gaussian process [16] with the measurement noise εt𝒩(0,σε2)\varepsilon_{t}\sim\mathcal{N}(0,\sigma_{\varepsilon}^{2}). Leveraging the local stationary property, we divide these signals into KK nonoverlapping stationary intervals of JJ samples, such that T=KJT=KJ. The segmented vectors for interval kk are denoted as Yk,Xk,εkJY_{k},X_{k},\varepsilon_{k}\in\mathbb{R}^{J}, with the jjth element as Yk,j=yJ(k1)+jY_{k,j}=y_{J(k-1)+j} for k=1,,Kk=1,\ldots,K and j=1,,Jj=1,\ldots,J.

To perform time-frequency analysis, we introduce the latent Zk=(Zk,1,,Zk,J)JZ_{k}=(Z_{k,1},\ldots,Z_{k,J})\in\mathbb{C}^{J}, where Zk,jZ_{k,j} is a complex Gaussian variable with the magnitude corresponding to the power at the normalized frequency ωj=2π(j1)/J\omega_{j}=2\pi(j-1)/J and interval kk [17]. To model the evolution of spectra across the windows, we assume that {Zk}k=1K\{Z_{k}\}_{k=1}^{K} follow a random walk prior

Zk=Zk1+vk,\displaystyle Z_{k}=Z_{k-1}+v_{k}, (2)

where the state noise vkv_{k} is a complex Gaussian process with a diagonal covariance I(σv,j2)I(\sigma^{2}_{v,j}), i.e., vk𝒞𝒩(0,I(σv,j2))v_{k}\sim\mathcal{CN}(0,I(\sigma^{2}_{v,j})). The prior encodes two important properties of {Zk}k=1K\{Z_{k}\}_{k=1}^{K}. First, the state variance σv,j2\sigma_{v,j}^{2} controls the smoothness of the process, with a large value indicative of non-smooth or fluctuating process. Second, Zk,jZ_{k,j} and Zk,jZ_{k,j^{\prime}} for jjj\neq j^{\prime} are independent a priori.

We can link the time-frequency process {Zk}k=1K\{Z_{k}\}_{k=1}^{K} with time-domain observation {Yk}k=1K\{Y_{k}\}_{k=1}^{K}, using the Fourier transform matrix FJ×JF\in\mathbb{C}^{J\times J} with Fj,l=J1/2exp(i2π(l1)j/J)F_{j,l}=J^{-1/2}\exp(-i2\pi(l-1)j/J) and the inverse Fourier matrix WJ×JW\in\mathbb{C}^{J\times J}, such that FW=IFW=I,

Yk=Xk+εk=WZk+εk\displaystyle Y_{k}=X_{k}+\varepsilon_{k}=WZ_{k}+\varepsilon_{k} (3)
FYk=FWZk+Fεk=Zk+εkF,\displaystyle\Rightarrow FY_{k}=FWZ_{k}+F\varepsilon_{k}=Z_{k}+\varepsilon_{k}^{F},

where εkF𝒞𝒩(0,I(σε2))\varepsilon_{k}^{F}\sim\mathcal{CN}(0,I(\sigma_{\varepsilon}^{2})), since FI(σε2)W=I(σε2)FI(\sigma_{\varepsilon}^{2})W=I(\sigma_{\varepsilon}^{2}).

Along with the state-space model, we incorporate the data tapers to further reduce the variance of the spectral estimates. Specifically, we use MM Slepian tapers, leading to the multitaper (MT) method that optimally balances the bias-variance trade-off via bandwidth adjustment [18, 19]. This essentially produces MM independent set of state-space models,

Yk(m),F\displaystyle Y^{(m),F}_{k} =Zk(m)+εk(m),F\displaystyle=Z^{(m)}_{k}+\varepsilon^{(m),F}_{k} (4)
Zk(m)\displaystyle Z_{k}^{(m)} =Zk1(m)+vk(m),\displaystyle=Z_{k-1}^{(m)}+v^{(m)}_{k}, (5)

where vk(m)𝒞𝒩(0,I(σv,j2,(m)))v_{k}^{(m)}\sim\mathcal{CN}(0,I(\sigma^{2,(m)}_{v,j})), εk(m),F𝒞𝒩(0,I(σε2,(m)))\varepsilon_{k}^{(m),F}\sim\mathcal{CN}(0,I(\sigma_{\varepsilon}^{2,(m)})), Yk(m)Y_{k}^{(m)} denotes the mthm^{\text{th}} Slepian taper applied to YkY_{k}, Yk(m),FY_{k}^{(m),F} denotes Fourier transform of Yk(m)Y_{k}^{(m)}, i.e., Yk(m),F=FYk(m)Y_{k}^{(m),F}=FY_{k}^{(m)}, and Zk(m)Z^{(m)}_{k} represents the mmth spectral eigen-coefficient of ZkZ_{k}. The variants of SSMT modify Eqs. (4) or (5) [11, 12, 13, 14].

We now focus on ωj\omega_{j} for simplicity. Based on Eqs. (4) and (5), we derive a Kalman filter algorithm for the estimation of Zk,j(m)Z^{(m)}_{k,j} using Kalman gain Ck,j(m)C^{(m)}_{k,j}

Zk|k,j(m)\displaystyle Z^{(m)}_{k|k,j} =(1Ck,j(m))Zk1|k1,j(m)+Ck,j(m)Yk,j(m),F\displaystyle=(1-C^{(m)}_{k,j})Z^{(m)}_{k-1|k-1,j}+C^{(m)}_{k,j}Y^{(m),F}_{k,j} (6)
σk|k,j2,(m)\displaystyle\sigma^{2,(m)}_{k|k,j} =(1Ck,j(m))(σk1|k1,j2,(m)+σv,j2,(m)),\displaystyle=(1-C^{(m)}_{k,j})(\sigma^{2,(m)}_{k-1|k-1,j}+\sigma^{2,(m)}_{v,j}), (7)

where the notation k|sk|s denotes the estimate on interval kk given the data observed up to interval ss and Ck,j(m)C^{(m)}_{k,j} is given as

Ck,j(m)=σk1|k1,j2,(m)+σv,j2,(m)σε2,(m)+σk1|k1,j2,(m)+σv,j2,(m).\displaystyle C^{(m)}_{k,j}=\frac{\sigma^{2,(m)}_{k-1|k-1,j}+\sigma^{2,(m)}_{v,j}}{\sigma_{\varepsilon}^{2,(m)}+\sigma^{2,(m)}_{k-1|k-1,j}+\sigma^{2,(m)}_{v,j}}. (8)

The spectrogram estimate at frequency ωj\omega_{j} on interval kk is

fkSSMT(ωj)=M1m=1M|Zk|k,j(m)|2.\displaystyle f^{\text{SSMT}}_{k}(\omega_{j})=M^{-1}\sum^{M}_{m=1}|Z^{(m)}_{k|k,j}|^{2}. (9)

The parameters {σv,j2,(m)}j,m=1J,M\{\sigma_{v,j}^{2,(m)}\}_{j,m=1}^{J,M} and {σε2,(m)}m=1M\{\sigma_{\varepsilon}^{2,(m)}\}_{m=1}^{M} are estimated with EM algorithm [15]. SSMT tracks nonstationarity in the data with the time-invariant parameters σv,j2,(m)\sigma^{2,(m)}_{v,j} and σε2,(m)\sigma_{\varepsilon}^{2,(m)}, since Zk|k,j(m)Zk1|k1,j(m)Z^{(m)}_{k|k,j}\neq Z^{(m)}_{k-1|k-1,j} and thus fkSSMT(ωj)fk1SSMT(ωj)f^{\text{SSMT}}_{k}(\omega_{j})\neq f^{\text{SSMT}}_{k-1}(\omega_{j}).

III Adaptive SSMT

Although SSMT can model slowly time-varying spectral dynamics, it is restrictive for highly fluctuating spectral dynamics. Such strong nonstationarity (or fluctuation) is common in EEG with external intervention to the brain or with the change of the brain state during anesthesia/sleep [20]. Fig. 1 shows a snippet of human anesthesia EEG in which SSMT (Fig. 1(b)) cannot track the apparent dynamics shown by the MT approach (Fig. 1(a)). However, the proposed ASSMT (Fig. 1(c)) is able to capture the spectral dynamics and remove non-relevant background activity.

Refer to caption
Figure 1: A spectrogram snippet estimated with (a) MT (b) SSMT (c) ASSMT.

The failure of SSMT is due to the time-invariant parameters, as the Kalman gain quickly converges to a steady-state value C,j(m)C_{\infty,j}^{(m)} [21]. Denoting the observation prediction error as ΔYk,j(m)=Yk,j(m),FZk1|k1,j(m),F\Delta Y^{(m)}_{k,j}=Y^{(m),F}_{k,j}-Z_{k-1|k-1,j}^{(m),F} and the latent estimate change as ΔZk,j(m)=Zk|k,j(m)Zk1|k1,j(m)\Delta Z_{k,j}^{(m)}=Z^{(m)}_{k|k,j}-Z^{(m)}_{k-1|k-1,j}, we can express Eq. (6) as ΔZk,j(m)=C,j(m)ΔYk,j(m)\Delta Z_{k,j}^{(m)}=C^{(m)}_{\infty,j}\Delta Y_{k,j}^{(m)}. If C,j(m)C_{\infty,j}^{(m)} is low (SSMT for Fig. 1(b)), Zk|k,j(m)Z^{(m)}_{k|k,j} fails to reflect fluctuation in Yk,j(m),FY_{k,j}^{(m),F}. Consequently, fkSSMTf_{k}^{\text{SSMT}} cannot reliably track the spectral dynamics.

To resolve this issue, adaptive SSMT (ASSMT) posits a time-varying state-space model, to allow the adaptive change of σv,j2,(m)\sigma_{v,j}^{2,(m)}. Specifically, we replace vk(m)𝒞𝒩(0,I(σv,j2,(m)))v_{k}^{(m)}\sim\mathcal{CN}(0,I(\sigma^{2,(m)}_{v,j})) in Eq. (5) with vk,j(m)𝒞𝒩(0,I(σv,k,j2,(m)))v^{(m)}_{k,j}\sim\mathcal{CN}(0,I(\sigma^{2,(m)}_{v,k,j})), to indicate the state variance’s dependence on time. This prevents Kalman gain from converging to a steady-state value, allowing the algorithm to produce more flexible spectrogram estimates. σε2,(m)\sigma_{\varepsilon}^{2,(m)} remains constant across windows because we assume stationary background noise.

With the modified generative model, we now address when and how ASSMT tracks varying degrees of nonstationarity. We first quantify the notion of nonstationarity, and then propose an adaptive parameter estimation approach.

III-A Measure of nonstationarity

We define Y~k,j(m)=𝔼|Yk,j(m),FYk1,j(m),F|2\widetilde{Y}_{k,j}^{(m)}=\mathbb{E}|Y_{k,j}^{(m),F}-Y_{k-1,j}^{(m),F}|^{2} as the measure of nonstationarity, which is the expected observation difference and a function of window, frequency, and taper. Intuitively, we use high spectral fluctuation as a proxy for high nonstationarity, that is, when Y~k,j(m)\widetilde{Y}_{k,j}^{(m)} exceeds a frequency-dependent threshold βj\beta_{j}.

To estimate Y~k,j(m)\widetilde{Y}_{k,j}^{(m)}, we use exponential moving average (EMA), often used as a simple, yet effective approach to estimate the expectation in the filtering literature [22]

Y~k,j(m)=(1α)Y~k1,j(m)+α|Yk,j(m),FYk1,j(m),F|2,\displaystyle\widetilde{Y}_{k,j}^{(m)}=(1-\alpha)\widetilde{Y}_{k-1,j}^{(m)}+\alpha|Y_{k,j}^{(m),F}-Y_{k-1,j}^{(m),F}|^{2}, (10)

where 0α10\leq\alpha\leq 1 is a smoothing factor chosen a priori. This approach allows us to use Y~k,j(m)\widetilde{Y}_{k,j}^{(m)} as a heuristic indicator of strong nonstationarity, decoupled from the generative model and hence the estimation procedure of Zk,j(m)Z^{(m)}_{k,j}. The choice of α\alpha reflects the belief on the impulsiveness of nonstationarity and the volatile nature of the state transitions. With large α\alpha, Y~k,j(m)\widetilde{Y}_{k,j}^{(m)} is sensitive to instantaneous fluctuation.

III-B Estimation of parameters & adaptive thresholding

We now examine how to use Y~k,j(m)\widetilde{Y}_{k,j}^{(m)} to set βj\beta_{j} and subsequently estimate the parameters. Using Eq. (4), we have

Y~k,j(m)=𝔼|Yk,j(m),FYk1,j(m),F|2\displaystyle\widetilde{Y}_{k,j}^{(m)}=\mathbb{E}|Y_{k,j}^{(m),F}-Y_{k-1,j}^{(m),F}|^{2} (11)
=𝔼|(Zk,j(m)Zk1,j(m))+(εk(m),Fεk1(m),F)|2\displaystyle\quad=\mathbb{E}|(Z^{(m)}_{k,j}-Z^{(m)}_{k-1,j})+(\varepsilon^{(m),F}_{k}-\varepsilon^{(m),F}_{k-1})|^{2}
=𝔼|Zk,j(m)Zk1,j(m)|2+𝔼|εk(m),Fεk1(m),F|2,\displaystyle\quad=\mathbb{E}|Z^{(m)}_{k,j}-Z^{(m)}_{k-1,j}|^{2}+\mathbb{E}|\varepsilon^{(m),F}_{k}-\varepsilon^{(m),F}_{k-1}|^{2},

where we use the uncorrelatedness of the two differences. Next, we use the fact that 1) 𝔼|Zk,j(m)Zk1,j(m)|2=σv,k,j2,(m)\mathbb{E}|Z^{(m)}_{k,j}-Z^{(m)}_{k-1,j}|^{2}=\sigma^{2,(m)}_{v,k,j} and 2) εk(m),F\varepsilon^{(m),F}_{k} and εk1(m),F\varepsilon^{(m),F}_{k-1} are independent with variance σε2,(m)\sigma^{2,(m)}_{\varepsilon}, hence 𝔼|εk(m),Fεk1(m),F|2=2σε2,(m)\mathbb{E}|\varepsilon^{(m),F}_{k}-\varepsilon^{(m),F}_{k-1}|^{2}=2\sigma_{\varepsilon}^{2,(m)}, which leads to

𝔼|Yk,j(m),FYk1,j(m),F|2=σv,k,j2,(m)+2σε2,(m).\displaystyle\mathbb{E}|Y_{k,j}^{(m),F}-Y_{k-1,j}^{(m),F}|^{2}=\sigma^{2,(m)}_{v,k,j}+2\sigma^{2,(m)}_{\varepsilon}. (12)

This establishes the connection between the nonstationarity metric, Y~k,j(m)\widetilde{Y}_{k,j}^{(m)}, and the two sources of variance in our model.

With Eq. (10) and Eq. (12), we can estimate σv,k,j2,(m)\sigma^{2,(m)}_{v,k,j}. For σε2,(m)\sigma_{\varepsilon}^{2,(m)}, we use σ^ε2,(m)\widehat{\sigma}^{2,(m)}_{\varepsilon} estimated from SSMT. This leads to

σ^v,k,j2,(m)=Y~k,j(m)2σ^ε2,(m).\displaystyle\widehat{\sigma}^{2,(m)}_{v,k,j}=\widetilde{Y}_{k,j}^{(m)}-2\widehat{\sigma}^{2,(m)}_{\varepsilon}. (13)

We further lower bound σ^v,k,j2,(m)\widehat{\sigma}^{2,(m)}_{v,k,j} for two reasons. First, we impose that the time-varying SNR, γk,j=σv,k,j2,(m)/σε2,(m)\gamma_{k,j}=\sigma^{2,(m)}_{v,k,j}/\sigma_{\varepsilon}^{2,(m)}, is greater than the baseline SNR of the system, i.e., γk,jγk,jbaseline\gamma_{k,j}\geq\gamma_{k,j}^{\text{baseline}}. Second, we require nonnegative σ^v,k,j2,(m)\widehat{\sigma}^{2,(m)}_{v,k,j}. Since SSMT estimates the baseline properties of the data, we set γk,jbaseline=σ^v,j2,(m),SSMT/σ^ε2,(m)\gamma_{k,j}^{\text{baseline}}=\widehat{\sigma}^{2,(m),\text{SSMT}}_{v,j}/\widehat{\sigma}_{\varepsilon}^{2,(m)}. This yields the estimation procedure

σ^v,k,j2,(m)=max(Y~k,j(m)2σ^ε2,(m),σ^v,j2,(m),SSMT).\displaystyle\widehat{\sigma}^{2,(m)}_{v,k,j}=\max(\,\widetilde{Y}_{k,j}^{(m)}-2\widehat{\sigma}^{2,(m)}_{\varepsilon},\,\widehat{\sigma}_{v,j}^{2,(m),\text{SSMT}}\,). (14)

This obviates the need for EM beyond the initial phase for estimating σ^ε2,(m)\widehat{\sigma}_{\varepsilon}^{2,(m)} and σ^v,j2,(m),SSMT\widehat{\sigma}_{v,j}^{2,(m),\text{SSMT}}. Although EM can estimate the time-varying parameters, it requires multiple forward/backward passes through the entire data, which is computationally expensive. In addition, with every new observation, σ^v,k,j2,(m)\widehat{\sigma}^{2,(m)}_{v,k,j} in the earlier windows need to be re-estimated.

III-C Estimation of nonstationary spectra

We use Kalman filter with σ^v,k,j2,(m)\widehat{\sigma}^{2,(m)}_{v,k,j} to estimate the spectrogram fkASSMT(ωj)f^{\text{ASSMT}}_{k}(\omega_{j}). ASSMT thus operates with two different modes, depending on βj=2σ^ε2,(m)+σ^v,j2,(m),SSMT\beta_{j}=2\widehat{\sigma}^{2,(m)}_{\varepsilon}+\widehat{\sigma}_{v,j}^{2,(m),\text{SSMT}}. If Y~k,j(m)βj\widetilde{Y}_{k,j}^{(m)}\geq\beta_{j}, ASSMT uses larger state variance to track high nonstationarity. Given σk1|k1,j2,(m)\sigma^{2,(m)}_{k-1|k-1,j} and σ^ε2,(m)\widehat{\sigma}_{\varepsilon}^{2,(m)} in Eq. (8), we observe that the increase in σ^v,k,j2,(m)\widehat{\sigma}^{2,(m)}_{v,k,j} leads to the increase in Ck,j(m)C_{k,j}^{(m)}. This agrees with our intuition, since we want the Kalman gain to increase such that ΔZk,j(m)\Delta Z_{k,j}^{(m)} explains a greater portion of ΔYk,j(m)\Delta Y^{(m)}_{k,j}. For Y~k,j(m)<βj\widetilde{Y}_{k,j}^{(m)}<\beta_{j}, ASSMT simply uses the baseline σ^v,j2,(m),SSMT\widehat{\sigma}_{v,j}^{2,(m),\text{SSMT}} and thus operates with a fixed Kalman gain. This explains how ASSMT with adaptive state variance is able to track strong nonstationarity.

IV Results

We apply ASSMT to two datasets: 1) nonstationary simulated data and 2) EEG data from a patient under propofol anesthesia. We compare the spectrogram estimates between MT, SSMT, and ASSMT. All spectrograms are in the dB scale.

IV-A Application to Simulated Data

We simulate the data as a superposition of amplitude-modulated process yt,1y_{t,1} and frequency-modulated process yt,2y_{t,2}. The process yt,1y_{t,1} is generated from an AR(6) process centered at 11 Hz. The process yt,2y_{t,2} is generated from ARMA(6,4) with varying pole loci. The observations are given by yt=yt,1cos(2πf0t)+yt,2+σvty_{t}=y_{t,1}\cos(2\pi f_{0}t)+y_{t,2}+\sigma v_{t}, where f0=0.02f_{0}=0.02 Hz, vt𝒩(0,1)v_{t}\sim\mathcal{N}(0,1) and σ\sigma is chosen to achieve an SNR of 30 dB. More details can be found in [12]. For MT, we use 6-second windows with 50% overlap and M=3M=3 Slepian tapers, and 6-second non-overlapping windows for both SSMT and ASSMT. For SSMT, we use the entire data to estimate the parameters. For ASSMT, we use the initial 300 seconds of the data to compute the baseline parameters and use α=0.95\alpha=0.95.

Fig. 2 shows the ground truth and the estimated spectrograms. Although MT captures the spectral dynamics reasonably well, it picks up background noise and spectral artifacts (i.e., vertical lines), and induces mixing of adjacent frequency bands due to low resolution. SSMT (Fig. 2 (c)) resolves these issues with sharper spectral localization and removal of spectral artifacts, benefitting from the state-space prior.

ASSMT also shares the artifact rejection and noise reduction properties of SSMT. Moreover, ASSMT performs better denoising, as evidenced in the 55 to 2020 Hz frequency band. The Itakura-Saito divergence (IS) [23] between the ground truth and the spectrogram estimate also confirms this observation, with ASSMT attaining the lowest value (ISMT=6.51\text{IS}^{\text{MT}}=6.51, ISSSMT=3.16\text{IS}^{\text{SSMT}}=3.16, ISASSMT=2.75\text{IS}^{\text{ASSMT}}=\mathbf{2.75}). We attribute this difference to SSMT’s state variance σ^v,j2,(m),SSMT\widehat{\sigma}_{v,j}^{2,(m),\text{SSMT}} and Kalman gain fixed to high values, between 55 to 2020 Hz. However, since ASSMT does not commit to a fixed value, it can adaptively change the Kalman gain at different regimes of the data.

Refer to caption
Figure 2: Spectrograms for simulated data (a) ground truth (b) MT (c) SSMT (d) ASSMT with α=0.95\alpha=0.95.

IV-B Application to EEG recorded during anesthesia

The EEG was recorded (fs=250f_{s}=250 Hz) from a volunteer receiving propofol administered with increasing rate, followed by the decreasing rate [20]. This setup induces altering states of unconsciousness (or brain states), resulting in varying levels of nonstationarity. We used M=5M=5 Slepian tapers, J=1,000J=1,000 samples. For SSMT, we estimate the spectrogram based on the parameters estimated from 1) the initial 4 minutes of data (Fig. 3 (b)) and 2) the entire data (Fig. 3 (c)). For ASSMT, we use the initial 4 minutes to compute the baseline parameters. Fig. 3 shows the estimated spectrograms.

Refer to caption
Figure 3: Propofol anesthesia EEG spectrograms (a) MT (b) SSMT with initial 4-min EM window (c) SSMT with full data EM window (d) ASSMT with α=0.95\alpha=0.95 (e) ASSMT with α=0.05\alpha=0.05. Both ASSMT use initial 4-min EM window. Red horizontal lines correspond to 10 and 15 Hz.
Refer to caption
Figure 4: Kalman gain and state noise variance for SSMT and ASSMT for the first taper m=1m=1. (a) Kalman gain and (b) State noise variance at 10 Hz. (c) Kalman gain and (d) state noise variance at 15 Hz.

SSMT vs. ASSMT: SSMT based on the initial 4 minutes of the data (Fig. 3 (b)) produces low estimates for σ^v,j2,(m)\widehat{\sigma}_{v,j}^{2,(m)} due to absence of spectral dynamics for the baseline state. Although it removes background noise well, as a result, it misses most of the strong spectral fluctuation, as evidenced in extreme denoising of the spectra from 40 min to 120 min.

This can be mitigated by applying EM to a different section of the data, or the entire data. Due to high estimates for σ^v,j2,(m)\widehat{\sigma}_{v,j}^{2,(m)}, SSMT better captures the strong nonstationarity (Fig. 3 (c)). However, it fails to denoise the baseline state (0 to 4040 min) due to high Kalman gain, as similarly observed in the simulation. These results demonstrate a drawback of the time-invariant assumption, as σ^v,j2,(m)\widehat{\sigma}_{v,j}^{2,(m)} estimated from different sections could yield significantly different spectrogram estimates.

In contrast, ASSMT adaptively denoises the spectrogram (Fig. 3 (d-e)) even with the initial baseline parameters, the same setting for which SSMT failed (Fig. 3 (b)). The time-varying nature of the model allows it to switch between low σ^v,k,j2,(m)\widehat{\sigma}_{v,k,j}^{2,(m)}, appropriate for removing background noise, and high σ^v,k,j2,(m)\widehat{\sigma}_{v,k,j}^{2,(m)}, appropriate for capturing the spectral fluctuation, without fully committing to either parameters.

Another difference is the computation time. The time for using EM on the entire data and Kalman filtering is 400400 seconds. For ASSMT, however, the procedure takes 66 seconds, demonstrating an appreciable reduction in computational time and thus, making it more suitable for real-time application.

Effect of Kalman gain: To further understand ASSMT, we analyze the evolution of state variance and the Kalman gain across time at the representative frequency bands (10 and 15 Hz), shown in Fig. 4. For both frequency bands, we observe that the state variance and consequently the Kalman gain is increased above the threshold, in tandem with the large spectral fluctuation, as desired. On the other hand, The SSMT Kalman gain stays fixed at either 0.1 (for initial 4-min estimation, solid black in Fig. 4) or 0.9 (for entire data estimation, dotted black in Fig. 4).

Effect of α\boldsymbol{\alpha}: We observe that the denoising performance of ASSMT is robust towards the choice of α\alpha. To further understand how α\alpha affects the filtering/denoising operation, we analyze how σ^v,k,j2,(m)\widehat{\sigma}^{2,(m)}_{v,k,j} and Kalman gain change over time. ASSMT with α=0.95\alpha=0.95 (red) is dominated by the heavy fluctuation, |Yk,j(m),FYk1,j(m),F|2|Y_{k,j}^{(m),F}-Y_{k-1,j}^{(m),F}|^{2}. In contrast, ASSMT with small α=0.05\alpha=0.05 (blue) shows smoother state variance and Kalman gain. In this case, ASSMT starts adapting when there is significant evidence for nonstationarity. This explains why α=0.05\alpha=0.05 performs more denoising than α=0.95\alpha=0.95, as it is less susceptible to instantaneous fluctuations (40 min., 0 to 10 Hz) and background noise (75 to 95 min., 2 to 10 Hz).

V Conclusion

We introduced an adaptive state-space multitaper (ASSMT) framework for estimating spectral dynamics in the nonstationary time series. By relaxing the time-invariant parameters assumption and proposing an adaptive parameter estimation scheme, we demonstrated that ASSMT was able to capture strong power fluctuations more reliably compared to SSMT.

References

  • [1] T. F. Quatieri, Discrete-Time Speech Signal Processing: Principles and Practice.   Prentice-Hall, 2008.
  • [2] J. S. Lim, Two-Dimensional Signal and Image Processing.   Prentice-Hall, 1990.
  • [3] W. J. Emery and R. E. Thomson, Data Analysis Methods in Physical Oceanography.   Elsevier, 2001.
  • [4] W. Truccolo, U. T. Eden, M. R. Fellows, J. P. Donoghue, and E. N. Brown, “A point process framework for relating neural spiking activity to spiking history, neural ensemble, and extrinsic covariate effects,” Journal of Neurophysiology, vol. 93, no. 2, pp. 1074–1089, 2005.
  • [5] P. Mitra and H. Bokil, Observed Brain Dynamics.   Oxford University Press, 2007.
  • [6] R. Dahlhaus, “Fitting time series models to nonstationary processes,” The Annals of Statistics, vol. 25, no. 1, pp. 1 – 37, 1997.
  • [7] T. H. Koornwinder, Wavelets : An Elementary Treatment of Theory and Applications.   World Scientific, 1995.
  • [8] A. V. Oppenheim, R. W. Schafer, and J. R. Buck, Discrete-time Signal Processing.   Prentice-Hall, Inc., 2009.
  • [9] D. B. Percival and A. T. Walden, Spectral Analysis for Physical Applications.   Cambridge Univ Press, 1993.
  • [10] S. Kim, M. K. Behr, D. Ba, and E. N. Brown, “State-space multitaper time-frequency analysis,” Proceedings of the National Academy of Sciences, vol. 115, no. 1, pp. E5–E14, 2018.
  • [11] D. Ba, B. Babadi, P. L. Purdon, and E. N. Brown, “Robust spectrotemporal decomposition by iteratively reweighted least squares,” Proceedings of the National Academy of Sciences, vol. 111, no. 50, pp. E5336–E5345, 2014.
  • [12] P. Das and B. Babadi, “Dynamic bayesian multitaper spectral analysis,” IEEE Transactions on Signal Processing, vol. 66, no. 6, pp. 1394–1409, 2018.
  • [13] A. H. Song, S. Chakravarty, and E. N. Brown, “A smoother state space multitaper spectrogram,” in 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), 2018, pp. 33–36.
  • [14] A. H. Song, D. Ba, and E. N. Brown, “Plso: A generative framework for decomposing nonstationary time-series into piecewise stationary oscillatory components,” in Proceedings of the Thirty-Seventh Conference on Uncertainty in Artificial Intelligence, vol. 161, 2021, pp. 1371–1381.
  • [15] A. P. Dempster, N. M. Laird, and D. B. Rubin, “Maximum likelihood from incomplete data via the em algorithm,” Journal of the Royal Statistical Society: Series B (Methodological), vol. 39, no. 1, pp. 1–22, 1977.
  • [16] S. Adak, “Time-dependent spectral analysis of nonstationary time series,” Journal of the American Statistical Association, vol. 93, no. 444, pp. 1488–1501, 1998.
  • [17] D. R. Brillinger, Time Series: Data Analysis and Theory.   SIAM, 1981.
  • [18] D. Thomson, “Spectrum estimation and harmonic analysis,” Proceedings of the IEEE, vol. 70, no. 9, pp. 1055–1096, 1982.
  • [19] B. Babadi and E. N. Brown, “A review of multitaper spectral analysis,” IEEE Transactions on Biomedical Engineering, vol. 61, no. 5, pp. 1555–1564, 2014.
  • [20] P. L. Purdon, E. T. Pierce, E. A. Mukamel, M. J. Prerau, J. L. Walsh, K. F. K. Wong, A. F. Salazar-Gomez, P. G. Harrell, A. L. Sampson, A. Cimenser, S. Ching, N. J. Kopell, C. Tavares-Stoeckel, K. Habeeb, R. Merhar, and E. N. Brown, “Electroencephalogram signatures of loss and recovery of consciousness from propofol,” Proceedings of the National Academy of Sciences, vol. 110, no. 12, pp. E1142–E1151, 2013.
  • [21] R. H. Shumway and D. S. Stoffer, Time Series Analysis and Its Applications.   Springer, 2017.
  • [22] J. G. Proakis and D. K. Manolakis, Digital Signal Processing.   Prentice-Hall, Inc., 2006.
  • [23] F. Itakura and S. Saito, “A statistical method for estimation of speech spectral density and formant frequencies,” 1970.

A. Application to EEG recorded during anesthesia

We apply ASSMT to EEG recording of another volunteer receiving propofol anesthesia. All parameter settings of SSMT and ASSMT are same as those used in the main paper.

Refer to caption
Figure 5: Propofol anesthesia EEG spectrograms (a) MT (b) SSMT with initial 4-min EM window (c) SSMT with full data EM window (d) ASSMT with α=0.95\alpha=0.95 (e) ASSMT with α=0.05\alpha=0.05. Both ASSMT use initial 4-min EM window.