This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Predicting Parkinson’s Disease with Multimodal Irregularly Collected Longitudinal Smartphone Data

Weijian Li1, Wei Zhu1, E. Ray Dorsey2,3, and Jiebo Luo1 1Department of Computer Science, University of Rochester, Rochester, NY, USA 2Center for Health + Technology and Department of Neurology, University of Rochester, Rochester, NY, USA Email:1{wli69, wzhu15, [email protected]}, 2{[email protected]}
Abstract

Parkinson’s Disease is a neurological disorder and prevalent in elderly people. Traditional ways to diagnose the disease rely on in-person subjective clinical evaluations on the quality of a set of activity tests. The high-resolution longitudinal activity data collected by smartphone applications nowadays make it possible to conduct remote and convenient health assessment. However, out-of-lab tests often suffer from poor quality controls as well as irregularly collected observations, leading to noisy test results. To address these issues, we propose a novel time-series based approach to predicting Parkinson’s Disease with raw activity test data collected by smartphones in the wild. The proposed method first synchronizes discrete activity tests into multimodal features at unified time points. Next, it distills and enriches local and global representations from noisy data across modalities and temporal observations by two attention modules. With the proposed mechanisms, our model is capable of handling noisy observations and at the same time extracting refined temporal features for improved prediction performance. Quantitative and qualitative results on a large public dataset demonstrate the effectiveness of the proposed approach.

Index Terms:
Parkinson’s Disease, Multimodal data, Smartphone, Neural Ordinary Differential Equations

I Introduction

As the second most prevalent chronic neurodegenerative movement disorder in the world, Parkinson’s Disease (PD) is on a remarkable increase over the past years [1] and continuously posing severe threats to patient health with symptoms such as stress, tremor, degradation of memory and physical activities. Medical treatments are available to mitigate the PD symptom effects. Thus, timely and correct diagnosis of PD is crucial for early interventions prior to serious deterioration. A typical way of diagnosing the Parkinson’s Disease is through in-person assessment with clinicians. However, PD symptoms could be variable over time [2] which influences the onsite diagnosis quality given sparsely obtained health condition records and potential unobservable health condition changes. In addition, diagnosis by clinicians are usually subjective and difficult to calibrate.

Refer to caption
Figure 1: Illustration on predicting the Parkinson’s Disease (PD) as well as a brief overview of the proposed approach. A time-series based learning approach with attention mechanisms on both temporal and modality features, adaptively aggregates multimodal activity information for final PD prediction.

Recent studies [3, 4, 5] develop device-based software that include remote health measurements. The mPower study [3], for example, proposes a smartphone-based App that provides clinical related PD tests, which involve interactions with the participants and can be conducted outside clinics and at any time. Such remote health access approaches show an opportunity for timely PD diagnosis as well as improving disease understanding with the enriched longitudinal quantitative records [6]. The nature of the obtained device signals could also facilitate normalized objective measurements. However, the out-of-lab measurements pose new challenges: (1) the uncontrollable test time-points and the combination of test subjects lead to irregularly distributed results in the temporal dimension; (2) the self-reported test results lack quality control which may introduce noisy observations and affect the overall prediction performance. These are common difficulties when dealing with real-world multidimensional time-series data [7, 8], especially in the medical domain [9, 10].

Refer to caption
Figure 2: Overview of the proposed framework. Our model mainly consists of three modules: (1) multimodal feature extractor, (2) temporal encoder with multimodal attention and (3) embedding Self-Attention Pooling.

Several methods have been proposed to tackle the irregularly longitudinally distributed samples [11, 12, 13, 14, 15, 16, 17, 18]. Among them, the Neural Ordinary Differential Equations (ODEs) [16, 17, 18, 19] are a group of continuous-time models with a series of hidden states in the latent space. The observed fix-interval time-series signal with possibly missing values can then be modeled by continuous latent space representations. Given the function dynamics and a numerical ODE solver, each hidden state can be computed, representing the latent trajectory. However, in the medical field, present Neural ODE based methods focus on in-hospital collected data, e.g. patient ICU measures [9] and EHR records [10], the effective way to deal with noisy self-reported multimodal data in the wild remains unclear.

To address the above issues, we present a novel end-to-end deep-learning based model for predicting Parkinson’s Disease with self-reported multimodal smartphone data collected in the wild. To be specific, the proposed method first extracts feature representations from different modalities using mode-related encoders. An ODE based time-series encoder is then introduced to map the observed signals into a latent space for continuous modeling. Finally, a state-wise self-attention mechanism is proposed to learn aggregate local features for the prediction task and, more importantly, for better model interpretability important to clinical practice.

In summary, our main contributions are three folds:

  • We predict Parkinson’s Disease based on sporadically-observed activity data in the wild collected from smartphones with Neural Ordinary Differential Equations (ODEs). To our best knowledge, this is the first attempt at time-series prediction of Parkinson’s Disease in an uncontrolled environment. The proposed model has the potential to be adapted to similar tasks.

  • We synchronize discrete observations into unified time-points and extract valuable multimodal representations from noisy data with a multimodal attention mechanism.

  • We aggregate temporal observations with a self-attention mechanism for an enriched joint local and global representation as well as improving the interpretability for clinical practice.

II Methods

II-A Overview

Problem Definition Our work is based on the data collected by a large-scale Parkinson’s Disease study named mPower [3], which contains activity tests conducted by the participants through smartphones. Besides, participants also report their PD diagnosis status. Thus, given this remotely collected dataset, our goal is to correctly classify each participant as a PD patient or a non-PD patient. The overview of the proposed framework can be found in Figure 2.

II-B Multimodal Feature Extractor

The goal of the multimodal feature extractor is to encode raw activity test results into enriched feature embeddings as the model inputs for end-to-end learning. Considering the sequential data samples conducted in each of the activity tests, e.g. the accelerometer sequence generated when the participant is walking, we decide to adopt the widely adopted Temporal Convolutional Networks (TCN) [20] as the raw feature extractors. In detail, given an observed test sequence xtm={x0,,xLmt}x_{t}^{m}=\{x_{0},...,x_{L_{mt}}\} with length LmtL_{mt} from modal mMm\in M at observation time tTt\in T , each layer of TCN applies a dilated convolution on the test sequence with a focus on the longitudinal causal relationship:

F(snl)=i=1k1f(i)sndil1F(s^{l}_{n})=\sum_{i=1}^{k-1}f(i)s^{l-1}_{n-d*i} (1)

where dd is the dilation rate, and kk is the filter size. Then the nnth element in the llth layer is computed with the convolutional kernel ff applied on the ndin-d*i elements in the l1l-1th layer. In this way, we obtain the embeddings vt={vt0,vtm}v_{t}=\{v_{t}^{0},...v_{t}^{m}\} with vtmD×1v_{t}^{m}\in\mathbb{R}^{D\times 1} for each modality mm at observation time tt.

II-C Temporal Encoder with Multimodal Attention

Traditional ways of handling irregularly sampled data rely on direct aggregations or data imputation strategies. However, these methods are likely to cause information loss at important sequential and continuous signal changes, which are crucial for health condition analysis. Given the extracted and aligned multimodal features at each time point, we propose to model continuous PD symptom changes in the latent space through the Neural Ordinary Differential Equations (ODEs) [16, 17].

Recall that for ODEs, time-series is represented by a latent trajectory determined by the initial state h0h_{0}. Given the observed time points t0,t1,,tTt_{0},t_{1},...,t_{T} and an initialized state h0h_{0}, an ODE solver computes h1,,hTh_{1},...,h_{T} representing the hidden states for each time point. Formally,

h0p(h0)\displaystyle h_{0}\sim p(h_{0}) (2)
h1,,hT=ODESolve(h0,f,θf,t0,,tT)\displaystyle h_{1},...,h_{T}=\mbox{ODESolve}(h_{0},f,\theta_{f},t_{0},...,t_{T}) (3)

where function ff produces the gradient h(t)t=f(h(t),θf)\frac{\partial h(t)}{\partial t}=f(h(t),\theta_{f}) which is parameterized with a neural network. Each hidden state hth_{t} is then obtained by integrating the gradient through time, which is achieved by an ODESolver. To incorporate the observations at each time point tt and adjust the latent trajectory accordingly, hidden states are updated by a network, e.g. an RNNCell:

ht=RNNCell(ht,ut)h_{t}=\mbox{RNNCell}(h_{t}^{\prime},u_{t}) (4)

where hth_{t}^{\prime} is the hidden state before the update and utu_{t} is the observation features at current time tt.

A simple way to construct input utu_{t} is by directly concatenating the embeddings vtv_{t} extracted from the original observations. However, since participants are free to choose the types of activity tests to perform at home, the observed test results at each time-point are usually incomplete, resulting in some of the modal features being missing. Therefore, we attach each of the modality features with a binary mask vtmv_{t}^{\prime m} with the same feature length, indicating its observation status: vtm[vtmvtm]v_{t}^{m}\leftarrow[v_{t}^{m}\cdot v_{t}^{\prime m}].

In addition, different modalities may contribute differently to the final PD prediction due to abnormal or noisy test results. Given the hidden state hth_{t}^{\prime}, a valid modality test should share common representations that consider the similar semantics, e.g. the identity of the same participant. Based on this observation, we propose to integrate an attention mechanism inside the GRU cell, named M-GRU, to further adaptively learn an aggregation function by assigning a weight to each of the modalities:

ut=m=1Mamvtm\displaystyle u_{t}^{\prime}=\sum_{m=1}^{M}a_{m}*v_{t}^{m} (5)

where

am=exp{em}m=1Mexp{em}\displaystyle a_{m}=\frac{\mbox{exp}\{e_{m}\}}{\sum_{m=1}^{M}\mbox{exp}\{e_{m}\}} (6)
em=wmTtanh(Whmht+Wvmvt+bm)\displaystyle e_{m}=w_{m}^{T}\mbox{tanh}(W_{hm}*h_{t}^{\prime}+W_{vm}*v_{t}+b_{m}) (7)

where wmw_{m}, WhmW_{hm}, WvmW_{vm} and bmb_{m} are learnable parameters for computing the transformed representaion eme_{m}. In this paper, we adopt the GRU unit as the RNN cell for updating the hidden state hth_{t}. The state-wise input is a concatenation of the original inputs and the attended representations: ut=[vtut]u_{t}=[v_{t}\cdot u_{t}^{\prime}]. Therefore the updating function can be written as follows:

zt=σ(Wz[ht,vt])\displaystyle z_{t}=\sigma(W_{z}*[h_{t}^{\prime},v_{t}]) (8)
rt=σ(Wr[ht,vt])\displaystyle r_{t}=\sigma(W_{r}*[h_{t}^{\prime},v_{t}]) (9)
ht~=σ(Wg[rtht,ut])\displaystyle\tilde{h_{t}}=\sigma(W_{g}*[r_{t}*h_{t}^{\prime},u_{t}]) (10)
ht=(1zt)ht+ztht~\displaystyle h_{t}=(1-z_{t})*h_{t}^{\prime}+z_{t}*\tilde{h_{t}} (11)

where WzW_{z}, WrW_{r} and WgW_{g} are learnable parameters, σ\sigma is the hyperbolic tangent function.

Refer to caption
Figure 3: Illustration on the multimodal attention mechanism based on a GRU cell. At each observation point, each of the modality is attended by the incoming hidden state hth_{t}^{\prime} prior to update. The attended value utu_{t}^{\prime} is then concatenated with hidden state hth_{t}^{\prime} and observational inputs xtx_{t} for the following hth_{t} computation. σ\sigma represents the hyperbolic tangent function.

II-D Embedding Self-Attention Pooling

Different from the observations in controlled environments, e.g. ICUs in hospitals, self-reported test results suffer from poor quality control. Adopting one single state as the user representation for prediction can be easily biased by certain noisy observations. To increase our model’s robustness and extract raw symptom clues from each modality, we adopt a self-attention mechanism [21] on all the encoded modality features at each step vtv_{t} to form a time-wise global representation:

h=t=1Tatvt\displaystyle h=\sum_{t=1}^{T}a_{t}*v_{t} (12)

where

at=exp{wTtanh(WvkT)}t=1Texp{wTtanh(WvkT)}\displaystyle a_{t}=\frac{\mbox{exp}\{w^{T}\mbox{tanh}(Wv_{k}^{T})\}}{\sum_{t=1}^{T}\mbox{exp}\{w^{T}\mbox{tanh}(Wv_{k}^{T})\}} (13)

where vtv_{t} is a concatenation of the extracted modal features vtmv_{t}^{m}, while ww and WW are learnable parameters. We then concatenate the time-wise representation with the last hidden state from the temporal encoder as user representation logits. The final prediction is obtained by applying the sigmoid function on the transformed logits:

y^=Sigmoid(wT[hhT]+b)\hat{y}=\mbox{Sigmoid}(w^{T}[h\cdot h_{T}]+b) (14)

II-E Training

We adopt the standard binary cross-entropy loss on the predicted logit y^\hat{y} and the target label yy:

=1Ni=1Nyilog(yi^)+(1yi)log(1yi^)\mathcal{L}=-\frac{1}{N}\sum_{i=1}^{N}y_{i}log(\hat{y_{i}})+(1-y_{i})log(1-\hat{y_{i}}) (15)

III Experiment

III-A Dataset Description

We first give a brief review on the mPower dataset [3]111https://www.synapse.org/#!Synapse:syn4993293/wiki/247859. It contains four types of PD-related activity test results from the participants conducted on their smartphones. In this study, we adopt three of them and leave the more complicated voice signals for future work:

Tapping Test It measures the impaired finger dexterity and tapping speed which are common signs of Parkinson’s Disease. In this test, participants are asked to place their smartphone on a flat surface and use the two fingers from the same hand to tap two buttons shown on their screen alternatively for 20 seconds.

Walking Test It evaluates participant’s gait and balance. During this test, participants need to carry the smartphone in the pocket and walk out-bounds, stand still then walk back.

Memory Test It focuses on evaluating participant’s short-term spatial memory. During this test, participants are shown an illuminated pattern on their smartphone screen and asked to replicate the pattern by touching the corresponding places in the correct order.

Each of the tests also asks participants to choose their medication points when conducting the test, namely: Immediately before Parkinson medication, Just after Parkinson medication (at your best), I don’t take Parkinson medications, and Another time. For the tapping and walking test, we adopt the accelerometer readings which contain (x,y,z)(x,y,z) coordinate sequences in Gs. For the memory test, we use both the tapping response sequences, the corresponding targets and the time spent for the response.

TABLE I: Statistics of the dataset after preprocessing and synchronizing the three modalities into common time periods. ±\pm represents a standard deviation following the mean value.
Properties Values
Samples (#) 1,236
Gender: Male & Female (%) 67.7% & 32.3%
PD & Non-PD (%) 62.4% & 37.6%
Age (#) 60.61 ±\pm 8.76
Total tests (#) & Missing rate (%) 122,790 & 57.6%
Sequence length per ID (#) 13.75 ±\pm 23.51
Walking test per ID (#) 6.49 ±\pm 14.21
Tapping test per ID (#) 13.26 ±\pm 20.62
Memory test per ID (#) 1.73 ±\pm 6.33

III-B Data Preprocessing

The mPower dataset is collected by the participants outside hospitals with limited quality control. To achieve our goal of multimodal time-series analysis, careful data pre-processing is crucial to remove noisy signals. Due to a large variation on the time when the test is conducted, temporally synchronize test results across different modalities to obtain multimodal observations at unified time points is also needed. We preprocess the collected data as follows:

Accelerometer Sequences For the tapping and walking tests, we adopt the accelerometer readings from the smartphone which contain sequences of (x,y,z)(x,y,z) coordinates. Each of the sequences is first processed with the low-pass filters [22] to remove the gravitational component. Since the tests are conducted in a highly uncontrolled environment, noisy observations, e.g. no tapping or not standing still, need to be removed. A publicly available222 https://github.com/deepcharles/ruptures change point detection algorithm [23] is then applied on the processed signal to segment the potential movements of interest. The longest segmented sequence with a signal standard deviation above a predefined threshold is extracted as the final observed sequence.

Memory Records For the memory tests, we adopt each participant’s actual button-tapping sequences and the corresponding target button sequences through time. If the participant play the memory game multiple times during the test, we concatenate the tests sorted by time. Game scores generated by the App are also attached to each of the touches in the game with a four dimensional representation for each touch: (time,actual,target,score)(time,actual,target,score).

Time Synchronization Notice that a different PD medication point may influence the test performance, e.g. just before medication is worse than at your best. To remove this effect, we first group the records by participant IDs and the medication point when the test is conducted. The test records with the Another time medication status is not used due to its ambiguous representation. The unique combinations of Participant ID + Medication Point are considered as the new unique IDs. To construct multimodal representations for each ID at unified time periods, the obtained records from different modalities within 24 hours are then grouped together. If there are duplicate records in the same time period, we sort them by the average observation time of the three modalities, and only the last observed one is kept.

Other Preprocessings Similar to previous studies [6, 11], we remove participants with ages below 45 who are less likely with PD symptoms. Participants perform less than 5 tests in total are also not included in the study. In the end, we obtain 1,236 samples containing their synchronized multimodal sequences. Detailed dataset statistics can be found in Table I.

III-C Methods for Comparison

We compare the proposed method with six baseline models. Three of them are traditional methods while the other three are deep learning based including the state-of-the-art time-series analysis methods RNN+Δt\Delta t, GRU-D, and ODE-RNN.

TABLE II: Evaluation results with 5-fold cross-validation. ±\pm represents a standard deviation following the mean value of the five folds.
Method AUC AUPR F1
LR [24] 0.556 ±\pm 0.028 0.665 ±\pm 0.058 0.521 ±\pm 0.071
SVM [25] 0.547 ±\pm 0.057 0.657 ±\pm 0.049 0.697 ±\pm 0.022
XGBoost [26] 0.631 ±\pm 0.042 0.726 ±\pm 0.030 0.730 ±\pm 0.029
RNN [27]+Δ\Deltat 0.726 ±\pm 0.026 0.811 ±\pm 0.035 0.771 ±\pm 0.012
GRU-D [13] 0.754 ±\pm 0.030 0.827 ±\pm 0.033 0.788 ±\pm 0.023
ODE-RNN [17] 0.767 ±\pm 0.022 0.845 ±\pm 0.031 0.797 ±\pm 0.023
Proposed 0.793 ±\pm 0.024 0.865 ±\pm 0.028 0.816 ±\pm 0.021
  • LR [24]: We leverage a standard logistic regression classifier for binary classification.

  • SVM [25]: We adopt a standard Support Vector Machine classifier with the RBF kernel for comparison.

  • XGBoost [26]: It stands for Extreme Gradient Boosting which is a tree-based boosting algorithm.

  • RNN+Δ\Deltat: We concatenates intervals Δ\Deltat to the input features and feed them into a standard RNN model [27]. The last hidden state is used for the final prediction.

  • GRU-D [13]: The GRU-D model is also designed for modeling trajectory changes with a hidden state exponentially decay through time. In addition, it concatenates observational masks and time intervals between the observations as additional clue into the inputs.

  • ODE-RNN [17]: The ODE-RNN model focuses on the continuous latent space trajectory modeling which captures inter-observation changes with ODE and at-observation hidden state updates with a GRUCell.

III-D Experimental Settings

Baseline methods LR and SVM are adopted from the Scikit-Learn library. The XGBoost algorithm is adopted from its publicly available python version. Since LR, SVM and XGBoost algorithms are not designed for handling irregular time-series inputs, we take the average of the features from all the steps as the global representation and feed them into these classifiers. For each step, we concatenate three modalities as a combined representation. The remaining methods use the same inputs and the same settings for the feature extractor and final prediction network. During training, the Adam optimizer is used with a learning rate initialized as 0.01 and decay by 0.96 for each epoch. All experiments are conducted with 5-fold cross-validation. Area Under the Curve (AUC), Area Under the Precision-Recall Curve (AUPR) and F1 scores are used as the evaluation metrics. We report the average and standard deviation values obtained from cross-validation. The proposed model is implemented in PyTorch and experimented on a single NVIDIA Geforce GTX 1080Ti GPU.

Refer to caption
Figure 4: Comparison on the performance of the proposed attention mechanisms.
Refer to caption
Figure 5: Visualization of the obtained attention values for a sequence of 16 (t0t15t_{0}-t_{15}) tests from a participant. Top chart: the attention weights on each of the temporal observations. Three middle charts: exemplar attention weights for three modalities at time t0t_{0}, t11t_{11} and t14t_{14}. Bottom charts: raw signals for walking, tapping and memory tests at the three observation times. Horizontal lines represent no test result for the modality.

III-E Quantitative Evaluations

Overall performance As shown in Table II, the proposed method in general achieves the best performance compared to all the strong baseline methods, e.g. RNN+Δt+\Delta t and ODE-RNN, over which our method achieves 10% and 5% performance gain in AUC respectively, demonstrating its effectiveness in the PD prediction task. LR, SVM and XGBoost perform worse than the deep-learning based methods. We think there are two main reasons. First, the features used for these methods are losing important temporal information by aggregating through time with average pooing. A trajectory that records a participant’s test performance is shown to be helpful for extracting the long-term PD symptom patterns that benefit our prediction task. The high model capacity of the deep learning based methods also provides higher power for dealing with high dimensional features. Encoding raw test signals and includes them into end-to-end learning also helps learn rich feature inputs. For the deep learning based methods, RNN+Δt+\Delta t performs better than non-deep-learning methods, indicating the benefits of time-series pattern learning as well as the integration of time-interval information. In addition to the updates at each observation, GRU-D also considers the dynamic changes between the observations by introducing an exponential hidden state decay mechanism that constructs a temporal relationship with respect to the time intervals. ODE-RNN further expands this idea by leveraging the ODE solvers to compute the derivatives for hidden state changes. Different from a predefined decay, ODE models are more flexible for handling continuous state changes with arbitrary time-intervals, leading to enriched latent space trajectory representation. The proposed method achieves the best results across all three evaluation metrics.

On proposed attention modules From Figure 4, we can see all the models with the proposed attention mechanism separately achieve better performance than the previous methods. Combining the two proposed attention mechanisms together, we achieve the best results which indicate a mutual improvement effect. When looking more closely, we find that our model with temporal attention brings the most improvements with a similar result when adopting both of the attention mechanisms. Recall that the temporal attention aggregates embedded multimodal features at each observation time to a unified representation. In this way, we believe our model not only preserves the original local modal representations but also learns to extract the most informative ones that provides our model with extra knowledge for decision making.

III-F Qualitative Evaluation

Attention visualization To further examine the effect of the proposed method, we visualize the attention weights learned by each of the attention modules. As shown in Figure 5, from a global temporal point of view, each step is assigned an attention value, with the largest one at time t14t_{1}4 and the second at time t11t_{1}1. This difference indicates our model is looking for certain patterns from each of the observations. Looking more into the details, we highlight three of the representative observation time points, namely t14t_{14}, t11t_{11} and a lower attended t0t_{0}. For t14t_{14}, we find that the memory test is being paid the most attention, following by the tapping test. The lowest weight is given to the walking test where no test result is presented. We consider some reasons are behind the memory test’s highest weight. One is that the memory test is less affected by potential noise because the task itself takes much fewer body movements than walking and tapping (continuously tapping the screen). Another is that the performance of the memory test is easier to quantify by directly comparing participant’s responses (actual tapping sequence, response speed) to clear targets (target tapping sequence, faster response speed), which helps measure the health status and improve the PD prediction accuracy. For t11t_{11} and t0t_{0}, we find in both cases the memory test is not conducted. Instead, our model focuses differently on tapping and walking. Yet notice that our model could be biased to memory tests if the existence of memory test is actually a reflection of PD existence. Future work could be directed on analyzing data bias problems for better generalizability. Looking into the signals, we find a more intense tapping sequence in t11t_{11} than the one in t0t_{0} which may contain richer behavior patterns for analysis.

IV Conclusion

In this paper, we present a novel time-series based deep learning approach to Parkinson’s Disease prediction based on remotely and irregularly collected data from smartphones. Different from previous methods, we synchronize discrete observations to unified observational time points to construct multimodal time-series representations using the Neural Ordinary Differential Equations. Two proposed attention mechanisms adaptively learn important features from noisy signals in both the temporal and modality dimensions. Insights and improved quantitative and qualitative results on a large public dataset demonstrate the effectiveness of the proposed approach.

V Acknowledgement

Research reported in this publication was supported by the National Institute Of Neurological Disorders And Stroke of the National Institutes of Health under Award Number P50NS108676. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. We also thank for the support from NSF through award IIS-1722847.

References

  • [1] A. Rossi, K. Berger, H. Chen, D. Leslie, R. B. Mailman, and X. Huang, “Projection of the prevalence of parkinson’s disease in the coming decades: Revisited,” Movement Disorders, vol. 33, no. 1, pp. 156–159, 2018.
  • [2] G. Becker, A. Müller, S. Braune, T. Büttner, R. Benecke, W. Greulich, W. Klein, G. Mark, J. Rieke, and R. Thümler, “Early diagnosis of parkinson’s disease,” Journal of neurology, vol. 249, no. 3, pp. iii40–iii48, 2002.
  • [3] B. M. Bot, C. Suver, E. C. Neto, M. Kellen, A. Klein, C. Bare, M. Doerr, A. Pratap, J. Wilbanks, E. R. Dorsey et al., “The mpower study, parkinson disease mobile data collected using researchkit,” Scientific data, vol. 3, no. 1, pp. 1–9, 2016.
  • [4] H. Zhang, C. Song, A. Wang, C. Xu, D. Li, and W. Xu, “Pdvocal: Towards privacy-preserving parkinson’s disease detection using non-speech body sounds,” in The 25th Annual International Conference on Mobile Computing and Networking, 2019, pp. 1–16.
  • [5] A. Zhan, M. A. Little, D. A. Harris, S. O. Abiola, E. Dorsey, S. Saria, and A. Terzis, “High frequency remote monitoring of parkinson’s disease via smartphone: Platform overview and medication response detection,” arXiv preprint arXiv:1601.00960, 2016.
  • [6] P. Schwab and W. Karlen, “Phonemd: Learning to diagnose parkinson’s disease from smartphone data,” in AAAI, vol. 33, 2019, pp. 1118–1125.
  • [7] T. Schneider, “Analysis of incomplete climate data: Estimation of mean values and covariance matrices and imputation of missing values,” Journal of climate, vol. 14, no. 5, pp. 853–871, 2001.
  • [8] Z. Cui, R. Ke, and Y. Wang, “Deep bidirectional and unidirectional lstm recurrent neural network for network-wide traffic speed prediction,” SIGKDD workshop, 2017.
  • [9] G. D. Clifford, C. Liu, B. Moody, D. Springer, I. Silva, Q. Li, and R. G. Mark, “Classification of normal/abnormal heart sound recordings: The physionet/computing in cardiology challenge 2016,” in 2016 Computing in Cardiology Conference (CinC).   IEEE, 2016, pp. 609–612.
  • [10] A. E. Johnson, T. J. Pollard, L. Shen, H. L. Li-wei, M. Feng, M. Ghassemi, B. Moody, P. Szolovits, L. A. Celi, and R. G. Mark, “Mimic-iii, a freely accessible critical care database,” Scientific data, vol. 3, p. 160035, 2016.
  • [11] J. Prince, F. Andreotti, and M. De Vos, “Multi-source ensemble learning for the remote prediction of parkinson’s disease in the presence of source-wise missing data,” IEEE Transactions on Biomedical Engineering, vol. 66, no. 5, pp. 1402–1411, 2018.
  • [12] C. Zhang, Z. Han, H. Fu, J. T. Zhou, Q. Hu et al., “Cpm-nets: Cross partial multi-view networks,” in NeurIPS, 2019, pp. 557–567.
  • [13] Z. Che, S. Purushotham, K. Cho, D. Sontag, and Y. Liu, “Recurrent neural networks for multivariate time series with missing values,” Scientific reports, vol. 8, no. 1, pp. 1–12, 2018.
  • [14] Q. Tan, M. Ye, B. Yang, S.-Q. Liu, and A. J. Ma, “Data-gru: Dual-attention time-aware gated recurrent unit for irregular multivariate time series,” 2020.
  • [15] I. M. Baytas, C. Xiao, X. Zhang, F. Wang, A. K. Jain, and J. Zhou, “Patient subtyping via time-aware lstm networks,” in SIGKDD, 2017, pp. 65–74.
  • [16] T. Q. Chen, Y. Rubanova, J. Bettencourt, and D. K. Duvenaud, “Neural ordinary differential equations,” in NeurIPS, 2018, pp. 6571–6583.
  • [17] Y. Rubanova, T. Q. Chen, and D. K. Duvenaud, “Latent ordinary differential equations for irregularly-sampled time series,” in NeurIPS, 2019, pp. 5321–5331.
  • [18] E. De Brouwer, J. Simm, A. Arany, and Y. Moreau, “Gru-ode-bayes: Continuous modeling of sporadically-observed time series,” in NeurIPS, 2019, pp. 7377–7388.
  • [19] J. Shi, J. Bi, Y. Liu, and C. Xu, “Cubic spline smoothing compensation for irregularly sampled sequences,” arXiv preprint arXiv:2010.01381, 2020.
  • [20] S. Bai, J. Z. Kolter, and V. Koltun, “An empirical evaluation of generic convolutional and recurrent networks for sequence modeling,” arXiv preprint arXiv:1803.01271, 2018.
  • [21] M. Ilse, J. M. Tomczak, and M. Welling, “Attention-based deep multiple instance learning,” ICML, 2018.
  • [22] R. Badawy, Y. P. Raykov, L. J. Evers, B. R. Bloem, M. J. Faber, A. Zhan, K. Claes, and M. A. Little, “Automated quality control for sensor based symptom measurement performed outside the lab,” Sensors, vol. 18, no. 4, p. 1215, 2018.
  • [23] C. Truong, L. Oudre, and N. Vayatis, “Selective review of offline change point detection methods,” Signal Processing, vol. 167, p. 107299, 2020.
  • [24] D. G. Kleinbaum, K. Dietz, M. Gail, M. Klein, and M. Klein, Logistic regression.   Springer, 2002.
  • [25] J. A. Suykens and J. Vandewalle, “Least squares support vector machine classifiers,” Neural processing letters, vol. 9, no. 3, pp. 293–300, 1999.
  • [26] T. Chen and C. Guestrin, “Xgboost: A scalable tree boosting system,” in SIGKDD, 2016, pp. 785–794.
  • [27] T. Mikolov, M. Karafiát, L. Burget, J. Černockỳ, and S. Khudanpur, “Recurrent neural network based language model,” in INTERSPEECH, 2010.