This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

A Comparison of Representation Learning Methods for Dimensionality Reduction of fMRI Scans for Classification of ADHD

Bhaskar Sen1, Member, IEEE 1 Independent Researcher, [email protected]
Abstract

This paper compares three feature representation techniques used to represent resting state functional magnetic resonance (fMRI) scans. The proposed models of feature representation consider the time averaged fMRI scans as raw representation of image data. The effectiveness of the representation is evaluated by using these features for classification of Attention Deficit Hyperactivity Disorder (ADHD) patients from healthy controls. The dimensionality reduction methods used for feature representation are maximum-variance unfolding, locally linear embedding and auto-encoders. The classifiers tested for classification purpose were neural net and support vector machine. Using auto-encoders with four hidden layers along with a support vector machine classifier yielded a classification accuracy of 61.25% along with 65.69% sensitivity and 52.20% specificity.

Index Terms:
fMRI, Representation Learning, ADHD

I Introduction

Statistical machine learning methods have recently permeated disciplines such as Psychiatry, which specializes in the diagnosis and treatment of neuropsychiatric disorders. The availability of large-scale functional Magnetic Resonance Image (fMRI) datasets have encouraged the application of advanced machine learning models to the diagnosis of neuropsychiatric disorders [1, 2, 3]. fMRI scans measure brain activity by detecting fluctuations in blood-oxygen levels over time [4]. Hence fMRI is a four dimensional image which scans 3-D regions of interest over time.Brain activations are represented digitally as voxels, the three-dimensional analogue of pixels.

Attention Deficit hyperactive disorder(ADHD) is a highly ubiquitous neuropsychiatric disorder affecting the lives of hundreds of millions of people. The existence of this disorder prevents a person leading a normal social life in the form of academic underachievement, inattentiveness, hyperactivities, unemployability. The causation of this disorder is attributed to structural abnormalities of specific brain regions [5] as well as atypical functional connectivity in brain. This encourages the application of functional magnetic resonance imaging for analysis of brain activities of ADHD patients and compare their difference with the fMRI scans of healthy controls. FMRI scans can capture the structural properties of the brain regions in the form of voxels of three dimensional image. Furthermore, an exciting and important extension of this analysis is to build a diagnostic model that can analyse the fMRI scan and predict with certain accuracy if the subject suffers from ADHD. In this paper, we will be building a classifier that can accurately predict ADHD with high sensitivity. This is an important incremental step for building a final classifier model that can predict ADHD both with high sensitivity and specificity.

One of the main challenges for analysis of fMRI scans is the high dimension of brain scans. A single fMRI scan may consist of hundreds of three-dimensional images over time, each of which is composed of approximately 500,000 voxels [6]. Therefore, extracting lower dimensional fMRI representations that retain discriminate features for diagnosis is an important step in the implementation of diagnosis algorithms.This paper investigates different representations of fMRI data, analyses three dimensionality reduction methods for feature representation and scrutinizes two classifiers for learning and classifications.

II Problem Statement

The high level objective of this paper is to be able to build a diagnostic model that can effectively classify ADHD patients from healthy controls. Ideally the model should have high sensitivity, specificity and classification accuracy. Sensitivity specifies the proportion of the ADHD patients classified by the classifier as ADHD patients. Specificity specifies the proportion of healthy subjects labeled by the classifier as true negatives and classification accuracy denotes the portion of the whole test-set correctly labeled by the classifier. Another objective is to find a good low dimensional feature representation for fMRI scans that can effectively be used for classification purposes.

III Related Work

Using the ADHD-200 competition dataset, which consists of 940 resting-state fMRI scans,111http://fcon_1000.projects.nitrc.org/indi/adhd200 Eloyan et al. [7](Winning team of ADHD-200 Competition) explored several different classifiers for ADHD diagnosis, including a support vector machine, gradient boosting, and voxel-based morphology. In addition, several feature extraction methods were investigated, including singular value decomposition and CUR matrix decomposition. The best classification accuracy was achieved by taking a weighted combination of these classifiers, which yielded 61.0%61.0\% accuracy on the test data.The specificity achieved is 94% with associated sensitivity of 21%.Also using the ADHD-200 competition dataset, Ghiassian et al. [8] extracted histogram of oriented gradient features from fMRI scans, which were then input to a support vector machine. The classifier yielded an accuracy of 62.6%62.6\% on the test dataset. These two methods report the highest classification accuracy on the ADHD-200 competition dataset. Sen et al [9] also used an automated learning of features from fMRI using an auto encoder and independent component analysis to classify ADHD.

Similar to previous works on fMRI for a compact representation of fMRI features, the proposed models will attempt to investigate the applicability of state-of-the art feature representation methods for fMRI and test their predictability for ADHD classification task.

IV fMRI Data

The dataset was taken from ADHD-200 competition.The dataset consists of functional (four Dimensional) and anatomical (three Dimensional) scans of 940 subjects. The functional scans were preprocessed following [10].After preprocessing each of the subject’s scans had size 79×95×68×9179\times 95\times 68\times 91.The ADHD competition also had provided anatomical scans(3D) of subjects. First,the functional scans were time averaged (79×95×6879\times 95\times 68) to create 3D dimensional image for each subject.Then these scans were downsampled (one-fourth) using Gaussian Pyramid technique [11].The reduced dimensional image has 20×24×1720\times 24\times 17 voxels.The anatomical images were also downsampled to the size 20×24×1720\times 24\times 17 using Gaussian Pyramid. The values were sqeezed between 0-1 using a squeezing function. Next,I also segmented regions of interests from fMRI scans for ADHD disease following [12] from cortex, cingulate gyrus, and thalamus regions of the brain using Harvard-Oxford Atlas. 222http://neuro.debian.net/pkgs/fsl-harvard-oxford-atlases.html. The voxel values for the regions are averaged and thus we get 12 time series of voxel values for each subject. I calculated the correlations between these time series for each subject and those values were converted to p-statistics. Thus we get 66 correlation values for each subject.

V Methods

After downsampling the scans contain 8160 voxels which is a very large number of features to work with. Hence we reduced the number of dimensions of the input data. Three different non-linear dimensionality reduction techniques, i.e., i) Locally-Linear Embedding ii) Maximum Variance Unfolding and iii) Autoencoder were explored for feature representation. For all these dimensionality reduction techniques the input is an observation matrix Ot×nO^{t\times n} where tt is number of observation and nn is dimension of each observation.The output will be a reduced dimensional representation Ot×mO^{t\times m} where m<nm<n.

V-1 Locally-linear Embedding

One simple way for non-linear dimensionality reduction is to approximate each data point as a combination of its neighboring datapoints. This is an unsupervised way of dimensionality reduction.The number of neighbors is chosen by the user.This algorithm approximates a data point 𝐱𝐢\bf{x_{i}} as a composition of its neighborhood points 𝐱𝐣\bf{x_{j}} and corresponding weights Wi,jW_{i,j} ,jNj\in N where NN is the neighborhood for 𝐱𝐢\bf{x_{i}}. The optimization problem then becomes

min𝐖i|𝐱𝐢jWi,j𝐱𝐣|2,\min_{\mathbf{W}}\sum_{i}|\mathbf{x_{i}}-\sum_{j}W_{i,j}\mathbf{x_{j}}|^{2}, (1)

Subject to jWi,j=1\sum_{j}W_{i,j}=1 and jNj\in N where NN is the neighborhood for 𝐱𝐢\bf{x_{i}}.This optimization problem can be solved by solving a least-squares problem[13].

V-2 Maximum Variance Unfolding

Maximum Variance Unfolding is a non-linear dimensionality reduction technique that maps the data to a non-linear feature space that is defined by a kernel function. The mapping attempts to unfold the data onto a lower dimensional manifold. We assume that the fMRI voxels lie on a lower dimensional manifold, which the algorithm will attempt to recover. In order to stretch the underlying lower dimensional manifold, the feature space should maximize the distance between two neighboring points while keeping the locality constraint intact. Formally, if 𝐱i\mathbf{x}_{i} is a datapoint and 𝐱j\mathbf{x}_{j} N\in N where NN defines neighborhood of 𝐱𝐢\bf{x_{i}} and their corresponding feature-space representation is Φi\Phi_{i} and Φj\Phi_{j} then the problem of manifold learning becomes

maxϕi,j|Φ(𝐱i)Φ(𝐱j)|2,\max_{\mathbf{\phi}}\sum_{i,j}|\Phi(\mathbf{x}_{i})-\Phi(\mathbf{x}_{j})|^{2}, (2)

subject to the constraint that |Φ(𝐱i)Φ(𝐱j)|2=|𝐱i𝐱j|2|\Phi(\mathbf{x}_{i})-\Phi(\mathbf{x}_{j})|^{2}=|\mathbf{x}_{i}-\mathbf{x}_{j}|^{2} and iΦ(𝐱i)=0\sum_{i}\Phi(\mathbf{x}_{i})=0 for all i,ji,j[14]. This optimization problem can be efficiently solved by semidefinite programming techniques.

V-3 Autoencoder

High-dimensional data can be converted to low-dimensional codes by training a multilayer neural network with a small central layer to reconstruct high-dimensional input vectors. Two methods for optimization, i) Conjugate Gradients(CG) and ii) Limited Memory BFGS were utilized. Usage of CG and L-BFGS for optimising weights in a deep network was empirically studied in [15]. In this paper, we investigated the usage of these two optimization as a fine tuning step after pretraining using contrastive divergence(CD). Using intelligent guessing of weights from contrastive divergence(CD),L-BFGS and CG are expected to find the solution fast. The transfer function used in each layer was sigmoid.Layerwise the objective function is convex but as we stack the layers,the objective function becomes non-convex.Hence an intelligent guess for the initial weight becomes important. The autoencoder was tested for non-sparse representation as well as columnwise sparse representation.

V-A Feature Selection

Feature selection becomes important if we have very low variation of a feature for all data. Before training a learner, we used PCA to choose number of required features.For doing that we used five fold cross validation to choose number of principal components based on classification accuracy on validation set.

V-B Classifer

For classification, both Support Vector Machine and Neural Net classifiers were tried. But Support Vetor Machine classifier was chosen based on the reason given in  VI-A. For evaluating the classification models, accuracy and sensitivity were given specific attention since diagnostic process involves reducing false positives. However, accuracy, specificity, sensitivity, specificity and J-statistics were shown for final results.

VI Results

The features were extracted using a two layer autoencoder (VisibleNodes\rightarrow30\rightarrowOutput nodes). The features in central layer were used for learning an SVM classifier(with rbf kernel with σ=1\sigma=1). The five-fold cross validation accuracy is given in Table I.

TABLE I: 5-fold cross-validation results for ADHD classification (healthy, ADHD) using various dataset
Dataset 5-Fold CrossValidation Accuracy
Time Averaged fMRI Data 53.33%
Anatomical data 51.80%
fMRI Data+Correltion Matrix 52.08%
Anatomical data+Correlation Matrix 51.93%

As clearly seen from the accuracy, time averaged fMRI data outperforms others for accuracy.

Number of layers in the autoencoder was chosen using layerwise pretraining and backpropagation (Conjugate gradient with three line searches). Number of iteration in pretraining was fixed at 10 and number of iteration during back-propagation was fixed at 5. These numbers (iterations) were empirically chosen that gave best accuracies on 5-fold cross validation. The features (70 for all the experiments) were used to train a SVM classifier with rbf kernel. The sigma values were tested from 0.01 to 10.σ=0.6\sigma=0.6 gave the best 5-fold cross validation accuracy and hence the result is shown just for σ=0.6\sigma=0.6. Here hidden layers upto central layer has been shown. The other half reciprocates the first half.

TABLE II: Accuracy and Sensitivity Using Different Layers.
Hidden Nodes Accuracy Sensitivity
8160\rightarrow4000\rightarrow1000\rightarrow500\rightarrow70 58.13% 50%
8160\rightarrow1000\rightarrow500\rightarrow70 55.60% 65%
8160\rightarrow200\rightarrow70 57.01% 62.26%

As stated in [9], the distribution of the ADHD and Healthy patients in the dataset is skewed. Hence 8160\rightarrow200\rightarrow70\rightarrow200\rightarrow8160 met the criteria even though the autoencoder with nine hidden layer gave best accuracy but poor sensitivity. On the other hand the autoencoder with seven hidden layers gave very high sensitivity but poor accuracy. The sensitivity of the representation decreases for a very deep layer.The accuracy is high because it just labels most as majority class. This answer is not definite though as we see some fluctuations in sensitivity with increase of depth of layers.Ideally many other possible depths should have been tested. Also experiments can be done with the width of each hidden layer to find a more definite answer. We used nodes of central hidden layer of 81602007020081608160\rightarrow 200\rightarrow 70\rightarrow 200\rightarrow 8160 as features of the SVM classifier. Effectively, 70 features were used for testing classifier accuracy along with sensitivity.Here also for learning and classification an SVM classifier along with rbf kernel was chosen.For testing purpose with Conjugate Gradient and L-BFGS,σ=0.6\sigma=0.6 was chosen.The result is given in III

TABLE III: Accuracy and Sensitivity Using Different Optimization
Hidden Nodes Accuracy Sensitivity
CG(With 3 Line Searches) 57.01% 62.26%
L-BFGS 57.90% 63.01%

Here clearly L-BFGS gave better accuracy as well as sensitivity.Hence the hypothesis that L-BFGS should give better representation was true.

Next, testing different sparse representations was necessary as the features learned with no regularization had some features always activated as shown. This was done introducing two sparse regularizers in the hidden layer.If XX is our data matrix and ϕ\phi=f(XV+𝟏.𝐚𝐓)f(XV+\bf{1.a^{T}}) is the hidden representation of the data then,

𝚁𝚎𝚐𝚞𝚕𝚊𝚛𝚒𝚣𝚎𝚛𝟷=||ϕ||1,1=1Tϕ1.{\tt Regularizer1}=\arrowvert\arrowvert\phi\arrowvert\arrowvert_{1,1}=1^{T}\phi 1. (3)

and

𝚁𝚎𝚐𝚞𝚕𝚊𝚛𝚒𝚣𝚎𝚛𝟸=12||(1t𝟏Tϕ)ρ𝟏T||22.{\tt Regularizer2}=\frac{1}{2}\arrowvert\arrowvert(\frac{1}{t}{\bf 1}^{T}\phi)-\rho{\bf 1}^{T}\arrowvert\arrowvert_{2}^{2}. (4)

where t is number of observations and ρ=0.1\rho=0.1. Regularizer1 in Eq. 3 introduces sparsity throughout the representation matrix.Whereas Regularizer2 in Eq 4 induces columnwise sparsity in the hidden representation matrix.Hence Regularizer2  4 encourages distinctive feature for different subjects.The fine tuning using L-BFGS was done layerwise and stacked to form a two layer model. Their result on classification accuracy was also compared with feature representation without any sparsity inducing regularizer.The results of the classification accuracy for no regularizer,regularizer1 regularizer2 is shown in Table IV

TABLE IV: Accuracy Using different Sparse Representations
Regularizers Accuracy
Noregularization 57.90%
Regularizer1 56.25%
Regularizer2 58.50%

From the result,the columnwise sparsity inducing matrix had the best accuracy for classification.This clearly make sense because if we compare the features and basis images learned from No-regularization and Regularizer2  1,the columnwise sparsity inducing matrix gave a better representation of features. On the other hand Regularizer1 fails to achieve good accuracy.

Next, we tested autoencoder representation with representation got from Locally Linear Embedding(LLE) and Maximum-Variance Unfolding(MVU) method.LLE was implemented following 333http://http://www.cs.nyu.edu/~roweis/lle/code.html. Also MVU was implemented using 444http://http://www.cse.wustl.edu/~kilian/code/files/mvu2012.zip.For comparison purpose, the reduced dimension was fixed at 70. For LLE and MVU,number of neighbors chosen empirically was N=30N=30. For classifier an SVM classifier with rbf kernel(σ=0.6\sigma=0.6) was used. The five fold cross validation accuracy is given below.

TABLE V: Accuracy and Sensitivity Using different Representation Methods
Hidden Nodes Accuracy Sensitivity Specificity J-Stat
LLE 55.65% 43.38% 65.66% 8%
MVU 52.21% 57.55% 64.93% 22%
Autoencoder 58.50% 63.21% 43.28% 6%

The autoencoder outperforms the other two methods.It achieves high accuracy and sensitivity but low specificity.On the other hand LLE gives better representation than MVU even though sensitivity is better for MVU.

The 70 features,which represents the input data,contain some features which do not vary for different subject.This is clear from the features and the basis image received from the autoencoder with nodes 8160\rightarrow200\rightarrow70(For L-BFGS). We chose to use PCA to select important features among these. From the spectral components of these features, as shown in Fig. 1,

Refer to caption
Figure 1: Spectral Values for Hidden Layer Features

The most useful information (above 90%) among the features lies in first few components.Hence we used PCA to find those components and used five-fold cross validation for choosing best number of components to be used for my algorithm.For training classifier again SVM classifer was used with rbf kernel.Here five-fold cross validation was used to choose best σ\sigma.Result is shown for σ=0.1\sigma=0.1

TABLE VI: Using PCA Components
PCA Components Accuracy Sensitivity Specificity J-Stat
2 58.33% 84.29% 21.12% 5%
5 61.25% 65.69% 52.20% 17%
7 57.31% 59.75% 54.21% 13%
10 60.01% 63.37% 48.51% 11%
15 58.33% 54.21% 53.55% 7%

Here choosing first five components gave best accuracy result. As shown in  VII the final algorithm achieved five-fold cross validation accuracy of 61.25% which is better than [7] but slightly less than [8].The algorithm achieved Sensitivity and J-Stat better than [7] but Specificity was lower.

TABLE VII: Comparison of the proposed Algorithm to ADHD-200 Competition
Algorithm Accuracy Sensitivity Specificity J-Stat
Our Result 61.25% 65.69% 52.20% 17%
ADHD-200 Result 61.0% 21% 94% 15%

Even though my model achieved close to state of the art performance for diagnosis of ADHD,it is far from satisfactory to be used for any practical purpose.The reason behind the lacklustre performance might be lack of important information in the fMRI scans for detecting ADHD.It might be possible that we are losing effective information while downsampling the fMRI scans.Hence, using the full images as the input of autoencoder can be tested.But this requires very high computation power.

VI-A Choice of Classifer

For choosing the classifer to be used,initially we had decided to test two classifiers:Neural Net and Support Vector Machine.But initial tests with Neural Net classifier showed that it was just following the majority class classification.Also using a neural net classifer did not permit to visualise the features learned.Hence we decided to use autoencoder for learning features and comparing with other feature representation method and use SVM for classification.

VII Conclusion

The development of automatic ADHD diagnostic algorithms from fMRI data is a challenging task. The application of statistical pattern recognition algorithms to this problem currently yield insubstantial results, rendering these classification systems unfit for practice in the health-care industry. However, much research is being done to improve these results and search for discriminating features for classifying ADHD amongst the plethora of voxel values present in a single fMRI scan. Apart from systems that aggregate fMR images over time to produce a single three-dimensional image of the brain that is then used for classification, in this paper we explored the applicability of different feature representation techniques that can be used for effective representation of fMRI scans. Specifically, we used ADHD-200 competition dataset for learning features using Autoencoder, Locally Linear Embeddings and Maximum Variance Unfolding.My results indicate that autoencoders followed by principal component analysis and support vector machine classifer, yields the best results with 61.25% 5-fold cross-validation accuracy.

Still, there is much work to be done in this area. Recurrent neural networks or convolutional neural network, should be explored in future work. Moreover, we propose that locating lower dimensional sets of fMRI features that retain discriminating power for ADHD classification is the heart of the problem and that the majority of future work should focus on this task.

References

  • [1] I. Rish, B. Thyreau, et al., “Discriminative network models of schizophrenia,” Advances in Neural Information Processing Systems, vol. 22, pp. 252–260, 2009.
  • [2] B. Sen, G. A. Bernstein, T. Xu, B. A. Mueller, M. W. Schreiner, K. R. Cullen, and K. K. Parhi, “Classification of obsessive-compulsive disorder from resting-state fmri,” in 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). IEEE, 2016, pp. 3606–3609.
  • [3] B. Sen, G. A. Bernstein, B. A. Mueller, K. R. Cullen, and K. K. Parhi, “Sub-graph entropy based network approaches for classifying adolescent obsessive-compulsive disorder from resting-state functional mri,” NeuroImage: Clinical, vol. 26, pp. 102208, 2020.
  • [4] M. Brown, J. Benoit, M. Juhás, E. Dametto, T. T. Tse, M. MacKay, B. Sen, A. M. Carroll, O. Hodlevskyy, P. H. Silverstone, et al., “fmri investigation of response inhibition, emotion, impulsivity, and clinical high-risk behavior in adolescents,” Frontiers in systems neuroscience, vol. 9, pp. 124, 2015.
  • [5] E. M. Valera, S. V. Faraone, et al., “Meta-analysis of structural imaging findings in attention-deficit/hyperactivity disorder,” Biological psychiatry, vol. 61, no. 12, pp. 1361–1369, 2007.
  • [6] B. Sen and K. K. Parhi, “Extraction of common task signals and spatial maps from group fmri using a parafac-based tensor decomposition technique,” in 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2017, pp. 1113–1117.
  • [7] A. Eloyan, J. Muschelli, et al., “Automated diagnoses of attention deficit hyperactive disorder using magnetic resonance imaging,” Frontiers in systems neuroscience, vol. 6, pp. 61, 2012.
  • [8] S. Ghiassian, R. Greiner, et al., “Using functional or structural magnetic resonance images and personal characteristic data to identify adhd and autism,” PloS one, vol. 11, no. 12, pp. e0166934, 2016.
  • [9] B. Sen, N. C. Borle, R. Greiner, and M. Brown, “A general prediction model for the detection of adhd and autism using structural and functional mri,” PloS one, vol. 13, no. 4, pp. e0194856, 2018.
  • [10] P. Bellec et al., “The neuro bureau adhd-200 preprocessed repository,” Neuroimage, vol. 144, pp. 275–286, 2017.
  • [11] C. H. Anderson et al., “Change detection and tracking using pyramid transform techniques,” in Intelligent Robots and Computer Vision IV. International Society for Optics and Photonics, 1985, vol. 579, pp. 72–78.
  • [12] G. Bush, E. M Valera, and L. J. Seidman, “Functional neuroimaging of attention-deficit/hyperactivity disorder: a review and suggested future directions,” Biological psychiatry, vol. 57, no. 11, pp. 1273–1284, 2005.
  • [13] S. T. Roweis and L. K. Saul, “Nonlinear dimensionality reduction by locally linear embedding,” science, vol. 290, no. 5500, pp. 2323–2326, 2000.
  • [14] K. Q. Weinberger et al., “Learning a kernel matrix for nonlinear dimensionality reduction,” in Proceedings of the twenty-first international conference on Machine learning, 2004, p. 106.
  • [15] Q. V. Le et al., “On optimization methods for deep learning,” in ICML, 2011.