This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Analysis of EEG frequency bands for Envisioned Speech Recognition

Ayush Tripathi Department of Electrical Engineering
Indian Institute of Technology Delhi
New Delhi, India
[email protected]
Abstract

The use of Automatic speech recognition (ASR) interfaces have become increasingly popular in daily life for use in interaction and control of electronic devices. The interfaces currently being used are not feasible for a variety of users such as those suffering from a speech disorder, locked-in syndrome, paralysis or people with utmost privacy requirements. In such cases, an interface that can identify envisioned speech using electroencephalogram (EEG) signals can be of great benefit. Various works targeting this problem have been done in the past. However, there has been limited work in identifying the frequency bands (δ,θ,α,β,γ\delta,\theta,\alpha,\beta,\gamma) of the EEG signal that contribute towards envisioned speech recognition. Therefore, in this work, we aim to analyze the significance of different EEG frequency bands and signals obtained from different lobes of the brain and their contribution towards recognizing envisioned speech. Signals obtained from different lobes and bandpass filtered for different frequency bands are fed to a spatio-temporal deep learning architecture with Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM). The performance is evaluated on a publicly available dataset comprising of three classification tasks - digit, character and images. We obtain a classification accuracy of 85.93%85.93\%, 87.27%87.27\% and 87.51%87.51\% for the three tasks respectively. The code for the implementation has been made available at https://github.com/ayushayt/ImaginedSpeechRecognition.

Index Terms:
Envisioned speech, EEG, CNN, LSTM, Frequency bands, Brain–computer interface

I Introduction

The emergence of ubiquitous electronic devices has increased the requirement for natural ways of interaction with electronic devices. This has in turn led to significant improvement in human computer interfaces (HCI) such as those driven by speech and gesture [1]. However, there are certain situations in which the use of such mediums of HCI are limited. This may be due to speech disabilities, conditions such as locked-in syndrome or situations in which privacy is of utmost priority to the user. In such cases, imagined speech recognition systems based on non-invasive Electroencephalography could be used as an exciting alternative. Through years, various Brain Computer Interfaces (BCIs) such as those on the principle of text entry or imagination of motor movements have been proposed. Most of these however are focused on a user who is required to focus on a particular image or letter for a significant duration of time in order to select it or imagine a particular movement for providing commands [2]. A BCI framework based on recognition of imagined speech can provide a user with a more natural way for interaction with electronic devices [3]. Imagined speech Recognition here may be defined as the automated recognition of a given object, word or a letter from brain signals of the user. There are various techniques to measure brain signals ranging from invasive methods such as surgically implanted electrodes to the non-invasive EEG. In this work, we focus on non-invasive brain signals acquired using EEG technique. Despite of the low spatial resolution of EEG, it has been widely popular owing to it’s simplistic nature and little to no discomfort to the user [4].

The recognition of imagined speech using EEG has been attempted at various levels such as word, syllable, vowel imagination [4, 5, 6]. In their work, Gonzalez-Castaneda et al. [7] proposed an imagined word classification system using EEG comprising of 55 words. They first applied a Common Average Reference method to improve the Signal to Noise ratio and then extracted features using the Discrete Wavelet Transform. In [3], classified 22 words by using Mel Frequency Cepstral Coefficients and a k-Nearest Neighbor classifier. The majority of current envisioned speech recognition systems are heavily reliant on extracting meaningful handcrafted features, which requires in-depth domain expertise [8]. With advancements in the field of deep learning, researchers switched attention towards automatic extraction of features from raw EEG signals. The features extracted from the signals are such that they contain discriminatory information about the classification task. Zhang et al. [8] developed a combination of cascade and parallel Convolutional Recurrent Neural Network (CRNN) architecture for intention recognition using EEG signals. The EEG signals were processed by using 2D-CNN by constructing a 2-dimensional mesh as per the spatial information utilizing the electrode distribution. The features thereby extracted were fed to a LSTM for classification. For a 55-class classification task, the authors reported an improvement of around 18%18\% over features extracted from frequency information.

In this work, we study the contribution of different EEG frequency bands (δ,θ,α,β,γ\delta,\theta,\alpha,\beta,\gamma) and signals acquired from different lobes of the brain (frontal, temporal, occipital and parietal) towards imagined speech recognition using EEG signals. The signals are first bandpass filtered according to different frequency bands as (a) δ:<4Hz\delta:<4Hz, (b) θ:47Hz\theta:4-7Hz, (c) α:815Hz\alpha:8-15Hz, (d) β:1631Hz\beta:16-31Hz, and (e) γ:>32Hz\gamma:>32Hz. These bandpass filtered signals are fed to a 1DCNN-LSTM based architecture. The purpose of the 1D-CNN block is to extract spatial features from the signals while the LSTM block is used in order to extract the temporal information. We also select specific lobes of the brain and signals from only the selected lobes are fed to the classifier and the classification performance is obtained. Based on our observation that the frequency bands θ\theta and α\alpha have least contribution towards the classification of envisioned speech, we filter out these bands by using a bandreject filter to feed the the classifier. We show that such an approach improves the classification accuracy compared to using all the entire raw EEG signal. We demonstrate the efficacy of this approach on a publicly available envisioned speech dataset [9] consisting of 33 tasks - digit, character and images. We achieve classification accuracy of 85.93%85.93\%, 87.27%87.27\% and 87.51%87.51\% for the three tasks respectively.

II Dataset

We use a publicly available envisioned speech dataset containing recordings from 23 participants aged between 15-40 years [9]. EEG was recorded using Emotiv EPOC+ [10] wireless neuro headset. The headset consists of 1414 channels namely AF3, AF4, F3, F4, F7, F8, FC5, FC6, T7, T8, P7, P8, O1 and O2. The electrodes were placed at the scalp of the subject according to International 10-20 system of electrode placement as depicted in Figure 6. In addition to the aforementioned electrodes, two reference electrodes CMS and DRL were placed above the ears. The signals were recorded at a sampling frequency of 20482048 Hz and then down-sampled to 128128 Hz.

Refer to caption

Figure 1: Pictorial representation of the electrode placement for the data collection. The figure has been adapted from [9].

As depicted in Figure 2, the subject is displayed an object on the screen. Then, subject is instructed to imagine the viewed object for 1010 seconds while keeping the eyes closed and in resting state. Before displaying the next image to the participant, a gap of 2020 seconds is maintained in order to ensure that the subject returns to the state of rest before displaying the next item. In this manner, three categories of prompts have been shown to the participants and the corresponding EEG has been recorded.

Refer to caption

Figure 2: Flow diagram of the envisioned speech recognition recording setup. The figure has been adapted from [9].

The first category comprises of digits 090-9, second comprises of 1010 uppercase English alphabets - A, C, F, H, J, M, P, S, T, Y. The last category consists of items used in daily life - Apple, Car, Dog, Gold, Mobile, Rose, Scooter, Tiger, Wallet and Watch. In total, 230(2310)230(23*10) EEG recordings per category (comprising of 1010 classes) have been collected.

III Experimental Setup

III-A Methodology

Classification was performed by taking the three subsets - digits, characters and images of the described dataset. The entire EEG signal recorded for a particular class is of 1010 seconds duration. As in [11], the EEG signals were divided into 250250 ms duration (i.e. 32 samples) with a sliding increment of 6464 ms (i.e. 8 samples). For analyzing each of the different frequency bands, the signals are filtered by using the following standards:

  • δ\delta : low pass filter with a cutoff 4 Hz.

  • θ\theta : bandpass filter between 474-7 Hz.

  • α\alpha : bandpass filter between 7157-15 Hz.

  • β\beta : bandpass filter between 153115-31 Hz.

  • γ\gamma : high pass filter with a cutoff 3131 Hz.

The filtered signals are separately fed to the 1DCNN-LSTM model for classification. Out of the 1414 electrodes from which EEG signal was recorded, we identify 88 channels corresponding to the frontal lobe and 22 channels corresponding to the temporal, parietal and occipital lobes. EEG signals corresponding to these different lobes are also separately used for classification in raw form as well as by obtaining different frequency band information.

III-B Model Architecture

Refer to caption

Figure 3: Model Architecture for the envisioned speech recognition system.

The entire model architecture is depicted in Figure 3. The input to the model is a matrix of dimension 32×Nc32\times N_{c} where 3232 represents the timestamp and NcN_{c} is the number of channels (1414 when all channels are considered, 88 if only frontal lobe channels are considered and 22 if either of temporal, occipital or parietal lobe channels are considered). For the 1DCNN layer, 6464 filters of size 1010 with a stride of 11 are taken in order to extract the spatial information from the input. Subsequently, maxpooling by 22 is done. The spatial features extracted from the input are then fed to a LSTM block consisting of 256256 units to extract meaningful temporal information. This is then fed to a Dense layer comprising of 128128 neurons activated by the ReLU activation function and finally to the output layer comprising of 1010 neurons activated by the softmax activation function.

III-C Experimental Details

The data in each of the three tasks is split into a training set and a test set in a ratio of 80:2080:20. We employ a mini-batch training process by keeping a batch size of 128128. In order to prevent overfitting, an early stopping with a patience of 1010 while monitoring the validation accuracy is employed and a dropout of 50%50\% is used between the output layer and the penultimate layer of the CNN-LSTM architecture. Since the dataset is balanced among the different classes, categorical cross entropy loss is minimized using the Adam optimizer [12].

IV Results

Table I lists the performance of EEG signals obtained from different lobes of the brain and bandpass filtered for different EEG frequency bands. In terms of different frequency bands, it can be seen that the low frequency δ\delta band contains the most information followed by the γ\gamma and β\beta bands. The performance of θ\theta and α\alpha bands are significantly low compared to the other frequency bands. Thus, when the signal is bandpass filtered and the θ\theta and α\alpha frequency bands are removed, we see a boost in classification performance. This is very well in line with the literature of different frequency bands and their normal occurrence with respect to the task at hand. Furthermore, it can also be seen that the frontal lobe contains most information and therefore the accuracy of signals when only the frontal lobe is considered is significantly high. In Tables II and III, the results of experiments performed on the character and image tasks of the dataset are presented. The trend observed in these two tasks are also similar to that observed for the digit classification task.

TABLE I: Recognition accuracies for digit task
Part/Band δ\delta θ\theta α\alpha β\beta γ\gamma δ+β+γ\delta+\beta+\gamma All
Frontal 0.6934 0.3283 0.2651 0.5823 0.6318 0.7402 0.7244
Temporal 0.2018 0.1149 0.1178 0.2824 0.3834 0.4692 0.4983
Occipital 0.1248 0.1021 0.1028 0.2749 0.2207 0.1266 0.2211
Parietal 0.1426 0.1055 0.1124 0.3496 0.3435 0.2917 0.2993
All 0.8433 0.5009 0.5354 0.6769 0.7478 0.8593 0.8582
TABLE II: Recognition accuracies for character task
Part/Band δ\delta θ\theta α\alpha β\beta γ\gamma δ+β+γ\delta+\beta+\gamma All
Frontal 0.6978 0.3441 0.3899 0.5678 0.6433 0.7871 0.7521
Temporal 0.1556 0.1034 0.1077 0.3154 0.3928 0.5221 0.5221
Occipital 0.1391 0.1084 0.1261 0.3024 0.3456 0.1861 0.2522
Parietal 0.1381 0.1089 0.1136 0.3191 0.3271 0.2445 0.2893
All 0.8511 0.5405 0.1311 0.6872 0.7197 0.8727 0.8735
TABLE III: Recognition accuracies for image task
Part/Band δ\delta θ\theta α\alpha β\beta γ\gamma δ+β+γ\delta+\beta+\gamma All
Frontal 0.7155 0.2021 0.3756 0.5634 0.6236 0.8042 0.7593
Temporal 0.1903 0.1021 0.1296 0.3144 0.3731 0.5153 0.4965
Occipital 0.1323 0.1156 0.1095 0.1906 0.3503 0.2079 0.2157
Parietal 0.1469 0.0962 0.2006 0.3361 0.3144 0.3364 0.3231
All 0.8396 0.4961 0.5908 0.6575 0.7262 0.8751 0.8699

Refer to caption

Figure 4: Confusion matrix for the digit task.

Refer to caption

Figure 5: Confusion matrix for the character task. 0-9 represents in order the characters A, C, F, H, J, M, P, S, T, Y.

Refer to caption

Figure 6: Confusion matrix for the image task. 0-9 represents in order the images of Apple, Car, Dog, Gold, Mobile, Rose, Scooter, Tiger, Wallet and Watch.
TABLE IV: Comparison with previous work. (The results of other approaches have been adapted from [11]
Author Approach Character Digit Image
Tirupattur et.al. [13] CNN 71.2%71.2\% 72.9%72.9\% 73.0%73.0\%
Jolly et.al. [14] CNN,GRU - - 77.4%77.4\%
Kumar et.al. [9] Random Forest 66.9%66.9\% 68.5%68.5\% 65.7%65.7\%
For Comparison [11] CNN alone 71.0%71.0\% 66.4%66.4\% 72.0%72.0\%
For Comparison [11] Stacked LSTM 84.2%84.2\% 75.7%75.7\% 82.4%82.4\%
For Comparison [11] CNN+LSTM 87.1%87.1\% 82.8%82.8\% 86.6%86.6\%
Kumar et.al. [11] CNN+LSTM+MV 90.1% 85.1%85.1\% 89.4%
Proposed (δ+β+γ\delta+\beta+\gamma) + CNN-LSTM 87.3%87.3\% 85.9% 87.5%87.5\%

V Discussion & Conclusion

Due to the rapid advancement in low-cost EEG measurement systems, there has been a proportionate increase in real-world Brain Computer Interface applications. The recognition of imagined speech is an important task for people suffering from voice disorders or those with conditions such as locked-in syndrome. Moreover, it is also of importance for people with high privacy needs such as in public places where speech is not an ideal medium of communication. In spite of the tremendous possible use cases, there has been scarce work in this field compared to other motor tasks such as grasp and lift, movement etc. Since speech is a natural way of communication among people and also for interaction with other electronic devices, it is expected that a system based on decoding imagined speech should be of extreme importance. Most work surrounding the recognition of imagined speech has been done by using traditional signal processing based approaches such as extracting features like common spatial patterns (CSP), auto-regression coefficients, low-level descriptors such as energy, entropy etc. along with machine learning classifiers such as SVM, Random Forest k-NN, and Logistic Regression. In more recent works, there has been some focus on leveraging the advancement in Deep Learning by using architectures such as CNN, LSTM, GRU etc. for learning the complexities in the EEG signals.

In this work, we aim to explore the role of signals from different lobes of the brain (frontal, occipital, parietal and temporal) along with different EEG frequency bands (δ,θ,α,β,γ\delta,\theta,\alpha,\beta,\gamma) for imagined speech recognition. We use parallel bandpass filters to filter the EEG signal into corresponding frequency bands and feed it to a 1DCNN-LSTM based deep learning architecture for imagined speech and object recognition. The 1DCNN block extracts spatial features from the raw EEG signals and the LSTM block models the temporal information over the extracted features. We observe that the frequency bands θ\theta and α\alpha do not significantly contribute towards imagined speech recognition and therefore, we suggest bandreject filter for preprocessing the EEG signals before using them for the classification of imagined speech. Such an approach has been found to boost the classification performance on comparison with feeding directly the raw EEG signals to the 1DCNN-LSTM architecture. Furthermore, we also observe that the frontal lobe is the most significant contributor compared to other lobes of the brain. Both the observations are well in line with the literature on the functions of the frequency bands and the lobes. We perform experiments on a publicly available imagined speech dataset and achieve classification accuracies of 85.93%85.93\%, 87.27%87.27\% and 87.51%87.51\% for the task of classifying digits, characters and images respectively.

References

  • [1] C. Herff and T. Schultz, “Automatic speech recognition from neural signals: a focused review,” Frontiers in neuroscience, vol. 10, p. 429, 2016.
  • [2] S. Sreeja, V. Joshi, S. Samima, A. Saha, J. Rabha, B. S. Cheema, D. Samanta, and P. Mitra, “Bci augmented text entry mechanism for people with special needs,” in International Conference on Intelligent Human Computer Interaction.   Springer, 2016, pp. 81–93.
  • [3] N. Hashim, A. Ali, and W.-N. Mohd-Isa, “Word-based classification of imagined speech using eeg,” in International Conference on Computational Science and Technology.   Springer, 2017, pp. 195–204.
  • [4] M. AlSaleh, R. Moore, H. Christensen, and M. Arvaneh, “Discriminating between imagined speech and non-speech tasks using eeg,” in 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC).   IEEE, 2018, pp. 1952–1955.
  • [5] P. Saha, S. Fels, and M. Abdul-Mageed, “Deep learning the eeg manifold for phonological categorization from active thoughts,” in ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2019, pp. 2762–2766.
  • [6] C. H. Nguyen, G. K. Karavas, and P. Artemiadis, “Inferring imagined speech using eeg signals: a new approach using riemannian manifold features,” Journal of neural engineering, vol. 15, no. 1, p. 016002, 2017.
  • [7] E. F. González-Castañeda, A. A. Torres-García, C. A. Reyes-García, and L. Villaseñor-Pineda, “Sonification and textification: Proposing methods for classifying unspoken words from eeg signals,” Biomedical Signal Processing and Control, vol. 37, pp. 82–91, 2017.
  • [8] D. Zhang, L. Yao, X. Zhang, S. Wang, W. Chen, R. Boots, and B. Benatallah, “Cascade and parallel convolutional recurrent neural networks on eeg-based intention recognition for brain computer interface,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, no. 1, 2018.
  • [9] P. Kumar, R. Saini, P. P. Roy, P. K. Sahu, and D. P. Dogra, “Envisioned speech recognition using eeg sensors,” Personal and Ubiquitous Computing, vol. 22, no. 1, pp. 185–199, 2018.
  • [10] “Emotiv epoc+,” https://www.emotiv.com/epoc/, accessed: 02-12-2021.
  • [11] P. Kumar and E. Scheme, “A deep spatio-temporal model for eeg-based imagined speech recognition,” in ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).   IEEE, 2021, pp. 995–999.
  • [12] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
  • [13] P. Tirupattur, Y. S. Rawat, C. Spampinato, and M. Shah, “Thoughtviz: Visualizing human thoughts using generative adversarial network,” in Proceedings of the 26th ACM international conference on Multimedia, 2018, pp. 950–958.
  • [14] B. L. K. Jolly, P. Aggrawal, S. S. Nath, V. Gupta, M. S. Grover, and R. R. Shah, “Universal eeg encoder for learning diverse intelligent tasks,” in 2019 IEEE Fifth International Conference on Multimedia Big Data (BigMM).   IEEE, 2019, pp. 213–218.