Analysis of EEG frequency bands for Envisioned Speech Recognition
Abstract
The use of Automatic speech recognition (ASR) interfaces have become increasingly popular in daily life for use in interaction and control of electronic devices. The interfaces currently being used are not feasible for a variety of users such as those suffering from a speech disorder, locked-in syndrome, paralysis or people with utmost privacy requirements. In such cases, an interface that can identify envisioned speech using electroencephalogram (EEG) signals can be of great benefit. Various works targeting this problem have been done in the past. However, there has been limited work in identifying the frequency bands () of the EEG signal that contribute towards envisioned speech recognition. Therefore, in this work, we aim to analyze the significance of different EEG frequency bands and signals obtained from different lobes of the brain and their contribution towards recognizing envisioned speech. Signals obtained from different lobes and bandpass filtered for different frequency bands are fed to a spatio-temporal deep learning architecture with Convolutional Neural Network (CNN) and Long Short-Term Memory (LSTM). The performance is evaluated on a publicly available dataset comprising of three classification tasks - digit, character and images. We obtain a classification accuracy of , and for the three tasks respectively. The code for the implementation has been made available at https://github.com/ayushayt/ImaginedSpeechRecognition.
Index Terms:
Envisioned speech, EEG, CNN, LSTM, Frequency bands, Brain–computer interfaceI Introduction
The emergence of ubiquitous electronic devices has increased the requirement for natural ways of interaction with electronic devices. This has in turn led to significant improvement in human computer interfaces (HCI) such as those driven by speech and gesture [1]. However, there are certain situations in which the use of such mediums of HCI are limited. This may be due to speech disabilities, conditions such as locked-in syndrome or situations in which privacy is of utmost priority to the user. In such cases, imagined speech recognition systems based on non-invasive Electroencephalography could be used as an exciting alternative. Through years, various Brain Computer Interfaces (BCIs) such as those on the principle of text entry or imagination of motor movements have been proposed. Most of these however are focused on a user who is required to focus on a particular image or letter for a significant duration of time in order to select it or imagine a particular movement for providing commands [2]. A BCI framework based on recognition of imagined speech can provide a user with a more natural way for interaction with electronic devices [3]. Imagined speech Recognition here may be defined as the automated recognition of a given object, word or a letter from brain signals of the user. There are various techniques to measure brain signals ranging from invasive methods such as surgically implanted electrodes to the non-invasive EEG. In this work, we focus on non-invasive brain signals acquired using EEG technique. Despite of the low spatial resolution of EEG, it has been widely popular owing to it’s simplistic nature and little to no discomfort to the user [4].
The recognition of imagined speech using EEG has been attempted at various levels such as word, syllable, vowel imagination [4, 5, 6]. In their work, Gonzalez-Castaneda et al. [7] proposed an imagined word classification system using EEG comprising of words. They first applied a Common Average Reference method to improve the Signal to Noise ratio and then extracted features using the Discrete Wavelet Transform. In [3], classified words by using Mel Frequency Cepstral Coefficients and a k-Nearest Neighbor classifier. The majority of current envisioned speech recognition systems are heavily reliant on extracting meaningful handcrafted features, which requires in-depth domain expertise [8]. With advancements in the field of deep learning, researchers switched attention towards automatic extraction of features from raw EEG signals. The features extracted from the signals are such that they contain discriminatory information about the classification task. Zhang et al. [8] developed a combination of cascade and parallel Convolutional Recurrent Neural Network (CRNN) architecture for intention recognition using EEG signals. The EEG signals were processed by using 2D-CNN by constructing a 2-dimensional mesh as per the spatial information utilizing the electrode distribution. The features thereby extracted were fed to a LSTM for classification. For a -class classification task, the authors reported an improvement of around over features extracted from frequency information.
In this work, we study the contribution of different EEG frequency bands () and signals acquired from different lobes of the brain (frontal, temporal, occipital and parietal) towards imagined speech recognition using EEG signals. The signals are first bandpass filtered according to different frequency bands as (a) , (b) , (c) , (d) , and (e) . These bandpass filtered signals are fed to a 1DCNN-LSTM based architecture. The purpose of the 1D-CNN block is to extract spatial features from the signals while the LSTM block is used in order to extract the temporal information. We also select specific lobes of the brain and signals from only the selected lobes are fed to the classifier and the classification performance is obtained. Based on our observation that the frequency bands and have least contribution towards the classification of envisioned speech, we filter out these bands by using a bandreject filter to feed the the classifier. We show that such an approach improves the classification accuracy compared to using all the entire raw EEG signal. We demonstrate the efficacy of this approach on a publicly available envisioned speech dataset [9] consisting of tasks - digit, character and images. We achieve classification accuracy of , and for the three tasks respectively.
II Dataset
We use a publicly available envisioned speech dataset containing recordings from 23 participants aged between 15-40 years [9]. EEG was recorded using Emotiv EPOC+ [10] wireless neuro headset. The headset consists of channels namely AF3, AF4, F3, F4, F7, F8, FC5, FC6, T7, T8, P7, P8, O1 and O2. The electrodes were placed at the scalp of the subject according to International 10-20 system of electrode placement as depicted in Figure 6. In addition to the aforementioned electrodes, two reference electrodes CMS and DRL were placed above the ears. The signals were recorded at a sampling frequency of Hz and then down-sampled to Hz.
As depicted in Figure 2, the subject is displayed an object on the screen. Then, subject is instructed to imagine the viewed object for seconds while keeping the eyes closed and in resting state. Before displaying the next image to the participant, a gap of seconds is maintained in order to ensure that the subject returns to the state of rest before displaying the next item. In this manner, three categories of prompts have been shown to the participants and the corresponding EEG has been recorded.
The first category comprises of digits , second comprises of uppercase English alphabets - A, C, F, H, J, M, P, S, T, Y. The last category consists of items used in daily life - Apple, Car, Dog, Gold, Mobile, Rose, Scooter, Tiger, Wallet and Watch. In total, EEG recordings per category (comprising of classes) have been collected.
III Experimental Setup
III-A Methodology
Classification was performed by taking the three subsets - digits, characters and images of the described dataset. The entire EEG signal recorded for a particular class is of seconds duration. As in [11], the EEG signals were divided into ms duration (i.e. 32 samples) with a sliding increment of ms (i.e. 8 samples). For analyzing each of the different frequency bands, the signals are filtered by using the following standards:
-
•
: low pass filter with a cutoff 4 Hz.
-
•
: bandpass filter between Hz.
-
•
: bandpass filter between Hz.
-
•
: bandpass filter between Hz.
-
•
: high pass filter with a cutoff Hz.
The filtered signals are separately fed to the 1DCNN-LSTM model for classification. Out of the electrodes from which EEG signal was recorded, we identify channels corresponding to the frontal lobe and channels corresponding to the temporal, parietal and occipital lobes. EEG signals corresponding to these different lobes are also separately used for classification in raw form as well as by obtaining different frequency band information.
III-B Model Architecture
The entire model architecture is depicted in Figure 3. The input to the model is a matrix of dimension where represents the timestamp and is the number of channels ( when all channels are considered, if only frontal lobe channels are considered and if either of temporal, occipital or parietal lobe channels are considered). For the 1DCNN layer, filters of size with a stride of are taken in order to extract the spatial information from the input. Subsequently, maxpooling by is done. The spatial features extracted from the input are then fed to a LSTM block consisting of units to extract meaningful temporal information. This is then fed to a Dense layer comprising of neurons activated by the ReLU activation function and finally to the output layer comprising of neurons activated by the softmax activation function.
III-C Experimental Details
The data in each of the three tasks is split into a training set and a test set in a ratio of . We employ a mini-batch training process by keeping a batch size of . In order to prevent overfitting, an early stopping with a patience of while monitoring the validation accuracy is employed and a dropout of is used between the output layer and the penultimate layer of the CNN-LSTM architecture. Since the dataset is balanced among the different classes, categorical cross entropy loss is minimized using the Adam optimizer [12].
IV Results
Table I lists the performance of EEG signals obtained from different lobes of the brain and bandpass filtered for different EEG frequency bands. In terms of different frequency bands, it can be seen that the low frequency band contains the most information followed by the and bands. The performance of and bands are significantly low compared to the other frequency bands. Thus, when the signal is bandpass filtered and the and frequency bands are removed, we see a boost in classification performance. This is very well in line with the literature of different frequency bands and their normal occurrence with respect to the task at hand. Furthermore, it can also be seen that the frontal lobe contains most information and therefore the accuracy of signals when only the frontal lobe is considered is significantly high. In Tables II and III, the results of experiments performed on the character and image tasks of the dataset are presented. The trend observed in these two tasks are also similar to that observed for the digit classification task.
Part/Band | All | ||||||
---|---|---|---|---|---|---|---|
Frontal | 0.6934 | 0.3283 | 0.2651 | 0.5823 | 0.6318 | 0.7402 | 0.7244 |
Temporal | 0.2018 | 0.1149 | 0.1178 | 0.2824 | 0.3834 | 0.4692 | 0.4983 |
Occipital | 0.1248 | 0.1021 | 0.1028 | 0.2749 | 0.2207 | 0.1266 | 0.2211 |
Parietal | 0.1426 | 0.1055 | 0.1124 | 0.3496 | 0.3435 | 0.2917 | 0.2993 |
All | 0.8433 | 0.5009 | 0.5354 | 0.6769 | 0.7478 | 0.8593 | 0.8582 |
Part/Band | All | ||||||
---|---|---|---|---|---|---|---|
Frontal | 0.6978 | 0.3441 | 0.3899 | 0.5678 | 0.6433 | 0.7871 | 0.7521 |
Temporal | 0.1556 | 0.1034 | 0.1077 | 0.3154 | 0.3928 | 0.5221 | 0.5221 |
Occipital | 0.1391 | 0.1084 | 0.1261 | 0.3024 | 0.3456 | 0.1861 | 0.2522 |
Parietal | 0.1381 | 0.1089 | 0.1136 | 0.3191 | 0.3271 | 0.2445 | 0.2893 |
All | 0.8511 | 0.5405 | 0.1311 | 0.6872 | 0.7197 | 0.8727 | 0.8735 |
Part/Band | All | ||||||
---|---|---|---|---|---|---|---|
Frontal | 0.7155 | 0.2021 | 0.3756 | 0.5634 | 0.6236 | 0.8042 | 0.7593 |
Temporal | 0.1903 | 0.1021 | 0.1296 | 0.3144 | 0.3731 | 0.5153 | 0.4965 |
Occipital | 0.1323 | 0.1156 | 0.1095 | 0.1906 | 0.3503 | 0.2079 | 0.2157 |
Parietal | 0.1469 | 0.0962 | 0.2006 | 0.3361 | 0.3144 | 0.3364 | 0.3231 |
All | 0.8396 | 0.4961 | 0.5908 | 0.6575 | 0.7262 | 0.8751 | 0.8699 |
V Discussion & Conclusion
Due to the rapid advancement in low-cost EEG measurement systems, there has been a proportionate increase in real-world Brain Computer Interface applications. The recognition of imagined speech is an important task for people suffering from voice disorders or those with conditions such as locked-in syndrome. Moreover, it is also of importance for people with high privacy needs such as in public places where speech is not an ideal medium of communication. In spite of the tremendous possible use cases, there has been scarce work in this field compared to other motor tasks such as grasp and lift, movement etc. Since speech is a natural way of communication among people and also for interaction with other electronic devices, it is expected that a system based on decoding imagined speech should be of extreme importance. Most work surrounding the recognition of imagined speech has been done by using traditional signal processing based approaches such as extracting features like common spatial patterns (CSP), auto-regression coefficients, low-level descriptors such as energy, entropy etc. along with machine learning classifiers such as SVM, Random Forest k-NN, and Logistic Regression. In more recent works, there has been some focus on leveraging the advancement in Deep Learning by using architectures such as CNN, LSTM, GRU etc. for learning the complexities in the EEG signals.
In this work, we aim to explore the role of signals from different lobes of the brain (frontal, occipital, parietal and temporal) along with different EEG frequency bands () for imagined speech recognition. We use parallel bandpass filters to filter the EEG signal into corresponding frequency bands and feed it to a 1DCNN-LSTM based deep learning architecture for imagined speech and object recognition. The 1DCNN block extracts spatial features from the raw EEG signals and the LSTM block models the temporal information over the extracted features. We observe that the frequency bands and do not significantly contribute towards imagined speech recognition and therefore, we suggest bandreject filter for preprocessing the EEG signals before using them for the classification of imagined speech. Such an approach has been found to boost the classification performance on comparison with feeding directly the raw EEG signals to the 1DCNN-LSTM architecture. Furthermore, we also observe that the frontal lobe is the most significant contributor compared to other lobes of the brain. Both the observations are well in line with the literature on the functions of the frequency bands and the lobes. We perform experiments on a publicly available imagined speech dataset and achieve classification accuracies of , and for the task of classifying digits, characters and images respectively.
References
- [1] C. Herff and T. Schultz, “Automatic speech recognition from neural signals: a focused review,” Frontiers in neuroscience, vol. 10, p. 429, 2016.
- [2] S. Sreeja, V. Joshi, S. Samima, A. Saha, J. Rabha, B. S. Cheema, D. Samanta, and P. Mitra, “Bci augmented text entry mechanism for people with special needs,” in International Conference on Intelligent Human Computer Interaction. Springer, 2016, pp. 81–93.
- [3] N. Hashim, A. Ali, and W.-N. Mohd-Isa, “Word-based classification of imagined speech using eeg,” in International Conference on Computational Science and Technology. Springer, 2017, pp. 195–204.
- [4] M. AlSaleh, R. Moore, H. Christensen, and M. Arvaneh, “Discriminating between imagined speech and non-speech tasks using eeg,” in 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). IEEE, 2018, pp. 1952–1955.
- [5] P. Saha, S. Fels, and M. Abdul-Mageed, “Deep learning the eeg manifold for phonological categorization from active thoughts,” in ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2019, pp. 2762–2766.
- [6] C. H. Nguyen, G. K. Karavas, and P. Artemiadis, “Inferring imagined speech using eeg signals: a new approach using riemannian manifold features,” Journal of neural engineering, vol. 15, no. 1, p. 016002, 2017.
- [7] E. F. González-Castañeda, A. A. Torres-García, C. A. Reyes-García, and L. Villaseñor-Pineda, “Sonification and textification: Proposing methods for classifying unspoken words from eeg signals,” Biomedical Signal Processing and Control, vol. 37, pp. 82–91, 2017.
- [8] D. Zhang, L. Yao, X. Zhang, S. Wang, W. Chen, R. Boots, and B. Benatallah, “Cascade and parallel convolutional recurrent neural networks on eeg-based intention recognition for brain computer interface,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32, no. 1, 2018.
- [9] P. Kumar, R. Saini, P. P. Roy, P. K. Sahu, and D. P. Dogra, “Envisioned speech recognition using eeg sensors,” Personal and Ubiquitous Computing, vol. 22, no. 1, pp. 185–199, 2018.
- [10] “Emotiv epoc+,” https://www.emotiv.com/epoc/, accessed: 02-12-2021.
- [11] P. Kumar and E. Scheme, “A deep spatio-temporal model for eeg-based imagined speech recognition,” in ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2021, pp. 995–999.
- [12] D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” arXiv preprint arXiv:1412.6980, 2014.
- [13] P. Tirupattur, Y. S. Rawat, C. Spampinato, and M. Shah, “Thoughtviz: Visualizing human thoughts using generative adversarial network,” in Proceedings of the 26th ACM international conference on Multimedia, 2018, pp. 950–958.
- [14] B. L. K. Jolly, P. Aggrawal, S. S. Nath, V. Gupta, M. S. Grover, and R. R. Shah, “Universal eeg encoder for learning diverse intelligent tasks,” in 2019 IEEE Fifth International Conference on Multimedia Big Data (BigMM). IEEE, 2019, pp. 213–218.