Document Type : Original Research

Authors

1 School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran

2 Medical Image and Signal Processing Research Center, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran

10.31661/jbpe.v0i0.2207-1521

Abstract

Background: The P300 signal, an endogenous component of event-related potentials, is extracted from an electroencephalography signal and employed in Brain-computer Interface (BCI) devices.
Objective: The current study aimed to address challenges in extracting useful features from P300 components and detecting P300 through a hybrid unsupervised manner based on Convolutional Neural Network (CNN) and Long Short-term Memory (LSTM).
Material and Methods: In this cross-sectional study, CNN as a useful method for the P300 classification task emphasizes spatial characteristics of data. However, CNN and LSTM networks are combined to modify the classification system by extracting both spatial and temporal features. Then, the CNN-LSTM network was trained in an unsupervised learning method based on an autoencoder to improve Signal-to-noise Ratio (SNR) by extracting main components from latent space. To deal with imbalanced data, an Adaptive Synthetic Sampling Approach (ADASYN) is used and augmented without any duplication.
Results: The trained model, tested on the BCI competition III dataset, including two normal subjects, with an accuracy of 95% and 94% for subjects A and B in P300 detection, respectively. 
Conclusion: CNN-LSTM, was embedded into an autoencoder and introduced to simultaneously extract spatial and temporal features and manage the computational complexity of the method. Further, ADASYN as an augmentation method was proposed to deal with the imbalanced nature of data, which not only maintained feature space as before but also preserved anatomical features of P300. High-quality results highlight the suitable efficiency of the proposed method.

Highlights

Ramin Afrah (Google Scholar)

Zahra Amini (Google Scholar)

Keywords

Introduction

In healthy or disabled individuals, Brain-Computer Interface (BCI) systems can send instructions, operated by brain activity, to external gadgets without the use of a neuro-muscular system [ 1 , 2 ]. Different types of BCI systems are categorized according to the brain activity measurement method, such as Near Infrared Spectroscopy (NIRS) [ 3 ], Electroencephalography (EEG) [ 4 ], and Magnetoencephalography (MEG) [ 5 ]. EEG-based BCIs [ 6 ] are considered the most popular method due to some advantages, such as ease of use, non-invasiveness, and low cost. Some studies have been conducted on P300-based BCIs, a well-known type of EEG-based BCIs, to modify the performance [ 7 ].

Event-related Potentials (ERPs) are a type of EEG signal, generated after specific stimulation, and P300 as one of the most important components of ERP signals, is a positive deflection elicited approximately 300-1000 ms after auditory, somatosensory, or visual stimulation [ 6 , 8 , 9 ]. Anatomically, P300 is more elicited over the midline scalp, and its magnitude increases from the frontal to the parietal lobe [ 10 ].

Steps in setting up a P300-based BCI system include user tasks, EEG signal recording, signal pre-processing, feature extraction and translation, signal classification, and feedback to the user interface [ 11 ]. The user task is critical for the occurrence of the P300 signal, and a well-designed user task not only provides a strong P300 signal response with a short delay duration but also protects the subject against eye fatigue [ 12 ]. After designing a suitable task, EEG signals are recorded, which required preprocessing procedures, such as noise reduction and artifact removal. In the feature extraction stage, the most useful and meaningful features are extracted from the data, and the dimensionality reduction technique is performed if needed. Finally, a proper classifier is used to elicit the P300 wave [ 13 ] in the signal classification step based on the obtained features. The created command must then be returned to the user. Typically, criteria, such as accuracy and time of detection, are applied to assess the overall performance of a P300-based BCI system [ 14 ].

Farwell and Donchin introduced a well-known P300-based BCI [ 15 ], which produces P300 based on the oddball paradigm. They also generated a speller matrix composed of a 6×6 matrix of symbols. Each time, a row or column is randomly intensified for a short time until each column and row has been intensified once. The user selects the target stimuli by a specific symbol as well as counting the flashing of the row and column that intersect at that desired character [ 16 ].

The most important obstacles in P300 detection are considered variability and low SNR [ 17 - 19 ]. Differentiating P300 from other ERP components is a major goal of many studies in this field [ 20 ]. Some studies have focused on this field for P300 classification because of the noticeable performance of deep learning methods [ 21 ]. Accordingly, deep-learning approaches are firstly surveyed, and conventional methods are also reviewed in the P300 detection studies.

A deep belief network was utilized for P300 classification with an accuracy of 91% on a dataset of 9 subjects [ 22 ]. Kshirsagar et al. [ 23 ] collected data on 10 subjects with 8 channels of EEG signal to classify P300 in Devanagari script using deep learning methods. They used autoencoder and CNN with an accuracy of 95.82%. Mingfei Liu et al. [ 24 ] proposed a deep learning method based on batch normalization for P300 detection on the BCI competition III dataset; a total of 79.09% accuracy for subject B was the best score among acquired results. A deep learning method based on deep belief networks was applied for P300 detection again on the BCI competition III dataset with an accuracy of 86.4% for subject B as the best result was reported [ 25 ]. A simple CNN was used, and an accuracy of 86.4% was achieved for the classification of P300 on the BCI competition III dataset [ 26 ].

Data were collected from the Fz, Cz, and Pz channels of 15 subjects, and different pipelines were examined for feature extraction plus classification, in which Bayesian linear discriminant analysis with 72.13% accuracy showed the best performance [ 27 ]. A method based on regularized group sparse discriminant analysis in [ 28 ] on Pz, Cz, and P8 channels of the BCI competition III dataset was implemented, with an accuracy of 90.00%. A combination of optimization and pattern recognition methods was used for P300 detection, tested on five datasets, with an accuracy of 96.70% as the best performance [ 29 ]. The decision tree was used as a classifier on an open dataset from an expert psychologist association, with an accuracy of 99.68% obtained [ 30 ].

This study introduces a method to extract spatial and temporal features as two main types of P300 component features by applying the Convolution Long-Short Term Memory AutoEncoder (CLSTM-AE), which is a combination of CNN for spatial and LSTM for temporal feature extraction. Also, an autoencoder was used as an unsupervised learning approach to reduce the time complexity and information redundancy of the method.

Material and Methods

In this cross-sectional study, a deep learning method was proposed based on both CNN and long short-term memory to extract efficient spatial and temporal features for P300 detection. A new method of augmentation was also used to deal with the imbalanced nature of the P300 dataset based on the Adaptive Synthetic Sampling Approach (ADASYN). The accuracy, sensitivity, recall, and F1-Score were considered to evaluate the performance of the proposed models.

Dataset and pre-processing

The brain signals were recorded according to the BCI P300 speller system. The P300 speller is based on the so-called oddball paradigm, showing that uncommon expected inputs cause a positive deflection in the EEG after about 300 ms, called the P300 component. Farwell and Donchin [ 15 ] established a P300 speller based on this paradigm by developing a protocol, in which a subject is presented with a 6×6-character matrix, as displayed in Figure 1. For the spelling of a single character, each of the matrix’s 12 rows and columns is then intensified in a random sequence (in the sequel, we refer to such a collection of 12 intensifications as a series). The participant was instructed to focus on the character they prefer to spell, in which an EEG P300-evoked potential appears in response to the intensification of a row or column containing the desired character. This series of intensifications is repeated 15 times for each character to cause the spelling method more reliable. The dataset for this competition, which is available on the competition homepage (https://www.bbci.de/competition/iii/), was collected from two distinct subjects.

Figure 1. P300 Speller Matrix

Before digitalization at 240 Hz, signals were bandpass filtered from 0.1 to 60 Hz [ 15 ]. The dataset was described in further detail in the BCI competition [ 30 ]. The classification of problems addressed is as follows: After stimulation and recording a 64-channel EEG signal for each subject, we wish to predict whether or not such a signal has a P300 component. Therefore, from the raw signal, a period of 1000 ms after stimulation was extracted. Most informative channels for each subject were then selected, including seven channels (Fz, Cz, Pz, C3, C4, PO7, and PO8) for subject A, and eight channels (Cz, C2, C3, T8, FC4, F2, F4, and F8) for subject B. According to the structure of the spelling matrix, only one-sixth of the recorded signals contain P300 (one row and one column from all 12 rows and columns), resulting in severely imbalanced sample sizes of P300 and non-P300 (nP300) data. In this case, the binary classification is prone to biasing towards nP300 samples.

Different types of augmentation methods were evaluated to address severely imbalanced sample sizes of P300 and non-P300 (nP300) data; firstly, random oversampling, as a simple method was implemented, the nature of this procedure is rather random. Although this method preserves the P300 structure and feature space, it merely copies the samples of the preliminary dataset, resulting in overfitting (Figure 2-I). To overcome this problem, other augmentation methods were applied based on synthetic sampling (SMOTE). As the basic one, random SMOTE (Figure 2-II) was applied. However, random SMOTE does not suffer from copying preliminary dataset samples, the output can be completely changed in each run because of its random nature. A version of SMOTE was used based on Support Vector Machine (SVM), whose random part is restricted to support vector samples (Figure 2-III) to control the random characteristics. In this method, augmentation is just applied on support vector samples with a substantial effect on P300 structure and distribution of features in comparison with primitive feature space. To overcome this challenge ADASYN was employed as a method, focusing on low-density parts of the data to synthesize new samples. However, ADASYN not only does not duplicate the existing dataset samples but also preserves the feature space as before. Further, the outputs of this method do not change in each run. Details of ADASYN are described in depth in [ 31 ]. P300 formation and feature space distribution before and after ADASYN augmentation are shown in Figure 2-IV. The proposed ADASYN augmentation method has presented a better result compared to other augmentation methods, depicted in Figure 2, such as SVM SMOTE and random SMOTE.

Figure 2. (a) A sample P300 signal structure before and after augmentation, (b) Feature space before and after augmentation in different methods: I. random oversampling, II. SMOTE, III. SMOTE based on SVM, IV. ADASYN. (SMOTE: Synthetic Sampling, SVM: Support Vector Machine, ADASYN: Adaptive Synthetic Sampling Approach)

The signal was sent through a bandpass Chebyshev filter with a cut-off frequency of 0.1-20 Hz to remove undesirable frequencies, and after all baselines were drifted. When the P300 component is extracted from the raw signal, each channel signal is a one-dimensional vector with a length of 240 samples, corresponding to the sampling frequency rate. All channels are cascaded along each other in a larger one-dimension signal to form each segment of data, as indicated in Figure 3a. The channels were transposed and then arranged behind each other as the frames of a video to prepare the data for the CLSTM-AE model (Figure 3b). Before augmentation, 20% of data from each class (750 P300 and 750 nP300) was selected as test data, and the remaining (1700 P300 and 12000 nP300) was augmented as train data in such a way that the two classes were balanced after augmentation. Only 20% of the training data was used for validation, while the remaining 80% was used for training of CLSTM-AE. The amount of data in each train, test, and validation group is mentioned in Table 1.

Figure 3. (a) One-dimension presentation of a segment of data. (b) Three-dimension presentation of a segment.

Data samples Number
train 19200
validation 4800
test 1500
Table 1.Splits of the dataset for train, testing, and validation of the proposed model

Method

The historical information is transmitted through a Recurrent Neural Network (RNN) procedure via the chain network structure. The RNN can learn long-term knowledge, but the growth of the vacant distance between the two memory units may cause memory deterioration and gradient vanishing. Hochreiter and Schmidhuber [ 32 ] attempted to solve the problem. The LSTM network was suggested for time series forecasting, considering the gate mechanism and controlling memory information using three separate gates’ functions. The internal portion of the LSTM utilizes a nearly complete linkage, causing an issue of duplication of information. The LSTM solely takes into account temporal connections and disregards the spatial correlation within the data. Conversely, CNN is constructed by layering multiple convolutional and pooling layers, which can efficiently extract spatial information from images without the need for temporal information [ 33 ]. However, LSTM uses matrix multiplication on its own, when it’s combined with two dimensions CNN, the input is a matrix, not a series. Accordingly, the proposed CLSTM replaces the matrix multiplication with the convolution operation in the LSTM gates to simultaneously extract spatial and temporal features. It captures the benefits of both CNN and LSTM and not only can extract temporal features by LSTM but can also characterize spatial features by CNN. Figure 4 depicts the CLSTM unit in detail.

Figure 4. The Convolution Long-Short Term Memory (CLSTM) unit structure

The enhanced CLSTM unit has three inputs for every gate: the memory information from the former unit, the output from the former unit, and the current time input. The upgraded CLSTM, which is made up of an input gate, a forget gate, an output gate, and a memory unit, can successfully learn time series information.

1) Gate of Forgetting: the gate of forgetting is responsible for selectively discarding unnecessary information from the memory unit, as follows:

ft=σ(Wf[ct-1ht-1xt]+bf) (1)

where σ is the function of activation, ⊙ represents the operation of convolution, and xt represents the input data at a specific point in time. ct-1 is the memory unit’s prior information, and ht-1 is the previous CLSTM unit output. Wf and bf are the forget gate’s weight and bias, respectively [ 34 ].

2) The gate of input determines whether or not to incorporate fresh information into the memory unit, consisting of a pair of steps: determining the information, updating via the sigmoid layer, and generating substitute information via the hyperbolic tangent function.

it=σ(Wi[ct-1ht-1xt]+bi) (2)

Ct~=tanh(Wc[ct-1ht-1xt]+bc) (3)

Here it and Ct~ refer to the output of the input gate and the memory unit’s substitute information, and hyperbolic tangent [ 34 ].

3) Memory Cell: It keeps the information up to date.

Ct=ft.Ct-1+it.Ct~ (4)

Which Ct is the memory unit’s stored information [ 34 ].

4) Ultimately, based on the prior computation, the gate of output computes the last output.

ot=σ(Wo[ct-1ht-1xt]+bo) (5)

ht=ot. tanh(Ct) (6)

In this context ot and ht denote the current time outputs of the gate of output and CLSTM unit, respectively [ 34 ].

Proposed CLSTM-AE Network

Figure 5a exhibits the proposed CLSTM-AE network structure. The encoder section includes CLSTM, BatchNormalization, and pooling layers for encoding input data. DeCLSTM (Deconvolutional LSTM), deconvolutional, BatchNormalization, and upsampling layers are used in the decoder to decode the encoded features. CLSTM-AE includes additional information on past units. This enhances the ability to learn from previous memory. The structure of the network enables efficient feature learning for intricate signal-processing applications.

Figure 5. (a) Architecture of proposed CLSTM-AE, (b) Artificial neural network as a classifier, (c) CNN-LSTM classifier(CLSTM-AE: Convolution Long-Short Term Memory AutoEncoder, CNN: Convolutional Neural Network, LSTM Long Short-term Memory)

1) Utilizing Equations, the CLSTM unit utilizes a memory unit and three gates which are employed to extract input data features (1)-(6). CLSTM generates:

hE=F(σ(W[ct-1ht-1xt]+b)). (7)

Here, F demonstrates the computation of the CLSTM memory cell and three gates. The CLSTM-AE encoder takes into account the CNN network structure, allowing it to compress and extract information from the CLSTM output using convolution and pooling layers. The decoder utilizes a deconvolution layer and an up-sampling layer to improve the quality of data reconstruction [ 34 ].

2) Convolution Layer: the result of the convolution layer is:

Ci=f(Xwi+bi (8)

The CLSTM output is represented as X, Wi denotes an activation function, and the rectified linear unit (ReLU), is utilized alongside the ith convolution kernel [ 34 ].

3) The pooling layer reduces data dimensionality to improve computational efficiency. The output of the pooling layer is used for the L-length feature in the convolution layers in an ith channel,

Pi(m)=m{Ci(mW,(m+1)W}0mLS (9)

where W denotes the window of pooling width, and S shows the size of the stride [ 34 ].

4) Layer of up-sampling: This is the inverse of the pooling layer. For the ith feature, the result of up-sampling is:

Uki={0kjkXik=jkk[t 2t]t=1 2…l (10)

In this case, the variable l denotes the length of input features. jk denotes the location of the maximum value obtained during the max-pooling process, and Uki is the kth element of Ui [ 34 ].

5) Deconvolution layer: This is the inverse operation of the convolution layer.

Di=ReLu(Xwi-+ci) (11)

Here ⊗ refers to the deconvolution computation, and wi- which refers to the kernel used for deconvolution [ 34 ].

6) DeCLSTM: DeCLSTM has a structure, which is similar to CLSTM, but deconvolution is utilized in place of convolution. The DeCLSTM output is:

hD=F(σ(W[ct-1ht-1xt]+b)) (12)

where F is the DeCLSTM computation which refers to the components of the memory cell and three gates. ch-1 is the former output of the DeCLSTM unit [ 34 ].

CLSTM-AE Training

Table 2 provides an overview of the CLSTM-AE training process, which aims to minimize the loss function by reconstructing data. In this process, the autoencoder utilizes the reconstructed error as its loss function. The Mean Square Error (MSE) (Equation (13)) is employed as the loss function in CLSTM-AE training [ 34 ], as follows:

E=1mn=1myn-xn (13)

where m is the number of data points, yn is the predicted data point, and xn is the original data.

Input: Training data x
1-Establish the hyperparameters
2- Randomly initialize weights and biases.
3- At iteration number N.
4- Enter the training data, X.
5- Minimize the loss function during the training of the CLSTM-AE model.
6- Compute the feature output during the encoding phase, h.
7- Compute the reconstructed output during the decoding phase, y.
8- Compute the reconstructed error, E.
9- Update the decoder's parameters, Wd and Bd.
10-update the encoder's parameters Wd Bd
11-End of For
Output: Feature h and reconstructed data y.
Table 2.The Convolution Long-Short Term Memory AutoEncoder (CLSTM-AE) Training Process [ 34 ]

In this research, the CLSTM-AE model is optimized using AdaGrad. Because of its capability of adaptive learning rate. The CLSTM-AE is divided into two parts: encoder and decoder. The data are abstracted into their main components at the end of the encoder section (named bottleneck of autoencoder), using max-pooling and convolution layers. A combination of convolution and LSTM layers is used to extract both the spatial and temporal features. The batch normalization layer is used during training to avoid overfitting. The autoencoder architecture was designed with the fewest layers to reduce the computational burden. The data are then reconstructed as input at the end of the decoder part, which is repeated until the minimum error between the data and its reconstructed version is obtained during 50 epochs. The network has learned to reconstruct the data with minimum error, and the encoder has learned how to extract the main components of data at its final layer. Following training, the main components of data are extracted from the autoencoder’s bottleneck. The autoencoder is then saved, and the encoder section is separated from the autoencoder. To classify data into P300 and nP300, test data are fed to the encoder to extract its main components. The extracted features of the final layer of the encoder are flattened and fed into an artificial neural network during 250 epochs for classification, showing dropout to prevent overfitting during the training process and sigmoid as an activation function at its last layer (Figure 5b).

Results

After extracting features of the test data by a trained encoder, they are fed into an artificial neural network in 250 epochs for 15 signal trials of each subject. The accuracy and loss of the model for subject A and subject B are depicted in Figure 6. The model is also evaluated according to other metrics, presented in Table 3. Furthermore, we also prepared the results of the other state-of-the-art works on the same dataset in Table 4, showing the proposed method has superiority over most of the last outstanding papers.

Figure 6. Accuracy (a and c) and loss (b and d) of the proposed method during training and validation for subject A, and subject B, respectively.

Accuracy (%) Precision (%) Sensitivity (%) F1-Score (%)
CNN-LSTM CLSTM-AE CNN-LSTM CLSTM-AE CNN-LSTM CLSTM-AE CNN-LSTM CLSTM-AE
Subject A 91 95 88 91 98 99 95 95
Subject B 90 94 85 90 98 99 91 94
Table 3.Results of CNN-LSTM and CLSTM-AE (the proposed method)(CNN: Convolutional Neural Network, LSTM Long Short-term Memory, CLSTM-AE: Convolution Long-Short Term Memory AutoEncoder)
Reference number Dataset (Subject) Accuracy (%)
[ 24 ] A 76.00
B 80.00
[ 35 ] A 94.20
B 94.20
[ 36 ] A 81.00
B 84.00
[ 37 ] A 86.39
B 86.39
[ 15 ] A 88.40
B 90.00
Table 4.Results of the state-of-the-art works on the same dataset

Further, a CNN-LSTM was implemented by the same dataset to identify the superiority of CLSTM-AE to CNN-LSTM. The architecture of the CNN-LSTM is shown in Figure 5c, and results are presented in Table 3 Regarding accuracy, precision, sensitivity, and F1-Score, CLSTM_AE shows better performance as compared to CNN-LSTM. Additionally, CLSTM-AE extracts an abstraction of the data reducing the computational burden.

Discussion

The current investigation aimed at employing a 3D input CLSTM_AE model for the P300 detection. The proposed model was implemented based on deep learning principles to redefine the problem of P300 classification from a signal classification into a video classification problem to use both spatial and temporal features of EEG signal in two stages. In the first stage, the autoencoder as an unsupervised feature extractor is used to not only extract but also abstract the principle features in its latent space. In the second stage, it’s checked whether a signal contains P300 or not using a neural network as a classifier. One of the main challenges of P300 classification in the P300 speller matrix is the imbalanced nature of data [ 38 ]; accordingly, the ADASYN approach was used for augmenting the P300 class. ADASYN has three advantages, including maintaining the primary distribution of features in the feature space after augmentation, maintaining the morphological structure of P300, and the non-randomness characteristic of this method as compared to other synthetic augmentation methods, which causes the stability of augmentation in each run [ 31 ]. Accuracy as a metric to survey the number of true positive and true negative classified samples, precision as a metric to assess the ratio of true positive samples on true and false positive samples, sensitivity to examine the performance of the classifier on true positive samples and false negative samples, and F1-score to evaluate the trade-off between sensitivity and precision are used for performance evaluation. Based on the results in Table 4, the proposed method has superiority over most of the last outstanding papers [ 15 , 24 , 36 - 38 ].

Conclusion

In this work, a CNN-LSTM model is designed for P300 classification to redefine the signal classification problem to video classification and explore the feasibility of exploiting all the features of the data. Despite the high-quality classification, this method in its original form was computationally intensive. By considering CNN-LSTM’s ability to extract both spatial and temporal features in the form of video classification, we sought to preserve this advantage of CNN-LSTM and find a solution to the aforementioned drawbacks. To this end, due to the unsupervised nature of autoencoders, we not only retain the benefits of CNN-LSTM but also abstract the data based on features extracted efficiently from CNN-LSTM, reducing the computational burden and SNR improvement and finally provide more accurate classification results. Also, to solve the problem of unbalanced data, conventional methods repeat the number of segments in classes with a small number of segments. These techniques copy the data that may cause overfitting which are replaced by ADASYN as a new method for P300 augmentation. Compared to other studies, this work involves changing the view of the problem from signal classification to video classification, using all features of the data in the time domain and extracting spatial and temporal features simultaneously, and using minimal preprocessing algorithms before feature extraction. Furthermore, in the field of ERP, we use a new data augmentation method that preserves the structure and feature space of data as before. Future work could explore more efficient channels, parameter tuning, and other network architectures.

Authors’ Contribution

R. Afrah, Z. Amini, and R. Kafieh conceived the idea and analyzed the results. The first draft was written by all authors. The final manuscript was revised by Z. Amini. All the authors read and approved the final version of the manuscript.

Ethical Approval

Approval of all ethical procedures was granted by the Institutional Review Board and Ethics Committees of the National Institute for Medical Research Development under Approval Nos. IR.MUI.REARCH.REC.1401.237, Isfahan University of Medical Sciences, Isfahan, Iran.

Funding

This work was supported in part by the Vice Chancellery for Research and Technology of Isfahan University of Medical Sciences under Grants 2401134.

Conflict of Interest

None

References

  1. Jamil N, Belkacem AN, Ouhbi S, Lakas A. Noninvasive Electroencephalography Equipment for Assistive, Adaptive, and Rehabilitative Brain-Computer Interfaces: A Systematic Literature Review. Sensors (Basel).. 2021; 21(14):4754. Publisher Full Text | DOI | PubMed [ PMC Free Article ]
  2. Amini Z, Abootalebi V, Sadeghi MT. Isfahan, Iran: IEEE; 2010.
  3. Eastmond C, Subedi A, De S, Intes X. Deep learning in fNIRS: a review. Neurophotonics. 2022; 9(4):041411. Publisher Full Text | DOI | PubMed [ PMC Free Article ]
  4. Hasbulah MH, Jafar FA, Nordin MH. Fundamental of Electroencephalogram (EEG) Review for Brain-Computer Interface (BCI) System. Int Res J Eng Technol. 2019; 6(5):1017-28.
  5. Kobler RJ, Hirata M, Hashimoto H, Dowaki R, Sburlea AI, Müller-Putz GR. Austria: Institute of Neural Engineering Graz BCI; 2019.
  6. Abiri R, Borhani S, Sellers EW, Jiang Y, Zhao X. A comprehensive review of EEG-based brain-computer interface paradigms. J Neural Eng. 2019; 16(1):011001. DOI | PubMed
  7. Kundu S, Ari S. Brain-Computer interface speller system for alternative communication: a review. IRBM. 2022; 43(4):317-24. DOI
  8. Shojaedini SV, Morabbi S, Keyvanpour MR. A New Method to Improve the Performance of Deep Neural Networks in Detecting P300 Signals: Optimizing Curvature of Error Surface Using Genetic Algorithm. J Biomed Phys Eng. 2021; 11(3):357-66. Publisher Full Text | DOI | PubMed [ PMC Free Article ]
  9. Saliasi E, Geerligs L, Lorist MM, Maurits NM. The relationship between P3 amplitude and working memory performance differs in young and older adults. PLoS One. 2013; 8(5):e63701. Publisher Full Text | DOI | PubMed [ PMC Free Article ]
  10. Li L, Gratton C, Yao D, Knight RT. Role of frontal and parietal cortices in the control of bottom-up and top-down attention in humans. Brain Res. 2010; 1344:173-84. Publisher Full Text | DOI | PubMed [ PMC Free Article ]
  11. Oralhan Z. A new paradigm for region-based P300 speller in brain computer interface. IEEE Access. 2019; 7:106618-27. DOI
  12. Li Y, Nam CS, Shadden BB, Johnson SL. A P300-based brain–computer interface: Effects of interface type and screen size. Int J Hum Comput Interact. 2010; 27(1):52-68. DOI
  13. Fazel-Rezai R, Ahmad W. InTech; 2011.
  14. Lu Y, Bi L. EEG signals-based longitudinal control system for a brain-controlled vehicle. IEEE Trans Neural Syst Rehabil Eng. 2018; 27(2):323-32. DOI
  15. Zhang Z, Yu X, Rong X, Iwata M. Spatial-temporal neural network for P300 detection. IEEE Access. 2021; 9:163441-55. DOI
  16. Qu J, Wang F, Xia Z, Yu T, Xiao J, Yu Z, Gu Z, Li Y. A novel three-dimensional P300 speller based on stereo visual stimuli. IEEE Trans Human-Machine Syst. 2018; 48(4):392-9. DOI
  17. Kim M, Kim J, Heo D, Choi Y, Lee T, Kim SP. Effects of Emotional Stimulations on the Online Operation of a P300-Based Brain-Computer Interface. Front Hum Neurosci. 2021; 15:612777. Publisher Full Text | DOI | PubMed [ PMC Free Article ]
  18. Bulat M, Karpman A, Samokhina A, Panov A. Playing a P300-based BCI VR game leads to changes in cognitive functions of healthy adults [Internet]. bioRxiv [Preprint]. 2020 [cited 2020 May 30]. Available from: https://www.biorxiv.org/content/10.1101/2020.05.28.118281v3
  19. Rasheed S. A review of the role of machine learning techniques towards brain–computer interface applications. Mach Learn Knowl Extr. 2021; 3(4):835-62. DOI
  20. Alzahab NA, Apollonio L, Di Iorio A, Alshalak M, Iarlori S, Ferracuti F, et al. Hybrid Deep Learning (hDL)-Based Brain-Computer Interface (BCI) Systems: A Systematic Review. Brain Sci. 2021; 11(1):75. Publisher Full Text | DOI | PubMed [ PMC Free Article ]
  21. Zhang X, Yao L, Wang X, Monaghan J, McAlpine D, Zhang Y. A survey on deep learning-based non-invasive brain signals: recent advances and new frontiers. J Neural Eng. 2021; 18(3):031002. DOI | PubMed
  22. Cortez SA, Flores C, Andreu-Perez J. Lima, Peru: IEEE; 2020.
  23. Kshirsagar GB, Londhe ND. Improving Performance of Devanagari Script Input-Based P300 Speller Using Deep Learning. IEEE Trans Biomed Eng. 2019; 66(11):2992-3005. DOI | PubMed
  24. Liu M, Wu W, Gu Z, Yu Z, Qi F, Li Y. Deep learning based on batch normalization for P300 signal detection. Neurocomputing. 2018; 275:288-97. DOI
  25. Lu Z, Gao N, Liu Y, Li Q. Beijing, China: IEEE; 2018.
  26. Shan H, Liu Y, Stefanov TP. IJCAI; 2018.
  27. Delgado JMC, Achanccaray D, Villota ER, Chevallier S. Riemann-Based Algorithms Assessment for Single- and Multiple-Trial P300 Classification in Non-Optimal Environments. IEEE Trans Neural Syst Rehabil Eng. 2020; 28(12):2754-61. DOI | PubMed
  28. Wu Q, Zhang Y, Liu J, Sun J, Cichocki A, Gao F. Regularized Group Sparse Discriminant Analysis for P300-Based Brain-Computer Interface. Int J Neural Syst. 2019; 29(6):1950002. DOI | PubMed
  29. Bianchi L, Liti C, Liuzzi G, Piccialli V, Salvatore C. Improving P300 Speller performance by means of optimization and machine learning. Ann Oper Res. 2022; 312:1221-59. DOI
  30. Srimaharaj W, Chaisricharoen R. A novel processing model for P300 brainwaves detection. J Web Eng. 2021; 20(8):2545-70. DOI
  31. He H, Bai Y, Garcia EA, Li S. Hong Kong: IEEE; 2008.
  32. IOS Press; 2022.
  33. Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y, Al-Shamma O, et al. Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J Big Data. 2021; 8(1):53. Publisher Full Text | DOI | PubMed [ PMC Free Article ]
  34. Yu J, Liu X, Ye L. Convolutional long short-term memory autoencoder-based feature learning for fault detection in industrial processes. IEEE Trans Instrum Meas. 2020; 70:1-5. DOI
  35. Sarraf J, Pattnaik PK. A study of classification techniques on P300 speller dataset. Mater Today Proc. 2023; 80(3):2047-50. DOI
  36. Ghazikhani H, Rouhani M. A deep neural network classifier for P300 BCI speller based on Cohen’s classtime-frequency distribution. Turkish J Electr Eng Comput Sci. 2021; 29(2):1226-40. DOI
  37. Ditthapron A, Banluesombatkul N, Ketrat S, Chuangsuwanich E, Wilaiprasitporn T. Universal joint feature extraction for P300 EEG classification using multi-task autoencoder. IEEE Access. 2019; 7:68415-28. DOI
  38. Lee T, Kim M, Kim SP. Improvement of P300-Based Brain-Computer Interfaces for Home Appliances Control by Data Balancing Techniques. Sensors (Basel).. 2020; 20(19):5576. Publisher Full Text | DOI | PubMed [ PMC Free Article ]