A Comprehensive Review for Emotion Detection Based on EEG Signals: Challenges, Applications, and Open Issues

A Comprehensive Review for Emotion Detection Based on EEG Signals: Challenges, Applications, and Open Issues

Awf Abdulrahman* Muhammet Baykara

Department of Information Technology, Duhok Polytechnic University, Duhok 42001, Iraq

Department of Software Engineering, Technology of Faculty, Firat University, Elazig 23119, Turkey

Corresponding Author Email: 
mbaykara@firat.edu.tr
Page: 
1189-1200
|
DOI: 
https://doi.org/10.18280/ts.380430
Received: 
11 March 2021
|
Revised: 
19 July 2021
|
Accepted: 
25 July 2021
|
Available online: 
31 August 2021
| Citation

© 2021 IIETA. This article is published by IIETA and is licensed under the CC BY 4.0 license (http://creativecommons.org/licenses/by/4.0/).

OPEN ACCESS

Abstract: 

Emotion classification based on physiological signals has become a hot topic in the past decade. Many studies have attempted to classify emotions using various techniques, to discover human emotions accurately. This study focused on listing the most recent studies that have classified emotions based on electroencephalogram (EEG) signals. This study also focused on solving the problems and challenges facing researchers in emotion classification and EEG applications used in several fields. The plan of this study is based on a strategy with three aspects within specific rules: The first aspect is the methods; we chose studies that included new methods to extract features. The second aspect is the data sets. We tried to choose a study that classified the same data set. The third aspect is applications; we have listed many applications of the EEG in several areas. We concluded from this study that detecting human emotions using the EEG signals is one of the most reliable and widely used methods of detecting emotions in the past few years. Also, we have noticed that the EEG can detect human emotions, especially in psychiatry, for example, for epileptic patients whose emotions cannot be extracted using traditional methods such as facial expressions and tone of voice.

Keywords: 

electroencephalogram, classification, emotion recognition, features extraction, EEG, FFT, DWT

1. Introduction

Emotions play an important role in our daily life [1]. Emotion is the human response to what is going on around them of events in every moment. Interest in recognizing emotions has increased in the past two decades [2, 3]. Because if a person can discover and organize his emotions, his life will become better [1]. Researchers have tried to develop theories of emotion detection and methods and techniques for detecting emotions [2, 4]. One of the most commonly used emotion detection methods is the emotion detection method based on physiological signals received from the brain (EEG) [1]. This method raised the efficiency of emotion detection due to the reliability obtained from physiological signals. A person cannot control his inner emotions and the signals collected from his brain. Therefore, the interaction between humans and machines in the field of emotion detection has become a very hot topic for psychologists, medical scientists, and researchers in computer science [1, 5]. There is no agreement about the nature of emotions among psychologists, neuroscientists, and other researchers, as there are two different views of emotions among scientists: the first opinion considers emotion as a general state of individuals, and the other opinion considers emotion as a physiological and physical interaction [6]. For example, if a person is in a forest and suddenly a predator appears in front of him, then according to the first opinion, he will feel fear because he knows that this animal will prey on him. As for the second opinion, the appearance of a predator in front of the person leads to an acceleration of the heartbeat and blood pressure, which generates a feeling of fear due to the brain's response [6]. Emotions can be classified according to three basic theories, namely: Plutchik's theory, Ekman's theory, and James Lange's theory [6-9]. Emotions are classified according to Pluchik's theory into two categories. The first section is called basic emotions, which contains the following emotions: anticipation, joy, confidence, sadness, fear, surprise, anger, disgust, and the second section is called secondary emotions, and it is a combination of basic emotions such as aggression, submission, contempt, dread, and others [6]. As for the classification of emotions based on Ekman's theory [8] (the discrete model theory), This theory presented 6 of the basic emotions: sadness, happiness, anger, disgust, fear, and surprise [6, 8]. As for James and Lange theory [9], it classifies emotions into a two-dimensional space called the arousal-valence model, in which valence is from unpleasant (negative) to pleasant (positive), and arousal from passive (low) to active (high) [2, 6, 9], as in Figure 1 [10].

Figure 1. James and Lange theory (2-D space) [10]

The rest of the sections of this study, the second section (An electroencephalogram), where a brief explanation was given about the device used to collect the data. The third section (previous Works), where the most important previous studies were listed. The fourth section (Public Data sets in the literature) summarizes a group of emotion databases. The steps for implementing this study are explained in the fifth section (stages of this study). The sixth section (methods in the literature) lists the most important methods used in the literature to detect emotions based on EEG. In the seventh section (Challenges and future works), more challenges and problems that may face researchers in classification emotions are listed. And the last section is a conclusion.

2. An Electroencephalogram (EEG)

An electroencephalogram (EEG) collects data from a person by placing a wearable device around the head in a non-invasive manner, as in Figure 2 [11]. This device contains electrodes that are placed around the head according to the 10-20 system [12, 13], as shown in Figure 3 [13]. This device monitors brain activity and records signals from the brain. These devices transfer the recorded signals to computers via wired, wireless, or other communication [12]. Then researchers and scientists analyze that data to make the conclusions they want.

Figure 2. EEG device components [11]

Figure 3. The 10-20 system a) front side b) topside

Electroencephalography (EEG) is an easy and inexpensive way that has been used to record brain activity and has also demonstrated that it is through brain activity that emotions can be detected [14]. Measuring brain activity via EEG devices has helped scientists understand the persons’ emotions with disabilities whose emotions are difficult to detect through their facial expressions [15].

The process of classification emotions via the EEG passes through several primary and sequential stages, as in Figure 4. In the first stage (data collection stage), the type of stimuli that will be used to classify emotions (Visual stimuli such as viewing pictures, aural stimuli such as listening to music, or aural/visual stimuli such as viewing videos) watching film clips) is chosen. Then the number of people whose brain activity will be recorded and the number of EEG channels through which the signal will be collected is determined. Next, data is recorded from the brain as the subjects interact with the stimuli we mentioned. In the second stage (pre-processing stage), after collecting data from people, raw data is processed (pre-processing), for example, removing noise from the signal, reducing or increasing the number of samples, and dividing the signal into training data and test data, this stage helps in filtering data to improve results. In the third stage (feature extraction stage), features are extracted from pre-processed data, the techniques are used to extract features from EEG pre-processing signals, for example, fast Fourier transform (FFT), discrete wave transform (DWT), and statistical feature method (SF), and so on. This stage helps extract features from the original data that gives the classifier the ability to classify emotion accurately. In the fourth stage (classification stage), the classifiers that will classify the emotions are selected. The number of classes to be classified must be determined at this stage (i.e. two classes, three classes, or four classes), depending on the model used (arousal-valence model, negative and positive emotions model). In the fifth stage (The results) in this stage, the model that has been trained is evaluated.

Figure 4. The 10-20 system a) front side b) topside

3. Previous Works

Alakus et al. introduced a new, freely available dataset based on physiological brain signals (EEG) and computer games as visual/aural stimuli called GAMEEMO to classify four emotions: boring, calm, horror, and funny. Their experiment was conducted with the participation of 28 subjects. Data were collected using a wearable and mobile device called EMOTIV EPOC with 14 channels. Then the research team tested a dataset data that they collected by using different methods to extract features, including statistical method, DWT, Hjorth features, Shannon entropy, and others, then used three machine learning classifiers: MLPNN, SVM, K-NN to classify two models of emotion detection models: Negative and positive model, and Valence-Arousal model [2]. Chao and Liu suggested a method using a deep belief-conditional random field (DBN-CRF), integrated with an improved deep belief network with glia chains (DBN-GC), to classify three emotion datasets: DEAP, SEED, and AMIGOS. Initially, the research team extracted the features in three ways: Time domain, frequency domain, and Time-Frequency domain [16]. Sharan et al. suggested a method for detecting personality traits using EEG signals. They used the IAPS image dataset and the English version of the FilmStim video dataset as stimuli to collect data from 18 individuals. Then, they used an inter-hemispheric asynchrony technique to extract the features, and then four machine learning classifiers were used (KNN, LG, NB, and SVM) to classify 16 personal traits [17]. Priya et al. introduced a method for classifying stress during work stress. The team selected a physio bank database in their experiment. First, they extracted the features using PSD technology and then used two machine learning classifiers (KNN and SVM) [18]. Tao et al. proposed a new method for classifying emotions using an attention-based convolutional recurrent neural network (ACRNN) where the team applied this method to the DEAP and DREAMER dataset [19]. Gao et al. proposed a new approach by combining the Granger Causality Method (GC) and the Histogram of Oriented Gradient (HOG) method to extract features and then using the Support Vector Machine Classifier (SVM) two types of emotions were classified: stress and calm [20]. Hassan et al. used the Unsupervised Deep Belief Network (DBN) to extract features from the DEAP physiological dataset. They then combined the features extracted from DBN with statistical features such as EDA, PPG, and zENG. They used the fine gaussian support vector machine (FGSVM) algorithm to classify five basic emotions [21]. Yang et al. present an approach based on a multi-column CNN algorithm to classify the emotions based on EEG signals taken from the DEAP database. The research team achieved a performance accuracy of 90% [22]. Xing et al., used a new approach to extract features from EEG signals, used the Long Short-Term Memory Recurrent Neural Network (LSTM-RNN) to build a linear EEG mixing model and an emotion timing model using a Stack AutoEncoder (SAE). This model was applied to the DEAP dataset to classify emotions, and the team obtained a performance accuracy of 81.10% in valence and 74.38% in arousal [23]. Martínez-Rodrigo et al. studied two nonlinear indices to classify emotion (emotional states of calm and negative stress) based on the brain's physiological signals: CMAAPE and CMQSE. This study applied the DEAP dataset to classify calm and negative stress, and they used two machine learning classifiers: SVM and DT. Then, they assessed the differences between the entropy values for each time scale of the calm states and negative stress for both indicators using the statistical method in addition to the analysis of variance (ANOVA) [24]. Salama et al., proposed a new, multimodal approach to classifying human emotions. They used a three-dimensional convolutional neural network (3D-CNN) to extract features from electroencephalogram (EEG) signals and video data of human faces. Then they built three methods of classifying emotions: the emotion recognition approach based on EEG signals, the face-based emotion recognition approach, and the fusion-based emotion recognition approach. In the first approach based on EEG signals, the team used 3D-CNN to obtain final predictions of EEG signals using the arousal-valence model [25]. Goshvarpour and Goshvarpour built a new approach to extract nonlinear features to classify emotions based on physiological brain signals (EEG). They studied a physiological DEAP dataset. This approach relied on feature minimization by using three algorithms to reduce features: random subset feature selection (RSFS), sequential floating forward selection (SFFS), and sequential forward selection (SFS), and an SVM classifier to classify valence-arousal emotions [26]. Liao et al. proposed an approach based on a convolutional recurrent neural network to classify emotions, in which two methods were used to identify features: the convolutional neural network for spatial representations of EEG signals and the LSTM algorithm for the temporal representations of EOG, EMG, GSR, RSP, BVP, and TMP signals. This study was applied to the DEAP dataset, which is of interest to us in this study is the first method based on the EEG signals, they used three consecutive two-dimensional convolutional layers (2D-CNN) to extract the features, and each layer has the same size of the nucleus 3 × 3 [27]. Cui et al, present a new approach based on the Regional-Asymmetric Convolutional Neural Network (RACNN) to classify valence-arousal emotions based on EEG signals downloaded from the DEAP and DREAMER dataset. Their study focused on extracting temporal, regional, and asymmetric features, where they used continuous one-dimensional Convolutional layers in the temporal feature extractor, and then two two-dimensional Convolutional layers to capture regional information. Then they proposed an asymmetric differential layer (ADL) to extract the asymmetric characteristic of emotion between the left and right lobes of the brain. Their model achieves an average accuracy of over 95% with both datasets [28]. Ma et al, presented a method based on a multimodal residual LSTM (MMResLSTM) to classify arousal-valence emotions based on EEG brain signals. The DEAP dataset was used, which is from the important emotion datasets. They reduced EEG samples, removed impurities, computed average signal, and so on. Then they applied some feature extraction methods. for example, db4 discrete wavelet transforms, wavelet entropy, and wavelet energy features. Finally, they obtained a classification accuracy of 92.87% for arousal and 92.30% for valence [29]. Tripathi et al. used two neural network techniques (deep learning) to classification emotions: a simple deep neural network (DNN) and a convolutional neural network (CNN). They also used a statistical method to extract features [30]. Alakus et al. analyzed a GAMEEMO dataset based on physiological brain signals (EEG) to classify the negative and positive emotions model. The research team used the spectral entropy values of all EEG channels to extract the features, then used the bidirectional long-short term memory algorithm for deep learning to classify the emotions, finally, using the receiver operating system. characteristic (ROC) curve to evaluate the performance of the algorithm used [31]. Tarnowski et al. presented an experiment to analyze the galvanic skin response (GSR) and electroencephalogram (EEG) signals. Twenty-seven people participated in their experiment, where they were shown 21 short film clips, and then their emotions were classified based on the arousal-valence model. The team used a fast Fourier transform (FFT) method to extract features from the EEG signal, and two machine learning classifiers were used: SVM and k-NN [32]. Al-Nafjan et al, the research team chose the Deep Neural Network (DNN) to classify emotions based on the EEG. they used the DEAP database with two feature extraction methods: power spectrum density (PSD) and frontal asymmetry features. The approach they chose was efficient at 82% accuracy [33].

As can be understood from the studies in the literature, it has been seen that human emotions based on EEG signals can be found with high accuracy with various deep learning algorithms. Similarly, machine learning methods often give high success rates. It is seen that the DEAP dataset is generally used in studies. This dataset has a usage rate of more than 85%. The repetition of studies with different methods based on a single data set and similar stimuli does not add any scientific innovation to the literature. For this reason, it should be emphasized that new EEG datasets like GAMEEMO are needed to overcome this weakness.

4. Public Datasets in the Literature

There are many emotion databases freely available to the public that anyone can download and analyze without the need for someone's approval or the participation of an organization in his work [6]. We will describe some general datasets for emotion detection based on EEG signals, as in Table 1.

4.1 GAMEEMO Dataset

The GAMEEMO database is an emotion dataset based on brain physiological signals (EEG). This dataset was recorded by Alakus et al. [2], conducted with the participation of 28 subjects at Firat University, Department of Software Engineering. The age range of the subjects used in the study is between 20-27 years old. It was not specified how many of the subjects used in the study were women and men. Data were collected from the participants using four computer games as stimuli to extract the four basic emotions (boring, calm, horror, and funny) for 5 minutes. A total of 20 minutes long EEG data was available for each subject. Data were collected through 14 channels using the EMOTIV EPOC + Mobile EEG wearable device. 38252 samples were collected from each participant and for each game [2].

4.2 DEAP Dataset

It is a dataset of emotions recorded by Koelstra et al. [34] in which an arousal-valence model was used with 40 video clips as stimuli. This experiment was carried out in two laboratories, Geneva and Twente, in 2012, where 32 people participated in this experiment. Data were collected for 32 EEG channels, in addition to 4 EMG channels, 4 EOG signals, 2-channel GSR signals, 2 ERG signals, temperature, respiratory rate, and blood pressure. Five emotions were recorded: arousal, valence, liked or disliked, familiarity, and dominance [6, 34].

4.3 MAHNOB-HCI Dataset

MAHNOB-HCI is a multimedia database recorded by Soleymani et al. In 2012 [35]. Data were collected from 27 individuals while they viewed 20 videos and photos. The data collected are EEG signals (32 channels), ECG signals (3 signals), ERG (2 channels), GSR (2 channels), respiratory capacity, and skin temperature signals. This experiment went through two parts. In the first session, the participants were asked to watch video clips, and then they were asked to fill out a questionnaire about their emotions after watching the videos. At the same time, in the second session, the videos and pictures were shown once with right and wrong labels and once without labels [35].

4.4 SEED Dataset

Zheng and Lu [36] recorded SJTU's EEG dataset (SEED), a physiological dataset collected from 15 individuals while they watched the videos. The dataset was recorded to classify three types of emotions, positive, negative, and natural. Participants were asked to fill out a questionnaire after watching the videos. The EEG was recorded in three separate sessions. There was a week or more between one session and another, and the EEG signals were recorded according to the international standard system 10-20 [6, 36].

Table 1. Different kinds of emotion datasets

Dataset

Stimuli

Sub.

EEG

Emotion

Ref.

GAMEEMO

4 Video Game

28

14 EEG signals

Arousal-valence

[2]

Positive and Negative

[31]

DEAP

40 Video clips

32

32 EEG, 4 EMG, 4 EOG, 1 GSR, 1 RR

Arousal-valence

[16, 19, 22, 23, 25]

stress and calm

[20]

negative stress

[24]

stress recognition

[37]

MAHNOB-HCI

20 Video clips and Pictures

27

16 EEG, 3 ECG, 2 GSR, 1 RR, 1 Temp

Arousal-Valence

[38]

SEED

15 video clips

15

15 EEG signals

Arousal-Valence

[16]

5. The Methods in the Literature

There are many methods and techniques used in analyzing a data set of different emotions based on the physiological signals in the literature. In this study, we will summarize the latest methods and techniques used to classifying emotions in the literature, as in Table 2.

5.1 Discrete Wavelet Transform (DWT)

DWT is an important and widely used method for analyzing time-frequency signals in different frequency bands with different resolutions due to its unstable properties. The DWT decomposition of a signal uses a high and low pass filter, where DWT analyzes the signal into detailed and approximate information. The first decomposition in the DWT is considered in the approximation and detailed coefficients (A1, D1) to get the level 1 band, then A1 is decomposed into the approximation and detailed coefficients (A2, D2) again. Thus decomposition continues to a specified number of times [39-42], as in Figure 5.

Figure 5. DWT sub-bands decomposition

5.2 Fast Fourier Transform (FFT)

The Fast Fourier Transform (FFT) is an algorithm that decomposes a complex signal into smaller transformations, using a discrete Fourier transform (DFT) for a sequence and an inverse (IDFT) [32, 43, 44]. FFT is one of the most used algorithms in signal analysis [45]. As for the features that are extracted based on FFT, they depend on decomposing the original signal into smaller transformations. Then, to obtain the resulting signal, the decaying signals are combined with the decaying signals and removing the low-frequency signals [43-45].

5.3 Statistical Features (SF)

The method of mathematical-statistical analysis is applied to the signal to obtain more information about this signal, such as the mean, median, deviation, kurtosis, standard deviation, skewness, minimum, maximum, and others [46]. To clarify the impact of statistical features on the final results of the model, we will use the study [47] as a reference. In this study, emotions were classified using the statistical feature method. The study was divided into three stages: In the first stage, three statistical features were extracted (minimum value, maximum value, and the mean). The model was trained using only the three features and achieved an accuracy of 2.03%. Then, four other features were extracted (Variance, Standard deviation, Wave entropy, and Power bandwidth). The model was trained again using the new features and the old ones, and it achieved an accuracy of 32.14%. We note the impact of adding additional features on the results. Then, in the third stage, two additional features were extracted (Skewness and Kurtosis), and the model achieved an accuracy of more than 92%. To calculate mean, median, and stander deviation, we will use the equations as in Figure 6.

5.4 Deep learning methods (1D, 2D, and 3D-CNN)

The Convolutional neural network (CNN) algorithm is used to extract additional features from the raw data of the EEG signal in several recent studies, for example:

Ding et al. [48] constructed a one-dimensional convolutional neural network (1D-CNN) to extract the features. This network consists of 3 sequential segments: the temporal learner consisting of an EEG input segment (4 EEG channels * 1024 data points for each channel * 1), the spatial learner consisting of (4 * 9), and the classifier [48].

Liao et al. [27] used three two-dimensional Convolutional neural networks (2D-CNN) to extract the features. Each of these layers is composed of a core of equal size (3 x 3). The first layer also consists of 64 feature maps, the second layer 128, and the third layer 256 feature maps [27].

Salama et al. [25] used a three-dimensional Convolutional neural network (3D-CNN) as a deep learning technique. Six basic layers are used to extract features and train the model. It is the input layer, followed by two 3D convolutional layers. After each 3D layer, there is a max-pooling layer and then the last layer, the feature extraction layer [25].

Wang et al. [49] presented a novel approach to automatically detect human emotions based on a Residual block-based deep convolutional neural network (CNN) to extract features and classify emotion using EFDMs automatically.

Wen et al. [50] proposed an end-to-end model based on convolutional neural networks (CNNs) to improve feature extraction performance based on EEG signals. Pearson's correlation coefficient first rearranged the original EEG channels. Then fed the rearranged EEGs into the CNN.

Figure 6. Some Statistical equations [46]

5.5 Short-Time Fourier Transform (STFT)

The STFT method is one of the methods for converting the time domain signal into the time-frequency domain. This method combines the time domain and frequency domain in the signal analysis [45] to track the low energy level present in the original signal. This method has the advantage of adapting the Fourier transform to analyze only a small section of the signal at a given time. STFT splits the signal into small sequential or nested data frames, in contrast to the FFT method applied to all data frames [51]. For this reason, the STFT method is more efficient than the FFT method if the amount of data is large [45].

5.6 Power Spectral Density (PSD)

The Power spectral density (PSD) method is an important method to extract features from EEG physiological signals. It has been extensively studied to classify emotions because it is a good way to process narrow-band fixed signals, as it expresses the activity of each of the EEG signal bands (delta, theta, alpha, beta, and gamma) [52, 53].

5.7 Asymmetry in frequency for electrode Pairs (Based on PSD)

The features are extracted in this way by subtracting the absolute PSD values of the poles of the right lobe of the brain from the absolute PSD values of the poles of the left lobe [54]. If we use, for example, 14 channels in the experiment, that is, we have seven pairs of electrodes, and the features that we will get are equal to 7 * the number of frequency points.

5.8 Histogram of oriented gradient method (HOG) + Transfer Entropy (TE)

It is a method by which a combination of TE and HOG is to extract features from the signal. Arrays are created using the TE technique that may contain an amount of duplicate data. Therefore, the HOG method extracts features from the TE matrix [20, 55].

Table 2. Different kinds of emotion methods in the literature

Methods

Emotion

References

DWT

Happy, and sad emotion

[41]

SF

Arousal-valence

[2, 38, 30]

FFT

Relax, Stress

[56]

stress recognition

[37]

Arousal-valence

[32]

STFT

Arousal-valence

[57]

PSD

Arousal-Valence

[16, 33]

Relax, Stress

[56]

5 Emotions

[21]

negative emotions

[58]

Asymmetry in frequency

negative emotions

[58]

HOG + TE

Arousal-valence

[20]

3D-CNN

Arousal-valence

[25]

2D-CNN

Arousal-valence

[59]

1D-CNN

Arousal-valence

[48]

6. Challenges, Applications, and Future Works

6.1 Applications

Many applications have used the EEG in several fields, including the brain-computer interface (BCI), such as robot control applications [60], and applications used in smart homes to monitor and control sensors [61]. Ergonomics and biometrics, such as applications of human voice and face measurement [62]. Neuroscience [63], such as epilepsy [64], schizophrenia [65]. Customized solutions, such as health applications, monitor patient health [65] and educational applications [66]. Neuromarketing is like applications that analyze customer sentiments for a particular product and not others [67].

6.2 Challenges and future works

The ability to detect emotions based on physiological signals is one of the most fundamental topics in emotion classification. There are many ways to detect emotions, including facial expressions, writing, voice, and EEG signals. Because EEG signals are internal signals from the brain that a person cannot control, the challenge has increased for psychologists, computer science researchers, and others. The big challenge is finding the best methods that can be used to classify feelings based on EEG signals accurately. And the second challenge is a noise problem in the signal and in the low-frequency bands, where raw data processing is very complex to remove noise from the signal. Several techniques have been used in the data filtering process: Finite Impulse Response (FIR) filters, Adaptive Filter, Bandpass Filter, etc.

Another challenge is to reduce the input features' dimensions, handle the data with ease, and the unstable performance of the EEG [49].

Table 3 shows the advantages and disadvantages of an EEG. Table 4 shows the advantages and disadvantages of some of the most frequently used feature extraction methods in the literature.

The final challenge is choosing a suitable classifier for detecting emotions because there is an apparent discrepancy between the performance of the classifier and the performance of another in the same data set. When we conducted this study, we noticed that deep learning techniques such as DNN, 3D-CNN, LSTM, BiLSTM and MMResLSTM were frequently used in the literature in recent years. We recommend using deep learning techniques with the WPD method to classify emotions in future work because WPD is suitable for nonstationary signals.

Table 3. The advantages and disadvantages of the EEG

Advantages

Disadvantages

Ref.

  • High Temporal resolution.

  • Safe and easy technique.

  • Sensitive to EOG and ECG signals.

  • Low spatial resolution.

  • Nonstationary signal.

[68]

  • Cheap and easy to use.

  • No multi-Tesla magnets or antimatter-producing radioisotopes are required just wires and signal processing gear.

  • An inverse problem plagues EEG.

[69]

  • Confidential.

  • Very difficult to mimic.

  • Impossible to steal.

  • EEG exists in every person and requires recording.

  • Noise

  • Nonstationary signal.

[70]

Table 4. The advantages and disadvantages of the features extraction methods based on EEG

Method

Advantages

Disadvantages

Independent Component Analysis (ICA)

  • Computationally efficient.

  • It shows High performance for large-sized data.

  • It decomposes signals into temporal independent and spatially fixed components.

  • Computationally efficient

  • High performance for large data sized

  • It can’t be applicable for underdetermined cases

  • Requires more computations for decomposition.

  • Require more computational for decomposition

Principal Component Analysis (PCA)

  • Dimension reduction without losing data.

  • It helps in the reduction of feature dimensions.

  • Ranking will be done and helps in the classification of data

  • Unable to process the complicated set of data

  • Not well as ICA.

Adaptive Filtering

  • Ability to modify the signal features according to signals being analyzed.

  • Works well for the signals and artefacts with overlapping spectra nature.

  • Works well for the signals with overlapping spectra

  • Needs two signals (including reference signal)

Wavelet Packet Decomposition (WPD)

  • Able to analyze the nonstationary signals

  • Increase computational time.

Fast Fourier Transform (FFT)

  • It works well for stationary signals.

  • A powerful method of frequency analysis.

  • Not applicable for nonstationary signals.

  • Unable to measure both time and frequency.

  • It suffers from large noise sensitivity.

  • Poor time localization makes it not suitable for all kinds of applications.

Wavelet Transformations (WT)

  • Suitable for nonstationary signals.

  • Able to analyze the signal in the time and frequency domain.

  • Capable of analyzing signals with discontinuities through variable window size.

  • It can analyze signals both in time and frequency domains.

  • Can extract energy, distance or clusters etc.

  • Lacking specific methodology to apply to pervasive noise

  • Lacking specific methodology to apply WT to the pervasive noise.

  • Performance limited by Heisenberg Uncertainty.

7. Conclusions

In this study, several studies have been reviewed on the classification of emotions based on physiological cues using brain signals (EEG), and the latest and most commonly used techniques in studies have been summarized for feature extraction, such as FFT, PSD, WT, DWT, 2D-CNN, etc. Also, most of the physiological sentiment datasets were summarized, including the DEAP dataset, the MAHNOB-HCI dataset, the SEED dataset, and the GAMEEMO dataset.

We focused on recent techniques and methods for classifying emotions, especially deep learning algorithms used to classify emotions. We also chose different types of emotion classification models: the negative and positive emotions model or the valence-arousal emotion model.

We noticed that deep learning techniques to extract features from EEG signals achieved better results than other traditional methods. For example: if we make a comparison of the results obtained in the articles we have reviewed, we will arrive at the following:

To make the comparison fair, we chose the same dataset (DEAP) and the same emotions model (arousal-valence model), so the studies [16, 23, 26, 29, 38] used traditional methods to extract the features and achieved the accuracy of (75.10%, 86.35%, 74.62%, 92.87%, 89.84, and 92.36%), respectively. While the studies [19, 22, 25, 27, 28] were used the deep learning methods to extract the features, and achieved the accuracy of (93%, 90.15%, 93%, 93.06%, and 97.11%), respectively. You can note the superiority of deep learning techniques in extracting features from EEG brain signals.

We noticed that the EEG helps in different fields and can be developed in many areas, especially in health applications or applications of human-computer interaction, despite the difficulty of processing the EEG signal. Still, many incentives may inspire researchers to develop this broad field.

We have made a great effort to cover the most recent studies in this field, especially feature extraction techniques. We have made hard efforts to cover the most recent methods that have been used in literature. We hope that this review will guide researchers working on EEG and signal processing (Table 5).

Table 5. Emotion recognition recent studies based on EEG

Ref.

Emotions

Method

Dataset

Stimuli

Recorded signals

Classifier

Year

Results

[2]

Negative-positive,

Valence-Arousal

Statistical features, DWT, Hjorth features, Shannon entropy, logarithmic energy entropy, sample entropy, multiscale entropy

GAMEEMO

Video game

EEGs from 28 individual

MLPNN, SVM, K-NN

2020

Good performance with both models, with an accuracy of 80% in some channels

[31]

Negative-positive

Spectral entropy

GAMEEMO

Video game

EEGs from 28 individual

bidirectional LSTM

2020

76.91%

[16]

Arousal-Valence

Time-domain: statistical methods, Frequency domain: PSD, Time-Frequency domain: HHS

DEAP

Video clips

EEGs from 32 individual

DBN, DBN-CRF

DBN-GC,

 

DBN-GC-CRF

2020

The best accuracy obtained from the DBN-GC-CRF classifier is 83.46% using the SEED dataset

SEED

films clips

EEGs from 15 individual

AMIGOS

short and long videos

Individual and 4 people group

[17]

Positive and negative

5 emotions from Images and 7 basics from videos

Spectral data,

inter-hemispheric asynchrony

IAPS,

English version of the FilmStim

Images, Videos, both

EEGs from 18 individual

K-NN, LR, NB, SVM

2020

The team obtained an accuracy of 95.49% by using the SVM classifier for combining images and videos as stimuli

[18]

Relax, Stress

Power spectral density using Fast Fourier Transform

sub-band power ratio

physio bank

during work

EEGs from 36 individual

KNN, SVM

2020

Accuracy reached 99.42% of the FP1 channel

[19]

Arousal-Valence

ACRNN technique

DEAP

Video clips

EEGs from 32 individual

ACRNN,

CNN,

 RNN,

A-CNN-RNN

2020

DEAP=93%

DREAMER=97%

DREAMER

film clips

EEGs from 23 individual

[20]

stress and calm

TE features, HOG+TE features

DEAP

Video clips

EEGs from 32 individual

SVM

2020

GC+HOG=88.93%

TE+HOG=95.21%

[48]

Low arousal and high arousal

1D-CNN, RP, DE

-

game-playing

EEGs from 18 individual

TSception, LSTM

SVM

2020

86.03 using TSception

[21]

5 Emotion

DBN, EDA, PPG, and EMG based on power spectral density (PSD)

DEAP

Video clips

EEGs from 32 individual

FGSVM

2019

Happy Emotions are classified 100% accurately.

[22]

Arousal-Valence

Multi-column structure

DEAP

Video clips

EEGs from 32 individual

multi-olumn CNN

2019

Valence=90.01%

Arousal=90.65%

[23]

Arousal-Valence

Signal Framing, Frequency Band Power (FBP), Pearson Correlation Coefficient (PCC)

DEAP

Video clips

EEGs from 32 individual

LSTM

2019

Valence=81.10%

Arousal=74.38%

[24]

negative stress

Amplitude-aware permutation entropy,

Quadratic sample entropy,

 Multiscale Entropy,

Composite Multiscale,

Statistical Study

CMQSE+CMAAPE

DEAP

Video clips

EEGs from 32 individual

SVM, DT

2019

86.35%

[25]

Arousal-Valence

3D-CNN

DEAP

Video clips

EEGs from 32 individual

3D-CNN, bagging, stacking, Mask-R-CNN

2020

the best accuracy from stacking technique

Valence= 96.13%

Arousal=96.79%

[26]

Arousal-Valence

Random subset feature selection (RSFS), sequential floating forward selection (SFFS), and sequential forward selection (SFS)

DEAP

Video clips

EEGs from 32 individual

SVM

2020

The results they obtained showed that RSFS performed

[27]

Arousal-Valence

2D-CNN, LSTM

DEAP

Video clips

EEGs from 32 individual

CRNN

2020

Valence=91.95%

Arousal=93.06%

[28]

Arousal-Valence

Regional and asymmetric Convolutional Neural Network (RACNN)

DEAP

Video clips

EEGs from 32 individual

RACNN, 2D-CNN, MLP, DE+SVM, DE+KNN

2020

Best accuracy obtained using a RACNN

DEAP dataset:

Valence=96.65%

Arousal=97.11%

DREAMER dataset:

Valence=95.55%

Arousal=97.01%

DREAMER

film clips

EEGs from 23 individual

[29]

Arousal-Valence

Discrete Wavelet, transform, wavelet entropy, wavelet energy

DEAP

Video clips

EEGs from 32 individual

MMResLSTM, LSTM,

residual LSTM

2019

MMResLSTM:

Valence=92.3%

Arousal=92.87%

[58]

negative emotional recognition

Frequency-Point Based on PSD, Asymmetry in Frequency Point of Electrode Pairs, Relative Band Power (Based on PSD), Asymmetry in Band Power of Electrode Pairs (Based on PSD), Relative Band Power Based on DWT, Asymmetry in Band Power of Electrode Pairs (Based on Wavelet), Entropy of Band Power (Based on Wavelet)

DEAP

Video clips

EEGs from 32 individual

SVM

MLP

2019

The proposed method achieved an accuracy of 94%.

[38]

Arousal-Valence

Statistical features (SF),

wavelet transform (WT),

power spectral entropy (PSE), higher-order crossing (HOC), and higher-order spectral (HOS).

DEAP

Video clips

EEGs from 32 individual

K-NN, NB, SVM

2018

The best accuracy using K-NN classifier and statistical method

DEAP dataset:

Valence=89.61%

Arousal=89.84%

MAHNOB-HCI dataset:

Valence=94.6%

Arousal=94%

MAHNOB-HCI

Movies and Pictures

EEGs from 30 individual

[37]

stress recognition

Time Domain, Frequency Domain, Neuroanatomical Features: Frontal EEG Asymmetry

DEAP

Video clips

EEGs from 32 individual

OMTL-VonNeuman

2018

DEAP=71.1%

Construction Workers’= 77.6%

Construction Workers’ EEG

During work

EEGs from 7 individual

[71]

four basic emotions (happy, calm, sad, and fear)

Approximate entropy

-

Video clips

EEGs from 10 individual

DBN-SVM

2018

83.34%

[30]

Arousal-Valence

Statistical parameter

DEAP

Video clips

EEGs from 32 individual

DNN, CNN

2017

The best performance was obtained using the CNN algorithm, with an accuracy of 81.41%.

[41]

Happy, and sad

Time-domain, frequency-domain, DWT

-

images

EEGs from 22 individual

ANN

2017

The best accuracy is 77.3% using the DWT method

[32]

Positive vs. negative vs. neutral,

High arousal vs. low arousal

FFT

-

film clips

EEGs from 27 individual

K-NN, SVM

2018

2018  Positive vs. Negative vs. neutral:

SVM=88%

K-NN=72%

High arousal vs. low arousal

SVM=96%

K-NN=84%

[57]

Eight emotional states

Multivariate synchro squeezing transform (MSST), Short-time Fourier transform (STFT), and wavelet transform (WT)

DEAP

Video clips

EEGs from 32 individual

SVM

2019

93%~

[33]

Arousal-Valence

PSD

Frontal asymmetry

DEAP

Video clips

EEGs from 32 individual

DNN

2017

Result: 82.0% for two classes

[72]

4 emotions (happy, sad, exciting, and hate)

Gabor wavelets, intrinsic mode functions (IMFs)

DEAP

Video clips

EEGs from 32 individual

SVM

2018

93.86% using Gabor wavelets method

[47]

Arousal-Valence

FFT, Statistical Features I, Statistical Features II, Statistical Features III

DEAP

Video clips

EEGs from 32 individual

SVM

2019

92.36%

  References

[1] Chen, G., Zhang, X., Sun, Y., Zhang, J. (2020). Emotion feature analysis and recognition based on reconstructed EEG sources. IEEE Access, 8: 11907-11916. https://doi.org/10.1109/ACCESS.2020.2966144

[2] Alakus, T.B., Gonen, M., Turkoglu, I. (2020). Database for an emotion recognition system based on EEG signals and various computer games-GAMEEMO. Biomedical Signal Processing and Control, 60: 101951. https://doi.org/10.1016/j.bspc.2020.101951

[3] Choudhury, R.R. (2018). Technical perspective: Is your WiFi a sensor? Communications of the ACM, 61(9): 90. https://doi.org/10.1145/3236619

[4] Fan, Y., Lu, X., Li, D., Liu, Y. (2016). Video-based emotion recognition using CNN-RNN and C3D hybrid networks. In Proceedings of the 18th ACM International Conference on Multimodal Interaction, pp. 445-450. https://doi.org/10.1145/2993148.2997632

[5] Alarcao, S.M., Fonseca, M.J. (2017). Emotions recognition using EEG signals: A survey. IEEE Transactions on Affective Computing, 10(3): 374-393. https://doi.org/10.1109/TAFFC.2017.2714671

[6] Soroush, M.Z., Maghooli, K., Setarehdan, S.K., Nasrabadi, A.M. (2017). A review on EEG signals based emotion recognition. International Clinical Neuroscience Journal, 4(4): 118. https://doi.org/10.15171/icnj.2017.01

[7] Liu, Y. J., Yu, M., Zhao, G., Song, J., Ge, Y., Shi, Y. (2017). Real-time movie-induced discrete emotion recognition from EEG signals. IEEE Transactions on Affective Computing, 9(4): 550-562. https://doi.org/10.1109/TAFFC.2017.2660485

[8] Ekman, P. (1992). An argument for basic emotions. Cognition & Emotion, 6(3-4): 169-200. https://doi.org/10.1080/02699939208411068

[9] Lang, P.J. (1995). The emotion probe: Studies of motivation and attention. American Psychologist, 50(5): 372. https://doi.org/10.1037/0003-066X.50.5.372

[10] Shu, L., Xie, J., Yang, M., Li, Z., Li, Z., Liao, D. (2018). A review of emotion recognition using physiological signals. Sensors, 18(7): 2074. https://doi.org/10.3390/s18072074

[11] Nagel, S. (2019). Towards a home-use BCI: fast asynchronous control and robust non-control state detection (Doctoral dissertation, Universität Tübingen). https://doi.org/10.15496/publikation-37739

[12] Casson, A.J., Yates, D.C., Smith, S.J., Duncan, J.S., Rodriguez-Villegas, E. (2010). Wearable electroencephalography. IEEE Engineering in Medicine and Biology Magazine, 29(3): 44-56. https://doi.org/10.1109/MEMB.2010.936545

[13] Bashivan, P., Rish, I., Heisig, S. (2016). Mental state recognition via wearable EEG. arXiv preprint arXiv:1602.00985.

[14] Horlings, R., Datcu, D., Rothkrantz, L.J. (2008). Emotion recognition using brain activity. In Proceedings of the 9th International Conference on Computer Systems and Technologies and Workshop for PhD Students in Computing, pp. II-1.

[15] Gao, Y., Lee, H.J., Mehmood, R.M. (2015). Deep learninig of EEG signals for emotion recognition. In 2015 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), pp. 1-5. https://doi.org/10.1109/ICMEW.2015.7169796

[16] Chao, H., Liu, Y. (2020). Emotion recognition from multi-channel EEG signals by exploiting the deep belief-conditional random field framework. IEEE Access, 8: 33002-33012. https://doi.org/10.1109/ACCESS.2020.2974009

[17] Sharan, R.V., Berkovsky, S., Taib, R., Koprinska, I., Li, J. (2020). Detecting personality traits using inter-hemispheric asynchrony of the brainwaves. In 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), pp. 62-65. https://doi.org/10.1109/EMBC44109.2020.9176108

[18] Priya, T.H., Mahalakshmi, P., Naidu, V.P.S., Srinivas, M. (2020). Stress detection from EEG using power ratio. In 2020 International Conference on Emerging Trends in Information Technology and Engineering (ic-ETITE), pp. 1-6. https://doi.org/10.1109/ic-ETITE47903.2020.401

[19] Tao, W., Li, C., Song, R., Cheng, J., Liu, Y., Wan, F., Chen, X. (2020). EEG-based emotion recognition via channel-wise attention and self attention. IEEE Transactions on Affective Computing. https://doi.org/10.1109/TAFFC.2020.3025777

[20] Gao, Y., Wang, X., Potter, T., Zhang, J., Zhang, Y. (2020). Single-trial EEG Emotion recognition using granger causality/transfer entropy analysis. Journal of Neuroscience Methods, 346: 108904. https://doi.org/10.1016/j.jneumeth.2020.108904

[21] Hassan, M.M., Alam, M.G.R., Uddin, M.Z., Huda, S., Almogren, A., Fortino, G. (2019). Human emotion recognition using deep belief network architecture. Information Fusion, 51: 10-18. https://doi.org/10.1016/j.inffus.2018.10.009

[22] Yang, H., Han, J., Min, K. (2019). A multi-column CNN model for emotion recognition from EEG signals. Sensors, 19(21): 4736. https://doi.org/10.3390/s19214736

[23] Xing, X., Li, Z., Xu, T., Shu, L., Hu, B., Xu, X. (2019). SAE+ LSTM: A New framework for emotion recognition from multi-channel EEG. Frontiers in Neurorobotics, 13: 37. https://doi.org/10.3389/fnbot.2019.00037

[24] Martínez-Rodrigo, A., García-Martínez, B., Alcaraz, R., González, P., Fernández-Caballero, A. (2019). Multiscale entropy analysis for recognition of visually elicited negative stress from EEG recordings. International Journal of Neural Systems, 29(2): 1850038. https://doi.org/10.1142/S0129065718500387

[25] Salama, E.S., El-Khoribi, R.A., Shoman, M.E., Shalaby, M.A.W. (2021). A 3D-convolutional neural network framework with ensemble learning techniques for multi-modal emotion recognition. Egyptian Informatics Journal, 22(2): 167-176. https://doi.org/10.1016/j.eij.2020.07.005

[26] Goshvarpour, A., Goshvarpour, A. (2020). A novel approach for EEG electrode selection in automated emotion recognition based on Lagged Poincare’s Indices and sLORETA. Cognitive Computation, 12(3): 602-618. https://doi.org/10.1007/s12559-019-09699-z

[27] Liao, J., Zhong, Q., Zhu, Y., Cai, D. (2020). Multimodal physiological signal emotion recognition based on convolutional recurrent neural network. In IOP Conference Series: Materials Science and Engineering 782(3): 032005. https://doi.org/10.1088/1757-899X/782/3/032005

[28] Cui, H., Liu, A., Zhang, X., Chen, X., Wang, K., Chen, X. (2020). EEG-based emotion recognition using an end-to-end regional-asymmetric convolutional neural network. Knowledge-Based Systems, 205: 106243. https://doi.org/10.1016/j.knosys.2020.106243

[29] Ma, J., Tang, H., Zheng, W.L., Lu, B.L. (2019). Emotion recognition using multimodal residual LSTM network. In Proceedings of the 27th ACM International Conference on Multimedia, pp. 176-183. https://doi.org/10.1145/3343031.3350871

[30] Tripathi, S., Acharya, S., Sharma, R.D., Mittal, S., Bhattacharya, S. (2017). Using deep and convolutional neural networks for accurate emotion classification on DEAP dataset. In Twenty-ninth IAAI conference.

[31] Alakus, T.B., Turkoglu, I. (2020). Emotion recognition with deep learning using GAMEEMO data set. Electronics Letters, 56(25): 1364-1367. https://doi.org/10.1049/el.2020.2460

[32] Tarnowski, P., Kołodziej, M., Majkowski, A., Rak, R.J. (2018). Combined analysis of GSR and EEG signals for emotion recognition. In 2018 International Interdisciplinary PhD Workshop (IIPhDW), pp. 137-141. https://doi.org/10.1109/IIPHDW.2018.8388342

[33] Al-Nafjan, A., Hosny, M., Al-Wabil, A., Al-Ohali, Y. (2017). Classification of human emotions from electroencephalogram (EEG) signal using deep neural network. International Journal of Advanced Computer Science and Applications, 8(9): 419-425. https://doi.org/10.14569/ijacsa.2017.080955

[34] Koelstra, S., Muhl, C., Soleymani, M., Lee, J.S., Yazdani, A., Ebrahimi, T. (2011). DEAP: A database for emotion analysis; Using physiological signals. IEEE Transactions on Affective Computing, 3(1): 18-31. https://doi.org/10.1109/T-AFFC.2011.15

[35] Soleymani, M., Lichtenauer, J., Pun, T., Pantic, M. (2011). A multimodal database for affect recognition and implicit tagging. IEEE Transactions on Affective Computing, 3(1): 42-55. https://doi.org/10.1109/T-AFFC.2011.25

[36] Zheng, W.L., Lu, B.L. (2015). Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks. IEEE Transactions on Autonomous Mental Development, 7(3): 162-175. https://doi.org/10.1109/TAMD.2015.2431497

[37] Jebelli, H., Khalili, M.M., Lee, S. (2018). A continuously updated, computationally efficient stress recognition framework using electroencephalogram (EEG) by applying online multitask learning algorithms (OMTL). IEEE Journal of Biomedical and Health Informatics, 23(5): 1928-1939. https://doi.org/10.1109/JBHI.2018.2870963

[38] Piho, L., Tjahjadi, T. (2018). A mutual information based adaptive windowing of informative EEG for emotion recognition. IEEE Transactions on Affective Computing, 11(4): 722-735. https://doi.org/10.1109/TAFFC.2018.2840973

[39] Amin, H.U., Malik, A.S., Ahmad, R.F., Badruddin, N., Kamel, N., Hussain, M., Chooi, W.T. (2015). Feature extraction and classification for EEG signals using wavelet transform and machine learning techniques. Australasian Physical & Engineering Sciences in Medicine, 38(1): 139-149. https://doi.org/10.1007/s13246-015-0333-x

[40] Chen, D., Wan, S., Xiang, J., Bao, F. S. (2017). A high-performance seizure detection algorithm based on Discrete Wavelet Transform (DWT) and EEG. PloS One, 12(3): e0173138. https://doi.org/10.1371/journal.pone.0173138

[41] Ang, A.Q.X., Yeong, Y.Q., Wee, W. (2017). Emotion classification from EEG signals using time-frequency-DWT features and ANN. Journal of Computer and Communications, 5(3): 75-79. https://doi.org/10.4236/jcc.2017.53009

[42] Subasi, A. (2007). EEG signal classification using wavelet feature extraction and a mixture of expert model. Expert Systems with Applications, 32(4): 1084-1093. https://doi.org/10.1016/j.eswa.2006.02.005

[43] Murugappan, M., Murugappan, S. (2013). Human emotion recognition through short time Electroencephalogram (EEG) signals using Fast Fourier Transform (FFT). In 2013 IEEE 9th International Colloquium on Signal Processing and its Applications, pp. 289-294. https://doi.org/10.1109/CSPA.2013.6530058

[44] Saxena, A., Tripathi, K., Khanna, A., Gupta, D., Sundaram, S. (2020). Emotion detection through EEG signals using FFT and machine learning techniques. In: Khanna A., Gupta D., Bhattacharyya S., Snasel V., Platos J., Hassanien A. (eds) International Conference on Innovative Computing and Communications. Advances in Intelligent Systems and Computing, vol 1087. Springer, Singapore. https://doi.org/10.1007/978-981-15-1286-5_46

[45] Zhu, C., Huang, H., Liu, H., Song, C., Ma, F. (2014). On-line vibration monitoring and diagnosing a multi-megawatt wind turbine gearbox. In Proc. of the International Gear Conference, pp. 1089-1098. https://doi.org/10.1533/9781782421955.1089

[46] Islam, M., Ahmed, T., Mostafa, S.S., Yusuf, M.S.U., Ahmad, M. (2013). Human emotion recognition using frequency & statistical measures of EEG signal. In 2013 International Conference on Informatics, Electronics and Vision (ICIEV), pp. 1-6. https://doi.org/10.1109/ICIEV.2013.6572658

[47] George, F.P., Shaikat, I.M., Ferdawoos, P.S., Parvez, M.Z., Uddin, J. (2019). Recognition of emotional states using EEG signals based on time-frequency analysis and SVM classifier. International Journal of Electrical & Computer Engineering, 9(2): 1012-1020. http://doi.org/10.11591/ijece.v9i2.pp1012-1020

[48] Ding, Y., Robinson, N., Zeng, Q., Chen, D., Wai, A.A.P., Lee, T.S., Guan, C. (2020). Tsception: A deep learning framework for emotion detection using EEG. In 2020 International Joint Conference on Neural Networks (IJCNN), pp. 1-7. http://doi.org/10.1109/IJCNN48605.2020.9206750

[49] Wang, F., Wu, S., Zhang, W., Xu, Z., Zhang, Y., Wu, C., Coleman, S. (2020). Emotion recognition with convolutional neural network and EEG-based EFDMs. Neuropsychologia, 146: 107506. https://doi.org/10.1016/j.neuropsychologia.2020.107506

[50] Wen, Z., Xu, R., Du, J. (2017). A novel convolutional neural networks for emotion recognition based on EEG signal. In 2017 International Conference on Security, Pattern Analysis, and Cybernetics (SPAC), pp. 672-677. https://doi.org 10.1109/SPAC.2017.8304360

[51] Abdullah, S., Nizwan, C.K.E., Nuawi, M.Z. (2009). A study of fatigue data editing using the Short-Time Fourier Transform (STFT). American Journal of Applied Sciences, 6(4): 565. https://doi.org/10.3844/ajas.2009.565.575

[52] Bastos-Filho, T.F., Ferreira, A., Atencio, A.C., Arjunan, S., Kumar, D. (2012). Evaluation of feature extraction techniques in emotional state recognition. In 2012 4th International Conference on Intelligent Human Computer Interaction (IHCI), pp. 1-6. https://doi.org/10.1109/IHCI.2012.6481860

[53] Ng, W.B., Saidatul, A., Chong, Y.F., Ibrahim, Z. (2019). PSD-based features extraction for EEG signal during typing task. In IOP Conference Series: Materials Science and Engineering, 557(1): 012032.

[54] Nguyen, T.H., Chung, W.Y. (2019). Negative news recognition during social media news consumption using EEG. IEEE Access, 7: 133227-133236. https://doi.org/10.1109/ACCESS.2019.2941251

[55] Zebari, R., Abdulazeez, A., Zeebaree, D., Zebari, D., Saeed, J. (2020). A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction. Journal of Applied Science and Technology Trends, 1(2): 56-70. https://doi.org/10.38094/jastt1224

[56] Priya, T.H., Mahalakshmi, P., Naidu, V.P.S., Srinivas, M. (2020). Stress detection from EEG using power ratio. In 2020 International Conference on Emerging Trends in Information Technology and Engineering (ic-ETITE), pp. 1-6. https://doi.org/10.1109/ic-ETITE47903.2020.401

[57] Ozel, P., Akan, A., Yilmaz, B. (2019). Synchrosqueezing transform based feature extraction from EEG signals for emotional state prediction. Biomedical Signal Processing and Control, 52: 152-161. https://doi.org/10.1016/j.bspc.2019.04.023

[58] Nguyen, T.H., Chung, W.Y. (2019). Negative news recognition during social media news consumption using EEG. IEEE Access, 7: 133227-133236. https://doi.org/10.1109/ACCESS.2019.2941251

[59] Liao, J., Zhong, Q., Zhu, Y., Cai, D. (2020). Multimodal physiological signal emotion recognition based on convolutional recurrent neural network. In IOP Conference Series: Materials Science and Engineering, 782(3): 032005. https://doi.org/10.1088/1757-899X/782/3/032005

[60] Shedeed, H.A., Issa, M.F., El-Sayed, S.M. (2013). Brain EEG signal processing for controlling a robotic arm. In 2013 8th International Conference on Computer Engineering & Systems (ICCES), pp. 152-157. https://doi.org/10.1109/ICCES.2013.6707191

[61] Nafea, M., Abdul-Kadir, N.A., Harun, F.K.C. (2018). Brainwave-controlled system for smart home applications. In 2018 2nd International Conference on BioSignal Analysis, Processing and Systems (ICBAPS), pp. 75-80. https://doi.org/10.1109/ICBAPS.2018.8527397

[62] Schelenz, P.D., Klasen, M., Reese, B., Regenbogen, C., Wolf, D., Kato, Y., Mathiak, K. (2013). Multisensory integration of dynamic emotional faces and voices: method for simultaneous EEG-fMRI measurements. Frontiers in Human Neuroscience, 7: 729. https://doi.org/10.3389/fnhum.2016.00624

[63] Srinivasan, N. (2007). Cognitive neuroscience of creativity: EEG based approaches. Methods, 42(1): 109-116. https://doi.org/10.1016/j.ymeth.2006.12.008

[64] Alotaiby, T.N., Alshebeili, S.A., Alshawi, T., Ahmad, I., Abd El-Samie, F.E. (2014). EEG seizure detection and prediction algorithms: A survey. EURASIP J. Adv. Signal Process, 2014: 1-21. https://doi.org/10.1186/1687-6180-2014-183

[65] Krishnan, P.T., Raj, A.N.J., Balasubramanian, P., Chen, Y. (2020). Schizophrenia detection using MultivariateEmpirical Mode Decomposition and entropy measures from multichannel EEG signal. Biocybernetics and Biomedical Engineering, 40(3): 1124-1139. https://doi.org/10.1016/j.bbe.2020.05.008

[66] Spüler, M., Krumpe, T., Walter, C., Scharinger, C., Rosenstiel, W., Gerjets, P. (2017). Brain-computer interfaces for educational applications. In Informational Environments, pp. 177-201. https://doi.org/10.1007/978-3-319-64274-1_8

[67] Oon, H.N., Saidatul, A., Ibrahim, Z. (2018). Analysis on Non-linear features of electroencephalogram (EEG) signal for neuromarketing application. In 2018 International Conference on Computational Approach in Smart Systems Design and Applications (ICASSDA), pp. 1-8. https://doi.org/10.1109/ICASSDA.2018.8477618

[68] Lakshmi, M.R., Prasad, T.V., Prakash, D.V.C. (2014). Survey on EEG signal processing methods. International Journal of Advanced Research in Computer Science and Software Engineering.

[69] Santhiya, P., Chitrakala, S. (2019). A survey on emotion recognition from EEG signals: Approaches, Techniques & Challenges. In 2019 International Conference on Vision Towards Emerging Trends in Communication and Networking (ViTECoN), pp. 1-6. https://doi.org/10.1109/ViTECoN.2019.8899671

[70] Pham, T., Ma, W., Tran, D., Nguyen, P., Phung, D. (2013). A study on the feasibility of using EEG signals for authentication purpose. In International Conference on Neural Information Processing, pp. 562-569. https://doi.org/10.1007/978-3-642-42042-9_70

[71] Chen, T., Ju, S., Yuan, X., Elhoseny, M., Ren, F., Fan, M., Chen, Z. (2018). Emotion recognition using empirical mode decomposition and approximation entropy. Computers & Electrical Engineering, 72: 383-392. https://doi.org/10.1016/j.compeleceng.2018.09.022

[72] Zamanian, H., Farsi, H. (2018). A new feature extraction method to improve emotion detection using EEG signals. ELCVIA: Electronic Letters on Computer Vision and Image Analysis, 17(1): 29-44. https://doi.org/10.5565/rev/elcvia.1045