Open access peer-reviewed chapter

Feature Extraction for Emotion Recognition: A Review

Written By

Neha Garg and Kamlesh Sharma

Submitted: 07 December 2022 Reviewed: 30 December 2022 Published: 21 March 2023

DOI: 10.5772/intechopen.109740

From the Edited Volume

Emotion Recognition - Recent Advances, New Perspectives and Applications

Edited by Seyyed Abed Hosseini

Chapter metrics overview

178 Chapter Downloads

View Full Metrics

Abstract

With the increasing role of Artificial Intelligence and ubiquitous computing paradigms, a stage has arrived where human being and machine is interacting seamlessly. However, the users may face issues while interacting with these systems. A more simple and reliable interaction with machines will be possible by recognizing user’s emotions. In order to develop a system that can respond effectively to user’s emotions can be modeled by utilizing the electroencephalogram (EEG) as a bio-signal sensor. The emotion recognition plays a vital role in the area of Human Computer Interaction (HCI) and Brain Computer Interaction (BCI) to provide good interaction between brain and machine. The emotion recognition comprises of three major phases feature extraction, feature selection and classifiers. The present chapter provides an overview of feature extraction techniques utilized by researchers in frequency domain analysis, time domain analysis and time-frequency domain analysis. The chapter also discusses the process, issues and challenges for feature extraction in EEG, the application area of the EEG.

Keywords

  • emotion recognition (ER)
  • human computer interaction (HCI)
  • brain computer interaction (BCI)
  • feature extraction
  • electroencephalogram (EEG)

1. Introduction

The world in the verge of paradigm change in the field of Intelligent Systems: moving from an era in which people control gadgets to one in which autonomous devices, capable of self-management and aware of their environmental and situational context [1]. As per the definition of Ubiquitous Computing conceived by Mark Weiser [2], the world is approaching to a level of automation and computing where human and computers are interacting with each other naturally without the awareness of users. Ironically, one of the major issue is growing complexity with this paradigm shift is that user find it difficult to interact with such systems [3]. As a result, it’s crucial to identify all potential interaction modalities and organize them according to problem domain. For example personalized interaction is highly required in computer- mediated interaction like virtual reality to maintain user’s interest and engagement with the cognitive activity. Task engagement includes both the user’s cognitive activity and engagement but it also requires an understanding of user’s emotional transfer. Therefore, the physiological computing system can be used to provide insights into the cognitive and emotional processes involved in completing tasks [4]. In particular, the display of level of brain activity by processing EEG signals may be of benefits when integrated with input modality [5]. ER can be performed with the help of physiological and non-physiological signals [6].

Facial expressions, speech, voice, actions etc., are the non-physiological signals that could not contribute very precisely for ER [7]. EEG signals have proved strong implications in finding emotions states [8]. The classification accuracy for emotion recognition is higher for EEG signals and the model can also the changes in mood too [9].

Advertisement

2. Broad area description

With the advancement of technology HCI is providing the support to BCI. Providing machines with the capacity to understand and discover the various emotional states of users can be of essential significance for the next generation. Endowing digital devices with logical reasoning abilities about user affective context will provide the facility to detect the present state of feelings. Such as signs of feeling low, frustrated, fear and allow the machine to react in a more intellectual and empathetic way [10]. As a result, this collectively along with other HCI traits like consistency, flexibility, and usability may give the base to produce more clever and more adaptive interface [11]. Now a day’s various input modalities have been utilized to gather the input data for emotion recognition. The very first of them is audiovisual based communication along with eye gazed, facial expression, speech evaluation, etc. The second physiological measure is sensor based signals along with EEG signals; galvanic skin response and electrocardiogram can also be used [12].

Advertisement

3. What is emotion?

The emotion can be the affective aspect of consciousness or mental reaction or can be states a strong feeling towards a particular object and it can have some associated physiological and behavioral changes in voice, expression or mood, etc. The signals of emotions can be categorized into two categories physiological signals and non-physiological signals [13].

The physiological reactions can be the increment or decrement the value of EEG signals, body temperature, heartbeats, blood pressure, breathing rate, etc. the studies states the more impact of physiological reactions on women in comparison with the men [13]. The non-physiological reactions include the expressions, action, voice etc. Here physiological signals are difficult to ignore during data acquisition stage and provides more accurate result [14].

The researchers have proposed various emotion models based on discrete model and dimension model theory [6]. The theory proposed by P. Ekman has six primary emotional states surprise, happiness, fear, anger, disgust, and sadness [15]. The other complex emotions are secondary one that is the composition of these basic emotions. P. Lang proposed a dimensional model called valence- arousal model [16]. The model maps emotions into two-dimensional space where valence represents positive and negative conditions and arousal represents the intensity of human emotions. This model maps different emotions at same place, which results in difficulty to distinguish. The model has been represented in Figure 1. To overcome this issue, A. Mehrabian has proposed model by adding dominance dimension to the valence-arousal model [18]. This model is called Pleasure-Arousal-Dominance (PAD) model.

Figure 1.

Arousal valence based emotion model [17].

Advertisement

4. Characteristics of EEG signals

The EEG different frequency bands are related to conscious human activities [19]. Based on the frequency of EEG signals, the signals are divided into five categories, alpha, beta, gamma, theta and delta.

Delta signal has frequency of range 0.5–4 Hz and occurs in the state of unconsciousness such as dreamless, deep sleep. Theta signals occur with frequency of 4–8 Hz and appear in the state of sub consciousness like sleepiness, dreaming etc. Alpha waves arise at frequency 8–13 Hz frequency when in consciousness human is in relaxed state. Alpha waves have higher oscillatory energy in neutral and negative emotions than beta and gamma waves.

Beta waves occur when human mind is highly concentrated and active with frequency 13–30 Hz. The active state of mind can be dictated by taking average power ratio of beta and alpha waves. Gamma waves occur at very high frequency greater than 30 Hz and show the hyperactivity of brain. As the results suggest, that emotional EEG is more evoked in lower frequency band in comparison to higher frequency band. Similarly negative emotions have higher intensity and wider distribution than positive ones [20].

Advertisement

5. Features of EEG signals

EEG signal is a weak physiological signal, which is highly utilized by researchers for emotion recognition due to the high accuracy of results [6]. For emotion recognition, the features of EEG signals categorized in categories spatial domain features, time domain features, frequency domain features and time-frequency domain features.

5.1 Spatial domain features

Spatial analysis is distribution of electrical signals at different regions of mind during acquisition of EEG signals. Different regions of brain respond in a different manner to emotions [21], considering how spectral, spatial and temporal aspects complement each other. To consider both spatial-temporal and spatial-spectral characteristics parallel streams are created.

5.2 Time domain features

EEG signals are often recorded in the form of time domain which is statistical in nature. Time domain analysis is commonly performed by using histogram analysis or statistical methods [6]. However the time-domain features include EEG signals with less information loss. But due to complex form of EEG signals, there is no standard unified method for analysis. So the analysts need to be rich in knowledge and experience.

5.3 Frequency domain features

Frequency domain features are obtained after converting the original time domain signal using Fourier Transformation. The aim of frequency domain is to find the frequency information of signals along with power characteristics of various frequency bands.

5.4 Time-frequency domain features

The time-domain and frequency domain signals are merged to examine EEG signals more accurately. As convergence process of signals from time domain to frequency domain does not lost the time information.

Time- frequency analysis has the ability to completely reflect the distinctive information in EEG signals. Continuous wavelet transforms (CWT) and discrete wavelet transform (DWT) are the two primary forms of wavelet transformation, which separate the low-frequency component of the signals. It is more advantageous than the conventional approaches for dealing nonlinear and non-stationary signals.

Advertisement

6. Methodology

The Emotion Recognition using EEG signals process is comprises of four phases as shown in Figure 2.

  1. EEG Signal Acquisition—The collection of EEG signals can be divided into two methods—invasive and non-invasive. Here, the data collected using invasive method has higher signal to noise ration while non-invasive has utilization in BCI. The collection of EEG signals can be divided into two methods- invasive and non-invasive. Here, the data collected using invasive method has higher signal to noise ration while non-invasive has utilization in BCI. The first part comprises of four ways to collect EEG signals [22]; the first technique is electroencephalogram i.e. is graph obtained by recording and amplifying the brain’s signals. The signals are collected with the help of electrodes which are fixed in scalp. The second method is Electrocorticogram (ECoG). In ECoG, the recording is performed by surgically implanted electrodes. ECoG provides better spatial resolution and accuracy. Another method of signal acquisition is Functional Magnetic Resonance Imaging (fMRI), is a neuroimaging technique. fMRI reflect the oxygen saturation and blood flow level through the measurement of MR signals. Depth electrode is the fourth method of EEG signal detection. Deep Brain Simulation (DBS) is performed by placing electrodes at specific regions of brain to deliver current.

  2. Pre-processing—the pre-processing phase comprises of removal of artifacts, thresholding of the output, amplifying the signals, edge detection and signal averaging etc. [23].

  3. Feature Extraction—the feature selection phase is used to find a feature vector. Here, feature denotes a distinctive or characteristic measurement from a segment of pattern which plays a useful part in classification. For linear analysis of one-dimensional signals for either frequency or time domain, the various method has been discussed below:

  1. Fast Fourier Transformation (FFT) method

    The fast Fourier Transformation (FFT) is a simple and fast way of Discrete Fourier Transformation (DFT). FFT utilized to filter signals from the time domain to the frequency domain. FET is good for stationary signals. However the overall EEG signals are not stationary in nature but for particular band it can be utilized.

    Welch’s method is one of the methods for transforming EEG signals using Fourier transformation [24]. The data sequencing is applied to data windowing, where data sequence xi (n) and Fourier transformation is represented as

    xin=1Lk=1LXkωLn1k1E1
    Xk=n=1LxnωLn1k1E2

  2. Wavelet Transformation (WT) method

    WT plays a vital role in the field of diagnostic and recognition, which compresses the time-varying signal, into few parameters that represents the nature of signal [25]. The model uses time-frequency domain with a variable size windows to provide more flexibility to the systems. WT can be categorized in two ways continuous WT and discrete WT.

  3. Eigenvectors

    The eigenvector methods are utilized to calculate signals frequency and power from artifact ruled measurements. The essence of the techniques is to correlate the corrupted signals using Eigen decomposition [26]. The techniques that employed eigenvector are MUSIC method, Pisarenko’s method and minimum norm method.

  4. Time-frequency distribution

    The Time-frequency domain is utilized with the stationary and noiseless signal, that’s why windowing process is required for pre-processing. The various methods are used for TFD model like Short-time Fourier Transform (STFT), wavelet packet transformation (WPT) and Hilbert-Huang transformation (HHT).

  5. Autoregressive method

    Autoregressive (AR) method is advantageous for short data segment analysis. It also limits the loss of spectral leakage and provides better frequency resolution. Yule-Walker method and Burg’s method are used in AR model for spectral estimation.

Figure 2.

Emotion classification process for EEG signals [6].

The comparison of various models along with their advantages and disadvantages has been tabulated below in Table 1.

Method nameAdvantagesDisadvantagesAnalysis methodSuitability
Fast Fourier transform
  1. Good tool for stationary signal processing

  2. It is more appropriate for narrowband signal, such as sine wave

  3. It has an enhanced speed over virtually all other available methods in real-time applications

  1. Weakness in analyzing nonstationary signals such as EEG

  2. It does not have good spectral estimation and cannot be employed for analysis of short EEG signals

  3. FFT cannot reveal the localized spikes and complexes that are typical among epileptic seizures in EEG signals

  4. FFT suffers from large noise sensitivity, and it does not have shorter duration data record

Frequency domainNarrowband, stationary signals
Wavelet transform
  1. It has a varying window size, being broad at low frequencies and narrow at high frequencies

  2. It is better suited for analysis of sudden and transient signal changes

  3. Better poised to analyze irregular data patterns, that is, impulses existing at different time instances

Needs selecting a proper mother waveletBoth time and freq. domain, and linearTransient and stationary signal
EigenvectorProvides suitable resolution to evaluate the sinusoid from the dataLowest eigenvalue may generate false zeros when Pisarenko’s method is employedFrequency domainSignal buried with noise
Time frequency distribution
  1. It gives the feasibility of examining great continuous segments of EEG signal

  2. TFD only analyses clean signal for good results

  1. The time-frequency methods are oriented to deal with the concept of stationary; as a result, windowing process is needed in the preprocessing module

  2. It is quite slow (because of the gradient ascent computation)

  3. Extracted features can be dependent on each other

Both time and frequency domainsStationary signal
Autoregressive
  1. AR limits the loss of spectral problems and yields improved frequency resolution

  2. Gives good frequency resolution

  3. Spectral analysis based on AR model is particularly advantageous when short data segments are analyzed, since the frequency resolution of an analytically derived AR spectrum is infinite and does not depend on the length of analyzed data

  1. The model order in AR spectral estimation is difficult to select

  2. AR method will give poor spectral estimation once the estimated model is not appropriate, and model’s orders are incorrectly selected

  3. It is readily susceptible to heavy biases and even large variability

Frequency domainSignal with sharp spectral features

Table 1.

Comparison of various feature extraction techniques [26].

As we have discussed, the various techniques for feature extraction of EEG signals, Table 2 representing a comparative analysis for frequency resolution and spectral leakage from where one can easily conclude that AR method is utilized where one need to avoid spurious features.

  1. 4. Classification- the final stage of EEG signals processing is classification. Machine learning classifiers are hugely deployed to use features to predict the corresponding class. The classification is performed in three categories: positive, neutral and negative [28]. Algorithms like Support Vector Machine (SVM), Naive Bayes (NB), Random Forest (RF), etc. Are hugely used by researchers due to their simplicity and accuracy [6]. But these techniques never consider the temporal information associated with EEG signals.

MethodFrequency resolutionSpectral leakage
FFTLOWHIGH
WTHIGHLOW
ARHIGHLOW

Table 2.

Comparison of frequency resolution and spectral leakage of feature extraction techniques [27].

Recently deep learning algorithms like Convolution Neural Network (CNN), Recurrent Neural Network (RNN), Artificial Neural Network (ANN), etc. are also applied by researchers for relatively high accuracy [29].

Advertisement

7. Applications of EEG signals

Few application of emotion detection using EEG signals has been discussed below:

  1. Medical diagnosis—the signals collected using EEG can be used to diagnose various diseases related to brain like brain edema, Parkinson’s disease, epilepsy scots, etc.

  2. Education—studies has shown that there are some emotional states that are helpful for learning [17]. So by given different educational task and acquire signal during them can help in studying the impact on human.

  3. Video gaming—video games are to entertain player and making them attached and involved [30]. By analyzing the mood and mind-set of users, the video games can engage users emotionally, which may result in good participation rate.

The application are not restricted to these three categories only, they are also applicable in Brain Computer Interaction (BCI), Patient Care, Driving Autonomous Car, etc.

Advertisement

8. Solutions and recommendations

In this chapter, we have presented different machine learning algorithm and techniques and their comparison that are used for emotion recognition. In emotion recognition there is no direct approach that suite for all kind of applications. There are multiple factors that affect the choice of machine learning algorithm. The usefulness of emotion recognition and its application has been discussed. Machine algorithms are used to analyze data used for classification. The algorithm basically filter data into categories, which is achieved by providing a set of training examples, each set marked as belonging to one or the other of the two categories.

Advertisement

9. Future scope and conclusion

The future of emotion recognition is very bright. Emotion recognition is already an incredibly powerful tool that helps to solve really hard classification of emotions problems. With the use of emotion recognition, a very hard problem can be solved and system can be work as per the mood and feelings of users.

References

  1. 1. Ark WS. A systematic approach to occupancy modelling in ambient sensor-rich buildings. IBM Systems Journal. 1999;38(4):504-507
  2. 2. Weiser M. The computer for the 21st century. Scientific American. 1991;256:94-104
  3. 3. Pantic M, Nijholt A, Pentland A, Huanag TS. Humancentred intelligent human-computer interaction (hci): How far are we from attaining it? International Journal of Autonomous and Adaptive Communications Systems. 2008;1(2):168-187
  4. 4. Fairclough SH, Gilleade K, Ewing KC, Roberts J. Capturing user engagement via psychophysiology: Measures and mechanisms for biocybernetic adaptation. International Journal of Autonomous and Adaptive Communications Systems. 2013;6(1):63-79. DOI: 10.1504/IJAACS.2013.050694
  5. 5. Szafir D, Mutlu B. Pay attention!: Designing adaptive agents that monitor and improve user engagement. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. New York, NY, USA: ACM; 2012. pp. 11-20. DOI: 10.1145/2207676.2207679
  6. 6. Wang J, Wang M. Review of the emotional feature extraction and classification using EEG signals. Cognitive Robotics. 2021;1:29-40
  7. 7. Zhang T, Zheng W, Cui Z, Zong Y, Yan J, Yan K. A deep neural network-driven feature learning method for multi-view facial expression recognition. IEEE Transactions on Multimedia. 2016;18(12):2528-2536
  8. 8. Zheng WL, Lu BL. Investigating critical frequency bands and channels for eeg-based emotion recognition with deep neural networks. IEEE Transactions on Autonomics Mental Diseaases. 2015;7(3):162-175
  9. 9. Hilborn O. Evaluating classifiers for emotion recognition using EEG. Lecture Notes in Computer Science. 2015;8027:492-501
  10. 10. Picard RW. Affective computing: Challenges. International Journal of Human-Computer Studies. 2003;59(1):55-64
  11. 11. Dix A, Finlay J, Abowd GD, Beale R. Human-computer Interaction. 2004. Available from: http://www.amazon.com/Human-Computer-Interaction3rd-Alan-Dix/dp/0130461091
  12. 12. Burle B, Spieser L, Roger C, Casini L, Hasbroucq T, Vidal F. Spatial and temporal resolutions of eeg: Is it really black and white? A scalp current density view. International Journal of Psychophysiology. 2015;97(3):210-220. on the benefits of using surface Laplacian (current source density) methodology in electrophysiology. [Online]. Available from: http://www.sciencedirect.com/science/article/pii/S016787601500186
  13. 13. Cao KX. The Research of the EEG Frequency Power Feature in Three Basic Emotions. Tianjin, China: Tianjin Medical University; 2019
  14. 14. Fathalla RS, Alshehri WS. Emotions recognition and signal classification: A state-of-the-art. International Journal of Synthetic Emotions (IJSE). 2020;11(1):1-16
  15. 15. Ekman P. An argument for basic emotions. Cognition & Emotion. 1992;6(3–4):169-200
  16. 16. Lang PJ. The emotion probe: Studies of motivation and attention. American Psychologist. 1995;50(5):372
  17. 17. Kołakowska A, Landowska A, Szwoch M, Szwoch W, Wróbel MR. Emotion recognition and its application in software engineering. In: 2013 6th International Conference on Human System Interactions (HSI). Vol. 10. IEEE; 2013. pp. 532-539
  18. 18. Mehrabian A. Comparison of the PAD and PANAS as models for describing emotions and for differentiating anxiety from depression. Journal of Psychopathology Behavioral Assessment. 1997;19(4):331-357
  19. 19. Liu Z, Xie Q, Wu M, Cao W, Li D, Li S. Electroencephalogram emotion recognition based on empirical mode decomposition and optimal feature selection. IEEE Transactions Cognition Developmental System. 2019;11(4):517-526
  20. 20. Zhang JX, Bo H. Research on EEG emotion recognition based on CNN. Modified Computer. 2018;8:12-16
  21. 21. Jia Z, Lin Y, Cai X. SST-EmotionNet: Spatial-spectral-temporal based attention 3D dense network for EEG emotion recognition MM ‘20. In: The 28th ACM International Conference on Multimedia. ACM; 2020. pp. 2909-2917
  22. 22. Li B, Cheng T, Guo Z. A review of EEG acquisition, processing and application. Journal of Physics: Conference Series. 2021;1907(1):012045
  23. 23. Garg N, Sharma K. Text preprocessing for sentiment analysis based on social network data. International Journal of Electrical and Computer Engineering (IJECE). 2022;12(1):776-784
  24. 24. Faust O, Acharya RU, Allen AR, Lin CM. Analysis of EEG signals during epileptic and alcoholic states using AR modeling techniques. IRBM. 2008;29(1):44-52
  25. 25. Cvetkovic D, Übeyli ED, Cosic I. Wavelet transform feature extraction from human PPG, ECG, and EEG signal responses to ELF PEMF exposures: A pilot study. Digital Signal Processing. 2008;18(5):861-874
  26. 26. Al-Fahoum AS, Al-Fraihat AA. Methods of EEG signal features extraction using linear analysis in frequency and time-frequency domains. International Scholarly Research Notices. 2014
  27. 27. Agarwal R, Gotman J, Flanagan D, Rosenblatt B. Automatic EEG analysis during long-term monitoring in the ICU. Electroencephalography and Clinical Neurophysiology. 1998;107(1):44-58
  28. 28. Garg N, Sharma K. Annotated Corpus creation for sentiment analysis in code-mixed Hindi-English (Hinglish) social network data. Indian Journal of Science and Technology. 2020;13(40):4216-4224
  29. 29. Chen D-W et al. A feature extraction method based on differential entropy and linear discriminant analysis for emotion recognition. Sensors. 2019;19(7):1631
  30. 30. Adams E. Fundamentals of Game Design. Pearson Education; 2009

Written By

Neha Garg and Kamlesh Sharma

Submitted: 07 December 2022 Reviewed: 30 December 2022 Published: 21 March 2023