Open access peer-reviewed chapter

Signal Processing and Classification Approaches for Brain-Computer Interface

By Tarik Al-ani and Dalila Trad

Published: January 1st 2010

DOI: 10.5772/7032

Downloaded: 6537

1. Introduction

Research on brain-computer interface(BCI) systems began in the 1970s at the University of California Los Angeles (UCLA) (Vidal, 1973; 1977). The author gave in his papers the expression "Brain Computer Interface" which is the term currently used in literature.

A BCIsystem is a direct communication pathway between a brain and an external artificial device. BCIsystems were aimed at assisting, augmenting or repairing human cognitiveor sensory-motor functions.

The BCIsystems (BCIs) allow control of an artificial device based on the features extracted from voluntary electric, magnetic, or other physical manifestations of brain activity collected from epi- or subdurally from the cortex or from the scalp or in invasive electrophysiological manner, i.e. brain signals recorded intracortically with single electrode or multi-electrode arrays (Dornhege et al., 2007). There is a variety of non-invasive techniques for measuring brain activity. These non-invasive techniques include, the electroencephalography(EEG), magnetoencephalography(MEG), positron emission tomography(PET), functional magnetic resonance imaging(fMRI), and optical imaging. However, for technical, time resolution, real-time, and price constraints, only EEGmonitoring and related techniques are employed in the BCIcommunity. For more details refer to (Wolpaw et al., 2002; Mason et al., 2007; Dobkin, 2007). The neuronal electrical activity contain a broad band frequency, so the monitored brain signals are filtered and denoised to extract the relevant information (see section 3) and finally this information is decoded (see section 6) and commuted into device commands by synchronous controlor more efficiently by self-pacedor asynchronous controlin order to detect whether a user is intending something or not (see chapter 7 in (Dornhege et al., 2007) for details), Fig. 1. For some specific BCItasks, raw brain signal serves as stimulus as well as a control interface feedback.

The direct BCIscan be seen as a new means of communication that may be used to allow tetraplegic or individuals with severe motor or neuromuscular diseases (e.g. Amyotrophic lateral sclerosis (ALS), brainstem stroke, brain or spinal cord injury, cerebral palsy, muscular

Figure 1.


dystrophies, multiple sclerosis) to have effective control over artificial devices or external environment in order to increase or improve their communication qualities or their independence. Recent studies have demonstrated correlations between EEGsignals and actual or imagined movements and between EEGsignals and mental tasks (Keirn & Aunon, 1990; Lang et al., 1996; Pfurtscheller et al., 1997; Anderson et al., 1998; Altenmüller & Gerloff, 1999; McFarland et al., 2000; Wessberg et al., 2000; Pfurtscheller et al., 2000b; Nicolelis, 2001; Pfurtscheller et al., 2003). The BCIscan be used also in therapeutic applications by neurofeedback for rehabilitation or functional recovery (Birbaumer & Cohen, 2007; Dobkin, 2007; Birbaumer et al., 1999; Dornhege et al., 2007).

The BCIis a communication system that does not require any peripheral muscular activity. It has been shown by (Pfurtscheller & Aranibar, 1977; Pfurtscheller, 1999c; Neuper & Pfurtscheller, 1999a ) that the imagination of either a left or right hand movement results in an amplitude attenuation (event-related desynchronization(ERD) of μ(8-13Hz) and central β(13-30Hz) rhythms at the contra-lateral sensori-motor representation area and, in an amplitude increase (event-related synchronization(ERS) within the γband (30-40Hz) at the ipsi-lateral hemishpere. The event related (de)synchronisation(ERD, ERS) (Pfurtscheller et al., 1999a), see Fig. 2 and Fig. 3.

Figure 2.

Grand averageERDcurves recorded during motor imagery from the left (C3) and right sensorimotor cortex (C4) (the electrodes C3 and C4 are placed according to the International 10-20 system). The ERD time courses were calculated for the selected bands in the alpha range for 16 subjects. Positive and negative deflections, with respect to baseline (second 0.5 to 2.5), represent a band power increase (ERD) and decrease (ERD), respectively. The gray bar indicates the time period of cue presentation (i.e. the imagination starts at second 3). Figure from (Pfurtscheller et al., 2000a) which is modified from (Neuper &Pfurtscheller, 1999a).

Figure 3.

ERD maps for a single subject calculated for the cortical surface of a realistic head model. Figure from (Pfurtscheller et al., 2000a) which is modified from (Neuper &Pfurtscheller, 1999a).

The direct BCIscan also be seen as a new means to extend communication for healthy subjects in many fields such as multimedia communication, control of robots, virtual reality and video games (Thomas, 1977; Friedman et al., 2004; Bell et al., 2008; Lécuyer et al., 2008).

There are in general two types of BCIsystems: endogenous tasksand exogenous tasksbased systems (Dornhege et al., 2007).

The endogenous tasks BCIsystems, which are based on spontaneous activity, use brain signals that do not depend on external stimuli and that can be influenced by concentrating on a specific mental task. In order to obtain an efficient task recognition system, several concentration trials of human are, in general, realized. The concentration constraint is a very tiring mental task especially for disabled subjects who might have difficulties in acquiring voluntary control over their brain activity and it must be reduced in order to obtain an efficient task recognition system.

The exogenous tasks BCIsystems, which are based on evoked activity, use brain signals that do depend on external stimuli. Particularly interesting are systems based either on the P300or on SSVEPs(see section 2). Advantages of these potentials are that they are relatively well understood from a neurophysiologic point of view and that they can be evoked robustly across different subjects. Moreover, feedback training is not necessary in these systems, as theses potentials appear "automatically" whenever subjects concentrate onto one out of several stimuli presented in random order (Hoffman et al., 2008). Note that the material presented in this chapter is strongly biased towards sensorimotor(Changes in brain rhythms(μ,β, andγ)) and P300electrophysiological activities using EEGrecords.

In order to improve the performance of the BCIsystem design, it is necessary to use a good method of signal processingto allow easier extraction of physiological characteristics and also to use a good classifieradapted to the specificities of the BCIsystem. This chapter presents a compact guide to different signal processingtechniques that have received more attention in BCIs. We introduce then some selected feature extraction and classificationapproaches in the context of BCIsystems. A more exhaustive and excellent surveys on signal processingand classificationalgorithms may be found in the papers (Bashashati et al., 2007; Lotte et al., 2007). Then this chapter describes the application of two classificationapproaches, hidden Markovmodels(HMMs) and support vector machines(SVM), in the context of exogenous tasks BCIsystems based on P300 evoked potential. The chapter ends with a global conclusions and perspectives.

The methods presented in sections 3.3, 4, 5 and 6 are based on the statistical results given in the comprehensive survey of 96 BCIdesigns using electrical signal recordings published prior to January 2006 by (Bashashati et al., 2007). Among these methods, we give here only a brief descriptions of the most applied methods. They are introduced here without referencing all the published papers for the 96 BCIdesigns. The reader may refer to the paper (Bashashati et al., 2007) to find a rich bibliographical work. However, we give only the original references corresponding to each proposed method.


2. Electrophysiological control activities in BCIs

Current BCIsystems fall into seven main categories, based on the neuromechanisms and recording technology they use to generate control signals (Bashashati et al., 2007). The following list give a short descriptions of these electrophysiological activities used in BCIdesigns. This list is borrowed and adapted (with the authorization of authors) from the paper (Bashashati et al., 2007). We omitted the references of the different approaches given in this list. Many of these references are given in (Bashashati et al., 2007).

  1. - Sensorimotor activity BCIdesigns that use sensorimotor activity as the neural source of control can be divided into three sub-categories:

  2. - Changes in brain rhythms(μ,β,andγ)

  3. μrhythms in the range of 8-12 Hz and βrhythms in the range of 13-30 Hz both originate in the sensorimotor cortex and are displayed when a person is not engaged in processing sensorimotor inputs or in producing motor outputs. They are mostly prominent in frontal and parietal locations. A voluntary movement results in a circumscribed desynchronization in the μand lower βbands. This desynchronization is called event-related desynchronization(ERD) (Pfurtscheller & Aranibar, 1977; Pfurtscheller, 1999c; Neuper & Pfurtscheller, 1999a ) and begins in the contralateral rolandic region about 2 s prior to the onset of a movement and becomes bilaterally symmetrical immediately before execution of movement. After a voluntary movement, the power in the brain rhythms increases. This phenomenon, called event-related synchronization(ERS), is dominant over the contralateral sensorimotor area and reaches a maximum around 600 ms after movement offset. γrhythm is a high-frequency rhythm in the EEG. Upon the occurrence of a movement, the amplitude of γrhythm in the range of 30-40 Hz increases. Gamma γare usually more prominent in the primary sensory area.

  4. - Movement-related potentials(MRPs)

  5. MRPsare low-frequency potentials that start about 1-1.5 s before a movement. They have bilateral distribution and present maximum amplitude at the vertex. Close to the movement, they become contralaterally preponderant.

  6. - Other sensorimotor activities

  7. The sensorimotor activities that do not belong to any of the preceding categories are categorized as other sensorimotor activities. These activities are usually not restricted to a particular frequency band or scalp location and usually cover different frequency ranges. An example would be features extracted from an EEGsignal filtered to frequencies below 30 Hz. Such a range covers different eventrelated potentials(ERPs) but no specific neuromechanism is used.

  8. - Slow cortical potentials (SCPs)

  9. - P300 Evoked potential

  10. - Visual evoked potentials (VEPs)

  11. VEPsare small changes in the ongoing brain signal. They are generated in response to a visual stimulus such as flashing lights and their properties depend on the type of the visual stimulus. These potentials are more prominent in the occipital area. If a visual stimulus is presented repetitively at a rate of 5-6 Hz or greater, a continuous oscillatory electrical response is elicited in the visual pathways. Such a response is termed steady-state visual evoked potentials (SSVEP). The distinction between VEPand SSVEPdepends on the repetition rate of the stimulation.

  12. - Response to mental tasks

  13. BCIsystems based on non-movement mental tasks assume that different mental tasks (e.g., solving a multiplication problem, imagining a 3D object, and mental counting) lead to distinct, task-specific distributions of EEGfrequency patterns over the scalp.

  14. - Activity of neural cells (ANC)

  15. It has been shown that the firing rates of neurons in the motor cortex are increased when movements are executed in the preferred direction of neurons. Once the movements are away from the preferred direction of neurons, the firing rate is decreased.

  16. - Multiple neuromechanisms (MNs)

  17. BCIsystems based on multiple neuromechanisms use a combination of two or more of the above mentioned neuromechanisms.

3. Signal pre-processing methods in BCIs

To extract features(see section 4), it is necessary to pre-process first the data. Three steps are necessary to achieve this goal: Referencing, Temporal filteringand signal enhancement.

3.1. Referencing

(Hagemann et al., 2001) have stated that the differences between results of different studies are partly due to the differences in referencing. In the case of EEGrecordings from the cortex or from the scalp, these recordings are obtained using, in general, different electrodes on different positions. Since the brain activity voltage measured by a given electrode is a relative measure, the measurement may be compared to another reference brain voltage situated on another site. This results in a combination of brain activity at the given electrode, brain activity at the reference site and noise. Because of this, the reference site should be chosen such that the brain activity at that site is almost zero. Typically, the nose, mastoids and earlobes are used (Dien, 1998). In general, there are three referencing methods

  1. - Common reference

  2. The common reference technique is widely used in BCIs. This method uses one common reference for all electrodes. In general, the site of this reference is situated at large distance from all electrodes. The activity at the reference site influences all measurements equally, and differences between electrode measurements still contain all information needed.

  3. - Average reference

  4. The average reference subtracts the average of the activity at all electrodes from the measurements. This method is based on the principle that the activity at the whole head at every moment sums up to zero. Therefore, the average of all activity represents an estimate of the activity at the reference site. Subtracting this average produces in principle a dereferenced solution. However, the relatively low density of the electrodes and the fact that the lower part of the head is not taken into account, bring some practical problems along (Dien, 1998).

  5. - Current source density (CSD)

  6. The current source density (CSD) is used in many BCIs. It is "the rate of change of current flowing into and through the scalp" (Weber, 2001). This quantity can be derived from EEGdata, and it may be interpreted as the potential difference between an electrode and a weighted average of their surrounding electrodes. The CSDcan be estimated by computing the laplacian. The laplaciancomputes the sum of the differences between an electrode and its neighbours. A problem with this estimation is that it is actually only valid when the electrodes are in a two dimensional plane and equally distant.

3.2. Temporal filtering in BCIs

The brain signals are naturally contaminated by many internal and external noises. They can be removed using simple filters. The relevant information in BCIsis found in the frequencies below 30Hz. Therefore, all noise with higher frequencies (e.g. noise from the electrical net has a fixed frequency of 50Hz or 60 Hz) can be removed using FIR low pass filter. Specific frequency bands may also be selected using FIR bandpass filters.

3.3. Signal enhancement methods in BCIdesigns

The choice of a suitable enhancement technique is dependent on several factors such as the recording technology, number of electrodes, and neuromechanism of the BCI(Bashashati et al., 2007). Among seventeen pre-processing methods given by (Bashashati et al., 2007), we describe here briefly only six methods which are the most applied in BCIdesigns:

  1. - Spatial filters - Referencing methods

  2. The proper selection of a spatial filter for any BCIis determined by the location and extent of the selected brain control signal and of the various sources of EEGor non-EEGnoise.

  3. - Common average referencing (CAR)

  4. Common-average or "reference-free" recording has been suggested as a solution to the problem of the reference electrode (Offner, 1950; Lehmann & Skrandies, 1984; Stanny, 1989). Common-average referencing involves recording in bipolar fashion from a number of electrodes, all referred to a single site. One then calculates the grand mean EEGwaveform, by averaging across electrodes, and subtracts the result pointwise from the EEGrecorded at each electrode. Activity recorded by the reference electrode is theoretically of equal magnitude in the mean and individualelectrode waveforms. Consequently, the effect of the reference electrode should be eliminated from each recording electrode's output when the common-average waveform is subtracted (Stanny, 1989).

  5. - Surface Laplacian (SL)

  6. The SLis defined as the 2nd order spatial derivative of the surface potential. Due to its intrinsic spatial high-pass filtering characteristics, the SL can reduce the volume conduction effect by enhancing the high-frequency spatial components, therefore can achieve higher spatial resolution than surface potentials.

  7. - Principal component analysis (PCA)

  8. The PCA(Pearson, 1901) is a linear mapping that transforms a number of possibly correlated variables into a smaller number of uncorrelated variables called principal components. The first principal component accounts for as much of the variability in the data as possible, and each succeeding component accounts for as much of the remaining variability as possible. Depending on the field of application, it is also named the discrete Karhunen-Loève transform(KLT), the Hotelling transformor proper orthogonal decomposition(POD). The PCAreveals the internal structure of the data in a way which best explains the variance in the data. If a multivariate dataset is visualised as a set of coordinates in a high-dimensional data space (1 axis per variable), ICAsupplies the user with a lower-dimensional representation.

  9. - Independent component analysis (ICA)

  10. The more important artefactsin BCIsare generated by muscles and eyes blink (Gupta & Singh, 1996). Classical automatic methods for removing such artefacts can be classified into rejectionmethods and subtractionmethods.

  11. - Rejectionmethods consist of discarding contaminated EEG, based on either automatic or visual detection can be used in the BCIapplications framework. Their success crucially depends on the quality of the detection.

  12. - Subtractionmethods are based on the assumption that the contaminated EEG is a linear combination of an original EEGand other independent artefactsignals generated by the muscles and eyes blink. The original EEGis hence recovered by either subtracting separately recorded artefact-related signals from the measured EEG, using appropriate weights or by applying recent approaches for artefacts rejection: such as independent component analysis(ICA) (Common, 1994; Hyvärinen & Oja, 2000), peak elimination(Nakamura et al., 1996), neural network(Urszula et al., 1999) and fixed bandpass FIR filterbased approach (Gupta & Singh, 1996).

  13. The ICA(Common, 1994; Hyvärinen & Oja, 2000) is the more used technique. It is a computational method for separating a multivariate signal into additive subcomponents supposing the mutual statistical independence of the non-Gaussian source signals. It is a special case of blind source separation(BSS). ICAis particularly efficient when the EEGand the artefactshave comparable amplitudes. For more details about their advantages, their limitations and their applications for the removal of eyes activity artefacts, refer to (Jung et al., 1998; 2000).

  14. - Common spatial patterns (CSP)

  15. The CSP(Koles, 1991; Müller-Gerking et al., 1999) is a technique used to find the common projection matrix that decomposes the different classes of single trial EEGdatasets, and more specifically to find spatial structures of event-related (de)synchronization (ERD/ERS) in a EEGcontext. Such matrix maximizes the differences between the classes. (Guger et al., 2000) demonstrated the efficiency of the CSPmethod for real-time EEGanalysis and concluded that only parameters that must be adjusted for the CSPare the time segment for the calculation of the CSPand, during on-line processing, the time window for the calculation of the variances. But the selection of these parameters is not very crucial. An advantage of the CSPmethod is that it does not require a priori selection of subject-specific frequency bands, as necessary for bandpower or frequency estimation methods (Pfurtscheller et al., 1996; McFarland et al., 1997b).

  16. The CSP method is very sensitive to artefacts. A single trial containing, for example, a movement artifact can cause severe changes in the CSP(Müller-Gerking et al., 1999). The reason is the sample covariance (nonrobust estimate), which is used to estimate the covariance for the calculation of the spatial filters. However, during on-line operation of the BCI, the spatial filters perform a weighted spatial averaging of the EEG, and this reduces the influence of artefacts (Guger et al., 2000).

  17. in some applications, many electrodes are needed, (e.g. more than 18 (Ramoser et al., 2000), which necessitates costly hardware.

  18. since the CSP method detects spatial patterns in the EEG, any change in the electrode positions may render the improvements in the classificationaccuracy gained by this method useless. Therefore, this method requires almost identical electrode positions for all trials and sessions which may be difficult to accomplish (Ramoser et al., 2000). (Guger et al., 2000) recommended not to apply the electrodes anew after setting up a new CSPfor the following feedback sessions. For long-term implications to analyze the EEGin real time, EEGdata of several sessions can be used for the calculation of the CSP. This allows the generation of a more robust filter in order to overcome the mentioned problems.

  19. - Common spatial subspace decomposition (CSSD)

  20. The CSSDcan extract signal components specific to one condition from multiple MEG/EEGdata sets of multiple task conditions. Signal matrices or covariance matrices are decomposed using spatial factors common to multiple conditions. The spatial factors and corresponding spatial filters are then dissociated into specific and common parts, according to the common spatial subspace which exists among the data sets. Finally, the specific signal components are extracted using the corresponding spatial filters and spatial factors. (Wang et al., 1999).

  21. - Frequency normalization (Freq-Norm)(Bashashati et al., 2005).

  22. Other methods are given by (Bashashati et al., 2007).The study of (Bashashati et al., 2007) showed that

  23. signal pre-processing algorithms have been used for EEG-based BCIsand the ANC-based BCIs, but no signal enhancement algorithms have been applied on electrocorticogram(ECoG)-based BCIs. Only PCAhas been used in both groups, and

  24. spatial filteringincluding referencing(CARand SL) methods and CSPare among the most used techniques that have become increasingly popular in EEG-based BCIs.

Fig.4 shows the statistical results of the study realised by (Bashashati et al., 2007) concerning pre-processingmethods in BCIdesigns. (Bashashati et al., 2007) concluded that 96 BCIdesigns that employ signal enhancement techniques before extracting the features from the signal, 32% use surface Laplacian (SL), 22% use either principal component analysis (PCA) or independent component analysis (ICA), 14% use common spatial patterns (CSP) and 11% use common average referencing (CAR) techniques.

Figure 4.

Signal enhancement methods inBCIdesigns. Figure modified from (Bashashati et al., 2007) with authorisation.

In the following, we give a breif description of the two most used methods: Spatial filtersand Common spatial patterns.

3.3.1. Spatial filters: (SL) and (CAR)

(McFarland et al., 1997a) showed that the variability of the EEGor non-EEGnoise sources within the different BCIdesigns and even within individuals make difficult the application of the spatial filters. For BCIsthat use the μand βrhythms, the SLand CARmethods are superior to the ear reference method. However, it was shown that the reference method (CAR, bipolar, large Laplacian, small Laplacian, and referenced to the ear(McFarland et al., 1997a)) had minor influence on the classificationaccuracy (Ramoser et al., 2000). Fast and continuous feedback can also enhance the performance of the system (Guger et al., 2001; Neuper et al., 1999b). In the following, we introduce only the principles of the CSPgiven in (Guger et al., 2000).

3.3.2. Common spatial patterns (CSP)

As described by (Guger et al., 2000), the CSPmethod uses the covariance to design common spatial patternsand is based on the simultaneous diagonalisation of two covariance matrices (Fukunaga, 1972). The decomposition (or filtering) of the EEGleads to new time series, which are optimal for the discrimination of two populations (or classes). The patterns are designed such that the signal resulting from the EEGfiltering with the CSPhas maximum variance for population and minimum variance for the second population and vice versa. In this way, the difference between the first and second populations is maximized, and the only information contained in these patterns is where the variance of the EEGvaries most when comparing two conditions.

Given Nchannels of EEGfor each trial Xof population 1 and population 2, the CSPmethod gives an NxNprojection matrix according to (Koles, 1991; Müller-Gerking et al., 1999; Ramoser et al., 2000; Guger et al., 2000). This matrix is a set of subject-specific spatial patterns, which reflect the specific activation of cortical areas during hand movement imagination. With the projection matrixW, the decomposition of a trial Xis described by


This mapping projects the variance of Xonto the rows of Zand results in new time series. The columns of W1are a set of CSPsand can be considered as time-invariant EEGsource distributions. After interpolation, the patterns can be displayed as topographical maps.

By construction, the variance for population 1 is largest in the first row of Zand decreases with the increasing number of the subsequent rows. The opposite is the case for a trial with population 2.

4. Feature extraction methods in BCI designs

This section describes briefly the common BCIfeatures extraction methods. Concerning the design of a BCIsystem, some critical properties of these features must be considered (Lotte et al., 2007):

  1. noise and outliers: the brain signals (e.g. EEGs) have a poor signal-to-noise ratio;

  2. high dimensionality: in BCIsystems, feature vectors are often of high dimensionality. Several features are generally extracted from several channels and from several time segments before being concatenated into a single feature vector;

  3. time information: BCIfeatures should contain time information as brain activity patterns are generally related to specific time variations of EEG;

  4. the brain signals are non-stationary in nature;

  5. the brain signals are non-linear in nature;

  6. non sufficient training sets: training process is time consuming and demanding for the subjects.

There are many methods used in BCI, depending of the type of the BCIsystems. In the following we describe some main and specific methods. More exhaustive details are given by (Bashashati et al., 2007). The feature extraction methods described here are: Band powers(BP), Cross-correlation betweenEEG band powers, frequency representation(FR), time-frequency representation(TFR), Hjorth parameters, parametric modelling, inverse modeland specific techniques used for P300and VEPsuch as Peak picking(PP) and Slow cortical potentials calculation(SCPs).

4.1. Band powers (BP)

The features may be extracted from the EEGsignals by estimating the power distribution of the EEGin predefined frequency bands. (Pfurtscheller et al., 1997) used the band powers(BP) and demonstrated that for each subject, different frequency components in the αand βband were found which provided best discrimination between left and right hand movement imagination. These frequency bands varied between 9 and 14 Hz and between 18 and 26 Hz.

4.2. Cross-correlation between EEG band powers

In the case of EEGmeasurements the cross-correlation coefficients between the EEGactivity may be calculated to obtain some information from comparing different locations and different frequency bands (Farwell & Donchin, 1988; Musha et al., 1997; Bayliss & Ballard, 1999; 2000a;b; Wang et al., 2004a;b).

4.3. Frequency representation (FR)

Frequency representation(FR) features have been widely used in signal processing because of their ease of application, computational speed and direct interpretation of the results (Wolpaw et al., 2000; Blankertz et al., 2006). Specifically, about one-third of BCIdesigns have used power-spectral density(PSD) features (Bashashati et al., 2007).

4.4. Time-frequency representation (TFR)

Due to the non-linearity and non-stationarity nature of the EEGsignal, the classical methods based on Fourier transform(FT) are, in general, not efficient for feature extraction because the obtained features do not provide any time domain information, i.e. these features do not analyze the time-varying spectral content of the signals.

Time-frequency methods decompose the EEGsinto a series of frequency bands, and the instantaneous power is represented by the envelop of oscillatory activity, which forms the spatial patterns for a given electrode montage at a time-frequency grid (Millán & Mouriño, 2003; Wang et al., 2004a).

Wavelet-based feature extraction algorithms (Qin & He, 2005; Xu & Song, 2008; Haibin et al., 2008) necessitate the choice of a particular wavelet called mother waveletin order to extract useful information. This choice of an appropriate mother waveletmay be simplified by the prior knowledge of the physiological activity in the brain.

(Huang et al., 1998) proposed a more fairly recent technique called the Empirical Mode Decomposition(EMD) was proposed for nonlinear and non-stationary time series data. The (EMD) is a data driven approach (i.e. one does not need to define a mother waveletbeforehand) that can be used to decompose adaptively a signal into a finite well-defined high frequency and low frequency components, which are known as intrinsic mode functions (IMFs) or modes. They consider signals at their local oscillations, but they are not necessarily considered in the sense of Fourier harmonics. Their extraction is non-linear, but their recombination for exact reconstruction of the signal is linear. We think that this approach might be useful in BCIdesign.

4.5. Hjorth parameters

(Hjorth, 1970) has described three parameters that characterize the temporal dynamics of EEGsignal, (X(t)), of length Nin terms of amplitude, time scaleand complexity. The parameters are measured in the time domain, as opposed to the other features, which are measured in the frequency domain. It has been shown that these parameters are capable of discriminating between different mental states (Vourkas et al., 2000). The parameters are:

  1. Activity: a measure of the mean power of the signal (variance ofX(VAR(X))). It is measured as the standard deviation.

  2. Mobility: represents the mean frequency in the signal. The mobility can be computed as the ratio of the standard deviation of the slope and the standard deviation of the amplitude.

  3. Complexity: tries to capture the deviation from the sine wave. It is expressed as the number of standard slopes actually seen in the signal during the average time required for one standard amplitude, as given by the mobility.


These parameters are mainly used as features for the classificationof motor imagery (Obermeier et al., 2001; Boostani & Moradi, 2004; Lee & Choi, 2003; Pfurtscheller & Neuper, 2001).

4.6. Parametric modelling

In statistics, a parametric modelor parametric familyor finite-dimensional modelrefers to a family of distributions which can be described using a finite number of parameters. These parameters are usually collected together to form a single k-dimensional parameter vectorΘ= (θ1,θ2, ...,θk). In system theory, parametric modelassume that the time series under analysis to be the output of a given linear mathematical model. They require an a priori choice of the structure and order of the signal generation mechanism model.

Among the more used parametric modelling in BCIsare the autoregressive parameters (AR)and their variants such as multivariate AR parameters (MVAR), AR parameters with exogenous input (ARMAX)and Adaptive AR parameters (AAR)(Anderson & Sijercic, 1996; Schlogl et al., 1997; Anderson et al., 1998; Roberts & Penny, 2000; Burke et al., 2005; Vidaurre et al., 2007).

ARmethods assume that a signalX(t), measured at timet, can be modeled as a weighted sum of the values of this signal at previous time steps, to which we can add a noise term Et (generally a Gaussian white noise):


where the weights ai are the autoregressive parameterswhich are generally used as features for BCI. AARassume that the weights ai can vary over time. It seems that (AAR) parameters would give better results than (AR) parameters for motor imagery classification(Schlögl et al., 1997; Pfurtscheller et al., 1998), whereas they would give worse results for the classificationof cognitive tasks such as mental computations, mental rotation of a geometric figure, etc. (Huan & Palaniappan, 2004a; Huan & Palaniappan, 2004b). It should be noted that it is possible to derive a frequential information from the ai coefficients (McFarland & Wolpaw, 2005).

4.7. Inverse model

Inverse models have shown to be promising feature extraction algorithms (Qin et al., 2004; Grave et al., 2005; Wentrup et al., 2005; Congedo et al., 2006). Such models are able to compute the activity in the whole brain volume, only using EEGand a head model that generally represents the brain as a set of volume elements(voxels). The activity thus calculated in some relevant brain regions or voxelsmay be used as efficient features for BCIsystems.

4.8. specific techniques

4.8.1. Peak picking (PP)

Peak picking(PP) method detects a specific pattern based on its peak value in a region associated with a specific cognitive component. It is used specifically for the evoked potential P300(or P3)-based BCIsystem (Meinicke et al., 2003; Garrett et al., 2003; Bayliss et al., 2004; Bayliss & Inverso, 2005; Salimi Khorshidi et al., 2007; Hoffman et al., 2008), Fig.5.

Figure 5.

Typical P300 wave. From (Hoffman et al., 2008).

PP is a simple algorithm to recognize a P300 component using the difference between the minimum and maximum amplitude in a trial. A trial with a prototypical evoked potential P300 component contains a large peak from 300-400 ms and PP recognizes the P300 signal when the amplitude difference is greater than or equal to a specified voltage difference between the minimum, min(x), and maximum, max(x), voltage points within a specified time window, where x is a vector which represents the data for a single P300 response. For recognition, the time window with the best results may be between three and six hundred milliseconds citepbayliss04.

4.8.2. Slow cortical potentials (SCPs) calculation methods

The SCPs amplitudes are extracted on-line from the regular electroencephalogram, filtered, corrected for eye movement artefacts and fed back to the patient with visual, auditory or tactile feedback (Birbaumer et al., 1999; 2000; Hinterberger et al., 2004). The TFR methods are also used to extract the features of the SCPs (Hinterberger et al., 2003; Bostanov, 2004 ). Fig.6 shows the statistical results of the study realised by (Bashashati et al., 2007) concerning feature extraction methods in BCI designs. (Bashashati et al., 2007) concluded that 41% of the BCIs are based on the sensorimotor activity use PSD features, 16% rely on parametric modelling of the data, 13% use TFR methods and 6% do not employ any feature extraction methods. 74% of the SCP-based BCI designs calculate SCP signals using low-pass filtering methods, and 64% of the VEP-based BCIs use power-spectral features at specific frequencies. 26% of the BCIs based on P300 calculate the PP; 22% use TFR-based methods, 22% use no feature extraction method, and 15% use cross-correlation with a specific template. 41% of the BCI designs that use mental tasks to control a BCI use power spectral features and 37% use parametric modelling of the input signal. As most of the BCI designs that are based on neural

Figure 6.

Feature extraction methods in BCI designs based on sensorimotor activity, VEP, P300, SCP, response to mental tasks, activity of neural cells, and multiple neuromechanisms. Taken from (Bashashati et al., 2007).

cortical recordings mainly try to model the direct relationship between the neural cortical recordings and movements, they do not use a feature-extraction algorithm. 45% of the BCI designs that are based on multiple neuromechanisms rely on power-spectral features, 17% use parametric modelling, and 17% use TFR methods.

5. Feature selection and dimensionality reduction methods in BCI designs

In BCI applications, several features are generally extracted from several brain activity channels (several electrodes in the case of EEG measurements) and from several time segments (or sessions), before being concatenated into a single feature vector. Hence, the BCIs are often affected by a problem known as curse of dimensionality (Bellman, 1961). It was demonstrated that the amount of data needed to properly describe the different classes increases exponentially with the dimensionality of the feature vectors (Friedman, 1997; Jain, 2000).

(Flotzinger et al., 1994) and (Pfurtscheller & Guger, 1999b) have shown that when feature selection is used, the classification accuracy is better than when all the features are used. If the number of training data is small relatively to the number of features, the classification algorithm which will use these features and data will very likely give bad results. It is recommended to use at least 5 to 10 times more training data per class than the number of features (Jain & Chandrasekaran, 1982; Raudys & Jain, 1991). Unfortunately this cannot be applied in all BCI systems as generally the dimensionality is high and the training set small.

Among fourteen feature selection and dimensionality reduction methods in BCI designs given by (Bashashati et al., 2007), we give here briefly the definitions of only three methods which are the most applied in BCI designs:

  1. - Genetic algorithm (GA)

  2. A Genetic algorithm (GA) (Goldberg, 1989; Flotzinger et al., 1994) is a search technique used in computing to find exact or approximate solutions to optimization and search problems. Genetic algorithms are categorized as global search heuristics. They are a particular class of evolutionary algorithms (EA) that use techniques inspired by evolutionary biology such as inheritance, mutation, selection, and crossover. These algorithms are based on a sequence of generations whereby the population in each generation produces the next while trying to optimize some fitness criterion (Brill, 1992), such as maximum ability to classify the training set in the classification stage. Each member of the current population is assigned a binay-valued chromosome of length n (for an n-dimensional classification problem), whereby the value of each bit within the chromosome defines whether this feature is to be used for classification or not. A chromosome 11 11 1...1, therefore, means that all parameters are to be used for evaluation of the member's fitness and a chromosome 10100... 0 means that only the first and third parameters are to be used. The accuracy which can be achieved using a specific chromosome is calculated using a clustering algorithm such as k-nearest-neighbour classifier (K-NN) or any other clustering algorithm. For k-nearest-neighbour classification the k nearest data vectors are found for a new input vector which is then classified to the label to which the majority of these k data vectors belong. For more details, refer to (Flotzinger et al., 1994; Brill, 1992).

  3. - Principal component analysis (PCA)

  4. Principal component analysis (PCA) may be used in pre-processing stage of BCI designs (see section 3.3). PCA may also be used as a dimensionality reduction technique in terms of capturing the variance of the data, and it accounts for correlation among variable. It gives lower-dimensional representations of the data which better generalize to data independent of the training set than using the entire dimensionality of the observation space (Scholkopf, 1999). The PCA transforms a set of mvariables into another set of kmuncorrelated variables, maintaining as many of the variance of original data as possible (Moghaddam, 2002).

  5. - Distinctive sensitive learning vector quantization (DSLVQ)

  6. The influence of distinctive features on the distance function in the standard learning vector quantisation (LVQ) (Kohonen, 1990) is equal. The Distinction Sensitive (DS) algorithm (DSLVQ) (Flotzinger et al., 1994) employs an adaptive weighted distance function where the influence of features which frequently contribute to misclassifications is reduced while the influence of features which are shown to be very significant for proper classification is increased. For the weighted distance function of DSLVQ a global weights vector w is used which stores the distinctiveness, i.e. the relevance, of every single feature. This weights vector is adapted interactively along with the codebook. The distance function used may be the Euclidean distance, or any other weighted distance functions. (Flotzinger et al., 1994) proposed the following Euclidean distance between two feature vectors x and y:


The weights vector w can be seen as a scaling transformation from the original feature space into a DS-feature space. This transformation increases distances for very distinctive features and decreases distances for common features. Despite the usage of a weighted distance function, the codebook learning for DSLVQ is the same as for the LVQ3 algorithm. Additionally, the weights vector w must be updated with every learning iteration. Learning weights and codebook settings in parallel facilitate a quick approximation of these related parameters. For more details, refer to (Flotzinger et al., 1994).

Fig. 7 shows the statistical results of the study realised by (Bashashati et al., 2007) concerning feature selection and dimensionality reduction methods in BCI designs. (Bashashati et al., 2007) concluded that thirty-eight of the reported BCI designs employ feature selection and dimensionality reduction algorithms; 26% of these 38 designs use genetic algorithms (GA), 24% use distinctive sensitive learning vector quantization (DSLVQ), and 13% use PCA.

Figure 7.

Feature selection and dimensionality reduction methods in BCI designs. Figure modified from (Bashashati et al., 2007) with authorisation.


6. Classification in BCIs

Brain activity patterns are considered as dynamic stochastic processes due both to biological and to technical factors. Biologically, they change due to user fatigue and attention, due to disease progression, and with the process of training. Technically, they change due to amplifier noises, ambient noises, and the variation of electrode impedances (Wolpaw et al., 2002). Therefore, the time course of the generated time series signals (e.g. EEG) should be taken into account during feature extraction (Wolpaw et al., 2002). To use this temporal information, three main approaches have been proposed (Lotte et al., 2007):

  1. concatenation of features from different time segments: extracting features from several time segments and concatenating them into a single feature vector (Pfurtscheller et al., 1997; Haselsteiner & Pfurtscheller, 2000);

  2. combination of classifications at different time segments: it consists in performing the feature extraction and classification steps on several time segments and then combining the results of the different classifiers (Penny & Roberts, 1999; Lemm et al., 2004);

  3. dynamic classification: it consists in extracting features from several time segments to build a temporal sequence of feature vectors. This sequence can be classified using a dynamic classifier (Haselsteiner & Pfurtscheller, 2000; Obermeier et al., 2001).

The first approach is the most widely used despite that the obtained feature vectors are often of high dimensionality.

6.1. Classifier selection criteria

In order to choose the most appropriate classifier for a given set of features, the properties of the available classifiers must be chosen according to the following four classifier taxonomy and two main problems in BCI design as described by (Lotte et al., 2007):

6.1.1. Classifier taxonomy

  1. - Generative or Informative classifier - Discriminative classifier

  2. Generative classifiers, e.g., Bayes quadratic, learn the class models. To classify a feature vector, generative classifiers compute the likelihood of each class and choose the most likely. Discriminative ones, e.g., support vector machines (SVM), only learn the way of discriminating the classes or the class membership in order to classify a feature vector directly (Ng & Jordan, 2002; Rubinstein & Hastie, 1997);

  3. - Static classifier - Dynamic classifier

  4. static classifiers, e.g., multilayer perceptrons (MP), cannot take into account temporal information during classification as they classify a single feature vector. In contrast, dynamic classifiers such as hidden Markov model (HMM) (Rabiner, 1989), FIR filters-multilayer perceptrons (FIR-MLP) (Haselsteiner & Pfurtscheller, 2000) and Tree-based neural network (TBNN) (Ivanova et al., 1995), can classify a sequence of feature vectors and thus catch temporal dynamics.

  5. - Stable classifier - Unstable classifier

  6. Stable classifiers, e.g., linear discriminant analysis (LDA), have a low complexity (or capacity (Vapnik, 1995; 1999)). They are said to be stable as small variations in the training set do not considerably affect their performance. In contrast, unstable classifiers, e.g., multilayer perceptron, have a high complexity. As for them, small variations of the training set may lead to important changes in performance (Breiman, 1998).

  7. - Regularized classifier

  8. Regularization consists in carefully controlling the complexity of a classifier in order to prevent overtraining. Regularization helps limit (a) the influence of outliers and strong noise, (b) the complexity of the classifier and (c) the raggedness of the decision surface (Müller et al., 2003). A regularized classifier has good generalization performances and is more robust with respect to outliers (Duda, 2001; Jain, 2000) (see section 6.1.2).

6.1.2. Main classification problems in BCI research

While performing a pattern recognition task, classifiers may be facing two main problems related to the feature properties. These problems are: the curse-of-dimensionality and the bias-variance tradeoff. The first problem is discussed in section 5. The second problem is a general one where three inherent errors may occur.

Independently of the applications of the classification, this stage consists in association of the true class label c*corresponding to a feature vector x using a mapping (or a model f that may be learnt from a training set T. Let f*the optimal unknown theoretical mapping that has generated the labels. If the mean square error (MSE) is considered to evaluate the inherent classification errors, then these errors can be decomposed into three terms (Breiman, 1998; Friedman, 1997):


where the expectation is with respect to the distribution over c*and the introduced term E[f(x)]is the mean value of the estimated class label. These three terms describe three possible sources of inherent classification error:

  1. - noise: represents the irreducible noise within the system;

  2. - bias: represents a systematic error which is the divergence between the estimated mapping (i.e. the estimated class label) and the best mapping (i.e. the true class label). This term depends strongly on the classification method that has been chosen to obtain f(linear, quadratic,... );

  3. - variance: reflects the sensitivity to the used training set T.

The only terms that may be minimised are the bias and the variance terms. Simple classifiers (e.g. linear classifiers) have a high bias but low variance whereas more complex classifiers (e.g. polynomial classifiers) have a low bias but a high variance. To select the optimal model complexity, we must solve this bias-variance dilemma (Geman et al., 1992).

Actually, stable classifiers tend to have a high bias and a low variance, whereas the inverse is true for unstable classifiers. This can explain why simple classifiers sometimes outperform more complex ones. A low variance can be a solution to cope with the stochastic nature of brain signals. In the presence of strong noise and outliers, even linear systems can fail. One way of overcoming this problem is to use techniques, known as stabilization techniques. These techniques can be used to reduce the variance, e.g. combination of classifiers (Breiman, 1998) and regularisation techniques (Duda, 2001; Jain, 2000).

The nonlinear classifiers have more number of parameters than those in the linear classifiers. In addition, these parameters must be chosen appropriately. However, when there are large amounts and limited knowledge of data, nonlinear classifiers are better suited in finding the potentially more complex structure in the data. In particular, when the source of the data to be classified is not well known, using methods that are good at finding nonlinear transformation of the data is suggested. In these cases, kernel-based and neural-networks based methods can be used to determine the transformations. Kernel-based classifiers are classification methods that apply a linear classification in some appropriate (kernel) feature space. Thus, all the beneficial properties of linear classification are maintained, but at the same time, the overall classification is nonlinear. Examples of such kernel-based classification methods are support vector machines (SVMs) (Vapnik, 1999) and kernel Fisher discriminant (KFD) (Müller et al., 2003).

Better performances might also be achieved by using group (committee) of classifiers rather than using a single classifier but only a few BCI designs have employed such an approach in classifying features and achieved performance improvements (Millan et al., 2000; Peters et al., 2001; Varsta et al., 2000; Millan et al., 2002).

6.2. Classifiers used in BCI research

There are many categories of classification algorithms used to design BCIs. Among the more used Classifiers in the BCI community are: linear discriminant classifiers (LDC), neural networks (NN), nonlinear Bayesian classifiers (NBC), nearest neighbours classifiers (NNC) and combinations of classifiers. As a linear discriminant classifiers (LDC), we introduce in this section only the linear discriminant analysis (LDA) and support vector machines (SVMs), and as nonlinear Bayesian classifiers (NBC), we introduce in this section only the hidden Markov models (HMMs). For an extensive description and tutorials on these and other approaches, refer to (Lotte et al., 2007; Bashashati et al., 2007; Hung et al., 2005).

6.2.1. linear discriminant classifiers (LDC)

LDC are the most popular in BCI design. Two main kinds of LDC have been used for BCI design, namely, linear discriminant analysis (LDA) and support vector machine (SVM). Linear discriminant analysis (LDA)

LDA or Fisher's LDA has been used with success in many of BCIs such as motor imagery based BCI (Pfurtscheller, 1999c), P300 applications (Bostanov, 2004; Hoffman et al., 2008), multiclass (Garrett et al., 2003) or asynchronous (Scherer et al., 2004) BCI. It has a very low computational requirement and it is simple to use and generally provides good results.

The idea of LDA (Fukunaga, 1990; Duda, 2001) is to find a weight vector w so that two projected clusters c1 and c2 of N1 and N2 training feature vectors {x1,x2,...,xN1}and{x1,x2,...,xN2}on w can be well separated from each other by hyperplanes while keeping small variance of each cluster. This can be done by maximizing the so-called Fisher's criterion. LDA assumes normal distribution of the data, with equal covariance matrix for both classes. The separating hyperplane is obtained by seeking the projection that maximizes the distance between the two classes' means and minimizes the between variance (Fukunaga, 1990).


with respect tow, where Sbis the between-class variance matrix:

and Swis the within-class variance matrix:

where mc1and mc2are the cluster mean of classes c1and c2respectively.

The optimal weight vector w is the eigenverctor corresponding to the largest eigenvalue ofSw1Sb. Once this vector is obtained using the training data xiis obtained by means of the training data, the classification then may be done by projecting the test feature vectors on it, and then the projected test feature vectors may be classified by employing a clustering rule such as k-nearest-neighbor decision rule.

In the case of multiclass separation problem(ci,i 2), several hyperplanes are used. The strategy generally used in this case is the one versus the rest (OVR) (Schlögl et al., 2005) which separate each class from all the others. This technique is suitable for the on-line BCIs because it has a very low computational requirement. It is simple to use and generally provides good results. The main drawback of LDA is its linearity that can provide poor results on complex nonlinear EEG data (Garcia et al., 2003). A regularized Fisher's LDA (RFLDA) has also been used in the field of BCI (Blankertz et al., 2002; Müller et al., 2004) and give better results for BCI than the non-regularized version (Blankertz et al., 2002; Müller et al., 2004). Support vector machine (SVM)

The generalization of the pattern recognition or regression results obtained from a limited sample constitutes the essential stake of the artificial training (machine learning). It is known that the only minimization of the empirical risk (the error of training) does not guarantee a weak error on a corpus of test. Thus the techniques of regularization, used since the years 1960, allow to carrying out a compromise between the capacity of the model to be learned (related to its complexity) and its aptitude to be generalized. From the point of view conceptual, the concept of structural risk introduced by Vladimir Vapnik into the years 1990 (Vapnik, 1995; Cortes & Vapnik, 1995 ; Vapnik, 1999) gives a bound of the error of test according to the error of training and of the complexity of the model. Support vector machine (SVM) have been applied in various domains for classification and regression.

In the following, we will outline briefly the idea of SVM for binary classification and then how this method can be applied to solve multi-class classification problems.

  1. - Binary classification SVM

Support vector machines were originally designed for binary classification (Cortes & Vapnik, 1995 ). The main idea of this approach is to construct a hyperplane in order to separate two classes (ci{1, 1} orci{0, 1})so that the margin (Friedman, 1996) (the distance between the hyperplane and the nearest point(s) of each classes) is maximal and with minimal error. The theoretical principle of the SVM comprises two fundamental steps:

  1. Make a nonlinear mapping Φ of the feature vectors from input space towards another space (redescription space) of higher dimension provided with a scalar product (Hilbert space), Fig. 8.

Figure 8.

The idea of the SVM: transform a nonlinear problem of separation in the representation space into a linear problem of separation in a redescription space of higher dimension.

  1. Construct an optimal hyperplane allowing an optimal linear separation in this redescription space.

The interest of this mapping is that in the redescription space the pattern recognition can prove easy task: intuitively, more the dimension of the redescription space is hight; more the probability of being able to find a hyperplane separating between the feature vectors is high, Fig. 9.

Figure 9.

Linear mapping: the optimal separating hyperplane is that which maximizes the margin in the redescription space.

From the mathematical point of view, the nonlinear mapping (Φ:2n)is realized via a Kernel function (Hilbert-Schmidt Kernel) easy to calculate. In practice, some families of Kernel functions are known and it is allocated to the user of SVM to carry out test to determine that which is appropriate for its application (it is a question of translating the maximum a priori knowledge of which one lays out on the studied problem and the data). Fig. 10 shows a non-linear separation example.

Figure 10.

Non-linear mapping: a complex shape. Figure fromSchölkopf & Smola (2002).

Let x denote a vector drawn from the input space, assumed to be of dimension N0and let {Φj(x)},j= 1, 2, ...,N1denote a set of nonlinear transformations from the input space to the redescription space, where N1is the dimension of the redescription space. Given such a set of nonlinear transformations, we may define a hyperplane acting as the decision surface which is computed in the redescription space in terms of the linear weights of the machine as follows:


where w= {w0,w1, ...,wN1}denotes a set of linear weights connecting the redescription space to the output space. And it is assumed that Φ0(x) = 1for allx, so that w0denotes the bias. Define the vectorΦ(x) = [Φ0(x),Φ1(x), ...,ΦN1(x)]T, then we can rewrite the decision surface in the compact form:


Given the training feature vectors, Φ(xi)corresponds to the input patternxi, and the corresponding desired responsedi, i= 1, 2, ...,N0, which is either 1 or -1 (or, 0 or 1), it has been shown that (Haykin, 1994) the optimal weight vector wcan be expressed as


where{αi}, i= 1, 2, ...,N0is the optimal Lagrange multipliers resulted from maximizing the subject function


subject to the constraints

0 aiCE16

where Cis a user-specified constant which may be considered as regulation parameter. Substituting equation (13) into (12), we obtain the optimal hyperplane


which will be used for linearly separating the testing data, i.e. for any testing samplex, if


then x is classified into the subset having the training responsedi= 1, otherwise it is classified into the other subset withdi=(or 0). The user may choose the radial basis function (RBF) in defining the inner product kernelK(x,xi) =ΦT(xi)Φ(x).

According to equation (16), once the number of nonzero Lagrange multipliers, αi, is determined, the number of radial basis functions and their centers are determined automatically. This differs from the design of the conventional neural network, such as the back-propagation neural network or radial-basis function network (Haykin, 1994), where the numbers of hidden layers or of hidden neuron are usually determined heuristically. The kernel generally used in BCI research is the Gaussian or radial basis function (RBF) kernel:


In this case, the SVM is known as Gaussian SVM or RBF SVM (Burges, 1998; Bennett & Campbell, 2000; Garrett et al., 2003; Kaper et al., 2004).

This classifier has been applied, with success, to a relatively large number of BCIs problems (Blankertz et al., 2002; Garrett et al., 2003; Rakotomamonjy et al., 2005; Helmy et al., 2008; Trad et al., 2009).

    To solve a given multiclass problem it is preferable to use a combination of several binary SVM classifiers and a decision strategy to decide the class of the input pattern. The most popular methods to achieve this goal are: one-versus-all method using Winner-Takes- All strategy (WTA-SVM); One-Versus-One method implemented by Max-Wins Voting (MWV-SVM) (Duan et al., 2007); Decision Directed Acyclic Graph (DAGSVM) (Platt et al., 2000); and error correcting codes (Dietterich & Bakiri, 1995).

    6.2.2. Nonlinear Bayesian Classifiers (NBC)

    This section introduces one Bayesian classifier used for BCI: hidden Markov models (HMMs). This Classifier produces nonlinear decision boundaries. Furthermore, they are generative, which enables them to perform more efficient rejection of uncertain samples than discriminative classifiers. As brain signals used to drive BCI have specific time courses, HMM have been applied to the classification of temporal sequences of BCI features or even raw EEG (Obermeier et al., 2000; 2001; Cincotti et al., 2003; Schlogl et al., 2005; Helmy et al., 2008; Trad et al., 2009).

    HMMs are very efficient nonlinear techniques used for the classification of time series (Rabiner, 1989). HMMs have been widely applied in automatic speech recognition and now being applied to many fields, e.g. signal processing, pattern recognition, modelling and control. Their use necessitates two stages: a training stage where the stochastic process models are estimated through extensive observation corpus and decoding or detection stage where the model may be used off/on-line to obtain the likelihoods of the given testing sequence evaluated by each model.

    In its conventional definition, a HMM is defined by the following compact notation to indicate the complete parameter set of the modelα= (Π,A,B), where Π,Aand Bare the initial state distribution vector, matrix of state transition probabilities and the set of the observation probability distribution in each state, respectively:Π= [π1,π2, ...,πN], with πi=P(q1=si), A= [aij], withaij=P(qt+1=sj|qt=si), i, jN, si,sj2S,S= {s1,s2, ...,sN},t{1, 2, ...,T}. The observation at time (or index) t, Ot, is considered in this paper as continuousOtK. For a continuous observation (CHMMs case), the state conditional probability of the observation bi(Ot)may be defined by a finite mixture of any log-concave or elliptically symmetric probability density function (pdf ), e.g. Gaussian pdf, with state conditional observation mean vector ¯iand state conditional observation covariance matrixΣi. In this paper we consider only a single Gaussian pdf, so Bmay be defined asB= {¯i,Σi},i=1, 2, ...,N. At each instant of time t, the model is in one of the statesi,1iN. It outputs Otaccording to a density function bi(Ot)and then jumps to statej, jNwith probabilityaij. The state transition matrix defines the structure of the HMM (Rabiner, 1989). The model λmay be obtained off-line by a given training procedure. In practice, given the observation sequenceOl= {O1,O2, ...,OT}, lLand a modelλ, the HMMs need three fundamental problems to be solved

    1. How to calculate the likelihoodP(Ol|λ)? The solution to this problem provides a scoren of how Olbelongs toλ.

    2. How to determine the most likely state sequence that corresponds toOl? The solution to this problem provides the sequence of the hidden states corresponding to the given observation sequenceOl.

    3. How to adjust the model λin order to maximizeP(Ol|λ)? This is the problem of estimating the model parameters given a corpus of training observations sequences.

    Problems 1 and 2 are solved in the decoding or detection stage using the forward or the Viterbi algorithms (Rabiner, 1989), while problem 3 is solved during the training phase using either a conventional algorithm such as the Baum-Welch algorithm (Rabiner, 1989) or another optimization based algorithm such as the simulated annealing based algorithm (Alani & Hamam, 2010).

    7. Hidden Markov models and support vector machines approaches applied to P300-based Brain-Computer interface - a comparative study

    This section reports on our contribution to investigate the use of Hidden Markov Models (HMMs) and Support Vector Machines (SVMs) approaches for tasks classification in P300-based Brain-Computer Interface (BCI) system. These approaches are applied to electroencephalogram (EEG) data sets of the École Polytechnique Fédérale de Laussane - BCI Groupe. These data are issued from five disabled subjects and four able-bodied subjects. For each subject, the HMMs and SVM are trained on different sets. Each set represents a specific P300 potential evoked task and contains multiple sequences (issued from different trials) of EEG records. For each task, the HMMs and the SVM that have been built take into account the variability of EEGs during different trials. Based on Bayesian Inference Criterion (BIC) (Schwarz, 1978), the proposed HMM training algorithm is able to select the optimal number of states corresponding to each task. This scheme makes the training procedure independent of the initialization problem and the a priori knowledge of the optimal number of HMM-related states needed in the training algorithm. The SVM is trained by introducing the different sets at the same time. We report training procedures and testing results for the HMMs and the SVM then we compare the performance of the two approaches on the same testing data sets. We also present promising preliminary comparative results based on the two classification approaches with correct classification rates for all subjects.

    7.1. Problem definition

    It has been shown by (Hoffman et al., 2008) that evoked potential (P300) in EEG data due to different images (tasks) appearing randomly on a computer screen in face of the subject may be efficiently used for exogenous mental task recognition. In Hoffman's work, extensions of LDA have been employed for pattern recognition. In our work we investigate the use of HMMs and SVM to classify the same P300 based tasks defined in (Hoffman et al., 2008) and using tasks data given in (Hoffman et al., 2009).

    7.2. Data structure and feature extraction

    For the clarity of the following sections, this section introduce briefly the experimental setup and the data given by Hoffman et al. (Hoffman et al., 2008), (Hoffman et al., 2009) that we adopted in our work.

    7.3. Hoffman's experimental setup

    In the experimental setup given by Hoffman et al., the users used a laptop screen displaying randomly six images (a television, a telephone, a lamp, a door, a window, and a radio), Fig. 11. Subjects were asked to count silently how often a prescribed image was flashed. The six images (Tasks) were displayed on the screen and a warning tone was issued.

    Figure 11.

    The display used for evoking the P300. One image was flashed at a time (Hoffman et al., 2008), (Hoffman et al., 2009).

    These images were selected in order to control some electrical appliances based on the BCI system. Each image flashed during 100 ms followed by a period of 300 ms of non flashing. The sampling rate of EEG was 2048 Hz using 32 electrodes according to the 10-20 international system.

    7.3.1. Subjects

    The recorded data given by Hoffman et al. (Hoffman et al., 2009) and used in our work are issued from five disabled subjects (subjects 1 to 5) and four able-bodied subjects (subjects 6 to 9). The disabled subjects were all wheelchair-bound but had varying communication and limb muscle control abilities. Subjects 1 and 2 were able to perform simple, slow movements with their arms and hands but they were unable to control other extremities. Spoken communication with subjects 1 and 2 was possible, although both subjects suffered from mild dysarthria. Subject 3 was able to perform restricted movements with his left hand but was unable to move his arms or other extremities. Spoken communication with subject 3 was impossible. However the patient was able to answer yes/no questions with eye blinks. Subject 4 had very little control over arm and hand movements. Spoken communication was possible with subject 4, although a mild dysarthria existed. Subject 5 was only able to perform extremely slow and relatively uncontrolled movements with hands and arms. Due to a severe hypophony and large fluctuations in the level of alertness, communication with subject 5 was very difficult, thus this subject was not taken into account by Hoffman at al. (Hoffman et al., 2008), (Hoffman et al., 2009). Subjects 6 to 9 were normal persons (all male, age 30 to 32). None of subjects 6 to 9 had known neurological deficits. In this study we use the data from all subjects.

    7.3.2. Preprocessing and feature extraction

    The data were preprocessed (referencing, bandpass-filtering, downsampling, and artefacts rejection). The EEG was downsampled from 2048 Hz to 32 Hz by selecting each 64th sample from the bandpass-filtered data. For each subject, the data were obtained during four sessions. One session comprised on average 810 trials, and the whole data for one subject consisted on average of 3240 trials. One trial duration takes 1000 ms. Single trials of duration 1000 ms were extracted from the data. Single trials started at stimulus onset, i.e. at the beginning of the intensification of an image, and ended 1000 ms after stimulus onset. Due to the Inter-stimulus interval of 400 ms, the last 600 ms of each trial were overlapped with the first 600 ms of the following trial.

    7.3.3. Electrode selection

    Hoffman et al. (Hoffman et al., 2008), (Hoffman et al., 2009) proved that the configuration with 8 electrodes gives the best result for the studying data Fig. 12.

    Figure 12.

    The best configuration with 8 electrodes given by Hoffman et al. (Hoffman et al., 2008), (Hoffman et al., 2009).

    7.3.4. Feature vectors construction

    The samples from the selected electrodes were concatenated to obtain matrices of feature vectors. Each matrix is composed of Nelines and Ntcolumns, where Nedenotes the number of electrodes and Ntdenotes the number of temporal samples (feature vectors) in one trial. Due to the trial duration of 1000 ms and the downsampling to 32 Hz in (Hoffman et al., 2008) and (Hoffman et al., 2009), Ntis always taken equal to 32. In this work, we consider only matrices of dimension 8 × 32, i.e. each matrix Ocontains a sequence of 32 feature vectorsO= {O1,O2, ...,O32},OtK,K= 8. For the purpose of our HMMs or SVM training and testing schemes, we divided the set of all matrices corresponding to each subject into two parts: training set and testing set. Each set contains six subsets of multiple observation sequences (feature matrices{O1,O2, ...,OL},L 1), one subset for each image (Task) issued from different trials. Fig. 13 summarizes the training and data structures used in this work

    Figure 13.

    Structures of the training and testing data sets for subjects, 1s8. There are six subsets of multiple observation sequences (feature matrices{O1,O2, ...,OL},L1), one subset for each image (Task). Each matrix is of dimension 8 ×32.

    7.4. Method

    In this section, we give the HMM and SVM training schemes and their implementations in the context of our work.

    7.5. HMM training algorithm

    Given a set of L training data sequences{O1,O2, ...,OL}, Ol= {O1,O2, ...,OTl}, lLissued from one of Zclasses of stochastic processespz,z1, 2, 3, ...,Z. The problem of the training in our work is to build ZHMMs for each processpz.

    We consider that each HMM contains an optimal state number Noptof internal states to be determined automatically between NminandNmax. In order to determineNopt, we developed an optimal state number selection algorithm based on the Bayesian Inference Criterion. This algorithm constructs a number of HMMs with different number of states between Nmin and Nmax given by the user. For every state number, each iteration is initialized by the most appropriate model using data clustering, and by the rejection of the least linked state or the rejection of the least probable state of the previous iteration. Consequently, every training iteration begins by a more precise model. After constructing these HMMs, the algorithm selects the optimal HMM with the number of states that maximize the BIC.

    Given NminandNmax, the algorithm may be summarized as follows:

    1. Initialization (clustering): estimate the initial HMM λ0with Nmaxstates using k-mean clustering algorithm.

    2. WhileNNmin, do:

    a. run a training algorithm such a Baum-Welch algorithm (used in this work) until some convergence criterion is satisfied

    b. calculate and save the selection criterion, BIC(N), of the current model


    FN is the total number of free parameters of the current estimated HMM (λN) and Tt be the total number of observations in all training observation sequences, i.e. Tt=l=1LTl

    c. determine the number of zeros in each line of the state transition matrix, A.

    d. select the best initialization method for the next HMM according to the following criterion:

    • if all the lines of A contain zeros, then make a clustering of the observation data.

    • if all the lines of A don't contain zeros but only some lines, then find and reject the state with minimal connections to other states.

    • else, calculate the least probable state and reject it.

    e. obtain a reduced model λ¯

    f. Fixe NN1and λN  λ¯and go to step 2

    In this work we consider only a single Gaussian pdf, so Bmay be defined asB= {¯i,Σi}, i= 1, 2, ...,N. In this case FNis calculated by


    whereOtK. In our case, we used Nmin= 2andNmax= 4. These values were chosen in order to not constraint the clustering algorithm because the number of feature vectors is always equal to 32 (see section 7.2). The training and testing data sets for each subject described in section 7.2 were used for the following training and testing schemes:

    7.5.1. Training scheme

    Using the above algorithm, the goal of this scheme is to construct, for each subject, six HMMs: HMM1, HMM2,.., HMM6 corresponding to six images (Tasks). Each HMM is constructed with either 2 states, 3 states or 4 states. The training data sequences for all subjectss,s 8and for the six images, all collected from different trials (see section 7.2), are given by the set{Tr_s_I}, s 8, I 6, where.Tr. and.I. denote training and image (or task) number respectively.

    7.5.2. Off-line testing scheme

    Our task is to classify six given testing feature sequences into classI,I 6. In order to make off-line (or on-line detection in future work) of these sequences, a stochastic dynamic programming based (Viterbi decoding algorithm) (Rabiner, 1989) has been used. This algorithm gives at the same time the solution to problems 1 and 2 of HMMs discussed in section 6.2.2. In this case, the likelihoods of the data sequence are compared with respect to all six built HMMs. The selected HMM (i.e. task) is the model that maximize the likelihood. The testing data sequences for all subjectss,s 8and for the six images, all collected from six trials (Hoffman et al., 2008), (Hoffman et al., 2009), are given by the set{Te_s_I}, s 8, I 6, where Teand Idenote testing and image (or task) number respectively. Fig 14 show the training and testing schemes for one subject and for six images.

    7.5.3. Multiclass SVM (M-SVM)

    In our work, we used a modified version of the multiclass SVM (M-SVM). The M-SVM method adopted in this work is the one-versus-all. This method is done by winner takes-all

    Figure 14.

    Left: Training scheme for one subject and for six images. Right: Testing scheme for one subject given one or more task-related sequences.

    combination strategy (Friedman, 1996), the outputs of the binary one-vs-all classifiers are compared and the most probable output among the other ones is chosen. The ithSVM is trained with all of the examples in the ithclass with positive labels, and the rest with negative labels. An optimal hyperplane is constructed to separate N/M positive examples from N(M – 1)/M negative examples where M is the number of SVMs and N is the number of the training examples used to construct an SVM for a given class. The same training and testing data structures, described in sections 7.2, 7.5.1 and 7.5.2, were used for SVM training and testing. To train or to test the SVM for a one subject, a number of ×Ltraining or testing feature matrices for all 6 tasks corresponding to this subject were concatenated and fed together to the M-SVM algorithm.

    7.6. Results

    The performance of BCI depends on different evaluation parameters. Among these evaluation parameters of interest in this work is the classification accuracy. This section reports experimental evaluations of the classification accuracy based on the testing scheme introduced in section 7.5.2. Using 60 testing feature matrices for all tasks corresponding to each subject (10 feature matrices for each task), the obtained classification results give the highest percentages of classification rates for the two applied approaches SVM and HMMs. The highest classification percent rates for all subjects are shown in the diagonals of the confusion matrices given in Table 1, see (Appendix). To conclude on the obtained results, we consider the case of comparison between the one of the more severely disabled subject, i.e. subject 3 (late-stage amyotrophic lateral sclerosis) and one of the able-bodied subjects, i.e subject 6 (subject had no known neurological deficits). Fig. 15 show the HMMs-SVM based comparative result for subject 3 and for subject 6. Fig. 15 and Table 1 show that the overall classification percent rates give a satisfactory and promising results for all the disabled subjects compared to the able-bodied ones. Fig. 16 show the HMMs-SVM based comparative results of the means and standard deviations of all the tasks generated by all subjects.

    Fig. 16 show that

      Figure 15.

      Classification percent rates versus task number for subject 3 (top): late-stage amyotrophic lateral sclerosis and for subject 6 (down): subject had no known neurological deficits.

      Figure 16.

      Means and standard deviations of the classification percent rates in all tasks versus subject number (top: HMMs case, down: SVM case).

      1. In both the HMMs and the SVM cases, the means of the classification percent rates for the disabled subjects (except subject 2) as well as for the able-bodied subjects (except subject 9) are approximately uniformly distributed.

      2. The standard deviations of the classification percent rates for all subjects in the SVM case are less than or equal to those in the HMMs case.

      These results give a way to study the ability of the different persons to learn the different P300-based BCI tasks.

      8. Appendix

      Table 1.

      Confusion matrices for the eight subjects and the six tasks (subject 5 is not taken into account). Left: HMMs-based, right: M-SVM-based. All the values are rounded to make the text readable.

      9. Conclusions

      This chapter have introduced a comprehensive survey of signal processing, feature extraction/ selection and classification methods used to provide the readers with guidelines on design brain-computer interfaces (BCIs). Based on the results given by (Bashashati et al., 2007) and (Lotte et al., 2007), this chapter shows which method have received more attention by the BCI community prior to 2006. However, other existing methods, not currently applied in BCI design, could be explored. The exploration of new methods would be strongly driven by the new properties that will have to be taken into consideration in the real future applications of BCIs.

      As a contribution to the classification methods, a Hidden Markov Models (HMMs) and support vector machines (SVM) approaches for task classification in P300-based BCI system is presented. In the HMMs case, we proposed a training algorithm which is able to select automatically the optimal number of HMM-related states corresponding to each set of EEG training records. In the SVM case, we applied a multiclass SVM based on the one-versus-all method. The confusion matrices give a correct classification rates for the able-bodied subjects as well as for the disabled subjects. The comparative results demonstrate that the two approaches are promising. If the HMMs are not constructed correctly or the training data set is not sufficient, the HMM approach gives in this case a low classification rates. In such cases, a hybrid HMM-SVM may be employed, where the HMM is used as a dynamic classifier and the SVM is used as a good discriminative classifier by considering all the training examples in the training data set and train all the task-related models simultaneously. The authors are currently working on extending this work to include off-line and on-line training and classification schemes by using this strategy. The information transfer rate (bit rate) will be taken into account to include both accuracy and speed of a BCI system.



      The authors would like to thank Dr. U. Hoffman and the EPFL-Brain-Computer team for the data and the software given in (Hoffman et al., 2008) that they were used in this work. The authors would like also to thank Dr. A. Bashashati for his authorization to use or modify some figures given in the paper (Bashashati et al., 2007) to illustrate some sections given in this chapter.

      © 2010 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike-3.0 License, which permits use, distribution and reproduction for non-commercial purposes, provided the original is properly cited and derivative works building on this content are distributed under the same license.

      How to cite and reference

      Link to this chapter Copy to clipboard

      Cite this chapter Copy to clipboard

      Tarik Al-ani and Dalila Trad (January 1st 2010). Signal Processing and Classification Approaches for Brain-Computer Interface, Intelligent and Biosensors, Vernon S. Somerset, IntechOpen, DOI: 10.5772/7032. Available from:

      chapter statistics

      6537total chapter downloads

      10Crossref citations

      More statistics for editors and authors

      Login to your personal dashboard for more detailed statistics on your publications.

      Access personal reporting

      Related Content

      This Book

      Next chapter

      Toward Mobile Sensor Fusion Platform for Context-Aware Services

      By Akio Sashima, Takeshi Ikeda, and Koichi Kurumatani

      Related Book

      First chapter

      Biosensor for Environmental Applications

      By Andrea Medeiros Salgado, Lívia Maria Silva and Ariana Farias Melo

      We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities.

      More About Us