Open access

Optimal Fractal Feature and Neural Network: EEG Based BCI Applications

Written By

Montri Phothisonothai and Katsumi Watanabe

Submitted: 25 June 2012 Published: 05 June 2013

DOI: 10.5772/55801

From the Edited Volume

Brain-Computer Interface Systems - Recent Progress and Future Prospects

Edited by Reza Fazel-Rezai

Chapter metrics overview

3,453 Chapter Downloads

View Full Metrics

1. Introduction

This chapter provides the theoretical principle of fractal dimension analysis applied for a brain-computer interface (BCI). Fractal geometry is a mathematical tool for dealing with complex systems. A method of estimating its dimension has been widely used to describe objects in space, since it has been found useful for the analysis of biological data. Moreover, fractal dimension (FD) is one of most popular fractal features. The term waveform applies to the shape of a wave, usually drawn as instantaneous values of a periodic quantity versus time. Any waveform is an infinite series of points. Aside of classical methods such as moment statistics and regression analysis, properties such as the Kolmogorov-Sinai entropy [1], the apparent entropy [2] and the FD [3] have been proposed to deal with the problem of pattern analysis of waveforms. The FD may convey information on spatial extent and self similarity and self affinity [4]. Unfortunately, although precise methods to determine the FD have already been proposed [5,6], their usefulness is severely limited since they are computer intensive and their evaluation is time consuming. Recently, the FD is relatively intensive to data scaling and shows a strong correlation with the human movement of EEG data [7-9]. The time series with fractal nature are to be describable by the functions of fractional Brownian motion (fBm), for which the FD can easily be set. Waveform FD values indicate the complexity of a pattern or the quantity of information embodied in a waveform pattern in terms of morphology, spectra, and variance. This chapter, we investigate most widely popular six algorithms to be feature patterns for a BCI in which algorithms are in time and frequency domain approaches. However, many methods in evaluating FDs have been developed. As the existing algorithms, they present different characteristics therefore the optimum condition respects to the requirements of BCI application should be seriously considered.

Advertisement

2. Experimental paradigm

By NLAB, under supervision of Prof.Masahiro Nakagawa, Department of Electrical Engieering, Nagaoka University of Technology, 1603-1 Kamitomioka, Nagaoka-shi 940-2188 Japan. E-mail: masanaka@vos.nagaokaut.ac.jp Http://pelican.nagaokaut.ac.jp/

In this study, we tested 7-male and 3-female. Their ages were 21-32 years (mean age, 25.3 years; SD ±3.6), with 180 trials per subject (30 trails per task). At the beginning of each EEG recording session, the subjects were instructed to relax with their eyes open for 30 s. The EEG data from this period was used as the baseline for the tasks. The duration of one trial was 6 s throughout the experiment. Each trail in the experiment was divided into two periods, a relaxing period of 0-3 s and an imaging period of 3-6 s. When the subjects had their eyes open on a blank screen, they would be in a relaxed state of mind (relaxing period). In the imaging period, the subjects imagined with eyes open, as represented by arrows. This study focuses on the motor movement functions in the brain. Therefore, imagination of motors movements are used as the tasks throughout this experiment. To assist the subject imagine the task without difficulty, we provide the acting stimuli in accordance to the given tasks. Explicit indicators are also shown on a monitor for informing subject during the experiment, all selected tasks are listed in Table 1.

Table 1.

Selected tasks in this study

Advertisement

3. Fractal analysis and methods

In this chapter, most widely popular six algorithms in time and frequency domain analysis have been addressed; the box-counting method (BCM), Higuchi‘s method (HM), variance fractal method (VFD), detrended fluctuation analysis (DFA), power spectral density analysis (PSDA), and critical exponent method (CEM). To measure the fractality in short time intervals of time-sequential data from one end of the waveform to the other sequentially, the dynamical changes in the FDs with respect the time series based on the time-dependent fractal dimensions (TDFD) were observed. For the classification process, two new feature parameters on the basis of fractal dimension; Kullback–Leibler (K-L) divergence and the different expected values were presented. In experimental results, DFA was selected to evaluate fraction dimensions on the basis of TDFDs [11, 36]. The reason is that, we found that the DFA provides fast computation time and also presents reasonable values in terms of accuracy and variability when comparison with each other.

3.1. Time series-based analysis

3.1.1. Box-Counting Method (BCM)

One of the most common methods for calculating the FD of a self-similar fractal is the box-counting method (BCM). The definition of BCM is a bounded set in the Euclidian n -space that it composes self-similar property [25], Nr1/rD  by covering a structure with boxes of radii, r, the FD can be determined by

DBCM=limr0log2Nrlog21/rE1

We then repeat this process with several different radii. To implement this algorithm, number of contained box, nr, is computed from the difference between the maximum and minimum amplitudes of the data divide by the changed radius, as

nri=maxxr-minxrr(i),   for r2k|k=1,2,,log2L-1E2
Nr=inr(i)E3

where Nr  is a total number of contained boxes, xr  represents the EEG time series with length L, ri is a radius by changing a step of k within the i-th subdivision window, and integer-part function denoted by . To obtain the FD, least-square linear fitted line corresponds to the slope of the plot log2Nr  versus log21/r is applied.

3.1.2. Variance Fractal Dimension (VFD)

This method is determined by the Hurst exponent, H, whose calculation was divided from the properties of the fBm data. The basic idea of calculation is based on the power law relationship between the variance of the amplitude increments of the input time series, which was produced by a dynamic process over time. The main advantage of VFD was its support of the real-time computation [17]. We also selected the VFD for estimating the FD of EEG data. The amplitude increments of a datum over a time interval t adhere to the following power law relationship Var[x(t2-t1)]t2-t12H , the Hurst exponent can be calculated by using a log-log plot then given by

H=limt0 12log2Varxtlog2t E4

The variance in each window per stage k can be calculated as follows

Varxt=1(Nk-1)j=1Nkxjk2-1Nkj=1Nkxjk2E5
xjk=xjnk-xj-1nk,    for j=1,2,,NkE6

The least-square linear fitted line corresponds to the slope of the plot log2nk and log2Varxk, the Hurst exponent is computed as H=0.5s where s is the obtained slope then the FD can be estimated as

DVFD=2-HE7

The process of calculating the FD essentially involves segmenting the entire input time series data into numerous subsequence (or window). The values k represents the integer range chosen such that each window of size NT  contains a number nk=2k of smaller windows of size Nk=NT/nk.

3.1.3. Higuchi’s Method (HM)

FD is another measure of data complexity, generally evaluated in phase space by means of correlation dimension. Higuchi proposed an algorithm for the estimation of FD directly in time domain without reconstructing the strange attractor [18]. The HM also gives reasonable estimate of the FD in the case of short time segment. The HM algorithm based on the given finite time series y=y1, y2, ,y(N), a new time series, ymk, are constructed by the following equation

ymk=ym, ym+k,ym+2k,,ym+N-mkk for m=1,2,,kE8

where both m and k are integers and they indicate the initial time and the time interval, respectively. The length, Lm(k), of each curves is computed as

Lmk=1ki=1N-mkym+ik-y(m+(i-1)k)N-1N-mkkE9

The length of curve for time interval k, Lm(k) is computed as the average of the m curves. A relationship of this algorithm is Lmkk-DHM therefore we apply the least-squares fitting line of log2L(k) versus log2k, the negative slope of the obtained line is calculated giving the estimate of the FD,  DHM.

3.1.4. Detrended Fluctuation Analysis (DFA)

The idea of DFA was invented originally to investigate the long-range dependence in coding and noncoding DNA nucleotide sequence [19]. Due to simplicity in implementation, the DFA is now becoming the most important method in the field of fractal analysis [20]. Therefore the DFA method was also applied to FD estimation in this study.

Xk=i=1kxi-x-E10

This integrated series is divided into non-overlapping intervals of length n. In each interval, a least squares line is fit to the data (representing the trend in the interval). The series Xk is then locally detrended by subtracting the theoretical values Xnk given by the regression. For a given interval length n, the characteristic size of fluctuation for this integrated and detrended series is calculated by

Fn=1Ni=1NXk-Xnk2E11

This computation is repeated over all possible interval lengths (in practice, the minimum length is around 10, and the maximum is a half-length of input data, giving two adjacent intervals). In this experiment, we use the power of 2 based length for input EEG data therefore in this case we can range n=2k for k=4,5,,log2L-1. A relationship is expected, as F(n)nα where αis expressed as the slope of a double logarithmic plot log2F(n) versus log2n. As PSD, Then α can be converted into the Hurst exponent H=α-1 and the estimated FD according asDDFA=3-α.

3.2. Frequency-based analysis

3.2.1. Power Spectral Density Analysis (PSDA)

This method is widely used for assessing the fractal properties of time series, the method based on frequency domain analysis, and also works on the basis of the 1/f-like power scaling which can be obtained by the fast Fourier transform (FFT) algorithm. It is called a power-law relationship of Mandelbrot and van Ness [25] can be expressed as follows

f(s)1s-βE12

where s is the frequency, f(s)  is the correspondence power at a given frequency, and β is the scaling exponent which is estimated by calculating the negative slope of the linear fit relating logf(s)  to logs. The PSDA was applied to raw EEG data to classify the EEG characteristics. This study used the enhanced version of PSDA that initially proposed by Fougère [26] and modified by Eke [27]. To improve the stability of spectral estimates, raw EEG data of each individual trial were first detrending by subtracted the mean of the data series from each value, and then each value is multiplied by a parabolic window is given as below

ψi=1-2iM+1-12,     for i=1,2,,ME13

Finally the scaling exponent was estimated by the least-square fitting of log-log domain for the frequencies lower than 1/8 of the maximum sampling rate, i.e., less than 64 Hz in this study. The Hurst exponent can be determined as H=β-1/2 where 1<β<3 and the estimated FD can be computed by the following equation

DPSDA=2+1-β2E14

3.2.2. Critical Exponent Method (CEM)

This method was initially proposed by Nakagawa [21]. The CEM has been applied in the physiological data analysis and featuring [22-24]. The PSD of observed fBm data, PH(υ), in the frequency domain is determined as

PHυυ2H+1=υ-βE15

The moment, Iα, and the moment exponent, α, of the PSD is determined as

Iα=1ΩPHυυαdυ,     (-<α<+)E16

We will consider the frequency bands as finite value and substitute Eq. (15) to Eq. (16) thus the equation was given as

Iα1Ωνα-βdν= 1ΩνA-1dν=1AΩA-1=2AexpuA2sinhuA2E17

where A=α-β+1, u=logΩ, and Ω is the upper frequency variable which was normalized to the lower bound of the integration region as 1. Hence taking the logarithm of moment and the third order derivative can be written as

d3logIαdα3=28u3cosech3uA2coshuA2-1A3E18

We then determine the critical value α=αc which satisfies for Eq. (18) is equal to zero. Finally, the FD can be estimated by

DCEM=2-αc2E19

The PSDA and CEM used the FFT algorithm for transforming the EEG time series data into frequency components.

Advertisement

4. Performance assessment

Since high computation time and low variability are the requirement of the BCI system, the optimal algorithm for evaluating fractal dimension should be examined carefully. In this experiment, we selected two signals that fractal dimension value can be easily set for assessing. The usefulness of output results can be helped us know which algorithm is suited for applying it as the feature extractor.

4.1. Assessing with artificially generated signals

  1. Fractional Brownian motion (fBm) signal

    fBm data were simulated by NLAB, Department of Electrical Engieering, Nagaoka University of Technology, 1603-1 Kamitomioka, Nagaoka-shi 940-2188 Japan.

    . General model of the fBm signal can be written as the following fractional differential equation [32]

dH+1/2BH(t)dtH+1/2w(t)E20

The mathematical model for solving the fBm signal in Eq.(20) was proposed by Mandelbrot and Van Ness [25], which is defined by

BHt=1Γ(H+1/2)-0t-sH-1/2--sH-1/2dBs+0tt-sH-1/2dBsE21

where Γ(∙) is the Gamma function Γa=0+xa-1e-xdx and H is the Hurst exponent where 0<H<1. B is a stochastic process. To simulate the fBm signal, we can simply implement the above definition to discrete time series at n-th index by Riemann’s type sums as follows:

BHn=1Γ(h+1/2)k=-b0n-kH-12--kH-12G1(k)+k=0nn-kH-12G2(k)E22

In case H=1/2, Eq.(4.22) can be reduced, we then obtain

B1/2 (n)=k=0nG(k)E23
For n=1,2,,N and b=N3. G is a normal distribution space.
  1. Synthetic (Syn) signal. This signal was produced using the deterministic Weierstrauss cosine function [28]. To simulate the Syn signal, we can simply implement the above definition to discrete time series at n-th index by [33]

Whn= i=0Mγ-iHcos(2πγin)E24

where γ>1, and we set M=26 and γ=5. The fractal dimension of this signal is given by D=2-H.

4.2. Conditions for assessing

A summary of the selected parameter values in this experiment is shown in Table 2. The selection of signal lengths that are powers of two was motivated by the requirements of frequency-based algorithms (PSDA and CEM). Output performances of each algorithm were shown in Fig. 1.

4.3. Results of performance assessing

In order to test the effect of signal lengths on H estimation and each algorithm was applied on the entire series. The estimated fractal dimension values are then averaged at the change of H. Before utilizing those of algorithms, we should regard as the following things:

  1. Computation time. Short computation time is the requirement of the BCI applications. The fBm and Syn signals are performed to complete the entire series on a laptop 1.6 GHz Pentium with memory of 512 MB. Table 3 shows the results of the average computation time.

  2. Accuracy. To compare with the theoretical FD value (true FD value), we used a mean-squared error (MSE) value, which can be defined by

ED=1Li=1LDi-D^i2E25

where D is a theoretical FD value (true value),  D^ is an estimated FD value at ith step of H, and L is a signal length. Results are shown in Tables 4 and 5.

  1. Variability (or robustness). The standard deviation (SD) value is computed to indicate that that variability of algorithms. Results are shown in Tables 6 and 7.

Algorithm Parameter Value
BCM Step size of radius k=1,2,,log2p-1
VFD Step size of window k=2,3,,log2p-1
HM Maximum interval length  kmax=p-3
DFA Step size of interval length k=4,5,,log2p-1
PSDA FFT point; Maximum frequency NFFT=p; fmax=64Hz
CEM FFT point; Step size of moment exponent NFFT=p; α=0.001

Table 2.

Selected parameters in the experiment.

Figure 1.

The results for fractal analysis, performed on a series of test EEG data. (a) BCM algorithm: a log-log plot of a total number of contained boxes versus radius. (b) VFD algorithm: a log-log plot of variance versus window length. (c) HM algorithm: a log-log plot of a curve’s length versus interval length. (d) DFA algorithm: a log-log plot of a total length F(n) versus interval length. (e) PSDA algorithm: a log-log plot of power spectra versus frequency. (f) CEM algorithm: a log-log plot of moment function versus moment exponent value [16].

Signal length tavr [ms]
BCM VFD HM DFA PSDA CEM
L = 27 2.60 0.77 14.90 10.05 2.07 140.40
L = 28 3.50 1.10 148.60 20.20 2.90 227.50
L = 29 6.80 2.80 385.20 40.20 4.90 1,767.80
L = 210 12.60 4.30 697.90 82.50 10.80 3,351.20
L = 211 26.00 7.10 1,400.70 168.80 20.30 6,538.50
L = 212 54.00 12.40 2,490.50 329.70 46.40 12,964.80
L = 213 114.90 61.10 4,160.70 895.10 137.80 25,318.10
L = 214 265.50 164.70 14,630.80 2,971.50 478.70 50,157.60
L = 215 676.10 607.90 47,082.50 11,210.20 1,882.30 101,809.70
Average 129.11 95.79 7,890.20 1,747.53 287.35 22,475.06

Table 3.

Average computation time in millisecond (best value is marked in bold).

Signal length ED(fBm signal)
BCM VFD HM DFA PSDA CEM
L = 27 0.091 0.018 0.003 0.004 0.027 0.039
L = 28 0.069 0.013 0.031 0.002 0.031 0.047
L = 29 0.054 0.008 0.033 0.001 0.096 0.040
L = 210 0.044 0.007 0.001 0.001 0.099 0.045
L = 211 0.038 0.005 0.019 0.002 0.102 0.042
L = 212 0.032 0.003 0.015 0.001 0.102 0.041
L = 213 0.023 0.001 0.001 0.004 0.101 0.046
L = 214 0.022 0.004 0.001 0.001 0.105 0.046
L = 215 0.019 0.007 0.002 0.001 0.103 0.053
Average 0.044 0.007 0.012 0.002 0.085 0.044

Table 4.

MSE value performed with the fBm signal (best value is marked in bold).

Signal length SD(fBm signal)
BCM VFD HM DFA PSDA CEM
L = 27 0.083 0.009 0.054 0.003 0.059 0.038
L = 28 0.067 0.023 0.004 0.002 0.048 0.048
L = 29 0.044 0.002 0.003 0.009 0.173 0.026
L = 210 0.030 0.003 0.002 < 0.001 0.107 0.033
L = 211 0.020 0.057 0.003 < 0.001 0.098 0.041
L = 212 0.018 0.083 0.002 0.005 0.089 0.036
L = 213 0.026 0.105 0.004 0.022 0.027 0.044
L = 214 0.037 0.182 0.004 0.042 0.051 0.035
L = 215 0.049 0.183 0.003 0.065 0.029 0.029
Average 0.034 0.072 0.009 0.016 0.076 0.037

Table 5.

MSE value performed with the Syn signal (best value is marked in bold).

Signal length SD(fBm signal)
BCM VFD HM DFA PSDA CEM
L = 27 0.056 0.201 0.109 0.112 0.248 0.069
L = 28 0.047 0.150 0.089 0.095 0.154 0.056
L = 29 0.061 0.116 0.040 0.069 0.090 0.047
L = 210 0.034 0.088 0.032 0.049 0.056 0.042
L = 211 0.032 0.078 0.019 0.039 0.036 0.037
L = 212 0.023 0.079 0.016 0.036 0.028 0.031
L = 213 0.023 0.062 0.014 0.028 0.019 0.027
L = 214 0.018 0.054 0.013 0.024 0.014 0.026
L = 215 0.015 0.021 0.011 0.022 0.013 0.014
Average 0.034 0.094 0.038 0.053 0.073 0.039

Table 6.

Deviation performed with the fBm signal (best value is marked in bold).

Signal length SD(Syn signal)
BCM VFD HM DFA PSDA CEM
L = 27 0.039 0.151 0.056 0.085 0.278 0.036
L = 28 0.031 0.116 0.026 0.044 0.236 0.025
L = 29 0.023 0.074 0.022 0.043 0.141 0.046
L = 210 0.017 0.068 0.021 0.016 0.148 0.038
L = 211 0.018 0.070 0.019 0.011 0.165 0.030
L = 212 0.019 0.091 0.018 0.010 0.175 0.022
L = 213 0.022 0.063 0.017 0.009 0.127 0.020
L = 214 0.019 0.049 0.015 0.014 0.121 0.017
L = 215 0.008 0.011 0.009 0.012 0.119 0.012
Average 0.022 0.077 0.022 0.027 0.168 0.027

Table 7.

Deviation performed with the Syn signal (best value is marked in bold).

4.4. Time-Dependent Fractal Dimensions (TDFD)

One of common measure of irregularity in fluctuations as time-sequential data may be the fractal dimension. Fractal dimensions estimate the degree of freedom of fluctuations in a signal. To measure the fractality in short time intervals of time-sequential data from one end of the waveform to the other sequentially, we may observe the dynamical changes in the fractal dimensions with respect the time series. These fractal dimensions, namely, the time-dependent fractal dimensions (TDFD) [22].

In this study, EEG signal was sampled at 512 Hz (or L=3,072 points) whose duration is 6 seconds. We set a window function is a rectangular type, window size = 512 points (1 s), moving window with intervals = 10 points (19.5 ms) because the temporal resolution of EEG is millisecond or even better [29]. The number of obtained points can compute from NFD=(L-Lw)/t +1 where L is a signal length, Lw is a window length (LwL), and t is an interval. Thus, we then obtained 257 points of fractal dimension. The time series-based algorithms of fractal dimension present more effective than frequency-based algorithms [16].

Figure 2.

Computation process of TDFDs.

4.5. Evaluation of features by Kullback-Leibler divergence

In probability theory and information theory, the Kullback–Leibler divergence [34] (or called K-L divergence) is a measure of the difference between two probability distributions: from an actual probability distribution P to an arbitrary probability distribution Q.

DKLPQ=-+pxlogp(x)q(x)dxE26

For probability distributions P and Q of a discrete random variable is defined to be

DKLPQ=iPilogPiQiE27

Important properties of the K-L divergence are

  1. The K–L divergence is always positive DKLPQ0,

  2. DKLPQ=0  if and only if P=Q,

  3. The K-L divergence is not symmetric DKLPQDKLQP.

Figure 3.

Feature extraction concepts on the basis of the K-L divergence.

The different expected values between imaging and relaxing periods are also proposed to use as the featuring parameter together with the K-L divergence. In general, if X is random variable defined on a probability space, then the different expected value of two random variables can be defined by

E{X}PQ=ipixi-jqixjE28

In this study, we applied the K-L divergence and different expected values for finding the features between relaxing and imaging periods. According to Eq.(28), Pi is regarded as relaxing period, and Qi is regarded as imaging period. The probability distributions can be approximated by normalizing histogram of the data. The output patterns of K-L divergences versus the different expected value are also shown in Fig. 4. These obtained results can be classified by the neural network. More explanations of this classification process are in chapter 5.

Figure 4.

Left-hand movement imagination (LH-MI).

Advertisement

5. Neural network classifier

This study is concerned with the most common class of neural networks (NNs) called backpropagation algorithm. Backpropagation is an abbreviation for the backwards propagation of error. The learning process is processed based on the Levenberg–Marquardt (LM) backpropagation method [30] the Hessian matrix can be approximated as the following equation

H2JTJE29

and the gradient can be determined as

J=2JTeE30

where J is the Jacobian matrix that contains first derivatives of the network error with respect to weights and biases, and  e is a vector of network errors.

J(w)=e1(w)w1e1(w)wneN(w)w1eN(w)wnE31

The LM algorithm can use for updating the vector of weights, w, by

wn+1=wn-JTJ+μI-1JTeE32
w=-JTJ+μI-1JTeE33

where I is an identity matrix, μ is a learning rate, and n is an iteration step. This algorithm is particularly well suited for training NN with MSE. Moreover, the LM algorithm has been shown to be much faster than backpropagation in a variety of applications [31].

Figure 5.

Classification diagram of the proposed method.

Task Desired vector
LH-MI 10   00T
RH-MI 01   00T
FT-MI 00   10T
TG-MI 00   01T

Table 8.

Desired vectors from the proposed network

5.1. Cross-validation procedure

Then, the featured parameters typically perform in terms of fractal properties. The NN was selected as classifier and it includes adaptive self-learning that was learnt to produce the correct classifications based on a set of training examples via cross validation process. A tenfold cross validation method was used to increase the reliability of the results [13]. For each experiment, we divided all the data into three sets, namely, the training set (50% of the entire data), cross validation set (20% of the entire data), and testing set (30% of the entire data). According to the tasks, a set of desired values for network training are shown in Table 8. The MSE rate is succeeded versus the iteration step is shown in Fig. 6. In Table 9, we compared the proposed method with the conventional feature, namely, band power estimation (BPE) [14]. The training process makes the network to define its decision boundaries. Figures 7 to 8 show a tight decision boundary after training the network.

Advertisement

6. Discussions

This chapter discusses the major contributions of this study, 12 electrode channels were used to measure EEG signal over the sensorimotor area of the cortex. The accuracy rates of motor imagery tasks are satisfactory. The error in this classification is due to the ambiguous data between the relaxing and imaging periods, particularly when the subjects have imagined the movements. However, in this experiment, throughout the period of recoding raw EEG signals without training and also without rejecting a bad trial. On contrary, they can be different for different individuals. This makes the EEG signal depend on the measured subject. Even if different persons perform the same tasks and the measurement was done under identical conditions the resulting signals could differ [35].

Figure 6.

Log-log plot of learning curve.

Method Accuracy [%]
LH-MI RH-MI FT-MI TG-MI
BPE 77.6 79.5 76.4 77.3
Proposed 82.5 84.3 83.6 82.6

Table 9.

A comparison result of average accuracy rates.

Figure 7.

Decision boundary of the trained network.

Figure 8.

Corresponding 3-D surface of decision boundaries for each task.

Figure 9.

Updated weight values of the NN after learning; (Top panel) Proposed method, (Bottom panel) BPE Method.

The proposed method provides 2880 features for each subject, whereas BPE method provides 21600 features. The current features obtained from the proposed method present an easy practicable and reliable method for training the classification algorithms. Nevertheless, the number of features from BPE method can be reduced by changing the number of electrodes. We used the NN to classify the features processed by BPE method. After learning the NN, learning curves and decision boundaries of BPE method is shown in Fig. 9. As the results, output decision functions of classifying features based on BPE method does not perform well and cannot made clearly separation boundaries.

The six algorithms present specific properties in terms of capability and variability. Some algorithms appeared inapplicable for the EEG time series data in this study, such as PSDA and CEM algorithms, for example. The main reason was that two algorithms based on frequency analysis since estimating the Hurst exponent can be utilized by means of the FFT method where the FFT can be substantially applicable in the long-time series data, and it easily declines when the EEG data has been analyzed in a short-time series. However, we can enhance the variability and avoid such drawbacks by increasing the sampling rate at the same period of experiment. In terms of variability, the HM algorithm presents the lowest average values of SD. The DFA provides the fast computation time and also presents reasonable values in terms of accuracy and variability when comparison with each other. Although, VDF gives extremely fast computation time but the VFD itself has the drawbacks in terms of accuracy and robustness. The VFD algorithm appropriates for real-time application, since its consumed computation time was less than 10 ms. Moreover, as proposed algorithms we suggest that the most appropriate algorithm depends on an application purpose; for instance, the HM algorithm was suited for evaluating the FD of EEG data with high precision, whereas HM algorithm extremely consumed the processing time in comparison with the other algorithms. To show the performance of all algorithms, we then compared all estimated FDs with the theoretical value especially computed by short interval, i.e., 27, 28 and 29 points; these results were shown in Figs. 10 to 15.

Figure 10.

Fractal dimension evaluated by BCM. (Right) fBm signal; (left) Syn signal.

Figure 11.

Fractal dimension evaluated by VFD. (Right) fBm signal; (left) Syn signal.

Figure 12.

Fractal dimension evaluated by HM. (Right) fBm signal; (left) Syn signal.

Figure 13.

Fractal dimension evaluated by DFA. (Right) fBm signal; (left) Syn signal.

Figure 14.

Fractal dimension evaluated by PSDA. (Right) fBm signal; (left) Syn signal.

Figure 15.

Fractal dimension evaluated by CEM. (Right) fBm signal; (left) Syn signal.

Advertisement

7. Conclusion

By selecting the best performance of fractal algorithm, we can obviously improve the rate of classification and can develop the novel methods in terms of fractal properties. Fractal features and the experimental framework can be applied not only binary-command BCIs, but also multi-command BCIs. The measurement of FD from EEG data can be capable not only in particular works, but also in publicity data. The FD characterizes the self-affine property of EEG data and has a direct relation to the different tasks.

References

  1. 1. GrassbergerPProcacciaIEstimation of the Kolmogorov Entropy from a Chaotic Signal. Phys. Rev. A 19832825912593
  2. 2. PincusS. MGladstoneI. MEhrenkranzR. AA Regularity Statistics for Medical Data Analysis. J. Clin. Monit. 19917335345
  3. 3. KatzM. JFractals and the Analysis of Waveforms. Comput. Biol. Med. 1998
  4. 4. BarnsleyM. FFractals Everywhere, Ch. II and III, Academic Press Professional; 1993
  5. 5. GrassbergerPProcacciaIMeasuring Strangeness of Strange Attractors. Phys. Lett. D 19831486368
  6. 6. BadiiRPolitiAStatistical Description of Chaotic Attractors: The Dimension Function. J. Stat. Phys. 198540725750
  7. 7. BoostaniRMoradiM. HA New Approach in the BCI Research based on Fractal Dimension as Feature and Adaboost as Classifier. J. Neural Eng. 2004l.1212217
  8. 8. PopivanovDJivkovaSStomonyakovVNicolovaGEffect of Independence Component Analysis on Multifractality of EEG during Visual-Motor Task. Sig. Pro. 200585211223
  9. 9. LutzenbergerWElbertTBirbaumerNRayW. JSchuppHThe Scalp Distribution of the Fractal Dimension of the EEG and its Variation with Mental Tasks. Brain Topo. 199252734
  10. 10. Preil HLutzenbergerWPulvermüllerFBirbaumerNFractal Dimensions of Short EEG Time Series in Humans. Neurosci. Lett. 19972257780
  11. 11. PhothisonothaiMNakagawaMEEG-based Classification of Motor Imagery Tasks using Fractal Dimension and Neural Network for Brain-Computer Interface. IEICE Trans. Inf. Syst. 2008E91D: 44-53.
  12. 12. PhothisonothaiMNakagawaMEEG Signal Classification Method based on Fractal Features and Neural Network. The 30th IEEE EMBC 200838803883
  13. 13. SetionoRFeedforward Neural Network Construction using Cross Validation. Neural Comput. 20051327832786
  14. 14. PalaniappanRUtilizing Gamma Band to Improve Mental Task based Brain-Computer Interface Design. IEEE Trans. Neural Net. Rehabil. Eng. 200614299303
  15. 15. MandelbrotB. BThe Fractal Geometry of Nature. W.H. Freeman, New York; 1983
  16. 16. PhothisonothaiMNakagawaMFractal-based EEG Data Analysis of Body Parts Movement Imagery Tasks. J Physiol Sci. 2007574217226
  17. 17. KinsnerWBatch and Real-time Computation of a Fractal Dimension based on Variance of a Time Series. Univ. Manitoba Canada. Tech. Report 1994del946
  18. 18. HiguchiTApproach to an Irregular Time Series on the Basis of the Fractal Theory. Physica D 19883127783
  19. 19. PengC. KMietusJHausdorffJ. MHavlinSStanleyH. EGoldbergerA. LLong Range Anticorrelations and Non-Gaussian Behavior of the Heartbeat. Phys. Rev. Lett. 19937013431346
  20. 20. PengC. KHavlinSStanleyH. EGoldbergerA. LQuantification of Scaling Exponents and Crossover Phenomena in Nonstationary Heartbeat Time Series. Chaos 199558287
  21. 21. NakagawaMA Critical Exponent Method to Evaluate Fractal Dimension of Self-affine Data. J. Phys. Soc. Jpn. 19936242334239
  22. 22. SabanalSNakagawaMThe Fractal Properties of Vocal Sound and their Application in the Speech Recognition Model. Chaos Soli. Frac. 1996718251843
  23. 23. PetryAAugustoDBaroneCSpeaker Identification using Nonlinear Dynamical Features. Chaos Soli Frac. 200213221231
  24. 24. PhothisonothaiMNakagawaMEEG-based Fractal Analysis of Different Motor Imagery Tasks using Critical Exponent Method. Int. J. Bio. Life Sci. 200613175180
  25. 25. MandelbrotB. BVan NessJ. WFractional Brownian Motions, Fractional Noises and Applications. SIAM Rev. 196810422437
  26. 26. FougèreP. FOn the Accuracy of Spectrum Analysis of Red Noise Processes using Maximum Entropy and Periodogram Methods: Simulation Studies and Application to Geographical Data. J. Geograp. Res. 19859543554366
  27. 27. EkeAHermannPKocsisLKozakL. RFractal Characterization of Complexity in Temporal Physiological Signals. Physio. Meas. 2002R138
  28. 28. TricotCCurves and Fractal Dimension. New York: Springer-Verlag;1995
  29. 29. KoenigTStuderDHublDMelieLStrikW. KBrain connectivity at different time-scales measured with EEG. Philos. Trans R. Soc. Lond. B Biol. Sci. 200536010151024
  30. 30. MarquardtD. WAn Algorithm for Least-Squares Estimation of Non-linear Parameters. J. Soc. Indust. Appl. Math. 196311431441
  31. 31. PrincipeJ. CEulianoN. RLefebvreW. CNeural and Adaptive Systems: Fundamentals through Simulations. John Wiley & Sons; 2000
  32. 32. NakagawaMChaos and Fractals in Engineering. World Sciencetific; 1999
  33. 33. EstellerRVachtsevanosGEchauzJLittBA Comparison of Waveform Fractal Dimension Algorithms. IEEE Trans. Circuits Syst. 200148177183
  34. 34. KullbackSLeiblerR. AOn Information and Sufficiency. Ann. Math. Statist. 1951227986
  35. 35. CurranEStokesMLearning to Control Brain Activity: A Review of the Production and Control of EEG Components for Driving Brain-Computer Interface (BCI) Systems. Brain Cogn. 200351326336
  36. 36. Phothisonothai, M, & Nakagawa, M. A Classification Method of Different Motor Imagery Tasks Based on Fractal Features for Brain-Machine Interface. J. Intre. Neurosci. (2009)., 8, 95-122.

Notes

  • By NLAB, under supervision of Prof.Masahiro Nakagawa, Department of Electrical Engieering, Nagaoka University of Technology, 1603-1 Kamitomioka, Nagaoka-shi 940-2188 Japan. E-mail: masanaka@vos.nagaokaut.ac.jp Http://pelican.nagaokaut.ac.jp/
  • fBm data were simulated by NLAB, Department of Electrical Engieering, Nagaoka University of Technology, 1603-1 Kamitomioka, Nagaoka-shi 940-2188 Japan.

Written By

Montri Phothisonothai and Katsumi Watanabe

Submitted: 25 June 2012 Published: 05 June 2013