Open access peer-reviewed chapter - ONLINE FIRST

ANFIS: Establishing and Applying to Managing Online Damage

By Sy Dzung Nguyen

Submitted: October 5th 2018Reviewed: December 10th 2018Published: March 11th 2019

DOI: 10.5772/intechopen.83453

Downloaded: 129

Abstract

Fuzzy logic (FL) and artificial neural networks (ANNs) own individual advantages and disadvantages. Adaptive neuro-fuzzy inference system (ANFIS), a fuzzy system deployed on the structure of ANN, by which FL and ANN can interact to not only overcome their limitations but also promote the ability of each model has been considered as a reasonable option in the real fields. With the vital strong points, ANFIS has been employed well in many technology applications related to filtering, identifying, predicting, and controlling noise. This chapter, however, focuses mainly on building ANFIS and its application to identifying the online bearing fault. First, a traditional structure of ANFIS as a data-driven model is shown. Then, a recurrent mechanism depicting the relation between the processes of filtering impulse noise (IN) and establishing ANFIS from a noisy measuring database is presented. Finally, one of the typical applications of ANFIS related to online managing bearing fault is shown.

Keywords

  • fuzzy logic
  • artificial neural networks
  • adaptive neuro-fuzzy inference system

1. Introduction

As well known, the mathematical tools FL and ANN possess both the advantages and disadvantages as their specific characteristics. The hybrid structure ANFIS, where ANN and FL can interact to not only overcome partly the limitations of each model but also uphold their strong points [1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25], is, hence, considered as a reasonable option in many technology applications such as identifying [1, 2, 4, 6, 12], predicting [9, 11, 17, 25], controlling [3, 5, 7, 18, 19, 20, 21, 22, 23, 24, 25, 26], and filtering noise [14, 15, 16, 27, 28, 29].

To build an ANFIS from a given database, firstly, an initial data space (IDS) expressing the mapping f:XYmust be created. A cluster data space (CDS) is then built from the IDS to form the ANFIS via a training algorithm. Being viewed as a popular technique for unsupervised pattern recognition, clustering is an effective tool for analyzing and exploring data structures to build CDSs [30, 31, 32, 33, 34, 35, 36, 37]. Reality shows that the accuracy and training time of the ANFIS depend deeply on the features of both the IDS and CDS [2, 3, 4, 5, 6]. In the process of building ANFIS, the two issues as follows should be considered: (1) What is the essence of the interactive relation between ANFIS’s convergence capability and CDS’ attributes? (2) How to exploit this essence for increasing ANFIS’s ability to converge to the desired accuracy with the improved calculating cost?

Many different clustering approaches have been discovered [2, 3, 4, 10, 30, 31, 32, 33, 34, 37]. Separating data in X and in Y distinctly with a mutual result reference, step by step, was described in [10]. The method, however, could not solve appropriately the above issues. Besides, the difficulty in deploying fuzzy clustering strategies along with the high calculating cost was their disadvantage. Generally, a hard relation could not reflect fully database attributes [31, 34]. The well-known method of fuzzy C-means clustering was seen as a better option in this case. It, however, was not effective enough for the “non-spherical” general datasets [30, 37]. Therefore, the idea of fuzzy clustering in a kernel feature space was then developed to deal with these cases [30, 31, 32, 33, 34, 37]. A weighted kernel-clustering algorithm could be referred to [30], or a method of weighted kernel fuzzy C-means clustering based on adaptive distances was detailed in [31]. In spite of owning considerable advantages, the identification and prediction accuracy of the ANFIS based on the CDS coming from [30, 31] are sensitive to attributes of the CDS due to the negative influence of noise.

Reality has shown that noise status including IN always exists in the measured IDSs [2, 4, 9, 16], which degrades violently the accuracy of ANFISs deriving from them. There are many reasons resulting in this, such as the lack of precision of the measurement devices, tools, measurement methods, or the negative impact of the surrounding environment. In [7], an ANFIS took part in the system in the form of an inverse MR damper (MRD) model to specify the time-verifying desired control current. To maintain the accuracy of the inverse MRD model, the ANFIS was retrained after each certain period due to the dynamic response of the MRD depending quite deeply on temperature. Another more active approach is filtering noise or preprocessing data [7, 9, 11, 17, 21, 38, 39, 40]. In [11, 17], where ANFISs were employed to predict the health of mechanical systems, vibration signal was always measured and filtered to update the ANFISs. Related to the preprocessing data to set up ANFIS, it can observe that to maintain the stability of the above online ANFIS-based applications, reducing time delay is really meaningful. One of the becoming solutions for this can be referred to [16] where filtering IN and building ANFIS were carried out synchronously via a recurrent mechanism. A recurrent strategy for forming ANFIS was carried out, in which the capability to converge to a desired accuracy of the ANFIS training process could be estimated and directed online. As a solution, increasing the quality of both the IDS and CDS was paid attention. Building an ANFIS via a filtered database and exploiting the ANFIS as an updated filter to refilter the database were depicted via an online and recurrent mechanism. The process was upholden until the ANFIS-based database approximation convergent to the desired accuracy or a stop condition appears.

Inspired by the ANFIS’s capability, in order to provide the readers with the theoretical basis and application direction of the model, this chapter presents the formulation of ANFIS and one of its typical applications. The rest of the chapter is organized as follows. Section 2 shows a structure of ANFIS as a data-driven model deriving from fuzzy logic and artificial neural networks. Setting up the CDS consisting of the input data clusters, output data clusters, and the CDS-based ANFIS as a jointed structure is all detailed. Deriving from this relation, a theoretical basis for building ANFIS from noisy measuring datasets is presented in Section 3. An online and recurrent mechanism for filtering noise and building ANFIS synchronously is clarified via algorithms for filtering IN and establishing ANFIS. A typical application of ANFIS related to online managing bearing fault status is shown in Section 4. Finally, some general aspects are mentioned in the last section.

2. Structure of ANFIS

Let’s consider a given IDS having P input–output data samples x¯iyi, x¯i=xi1xinn, yi1, and i=1P.With a data normalization solution and a used certain clustering algorithm, a CDS is then created. The kth cluster, signed Γk,k=1C,consists of one input cluster and one output cluster signed ΓkAand ΓkB, respectively. The CDS can be seen as a framework for establishing ANFIS. This section presents how to build the CDS as well as the CDS-based ANFIS structure.

2.1 Some related notions

Some notions shown in [16] are used in this chapter as follows.

Definition 1. Normalizing a given IDS to set up a normalized initial data space signed IDS¯is performed as follows:

x˜ij=xij/maxkxkj,i,k=1P,j=1n.E1

By this way, the ith data sample (also signed (x¯i,yi)) in the IDS¯is constituted as follows:

x¯i=x˜i1x˜inTyii=1P.E2

Definition 2. The root-mean-square error (RMSE) in Eq. (3) is used to evaluate accuracy rate of ANFIS. The required RMSE value is signed E. The absolute error, ε¯i,i=1P,between the data output yi=fx¯iand the corresponding ANFIS-based output ŷix¯iis defined in Eq. (4). The desired value of ε¯iis signed ε¯:

RMSE=P1i=1P ŷix¯ifx¯i2,E3
ε¯i=ŷix¯ifx¯i,i=1P.E4

Definition 3. Let’s consider x¯iIDS¯in which IDS¯depicts an unknown mapping f:XY. The ANFIS-based approximation of f:XYis called to be continuous at x¯pypIDS¯if.

ŷix¯iyp±ε¯whenx¯ix¯p.E5

Definition 4. Let’s consider an ANFIS-based approximation of a mapping expressed by an IDS¯. The ANFIS is said to be a uniform approximation with a required error ε¯if at ∀ x¯iX, by choosing any small constant εε¯, corresponding data sample x¯jIDS¯always exists such that.

x¯hIDS¯;ifx¯hx¯ix¯jx¯i,thenŷhx¯hfx¯iŷjx¯jfx¯i=ε.E6

Definition 5. Data cluster Γkand data sample x¯pypΓkin a CDS derived from an IDS¯are depicted in Figure 1. Let sign Γk\pto be a subset consisting of the data samples belonging to Γkexcept x¯pyp. The subset contains Qkpdata samples. It is assumed that all of data samples in ΓkAare distributed closely, while in ΓkB, most of them are located closely, except yp; it is far from the other and distributes at one side of ΓkB. This status is described in Eq. (7):

yp ≫maxyiΓkB\pyioryp ≪minyiΓkB\pyi.E7

Figure 1.

Two typical distribution types in data cluster Γ k : Impulse noise point IN x ¯ p y p ∈ Γ k causing the distribution at one side, the right side (a), and the left side (b).

and satisfies Eqs. (8) and (9):

dk1=ypmaxyiΓkB\pyi>PE2Qkpε¯20.5ifyp>maxyiΓkB\pyi.E8
dk2=minyiΓkB\pyiyp>PE2Qkpε¯20.5ifyp<minyiΓkB\pyi.E9

In this case, x¯pypis called a critical data sample in the CDS.

2.2 Setting up the input data clusters

Let’s consider the normalized initial data space IDS¯(see Def. 1). Many well-known clustering methods can be used to build a CDS from the IDS¯. Here, the CDS is built by using the clustering algorithm KFCM-K (kernel fuzzy C-means clustering with kernelization of the metric) presented in [31]. By this way, distribution of data samples in the CDS is established. The membership degree of the jth data sample belonging to the ith cluster is denoted by μij01i,jand j=1P,i=1C. Cluster centroids x¯10,,x¯C0in the CDS are specified such that the following objective function is minimized:

JKFCMUx¯0=i=1Cj=1Pμijmϕx¯jϕx¯i02E10

subjected to i=1C μij=1jand μij01i,j. In Eq. (10), x¯i0=x0i1x0innis the ith cluster center; ϕx¯jϕx¯i02denotes the squared distance between x¯jand x¯i0in the kernel space; ϕ.is the kernel function; U=UμijC×Pis the distribution matrix; and m>1is the fuzzy factor.

The objective function can be rewritten via Gaussian kernel function as follows:

JKFCMUx¯0=2i=1Cj=1Pμijm1expx¯jx¯i02/σ2.E11

Deriving JKFCMUx¯0in Eq. (11) with respect to x¯i0, at the optimal centers, the following must be obtained:

x¯i0JKFCMUx¯i0=4σ2j=1Pμijmx¯jx¯i0expx¯jx¯i02/σ2=0E12

From Eqs. (11) to (12) and the use of Lagrange multipliers with μij01i,jand i=1Cμij=1j, the following update laws are obtained:

x¯i0=j=1P μijmx¯jKx¯jx¯i0j=1P μijmKx¯jx¯i0,i=1C.E13
μij=h=1C1Kx¯jx¯i01Kx¯jx¯h01/m11ifx¯jx¯i01andμikkj=0ifx¯j=x¯i0i=1C;j=1P.E14

By using index tsas in Eq. (15), tsto be the required value of tsand r to denote the rth loop, the clustering phase is accomplished until tsts:

ts=JKFCMrJKFCMr1/JKFCMr1E15

Specification of the optimal centers and their relationship values as abovementioned is detailed in Appendix A of [12].

2.3 Setting up the output data clusters

The result of the clustering process in the input data space is an input cluster centroid vector x¯10x¯C0of corresponding data clusters, respectively, signed as Γ1,,ΓC.Let A1,,AC,respectively, be input fuzzy sets established via x¯10,,x¯C0[12, 16]. The membership value of x˜ilbelonging to Akis inferred from Eq. (14):

μ¯kix˜il=h=1C1Kx˜ilxkl01Kx˜ilxhl01/m11k=1Ci=1Pl=1n..E16

With following the product law, membership value of x¯qbelonging to Aiis

μkqx¯q=l=1nμ¯kqx˜ql,k=1C,q=1P,E17

and its normalized membership value is as follows:

Nkx¯q=μkqx¯q/h=1Cμhqx¯q,q=1P,k=1C.E18

The membership of a data sample in each cluster determined based on Eqs. (16)(18) is then used to specify the hard distribution status of the data samples in each cluster. It is then used to specify the index vector aof hyperplanes (or the output data clusters) wk.and k=1C. The ith data sample is hardly distributed into the kth data cluster if

Nkx¯i=maxh=1CNhx¯i,i=1P,k=1C.E19

Deriving from the tkdata samples hardly distributed in the kth data cluster, by using the least mean squares method, vector a=a0a1anT=a0a¯Tof wk.is specified which is the solution of Eq. (20):

ani=1tkx˜in+an1i=1tkx˜in1++a1i=1tkx˜i1+a0tk=i=1tkyiani=1tkx˜inx˜i1+an1i=1tkx˜in1x˜i1++a1i=1tkx˜i12+a0i=1tkx˜i1=i=1tkyix˜i1ani=1tkx˜in2+an1i=1tkx˜in1x˜in++a1i=1tkx˜i1x˜in+a0i=1tkx˜in=i=1tkyix˜inE20

Finally, the value of hyperplane wkcorresponding to x¯iis calculated in Eq. (21):

wkx¯i=a0+a¯Tx¯iE21

2.4 Structure of ANFIS

As mentioned in Eq. (1), the ANFIS for approximating the mapping  f:XYis derived from M fuzzy laws in Eq. (22):

Ri:IFx˜q1isA1ix˜q1,AND,ANDx˜qnisAnix˜qnTHENyqiisBix¯qi=1MMC,E22

where Alix˜ql,i=1M,l=1n,is the membership value of x˜qlbelonging to input fuzzy set Ai,meaning Alix¯qμ¯iqx˜qlin Eq. (16); Bix¯qis the corresponding output fuzzy set of data sample x¯q, x¯q=x˜q1x˜qnT,q=1P.

In the fuzzification phase, membership value of x¯qbelonging to input fuzzy set Aisigned Aix¯qμiqx¯qis specified by Eq. (17). For the defuzzification, if the center-average method is used, the output of the qth data sample is expressed via the membership values in the input fuzzy space of x¯qas follows:

ŷqx¯q=i=1Myqiμiqx¯q/i=1Mμiqx¯qE23

where yqi=wx¯qiis the value of hyperplane wicorresponding to data sample x¯qcalculated in Eq. (21).

Finally, all the above-mentioned contents can be depicted via the ANFIS with five layers signed D, CL, Π, N, and S in Figure 2. Layer D (data) has n input nodes corresponding to n elements of data vector x¯i=xi1xinT, i = 1…P, while its outputs are the corresponding normalized values using Eq. (1). Layer CL (clustering) expresses the clustering process. The result of this process is C clusters with C corresponding cluster centroids x¯10,,x¯C0 ,to which C fuzzy sets, A1,…, AC, are given. The output of this layer is the membership value of x¯icalculated for each dimension x˜i1x˜invia Eq. (16). Layer Π(product layer) specifies membership values based on Eq. (17). Layer N (normalization) estimates the normalized membership value of a data sample belonging to each fuzzy set upon Eq. (18). Layer S (specifying) is used to estimate the output of the ANFIS based on any well-known method. In case of using the center-average defuzzification, it is calculated by Eq. (23), while it is specified by Eq. (24) if the “the winner takes all” law is employed:

ŷi=wkx¯i,i=1P,E24

where wkx¯iis the value of the kth hyperplane corresponding to input data sample x¯i(21); k is the index of the data cluster where data sample x¯igets the maximum membership specified via N.x¯ias in Eq. (25):

Nkx¯i=maxh=1CNhx¯iE25

Figure 2.

Structure of the ANFIS.

3. Building ANFIS from a noise measuring database

This section presents the recurrent mechanism together with the related algorithms consisting of the one for ANFIS-based noise filtering and the one for building ANFIS showed in [16].

3.1 Convergence condition of the ANFIS-based approximation

Deriving from a given IDS having P input–output data samples x¯iyi, x¯i=xi1xinnand yi1, i=1P, with a data normalization solution as in Def. 1, the IDS¯is built, to which a CDS is created as depicted in Section 2. It should be noted that IN is often considered as disturbances distributed uniformly in a signal source which impacts negatively on the created CDS. In general, IN causes raising the number of critical data samples in the CDS. The negative impact of IN on the convergent ability of training ANFIS is formulated via Theorem 1 as follows.

Theorem 1 [16]: Let’s consider a given IDS¯deriving from an IDS and an ANFIS uniformly approximating an unknown mapping f:XYexpressed by the IDS. The ANFIS is built via a CDS built from the IDS¯. Assume that X is compact. The necessary condition for the approximation convergent to a desired error Eis that in the CDS there is not any critical data sample.

Proof: Let’s consider cluster Γkbelonging to the CDS. Assume that x¯pypΓkis a critical data sample (see Def. 5); it has to be proven that the ANFIS will not converge to E.

It can infer from Eq. (3) that

RMSEP0.5ŷpx¯pfx¯p2+i=1Qkpŷix¯ifx¯i20.5E26

Because the ANFIS is a uniform approximation of f:XYand X is the compact set, it can infer that the ANFIS is continuous in Γk\p, so Eq. (27) can be inferred from Eq. (26):

RMSEP0.5ŷpx¯pfx¯p2+Qkpε¯20.5E27

It should be noted that the ANFIS is a uniform approximation of the f:XYin Γk\p, x¯pypΓkis a critical data sample, and samples in ΓkAare distributed closely. As a result, Eq. (28) can be inferred:

ŷpx¯pminyiΓkB\pyimaxyiΓkB\pyiE28

Due to yp>maxyiΓkB\pyi(see Def. 5), the following can be obtained from Eqs. (27) to (28):

RMSEP0.5maxyiΓkB\pyiyp2+Qkpε¯20.5E29

From Eqs. (8) and (29), it can conclude that RMSE>E. Similarly, due to yp<minyiΓkB\pyi(see Def. 5), from Eqs. (27) to (28), the following can be also inferred:

RMSEP0.5minyiΓkB\pyiyp2+Qkpε¯20.5E30

From Eq. (9) to (30), RMSE>Ecan be implied.

Finally, it can conclude that if existing at least a critical data sample in the CDS, the ANFIS could not converge to the required error [E]. □.

3.2 Algorithm for filtering IN

An essential advantage of the clustering algorithms presented in [30, 31] is the convergent rate. However, the quality of the ANFIS based on the CDS deriving from them is sensitive to the IDS attributes. It can be observed that the main reason of this status via Theorem 1 is the appearance of critical data samples. Besides, regarding the preprocessing IDS shown in [9], in spite of the positive filtering effectiveness, the calculating cost of the method is quite high. A becoming solution for the above issues can be referred in [16] where the recurrent mechanism illustrated in Figure 3 was employed. The recurrent mechanism has two phases being performed synchronously: filtering IN in the database and building ANFIS based on the filtered database.

Figure 3.

Flowchart of the FIN-ANFIS consisting of the three main phases, the clustering, establishing and estimating ANFIS, and filtering IN, which are performed simultaneously.

Firstly, an adaptive online impulse noise filter (AOINF) is proposed. The recurrent mechanism is then depicted via the algorithm named FIN-ANFIS consisting of three main phases: filtering IN, clustering data, and building ANFIS. By this way, the filtered IDS¯is used to build the ANFIS, then the created ANFIS is applied as an updated filter to refilter the IDS¯, and so on, until either the process converges or a stop condition is satisfied. To get a guarantee of convergence and stability, an update law for the AOINF is discovered via Lyapunov stability theory.

Remark 1. ANFIS cannot converge to the required error [E] if there is at least one critical data sample in the CDS (see Theorem 1). The clustering strategy of the FIN-ANFIS therefore focuses on preventing the clustering process from appearing critical data samples, along with seeking to exterminate the critical data samples in the CDS having been taking form. As a result, in each loop of the ANFIS training process, the strategy well directs the clustering process to a new CDS where either there is not any critical data sample or there exist with a smaller amount. Theorem 2 shows the convergence condition of the training process.

Theorem 2 [16]: Following the flowchart in Figure 3, the ANFIS-based approximation of an unknown mapping  f:XYexpressed by the given IDS is built via a CDS which drives from the IDS¯(the normalized IDS). Let Q be the number of critical data points in the CDS at the rth loop. At these critical data samples, if the data output is filtered by law (31), then the RMSE (3) of the ANFIS will converge to [E]:

r+1yi=yriρsgnyiŷir,i=1Q.E31

In the above, ρ>0is the update coefficient to be optimized by any well-known optimal method; yiŷiris the error between the ith data output and the corresponding ANFIS-based output; and function sgn.is defined as

sgnz=1ifz>01otherwise.E32

Proof: A Lyapunov candidate function is chosen as in Eq. (33), to which expression (34) can be inferred:

eX=XTX.E33
ėX=2i=1PQXiẊi+2j=1QXjẊj.E34

In the above, Ξ̇=dΞ/dtexpresses derivative of Ξwith respect to time; Xis the vector of state variables deriving from IDS¯as follows:

Xi=yiŷi;X=X1XPTE35

From update law (31), Eq. (34) can be rewritten as in Eq. (36):

ėX=2i=1PQXiẊi+2j=1QXjẏj=2i=1PQXiẊi2ρj=1QXjsgnXj.E36

It should be noted that the update process is performed with respect to the critical data points; hence, Eq. (36) can be rewritten as follows:

ėX=2ρj=1QXjsgnXj=2ρj=1QXj<0E37

In addition, the following can be implied from (33) to (35):

e0=0;eX0X.E38

Finally, it can infer from Eqs. (37) to (38) that eX0is a stable Lyapunov process. Hence, from Eq. (3) one can infer the aspect needing to be proven:

RMSE=limrreXP1E.    □E39

Remark 2. (1) To enhance the ability to adapt to the noise status of the IDS¯, ρin Eq. (31) is specified as follows:

ρ=αyiŷir,E40

where α0is an adaptive coefficient chosen by the designer. Thus, ρ=ρXittakes part in adjusting the filtering level Δi=r+1yiyri. (2) It can infer from Theorem 1 that disposing of critical data samples in the CDS needs to be carried out. Therefore, the useful solution offered in Theorem 2 via update law (31) is employed to establish the filtering mechanism of the AOINF as shown below.

The algorithm AOINF for filtering IN:

  1. Look for critical data samples in the CDS to specify the worst data point (WP) where the continuous status of the ANFIS is worst:

WPx¯iWPyiWPsuchthatyiWPŷiWP=maxh=1Pyhŷh.E41
  1. Specify the data samples satisfying condition (42):

yqŷq1σyiWPŷiWP,q=1Q¯.E42

In the above, ŷiWPis the ANFIS-based output, while yiWPis the corresponding data output at the WP; σ>1is an adaptive coefficient (to be 1.35 for the surveys shown in [16]).

  1. Based on the updating law (43) to filter the data samples satisfying condition (42)

yr+1qyrq+αyqŷqrsgnyqŷqr,q=1Q¯.E43

3.3 Algorithm for building ANFIS

Figure 4 illustrates the establishment of the CDS from the IDS¯. It consists of (1) building fuzzy clusters with centroids x¯10x¯C0or the input data clusters (see Subsection 2.2), (2) estimating the hard distribution of samples in each input data cluster indicated by x¯10x¯C0, and (3) building the hyperplanes or the output data clusters (see Subsection 2.3) in the output data space using the specified hard distribution status. Based on the created CDS, Figure 3 shows the flowchart of the FIN-ANFIS consisting of three main phases: filtering IN, building the CDS driving from the filtered IDS¯, and forming ANFIS.

Figure 4.

A process of establishing the CDS driving from the IDS ¯ consists.

3.4 Algorithm FIN-ANFIS

Initializing: The initial index of the loop process, r = 1; the number of clusters C ≪ P1; JKFCMr=Ω, where Ωis a real number Ω>ts; and the initial cluster centroids corresponding to r = 1 chosen randomly:

x¯i0r=xi10xin0,1iCE44

Build the input data clusters:

  1. Establish the input data clusters:

    Based on the x¯i0rto be known, calculate μijvia Eq. (14) to update x¯i0rvia Eq. (13).

  2. Specify the stop condition of the clustering phase via tsin Eq. (15):

    If tsts: go to Step 3; ff ts>tsand r<r, setup rr+1and return to Step 1; if ts>tsand r=rand C<P1, set CC+1,r1,and return to Step 1; and if ts>tsand r=rand C=P1, stop (not converge).

    Build ANFIS:

  3. Build and estimate ANFIS:

    Establish ANFIS as presented in Subsection 2.4.

    Calculate RMSE=P1i=1Pŷix¯iyi20.5in which ŷix¯iis the ANFIS-based output, while yiis the data output. If RMSEE, stop (the ANFIS is the desired one); if RMSE>Eand C<P1, go to Step 4; and if RMSE>Eand C=P1, stop (not converge).

  4. Set up a new cluster centroid:

    Based on Eq. (41), seek the worst data point WPx¯iWPyiWP; set CC+1,r1,and set up a new cluster centroid x¯C0in the neighborhood of the WP; and go to Step 5.

  5. Filter IN:

    Call the algorithm AOINF and return to Step 1.

4. ANFIS for managing online bearing fault

An application of ANFIS to estimating online bearing fault upon the ability to extract meaningful information from big data of intelligent structures is shown in this section. Estimating online bearing status to hold the initiative in exploiting the systems is meaningful because bearing is an important machine detailed in almost mechanical structures.

In [17], an Online Bearing Damage Identifying Method (ASSBDIM) based on ANFIS, singular spectrum analysis (SSA), and sparse filtering (SF) was shown. The method consists of two phases: offline and online. In offline, the ANFIS identifies the dynamic response of the mechanical system in the individual bearing statuses. The trained ANFIS is then used to estimate its real status in the online phase. These aspects are detailed in the following paragraphs.

4.1 Some related theories

4.1.1 Singular spectrum analysis

By using SSA, from a given time series, a set of independent additive time series can be generated [41, 42, 43]. This work is clarified via the algorithm for SSA presented in [42] as follows.

  1. Embedding:

    Let’s consider a given time series of N0 data points z0z1zN01. From selected window length L0, 1 < L0 < N0, sliding vectors Xj=zj1zjzj+L02T, j = 1,…,K=N0 − L0 + 1, and matrix Xas in Eq. (45) are built:

    X=z0z1zN0L0z1z2zN0L0+1zL02zL01zN02zL01zL0zN01.E45

  • Building the trajectory matrix:

    From Eq. (45), one builds matrix S=XXTL0×L0. Vectors Viare then constructed, Vi=XTUi/λi,i=1d, in which λ1,,λdare the non-zero eigenvalues of S arranged in the descending order and U1,…, Ud are the corresponding eigenvectors. A decomposition of the trajectory matrix into a sum of matrices X=i=1dEiis then established, where Ei=λiUiViT.

  • Reconstruction:

    Each elementary matrix is transformed into a principal component of length N by applying a linear transformation known as diagonal averaging or Hankelization. Let ZL0×Kbe a matrix of elements zi,j. By calculating L=minL0K,K=maxL0K, then Z can be transformed into the reconstructed time series g0,g1,,gN1as in Eq. (46):

    gk=1k+1m=1k+1zm,km+2,0k<L11Lm=1Lzm,km+2,L1k<K1N0km=kK+2NK+1zm,km+2,Kk<N0.E46

  • 4.1.2 Sparse filtering

    In this work SF is used to extract features from a given time series-typed measured database. Relying an objective function defined via the features, the method tries to specify the good features such that the objective function is minimized [11, 44, 45]. To deploy SF effectively, a process with the two following phases is operated. Preprocessing data based on the whitening method [46] is carried out in the first phase. A H-by-L matrix signed Fof real numbers depicting the relation between each of the H training data samples and the L selected features is established in the second phase. SF presented in [11, 45] is detailed as follows.

    In the first phase, a training set of the H data samples xi1×N, i=1H,in the form of a matrix signed SH×Nis established from the given time series-typed measuring dataset. By adopting the whitening method [46], it then tries to make the data samples less correlated with each other and speed up the convergence of the sparse filtering process which employs the eigenvalue decomposition of the covariance matrix covS=Z¯DZ¯T.In the expression, Dis the diagonal matrix of its eigenvalues, and Z¯is the orthogonal matrix of eigenvectors of covS. Finally, the whitened training set signed Swhiteis formed as in Eq. (47):

    Swhite=Z¯D1/2Z¯TS.E47

    Subsequently, in the second phase, SF maps the data sample xi1×Nof Swhiteonto Lfeatures fi,i=1L, relied on a weight matrix signed W¯N×L. A linear relation between data samples in Swhiteand the Lfeatures is expressed via W¯as in Eq. (48), in which FH×Lis called the feature distribution matrix:

    F=SwhiteW¯.E48

    Optimizing the feature distribution in Fis then performed as detailed in [45]. The features in each column of Fis normalized by dividing them by their l2-norm, f˜l=fl/fl2,l=1L. For each row of the obtained matrix, these features per example are normalize by computing f̂i=f˜i/f˜i2,i=1H, by which they lie on the unit l2-ball. The features normalized after the two above steps are optimized for sparseness using the l1-penalty to get a matrix signed F̂H×L. A loop process is then maintained via Eq. (48), in which F̂takes the role of F, until the optimal weights of W¯are to be established that make the objective function JSFW¯of Eq. (49) be minimized, to which, finally, F̂is resigned F:

    JSFW¯=i=1Hj=1LF̂ij.E49

    4.2 The ASSBDIM

    The ASSBDIM focuses on online bearing fault estimation. The aim is detailed in this subsection consisting of the way of setting up the databases and the algorithm ASSBDIM for online bearing fault estimation upon the built databases.

    4.2.1 Building the databases for the ASSBDIM

    A measuring dataset deriving from the mechanical system vibration is established for each surveyed bearing fault type. Regarding Q fault types, one obtains Q original datasets as in Eq. (50):

    D1D2DQT,E50

    where Diis corresponding to the ith bearing fault type 1iQ.

    By using SSA for Di, m time series as in Eq. (51) are set up:

    Di1Di2Dim,i=1QE51

    where m is parameter selected by the designer. This work is carried out by the three steps as presented in Subsection 4.1.1, in which Diis used in the first step as the given time series of N0 data points z0z1zN01for building the trajectory matrix Xin Eq. (45). Because the mechanical vibration signal is prone to the low frequency range [42], among the m time series, the (m-k) ones owning the highest frequencies are considered as noise. The k remainder time series as in Eq. (52) is hence kept to build the databases:

    Di1Di2Dik,i=1QE52

    Specifying the optimal value of both k and m will be mentioned in Subsection 4.2.2.

    For each time series in Eq. (52), for example, Dij, j=1k, based on SF one obtains the feature distribution matrix as in Eq. (48) which is resigned FijωH×L. By using this result for all the time series in Eq. (52), a new data matrix D¯ias in Eq. (53) is formed which is the input data space of the ith data subset corresponding to the ith bearing fault type:

    D¯i=Fi1ωFi2ωFikωH×kL.E53

    By employing this way for Q, the surveyed bearing fault types, an input data space in the form of matrix (54), are established, which relates to building two offline databases signed Off_DaB and Off_testDaB as well as one online database signed On_DaB used for the algorithm ASSBDIM as follows:

    D¯=ω11ω12ω1kLω21ω22ω2kLωQH11ωQH12ωQH1kLωQH1ωQH2ωQHkLQH×kLE54

    Namely, matrix D¯relates to the input data space (IDS), to which the databases for identifying the bearing status are built as follows. Firstly, by encoding the ith fault type by a real number yi, the output data space (ODS) of the ith subset can be depicted by vector y¯iof H elements yias in Eq. (55):

    y¯i=yiyiTH×1,i=1QE55

    Then, by combining Eq. (55) with Eq. (54), the input-output relation in the three datasets Off_DaB, Off_testDaB, and On_DaB can be described as in Eq. (56):

    databaseIDSODSD¯y¯E56

    In the above, the input space D¯comes from Eq. (54), while the output space y¯as in Eq. (57) is constituted of y¯iH×1in Eq. (55):

    y¯=y1,y1HyQ,yQHTQH×1E57

    4.2.2 The algorithm ASSBDIM for estimating health of bearings

    In the offline phase, by initializing the parameters in vector psin Eq. (58), together with applying SSA and SF to the measuring data stream, the Off_DaB and Off_testDaB are built as in Eq. (56):

    ps=L0N0mkHLE58

    where L0,N0come from Eq. (45); m and k relate to Eqs. (51) and (52), respectively; H and L derive from Eq. (53).

    An ANFIS built by the algorithm FIN-ANFIS (see Subsection 3.3) is utilized to identify dynamic response of the mechanical system corresponding to the bearing damage statuses reflected by the Off_DaB. Optimizing the parameters in psin Eq. (58) is then performed using the percentage of correctly estimated samples (Ac) as in Eq. (59) and the mean accuracy (MeA) as in Eq. (60) and the algorithm DE [47]:

    Ac=100×cr_samplesn/to_samplesn%,E59
    MeA=100×n=1Qcr_samplesn/n=1Qto_samplesn%,E60

    where corresponding to the nth damage type, n=1Q, cr_samplesnis the number of checking samples expressing correctly the real status of the bearing, while to_samplesnis the total of checking samples used in the survey; Qis the number of surveyed bearing fault types as mentioned in Eq. (50).

    Following the MeA, an objective function is defined as follows:

    J=MeAASSBDIML0N0mkHLmax.E61

    The Off_testDaB, function J, and DE [47] are then employed to optimize the parameters in vector ps, to get L0N0mkHLopt.

    Namely, by using the input of the Off_testDaB for the ANFIS which has been trained by the Off_DaB, one obtains the outputs ŷi,i=1H. These outputs are then compared with the corresponding encoded outputs to estimate the bearing real status, which is the one encoded by “q” satisfying Eq. (62):

    i=1Hŷiyq=minh=1Qi=1Hŷiyh.E62

    The completion of the offline phase as above can be seen as the beginning of the only phase. During the next operating process, first, by the way similar to the one for building the offline database Off_DaB, the online dataset On_DaB in the form D¯OND¯iH×kLas in Eq. (53) is built. By using the On_DaB for the ANFIS trained in the offline, the bearing real status at this time is then specified based on Eq. (62).

    The ASSBDIM is hence can be summarized as follows.

    The offline process:

    Initialize vector psin Eq. (58):

    1. Build the Off-DaB and Off-testDaB in the form of Eq. (56).

    2. Train an ANFIS to identify the Off-DaB using the algorithm FIN-ANFIS.

    3. Accomplish the system.

      The Off-testDaB is used as database of the trained ANFIS, using the condition (62) to calculate MeAin Eq. (60). If MeAMeA, then go to Step 4; otherwise, adjust the value of the elements in vector psin Eq. (58) using the algorithm DE [47], and then return to Step 1.

      The online process:

    4. Establish online database On-DaB D¯OND¯iH×kLas in Eq. (53).

    5. Estimate online bearing fault status based on the On-DaB, trained ANFIS, and condition (62); check the stop condition: if it is not satisfied, then return to Step 4; otherwise, stop.

    4.3 Some survey results

    4.3.1 Experimental apparatus and estimating way

    The experimental apparatus for measuring vibration signal is shown in Figure 5. The apparatus consists of the motor (1), acceleration sensors (2) and (4), surveyed bearings (3) and (5), module for processing and transforming series vibration signal incorporating software-selectable AC/DC coupling (Model: NI-9234) (6), and computer (7).

    Figure 5.

    Experimental apparatus for measuring vibration signal.

    In Table 1, “encoding value” is abbreviated to “EV.” The three cases listed in Table 1 related to nine of the widespread single-bearing faults as in Table 2 are surveyed. In the above, Q = 7 (see Eq. 50) for the Cases 1–2, while Q = 10 for Case 3; the damaged location is the inner or outer or balls (signed In, or Ou, or Ba, respectively); damaged degrees are from 1 to 3 (signed D1 or D2 or D3); the load impacting on the system at the survey time consists of Load 1 or 2 or 3 (signed L1 or L2 or L3). For example, LmUnd shows the load degree to be m and the bearing to be undamaged, or LmDnBa expresses the load degree to be m (1,…,3), the damage level to be n (1,…,3), and the damage location to be the ball.

    Table 1.

    Surveyed cases and the corresponding encoding values (EV).

    FaultsWidth (mm)Depth (mm)
    D1Ou0.200.3
    D2Ou0.300.3
    D3Ou0.460.3
    D1In0.200.3
    D2In0.300.3
    D3In0.400.3
    D1Ba0.150.2
    D2Ba0.200.2
    D3Ba0.250.2

    Table 2.

    The size of bearing single fault types used for surveys.

    The ASSBDIM with H = 303, m = 30, k = 7 along with four other methods [48, 49, 50, 51] is employed to be surveyed. The first one [48] (Nin = Nout = 100; number of segments to be 20×103and λ=1E5) is the intelligent fault diagnosis method using unsupervised feature learning toward mechanical big data. The second one [49] employs the energy levels of the various frequency bands as features. In the third one [50], a bearing fault diagnosis upon permutation entropy, empirical mode decomposition, and support vector machines is shown. In the last one [51], a method of identifying bearing fault based on SSA is presented.

    For the surveys, along with Ac and MeA, the root-mean-square error as in Eq. (63) is also employed, where yiand ŷi, respectively, are encoding and predicting outputs:

    LMS=i=1Hyiŷi2/H.E63

    4.3.2 Some survey results

    The measured databases from Cases 1 to 3 with Q = 7 as in Table 1 along which the methods consist of the ASSBDIM [17] and the ones from [48, 49, 50, 51] were adopted to identify the status of the bearing. The obtained results were shown in Figures 6, 7, 8, 9 and Tables 3 and 4.

    Figure 6.

    The predicting (−pre) output y ̂ i of the ASSBDIM in Case 1 and the corresponding encoded (−enc) output y i .

    Figure 7.

    The y ̂ i and y i depicted by lines (6) in Figure 6 to be zoomed in.

    Figure 8.

    The error reflecting the difference between y i and y ̂ i in Figure 6.

    Figure 9.

    Ac and MeA (mean accuracy) of the ASSBDIM in Case 2.

    Surveyed casesAc (%)
    [48][49][50][51][17]
    L2UnD99.6793.7398.6899.67100
    L2D1In98.3595.0592.7495.3899.67
    L2D2In99.6798.6899.3497.3699.01
    L2D3In99.0193.0795.3892.74100
    L2D1Ba98.6891.0996.3796.7097.36
    L2D2Ba99.6792.0894.7299.0198.68
    L2D3Ba97.3698.6810098.68100
    MeA (%)98.9294.9396.7597.0899.26

    Table 3.

    The accuracy of the methods in Case 2.

    Surveyed casesAc (%)
    [48][49][50][51][17]
    L1UnD95.0587.4685.81100100
    L1D1In94.7290.1089.7783.1799.67
    L1D2In92.0892.4188.1285.4899.34
    L1D3In93.4092.7492.4194.7299.34
    L1D1Ou95.3389.7784.1685.1582.51
    L1D2Ou92.4192.4184.8284.1694.39
    L1D3Ou95.0588.7888.7899.3489.11
    L1D1Ba86.4789.4490.4383.1794.72
    L1D2Ba87.7990.1096.0497.3692.74
    L1D3Ba88.1286.1410088.4584.82
    MeA (%)92.0489.9490.0390.1093.66

    Table 4.

    The accuracy of the methods in Case 3.

    4.3.3 Discussion

    Following the above results, it can observe that among the surveyed methods, the ASSBDIM which is based on ANFIS gained the best accuracy. This aspect can be recognized via the quite equivalent values between the encoding and predicting outputs from the tested data samples. The small difference depicted by the zooming in in Figure 7 and the root-mean-square error in Figure 8 as well as the high/higher values of Ac and MeA deriving from the ASSBDIM in Tables 3 and 4 and Figure 9 reflect clearly the ANFIS’s identification ability.

    It should be noted that the methodology shown via the algorithm ASSBDIM can be also used to discover the method of managing damage of mechanical structures as well.

    5 Conclusion

    The hybrid structure ANFIS, where ANN and FL can interact to not only overcome partly the limitations of each model but also uphold their strong points, has been seen as a useful mathematical tool for many fields. Inspired by the ANFIS’s capability, in order to provide the readers with the theoretical basis and application direction of the model, this chapter presents the formulation of ANFIS and one of its typical applications.

    Firstly, the structure of ANFIS as a data-driven model deriving from fuzzy logic and artificial neural networks is depicted. The setting up the input data clusters, output clusters and ANFIS as a joint structure is all detailed. Deriving from this relation, the method of building ANFIS from noisy measuring datasets is presented. The online and recurrent mechanism for filtering noise and building ANFIS synchronously is clarified via the algorithms for filtering noise and establishing ANFIS. Finally, the application of ANFIS coming from the online managing bearing fault is presented. The compared results reflect that among the surveyed methods, the ASSBDIM which exploited the identification ability of ANFIS gains the best accuracy. Besides, the methodology shown via this application can be also used as appropriate solution for developing new methods of managing damage of mechanical structures.

    In addition to the above identification field, it should be noted that (1) ANFIS has also attracted the attention of many researchers in the other fields related to prediction, control, and so on, as mentioned in Section 1 and (2) ANFIS can collaborate effectively with some other mathematical tools to enhance the effectiveness of technology applications.

    Download

    chapter PDF

    © 2019 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution 3.0 License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

    How to cite and reference

    Link to this chapter Copy to clipboard

    Cite this chapter Copy to clipboard

    Sy Dzung Nguyen (March 11th 2019). ANFIS: Establishing and Applying to Managing Online Damage [Online First], IntechOpen, DOI: 10.5772/intechopen.83453. Available from:

    chapter statistics

    129total chapter downloads

    More statistics for editors and authors

    Login to your personal dashboard for more detailed statistics on your publications.

    Access personal reporting

    We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities.

    More About Us