Open access peer-reviewed chapter

Novel Direct and Accurate Identification of Kalman Filter for General Systems Described by a Box-Jenkins Model

Written By

Rajamani Doraiswami and Lahouari Cheded

Submitted: 17 March 2018 Reviewed: 02 October 2018 Published: 30 April 2019

DOI: 10.5772/intechopen.81793

From the Edited Volume

Introduction and Implementations of the Kalman Filter

Edited by Felix Govaers

Chapter metrics overview

1,301 Chapter Downloads

View Full Metrics

Abstract

A novel robust Kalman filter (KF)-based controller is proposed for a multivariable system to accurately track a specified trajectory under unknown stochastic disturbance and measurement noise. The output is a sum of uncorrelated signal, disturbance and measurement noise. The system model is observable but not controllable while the signal one is controllable and observable. An emulator-based two-stage identification is employed to obtain a robust model needed to design the robust controller. The system and KF are identified and the signal and output error estimated. From the identified models, minimal realizations of the signal and KF, the disturbance model and whitening filter are obtained using balanced model reduction techniques. It is shown that the signal model is a transfer matrix relating the system output and the KF residual, and the residual is the whitened output error. The disturbance model is identified by inverse filtering. A feedback-feedforward controller is designed and implemented using an internal model of the reference driven by the error between the reference and the signal estimate, the feedforward of reference and output error. The successful evaluation of the proposed scheme on a simulated autonomously-guided drone gives ample encouragement to test it later, on a real one.

Keywords

  • identification
  • Box-Jenkins model
  • Kalman filter
  • whitening filter
  • signal estimation
  • model reduction
  • robust controller
  • feedback controller
  • feedforward controller
  • internal model principle
  • autonomous vehicles
  • drones

1. Introduction

In conventional Kalman filter applications, the system involved is typically linearized and then identified. Based on the identified system, the Kalman filter is then identified. In this chapter, we propose a novel approach in that (a) the system is represented by a more general model, termed multi-input multi-output Box-Jenkins (MIMO BJ model) which subsumes all previous classical models, such as ARMA models and their derivatives, and (b) the associated Kalman filter identification is carried out directly, i.e. it does not necessitate the prior identification of the system involved. The various tools involved in our proposed approach are all explained below.

1.1 Box-Jenkins model and its applications

Identification of a class of system described by MIMO BJ model, and the associated Kalman filter directly from the input-output data is proposed [1, 2]. There is no need to specify the covariance of the disturbance and the measurement noise, thereby avoiding the use the Riccati equation to solve for the Kalman gain. The output is the desired waveform, termed signal, corrupted by a stochastic disturbance and zero-mean white measurement noise. The state-space BJ model is an augmented system formed of the signal and disturbance model. The signal model and the disturbance models are driven respectively by a user-defined accessible input, and an inaccessible zero-mean white noise process. The signal model is generally a cascade, parallel and feedback combinations of subsystems such as controllers, actuators, plants, and sensors [3]. Unlike the ARMA model, the Box-Jenkins model is observable but not controllable while the signal model is both controllable and observable. In other words, the transfer matrix of the system is non-minimal whereas that of the signal is minimal. This issue will need to be addressed in the identification and implementation of the Kalman filter.

1.2 Kalman filter and its key properties

The structure of the Kalman filter is determined using the internal model principle which establishes the necessary and sufficient condition for the tracking of the output of a dynamical system [3, 4]. In accordance with this principle, the Kalman filter consists of (a) a copy of the system model driven by the residuals, and (b) a gain term, termed the Kalman gain, to stabilize the filter. The Kalman gain is determined such that the residual of the Kalman filter is a zero-mean white noise process with minimum variance. The Kalman filter enjoys the following key properties:

Tracking a signal: The estimate of the Kalman filter tracks a given signal if and only if the model that generates the signals including those of the noise and disturbances is embodied in the Kalman filter. In other words, the Kalman filter tracks the input, thanks to its internal model-based structure [3, 4].

Model matching: The residual is a zero-mean white noise process if and only if there is no mismatch between the actual model of the system and its identified version embodied in the Kalman filter, and its variance is minimum [4].

Optimality: The estimate is optimal in the sense that it is the best estimate that can be obtained by any estimator in the class of all estimators that are constrained by the same assumptions [5].

Robustness: Thanks to the feedback (closed-loop) configuration of the Kalman filter with residual feedback, the Kalman filter provides the highest robustness against the effect of disturbance and model variations [5].

Model-mismatch: If there is a model mismatch, then the residual will not be a zero-mean white noise process and an additive term termed fault-indicative term will occur. The fault-indicative term is a filtered version of the deviation in the linear regression model of the system or that of the signal [6, 7, 8].

1.3 Identification using residual model

The equation error in the regression model of the system is a colored noise process, and hence a direct identification of the system model from the input-output data by minimizing the equation error will not ensure that the estimates are consistent, unbiased, and efficient. The fundamental requirement of identification is that the leftover signal from identification, namely the residual is a zero-mean white noise process that contains no information. To meet this requirement, both the input and output of the system are filtered. Among the class of all linear whitening filters, the Kalman filter is the best. The system model is indirectly identified by minimizing the residual generated by the Kalman filter instead of the equation error. The Subspace Method (SM) uses the structure of the state-space model of the Kalman filter, whereas the prediction error method (PEM), which is the gold standard in system identification, is developed from the residual model [1, 9, 10, 11].

1.4 Emulator-based two-stage identification

The static and dynamic behavior of a physical system change as a result of variations in the parameters of some of its subsystems such as sensors, actuators, plant, disturbance models, and controllers. As the parameters of these subsystems are not generally accessible to generate data, instead, emulators, which are hardware or software devices, are connected in cascade to the output, input or both, of the subsystems. An emulator is a transfer function block which mimics the variations in the associated subsystems including the disturbance model. An emulator takes the form of a static gain or an all-pass filter to induce gain or phase variations in the subsystem it is connected to. Emulator parameters are perturbed to mimic various normal and abnormal, or faulty, operating scenarios resulting from variations in these subsystems. The emulator-generated data is employed in (a) the identification of robust systems and signal models and their associated Kalman filters using the two-stage identification scheme [2, 3, 6, 7, 8].

A two-stage identification is used in various applications including the non-parametric identification of impulse response, estimation of Markov parameters in the SM, in model predictive control, identification of a signal model and in system identification. The use of the two-stage identification is inspired by the seminal paper by [12] for an accurate estimation of the parameters of an impulse response from measurements in an additive white noise. It is shown via simulation that the variance of the parameter estimation error approaches the Cramer-Rao lower bound [13]. Further, it is shown analytically that using a high-order model (with an order several times larger than the true order) improves significantly the accuracy of the parameter estimates. The two-stage scheme has not received much attention in system identification although it has been mentioned as an alternative scheme to the PEM [1, 14], and has been successfully employed in identification in [15, 16, 17].

It should be emphasized that the prediction error method (PEM), viewed as a gold standard for system identification, is not geared for the estimation of the signal buried in the output, i.e. it is developed for the ARMA model and not for the Box-Jenkins one. A two-stage identification of the Box-Jenkins model is proposed as the system model is observable but not controllable while the signal model is both controllable and observable:

  • In the first stage, the robust system model and the associated Kalman filter are identified using the emulator-generated data using PEM, and the signal and the output error are both estimated. Further, the whitening filter that relates the output error and the residual is obtained.

  • In the second stage, minimal realizations of the signal model and the associated Kalman filter are obtained using model reduction method [18].

The high- order for the first stage and the reduced-order for the second are both selected using the Akaike information criterion (AIC) and are cross-checked by verifying the whiteness of the associated residual. The two-stage identification has also been successfully employed in the identification model.

The question arises as to how to obtain the system model and the signal model from the identified high-order model Kalman filters. A key property of the Kalman filter is established here, namely that the transfer matrix of the signal and the system is the matrix fraction description model derived from the Kalman filter residual model of the system. This property is exploited to derive the signal and the system transfer matrices. The state-space models of the signal and the system models are derived from the identified state-space models of the Kalman filters. Thus, the proposed scheme identifies (a) the Kalman filter for the system, (b) the Kalman filter for the signal first. Then, the system model and the signal model are separately obtained.

The proposed scheme is further extended to identify the signal model to complement the PEM. In the first stage, a very high-order model is identified using PEM. In the second stage, the signal model is identified using a balanced model reduction of the high-order identified model obtained in the first stage. The PEM and state-space method (SM) are both tailored to identify the signal model and estimate the signal by employing the proposed version of the two-stage identification scheme. The results of the comparison of the performance of these methods in identifying the system and signal models are presented.

1.5 Highlights of the contributions

  • The Auto-Regressive (AR), The Moving Average (MA), and the Auto-Regressive and Moving Average (ARMA) models are all special cases of the proposed Box-Jenkins model. As this model is more general and hence has wider applications, including robust controller design; estimation of latent variables; monitoring of the status of the system, fault diagnosis, development of condition-based maintenance programs and design of fault-tolerant systems; filtering of signals, speech enhancement, noise and echo cancelation in communication; 2-D image filtering and tracking of moving objects.

  • The state-space models of the system and the signal models are derived from the identified Kalman filters, by invoking the (causal) invertibility of the output error and the residual [5].

  • An efficient scheme to monitor the status of the system may be implemented from the proposed scheme. First the status of the system is monitored by analyzing the residual of the Kalman filter of the system model. If there is a variation, then the residual of the Kalman filter of the signal model is analyzed to ascertain whether a fault has occurred.

  • In practice, disturbances are inevitable, and can negatively affect the system performance. When the system is in an abnormal state, it is not in general easy to determine whether the abnormal operation is the result of variations in the disturbance or the occurrence of a fault. The proposed scheme provides a simple solution by analyzing the residuals of both Kalman filters as the residual of the Kalman filter for the system captures the variations in both the system and disturbance models, while that of the Kalman filter for the signal, and captures only the variations in the signal model. This is crucial for reducing false alarm, and all its concomitant risks and costs, resulting from variations in the disturbance and not in the signal model [19].

  • The PEM (SM) may be tailored to identify the signal model and estimate the signal itself by using the proposed two-stage identification scheme.

1.6 Applications

Applications include monitoring the status of the system and the signal models, distinguishing between the variations in the disturbance model and those in the signal model to help diagnose a fault in the system and ensure a low false alarm probability, estimating the latent variable, namely the signal, developing a framework for applications including robust controller design; fault diagnosis; speech and biological signal processing; tracking of moving objects, design of soft sensors to replace maintenance-prone hardware sensors, evaluate and monitor product quality, meeting the ever-increasing need for fault-tolerant systems for mission-critical systems found in aerospace, the nuclear power systems, and autonomous vehicles.

Advertisement

2. Problem formulation

The output y k R q is an additive sum of the signal s k R q , disturbance, d k R q and the measurement noise v k R q where R is real scalar field.

y k = s k + d k + v k E1

Where the signal and the disturbance models are:

s z = G s z u z E2
d z = G w z w z E3

Where u k R q is the input; w k R p is zero-mean white noise process that generates the disturbance d k R p , and is uncorrelated with the measurement noise v k ; G s z = D s 1 z N s z and G w z = D w 1 z N w z are qxp transfer matrix of order n s and n w respectively; ϑ k = d k + v k is the output error.

The signal model G s z is formed of cascade and parallel combinations of the subsystems such as actuators, plant and the sensors. Let the state space model of the signal and the disturbance models be respectively A s B s C s and A w B w C w .

Figure 1 shows the input-output model relating the input, the signal model, the signal, the disturbance model, the disturbance, the measurement noise and the output.

Figure 1.

System: signal, the disturbance and the measurement noise.

Linear regression model:

Using, (1)–(3), the expression for the linear regression becomes:

D sw z y z = D w z N sw z u z + υ z υ z = D s z N w z w z + D sw z v z E4

Where υ z = D z ϑ z is the equation error; D sw z = D s z D w z and N sw z = D w z N s z are respectively the denominator and numerator polynomials. The model is termed Box-Jenkins model.

Note that the model that generates the equation error υ z is a Moving Average (MA) model, whereas the one that generates the output error ϑ k is an Auto-Regressive Moving-Average (ARMA) model.

Augmented state-space model: The augmented state-space representation of the multi-input and multi-output (MIMO) system A B C D formed of the signal model A s B s C s D s and A w B w C w D w representing a p-input, q-output system, is given by:

x k + 1 = Ax k + Bu k + E w w k s k = C s x k + D s u k y k = Cx k + Du k + v k E5

Where A = A s 0 0 A w ; B = B s 0 ; E w = 0 B w ; C = C s C w ; A R nxn is an augmented state transition matrix formed of A s n s xn s and A w n w xn w ; B = B 1   B 2   .   B p R nxp ; C = C 1   C 2   .   C q T R qxn ; E w R nxp is a disturbance entry matrix;

x k = x 1 k x 2 k x 3 k x n k T R n ; s k = s 1 k s 2 k s 3 k s q k T R q ;

u k = u 1 k   u 2 k   u 3 k     u p k T R p ; y k = y 1 k   y 2 k   y 3 k     y q k T R q are respectively the state, the input and output; n = n s + n w is the order; w k R n and v k R q are respectively the disturbances and measurement noise; D sw z = z I A s z I A w where . is the determinant of . . Using D sw z = D s z D w z and N sw z = D w z N s z we get:

G z = C z I A 1 B + D = D sw 1 z N sw z G z = D s 1 z N s z = G s z E6

The augmented transfer matrix is not a minimal realization of the system output model as there is (stable) pole-zero cancelation since the polynomial D w z , which is common to both the numerator N sw z and the denominator D sw z . In other words, N sw z and D sw z are not coprime. The signal model A s B s C s D s associated with signal model G s z is controllable and observable while A B C D , associated with G z , is merely an observable. A s B s C s D s ( G s z ) is a minimal realization of A B C D ( G z ).

Assumptions: It is assumed that (a) the disturbance w k and the measurement noise v k are independent zero-mean Gaussian white noise processes with unknown but finite covariance, Q = E w k w T k and R = E v k v T k , respectively, and are inaccessible, (b) A C is observable, (c) the signal and disturbance models are both minimal, A s B s C s D s and A w B w C w D w are both controllable and observable, (d) The initial conditions x 0 , w k and v k are mutually uncorrelated. However, the signal s z and the disturbance w z may have spectral overlap, (e) the output error is bounded.

2.1 Kalman filter

Predictor form: A robust Kalman filter of the identified system A 0 B 0 C 0 D 0 relating the system input u z and system output y k to the estimated output y ̂ k is:

x ̂ k + 1 = A 0 K 0 C 0 x ̂ k + B 0 K 0 D 0 u k + K 0 y k y ̂ k = C 0 x ̂ k + D 0 u k e k = y k y ̂ k E7

Where x ̂ k = x ̂ 1 k x ̂ 2 k x ̂ 3 k x ̂ n k T R n and y ̂ k = y ̂ 1 k y ̂ 2 k y ̂ 3 k y ̂ q k T R q are respectively the best estimate of the state x k , and of the output y k ; e k = e 1 k e 2 k e 3 k e q k T R q is the residual or the innovation sequence; the Kalman gain K 0 R nxq ensures the asymptotic stability of the Kalman filter, i.e. ( A 0 K 0 C 0 ) is strictly Hurwitz having all its eigenvalues strictly inside the unit circle.

Innovation form: There is duality between the predictor, and the innovation forms of the Kalman filter [5]. The output y k and the residual e k are (causally) invertible. In other words, e k can be generated from the output y k and r k using the (causal) predictor form, and y k can be generated from e k and r k using the innovation form. The Kalman filter given by (7) is termed the predictor form and can be expressed in an alternative form, termed the innovation form, given by:

x ̂ k + 1 = A 0 x ̂ k + B 0 u k + K 0 e k y ̂ k = C 0 x ̂ k + D 0 u k E8

Figure 2 shows the system and the Kalman filter which embodies the system model A B C . The inputs to the Kalman filter are the input r k and the output y k which is corrupted by the noise v k and affected by the disturbance w k .

Figure 2.

The system and the Kalman filter model.

2.2 Residual model

The frequency-domain expression relating the input u z R p and the output y z R q to the residual e z R q is given by the following model termed the residual model:

e z = F 1 z D ¯ z y z F 1 z N ¯ z u z E9

where D ¯ z and N ¯ z are matrix polynomials, F z is the scalar characteristic polynomial termed Kalman polynomial, F z = z I A 0 + K 0 C 0 ; D z = z I A 0 ; D ¯ z = F z I C 0 z I A 0 + K 0 C 0 1 K 0 is qxq matrix; N ¯ z = F z C 0 z I A 0 + K 0 C 0 1 B 0 K 0 D 0 + D 0 is qxp matrix; I R qxq is an identity matrix;

D ¯ z = D ¯ 1 z D ¯ 2 z . D ¯ q z = D ¯ 11 z D ¯ 12 z . D ¯ 1 q z D ¯ 21 z D ¯ 22 z . D ¯ 2 q z . ; . . D ¯ q 1 z D ¯ q 2 z . D ¯ qq z ; N ¯ z = N ¯ 1 z N ¯ 2 z . N ¯ q z = N ¯ 11 z N ¯ 12 z . N ¯ 1 p z N ¯ 21 z N ¯ 22 z . N ¯ 2 p z . ; . . N ¯ q 1 z N ¯ q 2 z . N ¯ qp z E10

D ¯ ij z = = 0 n a ¯ ij z ; N ¯ ij z = = 1 n b ¯ ij z ; a ¯ ij and b ¯ ij are the coefficients of the polynomials D ¯ ij z and N ¯ ij z , respectively. The rational polynomials F 1 z D ¯ z and F 1 z N ¯ z associated with the system output y z and the input u z are termed as an output IIR filter, and an input IIR filter, respectively. The estimate of the Kalman filter y ̂ k is:

y ̂ z = I F 1 z D ¯ z y z + F 1 z N ¯ z u z E11

The residual model of the Kalman filter forms the backbone of the proposed identification scheme.

2.3 The key properties of the Kalman filter

The map relating the signal and its model, and the output IIR filter and an input IIR filter of residual model is developed next.

The following lemmas are developed by invoking the key property namely that the residual is a zero-mean white noise process if and only if there is no mismatch between the actual model of the system and its identified model embodied in the Kalman filter [4], that is, the identified model embodied in the Kalman filter is identical to that of the actual model:

2.3.1 Derivation of the signal and the signal model

The following Lemma 1 shows that (a) the estimate of the signal model is the matrix fraction description relating the transfer matrices relating the residual of the Kalman filter to the input, and the output of the system; (b) the estimate of the signal is its output generated by the system input; and the Kalman filter whitens the output error.

(a) The left-matrix description of the MIMO signal model derived from the state-space model A s 0 B s 0 C s 0 D s 0 , namely, G s z = C s 0 z I A s 0 1 B s 0 = D s 1 z N s z and the left-matrix description of the Kalman filter derived from the residual model, G ¯ z = D ¯ 1 z N ¯ z are identical. The signal model G s z and the signal s z are:

G ̂ s z = G ̂ z = G ¯ z s ̂ z = G ̂ s z u z = G ¯ z u z E12

Proof:

(a) Consider the residual model (9). Substituting for y z yields:

F 1 z D ¯ z D s 1 z N s z u z D ¯ 1 z N ¯ z u z + ϑ z = e z E13

Since the residual is a zero-mean, white noise process and is uncorrelated with u z and υ z , correlating both sides with the input u z 1 yields:

F 1 z D ¯ z D ̂ s 1 z N ̂ s z F 1 z N ¯ z E u z u z 1 = 0 E14

Assuming that the input correlation is not identically equal to zero, i.e. E u z u z 1 0 yields:

F 1 z D ¯ z D ̂ s 1 z N ̂ s z F 1 z N ¯ z = 0 E15

Simplifying we get:

D ̂ s 1 z N ̂ s z = D ¯ 1 z N ¯ z E16

Hence G ̂ s z = G ¯ z holds. Since D sw z = D s z D w z and N sw z = D w z N s z are not coprime as D w z is a common factor, then G ̂ s z = G ̂ z = G ¯ z .

(b) Substituting (16) in (13) we get

e z = F 1 z D ¯ z ϑ z = F 1 z D 1 z D ¯ z υ z E17

where F 1 z D ¯ z = I C z I A + KC 1 K .

Proof: follows from ϑ k = d k + v k , (9) and

2.3.2 Derivation of the output error and its model

The following Lemma 2 shows that the output error is the difference between the output and the estimate of the signal; the estimate of the output error model is the matrix fraction description of the transfer matrices relating the residual of the Kalman filter to the input; the estimate of the output error is obtained as its output when its input is the residual. It is assumed that the transfer matrix of the Kalman filter F 1 z D ¯ z = I C 0 z I A 0 + K 0 C 0 1 K 0 relating the output y z and the residual e z is minimum-phase, that is both the numerator and the denominator polynomials are asymptotically stable.

If the matrix D ¯ z is invertible, the output error ϑ z = d z + v z is given by:

ϑ ̂ k = y k s ̂ k ϑ ̂ z = F z D ¯ 1 z e z E18

where F z D ¯ 1 z = I C 0 z I A 0 + K 0 C 0 1 K 0 1 is termed the disturbance model estimate which generates the output error when excited by the residual.

Proof: Using (17) we get (18).

As the input w k driving the disturbance model is not accessible, then by substituting the actual input w k by the residual e k , although both are zero-mean white noise processes, only the denominator polynomial of the disturbance model can be identified. Hence the term “disturbance model estimate”.

Minimum realization of the output error model is obtained using balanced model reduction method by treating e z as the input and ϑ ̂ k as the output of a model [7].

2.3.3 Minimal realization of the signal model

There are two approaches to identifying the signal model and the signal. One approach is by deriving them from the residual model of the Kalman filter as shown in Lemma 1 given by (12) and the other approach is to invoke the duality between the predictor form (7) and the innovation form (8) of the Kalman filter. The latter approach may be more convenient.

In view of (12), the system model G z and the signal model G s z is derived from the identified Kalman filter (7) by simply replacing the transition matrix of the Kalman filter A 0 K 0 C 0 by the system transition matrix A 0 .

A 0 B 0 C 0 D 0 = A 0 + K 0 C 0 , B 0 C 0 D = A s B s C s D s = minreal A 0 B 0 C 0 D 0 E19

Where minreal A 0 B 0 C 0 D 0 is the minimal realization of A 0 B 0 C 0 D 0 .

Proof

There is duality between the predictor form (7) and the innovation form (8) of the Kalman filter [5]. The output y k and the residual e k are (causally) invertible. In other words, e k can be generated from the output y k and r k using the (causally) predictor form, and y k can be generated from e k and r k using the innovation form [3]. Moreover, A 0 B 0 C 0 D 0 and A s 0 B s 0 C s 0 D s 0 are associated with the system transfer matrices G z = D 1 z N z and G s z = D s 1 z N s z respectively, as shown in (4), implying that A s 0 B s 0 C s 0 D s 0 is a minimum realization of A 0 B 0 C 0 D 0 .

The minimum realization of the system A s 0 B s 0 C s 0 D s 0 is obtained from the balanced model reduction method by treating u k as the input and s ̂ k as the output of a model [7].

Advertisement

3. Emulator-based two-stage identification

An identified model at each operating point characterizes the behavior of the system in the neighborhood of that point. In practice, however, the system model may be perturbed because of variations in the parameters of that system. To overcome this problem, the system model is identified by performing a number of emulator parameter-perturbed experiments proposed in [7, 8, 9]. Each experiment consists of perturbing one or more emulator parameters. A robust model is identified as the best fit to the input–output data from the set of emulated perturbations. The robust model thus obtained characterizes the behavior of the system over wider operating regions (in the neighborhood of the operating point) whereas the conventional model characterizes the behavior merely at the nominal operating point (that is, the conventional approach assumes that the model of the system remains unperturbed at every operating point). In [7, 8, 9], it is theoretically shown that the identification errors resulting from the variations in the emulator parameters are significantly lower compared to those of the conventional ones based on performing a single experiment (that is, without using emulators). The emulator-based identification scheme is inspired from the model-free artificial neural network approach which captures the static and dynamic behaviors by presenting the neural network with data covering likely operating scenarios. The PEM identifies the robust model of the plant, and the Kalman filter associated with the plant is then derived from the identified model without any a-priori knowledge of the statistics, such as covariance of the disturbance and measurement noise affecting the input-output data.

An accurate emulator-based model identification scheme is proposed and employed here. An emulator, which is modeled as a product of first-order all-pass filters and which induces phase and gain changes, is connected in cascade to the input, output or both, of the signal model to emulate a set of likely operating regimes around the nominal operating point. The identified model is obtained as the best fit over all emulated operating regions, thereby ensuring both accuracy and robustness of the identified model.

3.1 Two-stage identification

  • In the first stage, a robust model of the system A 0 B 0 C 0 D 0 and its associated Kalman filter A 0 K 0 C 0 , B 0 K 0 D 0 K 0 , C 0 D 0 are identified using PEM from the set of the emulator-generated input-output data. Then the estimate s 0 k of the signal s k and the estimate ϑ ̂ k of the output error ϑ k are derived.

  • In the second stage, using the key properties established in Lemmas 1–3, the robust signal model A s 0 B s 0 C s 0 D s 0 and its associated Kalman filter A s 0 K s 0 C s 0 B s 0 K s 0 D s 0 K s 0 C s 0 D s 0 are obtained using balanced model reduction method and the PEM.

Akaike Information Criterion: To select an appropriate order for the identified system model in the first stage, and for the signal model in the second stage, the widely popular Akaike Information Criterion (AIC) is used, which weights both the parameter estimation error and the complexity of the model so as to arrive at an optimal order [1].

3.2 Signal model and the Kalman filter

Similar to the Kalman filter for the system (7), the Kalman filter for the signal is:

x ̂ s k + 1 = A s 0 K s 0 C s 0 x ̂ s k + B s 0 K s 0 D s 0 u k + K s 0 s 0 k s ̂ k = C s 0 x ̂ s k + D s 0 u k e s k = s 0 k s ̂ k E20

Where x ̂ s k R n s ; s ̂ k R n s ; the residual e s k = e s 1 k e s 2 k e s 3 k e sq k T R q is the residual; and K s R n s xq is the Kalman gain.

Status monitoring: The residuals e k and e s k of the Kalman filters (7) and (20) are employed to monitor the status of the overall system and to detect and isolate faults in the signal and disturbance models and the sensors. The proposed scheme provides a sound framework for developing fault-tolerant systems and condition-based maintenance systems as well.

Advertisement

4. Evaluation on the illustrative example

The proposed two-stage identification scheme and the key properties of the Kalman filter established in the lemmas in Sections 2.3.1–2.3.3 are verified using the illustrative example given in Section 3.1. The results of this illustration are shown below in Figure 3a and b.

Figure 3.

(a) Signal and its estimate; output error and (b) autocorrelations of equation error and the residual and its estimate.

Subfigures A and B, of Figure 3a compare the true step response of the signal and its Kalman filter estimate; subfigures C and D show the output error ϑ k and its estimate.

Remarks: These subfigures confirm the accuracy of the estimates of the signal and the output error (18) established in Lemmas 1 and 2. Subfigures A and B, of Figure 3b show the autocorrelation of the equation error whereas subfigures C and D show the autocorrelations of the residual.

Moreover, these subfigures clearly confirm that the equation error is a colored noise that is whitened by the KF, thus confirming (17) of Lemma 1 and making the KF residual a zero-man white noise process.

Table 1 compares the true and estimated poles of the signal and disturbances models. The estimated poles are obtained from the model reduction techniques employed in the second stage of the two-stage identification scheme.

True poles Identified poles
Signal G ̂ s z 0.7500 ± j 0.3708
0.8500 ± j 0.2784
0.7510 ± j 0.3715
0.8483 ± j 0.2769
Disturbance G ̂ w z 0.1980 ± j 0.8737
0.5663 ± j 0.4114
0.2031 ± j 0.8752
0.5822 ± j 0.3746

Table 1.

Poles of the signal and disturbance models.

Remarks: The estimated poles are close to the true ones, especially those of the signal.

Advertisement

5. Evaluation of the proposed scheme

The management of leakage faults in fluid systems is becoming increasingly important in recent years from the point of view of economy, potential hazards, pollution, and conservation of scarce resources. Leakage in pipes and storage tanks occurs due to faulty joints, aging, excessive loads, holes caused by corrosion and accidents and the like. The process control system is a MIMO system that exhibits turbulence and is modeled as a combination of a signal, (which includes an ideal noise-free height, flow rate, and control input), a disturbance that includes effects of turbulence, and a measurement noise. The augmented model of the signal, and the disturbance, whose output is a sum of the signal, the disturbance, and the measurement noise described by Box-Jenkins model. The transfer matrices of the signal and the disturbance may be totally different from those of the ARMA model, where the signal and the disturbance have identical denominator polynomials.

Physical systems are subject to model uncertainties and are affected to unknown stochastic disturbances such as turbulence and measurement noise. The proposed scheme covers a wider class of systems compared to the laminar flow model proposed. The laminar flow is the flow of a fluid when each particle of the fluid follows a smooth path which results in the velocity of the fluid being constant. The turbulent flow is an irregular flow that exhibits tiny whirlpool regions.

It is assumed that the disturbance is a Gaussian stochastic process and the measurement noise is a zero-mean Gaussian white noise process. The measurement output is, in general, an additive combination of the signal, disturbance and measurement noise. The output error, which is a sum of the disturbance and measurement noise, is assumed to be bounded. The signal and the disturbance are both modeled as outputs of linear time-invariant systems driven by some known input, and a Gaussian zero-mean white noise process, respectively. It is assumed that the signal, disturbance and measurement noise are mutually uncorrelated with each other.

5.1 Physical two-tank fluid system

A benchmark model is a cascade connection of a dc motor and a pump relating the reference input r t and the flow rate f t , the outflow q 0 t and leakage q t is a fourth-order system. The linearized signal model of the nonlinear SIMO system is:

h ̇ h ̇ 2 u ̇ f ̇ = a 1 α a 1 0 b 1 a 2 a 2 β 0 0 1 0 0 0 b m k p 0 b m k I a m h h 1 u f + 0 0 1 b m k p r t E21

Where h , h 2 , u and f are respectively the height of tank 1, the height of tank 2, the control input and the flow rate; a 1 , a 2 , α and β are parameters associated with the linearization process, α is the leakage flow rate, q = αh , and β is the output flow rate, q o = β h 2 . The output is given by:

s k = h k   f k   u k T y k = s k + d k + v k E22

5.2 Simulation of faults in the system

The process control system is interfaced to National Instruments LABVIEW as shown below in Figure 4a. The controller is implemented in LABVIEW.

Figure 4.

(a) Two-tank fluid system controlled by LABVIEW interfaced to a PC and (b) block diagram of process control system.

The two-tank system formed of subsystems and whose faults are to be isolated is shown in Figure 4b. There are four subsystems whose faults are to be isolated. Subsystem 1 is the flow rate sensor γ s 1 , subsystem 2 the height sensor γ s 2 , subsystem 3 the actuator G 1 = G 1 0 γ a where G 1 0 is the fault-free transfer function, and subsystem 4 the leakage fault sensor gain γ . The fault-free cases correspond to γ si = 1 : i = 0 , 1 , 2 , γ a = 1 and γ = 1 . The various subsystems and sensor blocks are all shown in Figure 4b. The first two blocks G 0 and G 1 = G 1 0 γ a , represent the controller and the actuator sub-systems, respectively. The leakage is modeled by the gain γ which is used to quantify the amount of flow lost from the first tank. Thus, the net outflow from tank 1 is quantified by the gain ( 1 γ ). Since the two blocks G 2 0 and ( 1 γ ) cannot be dissociated from each other, they are fused into a single block labeled G 2 = G 2 0 1 γ . The physical two-tank system is controlled using LABVIEW which acquires the flow rate, and the height sensor outputs. The controller is implemented in LABVIEW and the controller output drives the actuator, namely the DC motor and pump combination. A fault in the sensor is introduced by including the emulator block, γ si : i = 0 , 1 , 2 in the control input, flow rate, the height sensors, respectively in LABVIEW software. Similarly, an actuator fault is introduced by including an emulator γ a between the controller output and the input to the DC motor. The leakage fault is simulated by opening the drainage valve of the first tank. The amount by which the valve is opened is modeled by the emulator γ .

The height, flow rate, and control input profiles under various types of faults, are shown in Figure 5. Subfigures A, B and C show profiles for the leakage; subfigures D, E and F show the profiles for the actuator fault; subfigures G, H, and I show the profiles for sensor fault. The fault was simulated by varying the appropriate emulator parameters γ , γ a and γ s 2 , by 0.25, 0.5 and 0.75 times their nominal values representing ‘small’, ‘medium’ and ‘large’ fault sizes respectively.

Figure 5.

(a) Height, flow rate, control: nonlinear and (b) height, flow rate, control: linearized.

The height, the flow rate and the control input profiles under various types of faults are shown in Figure 5a for the nonlinear model, namely the dead-band effect of the actuator on the flow measurements. The measurement outputs are corrupted by the disturbance and measurement noise. Figure 5b show the outputs of the two-stage identification of the linearized signal model. Subfigures A, B and C on the top show height profiles, and subfigures D, E and F in the middle show the flow rate profiles, and G, H, and I at the bottom show the control input profiles under leakage, actuator and sensor faults, respectively. The faults are induced by varying the appropriate emulator parameters to 0.25, 0.5 and 0.75 times the nominal values to represent ‘small’, ‘medium’ and ‘large’ faults. However, by its control design objective, the closed-loop PI controller will hide any fault that may occur in the system and hence will make it difficult to detect it. Also, the physical system exhibits a highly-nonlinear behavior. The flow rate saturates at 4.5 ml/s. The dead-band effect in the actuator exhibits itself as a delay in the output response and saturation of the flow.

Remarks: The two-stage identification is employed to estimate the height, flow rate and the control input; their estimates are shown in Figure 5b. Comparing subfigures D, E, F confirms the superior performance of the identified estimates, thanks to the use of emulators.

Figure 6a shows the residuals and their test statistics, and Figure 6b shows the autocorrelations of the residuals when the system is subject to leakage, actuator, and sensor faults of various degrees such a small, medium and large fault sizes. Subfigures A, B, and C; D, E, and F; and G, H, and I of Figure 6a show the residuals and their statistics when there is a leakage, actuator and sensor faults, respectively. Subfigures A, B, and C; D, E, and F; and G, H, and I of Figure 6b show the corresponding auto-correlations for different fault types.

Figure 6.

(a) The residuals and test statistics and (b) autocorrelations of the residuals.

Remarks: The Bayes decision strategy was employed to assert the fault type, i.e., to decide whether it is either a leakage or an actuator or sensor fault, respectively, using the fault isolation scheme proposed in [8]. The variance of the residual, which is the maximum value of the autocorrelation function evaluated at the origin (i.e. at zero delay), indicates the fault size.

The proposed Kalman-filter-based scheme can detect and isolate small and nascent faults and estimate the fault size. Thanks to the emulator-generated data, it can also provide an accurate prognosis of the status of the system.

Advertisement

6. Conclusions

Emulator-based identification of a wider class of multiple-input and multiple-output system governed by Box-Jenkins model and the associated Kalman filter directly from the input-output data without a-priori knowledge of the disturbance and measurement noise statistics, and the establishment of key properties of estimation of the signal, the output error and their models are developed. The applications include monitoring the status of the system including faults, distinguishing between the variations in the disturbance model and those in the signal model to help diagnose a fault in the system and ensure a low false alarm probability, developing a framework for controlling autonomous vehicles, and meeting the ever-increasing need for fault-tolerant systems. The proposed emulator-based two-stage identification and estimation of the signal and its model were evaluated physical laboratory-scale process control system so as to estimate the signal corrupted by disturbance such as turbulence. Thanks to emulator-based identification, the estimates of the signal were accurate, the detection and isolation of leakage faults were promising and, as such, provide sufficient encouragement and impetus to try the proposed scheme on real-life processes in our future work.

Advertisement

Acknowledgments

The first author acknowledges the support of the Department of Electrical and Computer Engineering, The University of New Brunswick, and both authors acknowledge the help and suggestions of Dr. Haris Khalid.

References

  1. 1. Ljung L. System Identification: Theory for the User. New Jersey: Prentice-Hall; 1999
  2. 2. Doraiswami R, Cheded L. Robust Kalman filter-based least squares identification of a multi variable system. In: IET Control Theory and Applications. The Institution of Engineering and Technology; 2018. pp. 1-11. ISSN 1751-8644, www.ietdl.org
  3. 3. Doraiswami R, Diduch C, Stevenson M. Identification of Physical Systems: Applications to Condition Monitoring, Fault Diagnosis, Soft Sensor and Controller Design. United Kingdom: John Wiley and Sons; 2014. ISBN 9781119990123
  4. 4. Doraiswami R, Cheded L. Kalman filter for fault detection: An internal model approach. IET Control Theory and Applications. 2012;6(5):1-11
  5. 5. Mendel J. Lessons in Estimation Theory in Signal Processing, Communications and Control. Englewood Cliffs, New Jersey: Prentice-Hall; 1995. ISBN0-13120981-7
  6. 6. Doraiswami R, Cheded L. A unified approach to detection and isolation of parametric faults using a Kalman filter residuals. Journal of Franklin Institute. 2013;350(5):938-965
  7. 7. Doraiswami R, Cheded L. Linear parameter varying modelling and identification for condition-based monitoring of systems. Journal of Franklin Institute. 2015;352(4):1766-1790
  8. 8. Doraiswami R, Cheded L. Robust fault tolerant control using an accurate emulator-based identification technique. International Journal of Control. 2017. ISSN: 0020-7179 (Print) 1366-5820 (Online). Available from: http://www.tandfonline.com/loi/tcon20. http://dx.doi.org/10.1080/00207179.2017.1318452
  9. 9. Di Ruscio D. Closed and open loop subspace identification of the Kalman filter. Modelling Identification and Control. 2009;30(2):71-86. ISSN: 1890-1328
  10. 10. Qin JS. Overview of subspace identification. Computer and Chemical Engineering. 2006;30:1502-1513
  11. 11. Jansson M. A New Subspace Method for Open and Closed Loop Data. Prague, Czech Republic: IFAC World Congress; 2005
  12. 12. Kumaresan R, Tufts DW. Estimating exponentially damped sinusoids and pole-zero mapping. IEEE Transactions on Acoustics, Speech and Signal Processimng. December 1982;30(6):833
  13. 13. Porat B, Friedlander B. On the accuracy of the Kumaresan-tufts method for estimating complex damped exponentials. IEEE Transactions on Acoustics, Speech, and Signal Processing. 1987:35(2):231-235
  14. 14. Forssell U, Ljung L. Closed loop identification revisited. Automatica. 1999;35(7):1215-1241
  15. 15. Mallory GWJ, Doraiswami R. A frequency domain identification scheme for control and fault diagnosis. ASME Transactions on Dynamics, Measurement, and Control. 1997;119:48-56
  16. 16. Doraiswami R. A two-stage identification with application to control, feature extraction, and spectral estimation. IEE Proceedings: Control Theory and Applications. 2005;152(4):379-386
  17. 17. Doraiswami R, Cheded L. High-order least squares identification: A new approach. In: Proceedings of the International Conference of Control, Dynamic Systems and Robotics. Ottawa, ON, Canada; 2014
  18. 18. Gugercin S, Antoulas A. A survey of model reduction by balanced truncation some new results. International Journal of Control. 2004;47(8):748-766
  19. 19. Han J, Zhang H, Wang Y, Liu Y. Disturbance observer-based fault estimation and dynamic output feedback with local nonlinear models. ISA Transactions. 2015;59:114-124

Written By

Rajamani Doraiswami and Lahouari Cheded

Submitted: 17 March 2018 Reviewed: 02 October 2018 Published: 30 April 2019