Open access peer-reviewed chapter - ONLINE FIRST

Application of the Kalman Filter in Monitoring, Diagnosis, and Fault Parrying Problems for Observable Dynamical Systems

Written By

Alexander Chernodarov

Submitted: 28 December 2023 Reviewed: 21 January 2024 Published: 19 June 2024

DOI: 10.5772/intechopen.1005411

Applications and Optimizations of Kalman Filter and Their Variants IntechOpen
Applications and Optimizations of Kalman Filter and Their Variant... Edited by Asadullah Khalid

From the Edited Volume

Applications and Optimizations of Kalman Filter and Their Variants [Working Title]

Asadullah Khalid, Arif Sarwat and Hugo Riggs

Chapter metrics overview

5 Chapter Downloads

View Full Metrics

Abstract

This section is devoted to approaches for the formation, estimation and application of diagnostic parameters, as well as stochastic models in monitoring problems. The procedures for increasing the reliability and depth of dynamic systems monitoring are substantiated based on the decomposition of diagnostic models, the use of combined statistical criteria and the processing of observations in “forward” and “backward” time using the extended Kalman filter. The possibility of detecting and parrying anomalous observations using robust filtering procedures is shown. Such procedures are based on the use of an influence function that establishes the level of confidence for formed observations. U-D modification of the Kalman filter with an influence function in the observation selection loop is presented. Here U is an upper triangular matrix with unit diagonal elements and D is a diagonal matrix. Results of the mathematical simulation are given.

Keywords

  • dynamic system
  • monitoring
  • diagnosis
  • Kalman filter
  • consent criteria
  • decisive rules

1. Introduction

At present, the problem of increasing the reliability of dynamical systems (DSs), which are part of objects of various physical natures, still remains topical. The solution to the above problem can be based on the application of analytical approaches to fault detection and isolation (FDI). In observable dynamic systems, the FDI problem is solved using external information, diagnostic models, and decision rules. When constructing diagnostic models, a mathematical description of the reference (unperturbed) DS and the actual (perturbed) DS is used. The reference vector Y(t) and the actual vector Yat of state parameters correspond to such DSs. The dynamics of changes in these vectors are described by the following differential equations:

dYt/dt=Ẏt=FYt,E1
Ẏat=FYat+Gtξt,E2

where F(…) is a matrix of variable coefficients characterizing the dynamics of changes in DS parameters; ξt is a vector of random deviations of DS parameters from the required values; EξtξTtτ=Qtδtτ is the covariance matrix; Q(t) is a matrix of weighting coefficients; δtτ is the delta-function; E […] is the operator of mathematical expectation; Gt is a matrix of variable coefficients characterizing the dynamics of changes in random deviations.

Taking into account relations (1) and (2), the equation for DS errors will have the form

dxt/dt=ẋt=Atxt+Gtξt,E3

where xt=ΔYt=YatYt is the vector of DS errors; At=FYt/YYt=Yat is a matrix of variable coefficients characterizing the dynamics of changes in DS errors.

The estimates x̂t of DS errors can be obtained by processing observations z(t) using the extended Kalman filter (EKF) [1]:

zt=hYathYtSEI,E4

where SEI is the sensor of external information; hYtSEI is the observable parameters formed by SEI and having a model hYtSEI=hYt+ϑt; and ϑt is a vector of random deviations of SEI parameters from the reference with covariance matrix EϑtϑTtτ=Rtδtτ.

The mathematical model of observation (4) has the form:

zt=Htxt+ϑt,E5

where Ht=hYt/YYt=Yat is the matrix for the relation of observed parameters and the vector of DS errors.

The extension of the traditional linear Kalman filter is associated with its application in optimization problems of nonlinear dynamic systems of the form (2)(5). The observable DS with EKF in the errors estimation loop is shown in Figure 1. Ẑ is the predicted estimate of observations; ^ is the symbol for estimate; and ν is the vector of residuals. Such a vector can be used to form diagnostic parameters for monitoring observed dynamical systems with EKF in the fault detection loop.

Figure 1.

An observable dynamical system with the EKF in an error estimation loop.

Discrete observations Zi=Zti are related with the following DS error equation:

xi=Φixi1+Γiξi1,E6

where xi=xti; Φi matrix is the solution of the homogeneous part of Eq. (3).

dΦtti1/dt=Φ̇i=AtФtti1;Фti1ti1=I;

I is an identity matrix; Γi matrix is the solution of the inhomogeneous part of Eq. (3):

dΓtti/dt=Г̇i=AtГtti1+Gt;Γti1ti1=0.

The use of EKF for FDI problem can be based on the estimation of DS errors. However, in this case, it is necessary to associate each DS state with its own equation of the form (3) and its own filter. There is a need to create a bank of filters [2], which is difficult to implement in practice. Therefore, it is advisable to form FDI algorithms based on the serviceable states of the DS. Then, the diagnostic parameters will reflect the deviation of the actual state of the DS from the serviceable one.

Advertisement

2. Fault detection by the χ2 Chi-Square criterion

The functioning of the EKF is based on the use of a vector of residuals (see Figure 1)

νi=ZiẐi=ν1iν2iνjiνliT,E7

where Ẑi=Hix̂i/i1 and x̂i/i1=Φix̂i1/i1 is the predicted values of estimates of the DS errors at the instant t=ti of time after Zi1 observations have been processed.

The statistical properties of the residuals make it possible to form diagnostic parameters based on them. It is known [2] that the serviceable state of the DS and acceptable error values correspond to a vector of residuals that has a Gaussian distribution with zero mathematical expectation and a covariance matrix αi, that is,

νiN0αi2.E8

Taking into account the principle of orthogonality of optimal estimates [1] EeiϑiT=0, it can be shown that

αi2=EνiνiT=HiPi/i1HiT+Ri,E9

where Pi/i1=Eei/i1ei/i1T=ΦiPi1/i1ΦiT+ΓiQi1ΓiT is the predicted covariance matrix; ei/i1=xix̂i/i1.

In traditional EKF, all elements of the νi vector are processed simultaneously. Taking into account such processing, to monitor the state of the DS, it is necessary to check the l-dimensional vector of residuals for Gaussianity. In practice, the solution of this problem is a matter of some difficulty. Therefore, using the residual vector, more compact diagnostic parameters are formed. One of these parameters is formed by convolving the vector (7) and the covariance matrix (9)

Ji=νiTαi2νi.E10

In the quadratic form (10), the elements of the αi matrix are the normalizing coefficients that take into account the required statistical characteristics of the residuals. It can be shown [2] that if the νi vector has a Gaussian distribution, then the quadratic form (10) is distributed according to the χ2 law with l degrees of freedom, namely,

Jiχ2l2l,E11

that is, the l dimension of the νi vector is equal to the mathematical expectation of the Ji parameter and to one-half of its variance.

Relation (11) reflects the necessary condition for the absence of violations in the DS. Quadratic form (10) includes all l components of the νi vector, so it can be considered as a DS generalized state parameter. It is believed that if the Ji parameter is within acceptable limits, then there are no violations in the DS. The tolerances for the Ji parameter are formed taking into account the numerical characteristics of the χ2 distribution and a given level of significance for the goodness-of-fit criterion. Such conditions can be performed taking into account the properties of the a(l) quantile for the χ2 distribution, namely,

PJi>JT,al,E12

where P is a probability measure; JT,al is a tabular value of the Ji parameter for the quantile a(l) with the number l of degrees of freedom; a(l) is the quantile of a order, a01.

An analysis of the tabular data shows that quantile (12) with highly accuracy reflects the three mean square deviations rule (the 3σ rule). This rule finds application in the processing of scalar parameters ξ distributed according to the Gauss law and boils down to the following. The probabilities of deviations of such parameters from the mathematical expectation are characterized by the following values [3]:

PξEξkσ=0.3173,k=1;0.0455,k=2;0.0027,k=3.E13

For k = 3, relation (13) reflects the 3σ rule for the Gaussian distribution law. Taking into account the probabilistic characteristics (13) and the 0.02(l) quantile of [4], it can be argued that, with a confidence probability of 0.98, the necessary condition for the Ji parameter to belong to the χ2 distribution is associated with the fulfillment of the following inequality:

Jiγl2=EJi+3VJi=l+32l,E14

where V […] is an operator for variance.

Taking into account the γl2 tolerance, the monitoring of DS using the diagnostic parameter Ji for the χ2 criterion is reduced to checking the following conditions:

ifJiγl2,therearenofaults in theDS;ifJi>γl2,therearefaults in theDS.E15

Monitoring using parameter (10) allows you to determine the state of the DS as a whole without diagnosing which element of the observation vector the violation is associated with. This approach is traditional [5, 6] for DS monitoring. In practice, there is a need to diagnose DS with a depth up to the element of the observation vector.

Advertisement

3. Fault diagnosis by the χ2 Chi-Square criterion

The problem of DS diagnosis with a depth of up to elements of the vector Zi1 can be solved if the observation errors ϑi are statistically independent (uncorrelated), that is, if the matrix Ri in relation (9) is diagonal. If the observation errors are mutually correlated, their preliminary decomposition can be performed. Such a decomposition can be performed taking into account the structure of the covariance Ri. As the specified matrix is square, positive definite and symmetric, it can be represented in the following form [7]:

Ri=SiDiSiT,E16

where Si is an upper triangular matrix with unit diagonal elements and Di is a diagonal matrix.

Then, the vector of observations Si1Zi will have elements with mutually uncorrelated noise. Indeed, the following model corresponds to such observations:

Si1Zi=Si1Hixi+Si1ϑiE17

or

Z˜i=H˜ixi+ϑ˜i,E18

where

Z˜i=Si1Zi;H˜i=Si1Hi;ϑ˜i=Si1ϑi.

The covariance matrix has the form:

R˜i=Eϑ˜iϑ˜kT=Si1EϑiϑkTSiT=Si1RiSiT=Si1SiDiSiTSiT=Di,E19

where Di=diagD1iDjiDmi; SiT=Si1T.

Expression (19) shows that Z˜i observations have uncorrelated errors. Next, we will assume that the matrix Ri is diagonal. Then, such observations can be processed using a sequential modification of the EKF [8].

Prediction:

m0=x̂i/i1=Φix̂i1/i1;E20
M0=Pi/i1=ФiPi1/i1ΦiT+ΓiQi1ΓiT.E21

Updating:

νj=zjНjmj1;E22
αj2=HjMj1HjT+Rj;E23
Kj=Мj1HjT/αj2;E24
Mj=KjHjIMj1KjHjIT+KjRjKjT;j=1,l¯;E25
x̂i/i=ml;Pi/i=Ml,E26

where x̂i/i1 and x̂i/i, respectively, are the predicted and updated estimates of the DS error vector at the ith time point, obtained after processing observations Zi1 and Zi;

Pi/i1 and Pi/i covariance matrices of these estimates; mj and Mj, respectively, estimates of the DS error vector and their covariance matrix obtained after processing the jth observation; Hi=H1THjTHlTT is the matrix for the relation of observed parameters with the DS error vector; Hj is the row vector for the relation of the zj observation with the DS errors vector; Rj is the variance of the zj observation error; Qi is a matrix of weighting coefficients characterizing the noise intensities.

Unlike the traditional EKF, in algorithm (22)(26), when calculating the amplification coefficient (24), it is not necessary to invert the covariance matrix (9) for the residual vector (7). Such inversion is replaced by l division operations in Eq. (24). In addition, based on the presented algorithm, it seems possible to form statistical criteria of agreement for scalar diagnostic parameters.

When processing the residual vector (7) element-by-element, algorithm (22)(26) can be supplemented with procedures for analyzing the DS state for each of the l observations. For example, the checking of DS on the jth observation can be performed using a normalized residual βj=νj/αj. The statistical properties of such residuals make it possible to form scalar goodness-of-fit criteria. By analogy with the diagnostic parameter (10), for l = 1, the necessary condition for the absence of violations in the DS for each of the observations will have the form:

βj2χ212E27

or by the rule of 3σ, for the quantile a(1) = 0.02,

βj2γ12=Eβj2+3Vβj2=1+325.2.E28

Based on reference data [4], quantiles a˜1=0.01 and a¯1=0.001 can be matched to the extended γ˜127.6 and γ¯1211.8 tolerance values. However, when the tolerances are expanded, the probability of fault detection errors associated with a “false alarm” increases.

Taking into account the γ12 tolerance, the monitoring of DS using the diagnostic parameter βj2 for the χ2 criterion is reduced to checking the following conditions:

ifβj2γ12,therearenofaults in thejthchannel of theDS;ifβj2>γ12,therearefaults in thejthchannel of theDS.E29

Thus, the decomposition of the residual vector allows the formation of scalar goodness-of-fit criteria. The use of such criteria makes it possible to increase the depth of diagnosis.

Advertisement

4. Fault diagnosis by the ϑ2 fisher criterion

The use of the χ2 criterion makes it possible to detect current DS faults in real time. In practice, there is also a need to analyze the DS operation over a certain period of time.

Appropriate diagnostic parameters can be formed based on retrospective data. Technology (22)(26) for sequential processing of observations allows the formation of scalar diagnostic parameters based on a sample of residuals on a sliding time interval. To do this, you can use the ergodic properties of the estimates obtained by the Kalman filter, the parameters of which are a priori adjusted for the proper operation of the DS. These parameters are the variances of the innovation sequences (22) in each of the l observation channels. The predicted value of the variance of the residual (22) at the ith moment of time is determined by the ratio (23), and its estimate α̂ji is based on a sample of N values, that is,

α̂ji2=1N1k=iN+1iνjkν¯ji2;j=1,l¯,E30

where ν¯ji=1Nk=iN+1iνjk is an estimate of the mathematical expectation of the jth residual at the ith instant of time. Taking into account condition (8), the absence of faults can be associated with the desired value of the mathematical expectation of the residuals, namely, ν¯ji=0. As a parameter that characterizes the DS state on the time interval T=tiN+1ti, the ratio of the actual variance α̂j2 to its predicted value αj2 can be accepted, that is,

Fj=α̂j2/αj2.E31

It is known [4] that when the condition (32) presented below is satisfied, parameter (31) has ϑ2 or Fisher distribution, namely,

Fjϑ2bc,E32
whereb=N/N2andc=4NN1/N22N4E33

are tabular values, respectively, of the mathematical expectation and variance for the Fj parameter.

Taking into account the 3σ rule and a = 0.02 quantile, condition (32) can be represented as

Fjη12=EFj+3VFj=b+3c.E34

When the tolerance η12 is taken into account, the DS diagnosis according to the ϑ2 criterion using the Fj parameter is reduced to checking the following conditions:

ifFjη12,E35

there were no faults in the DS for the jth channel of observations on the time interval T=tiN+1ti;

ifFj>η12,E36

there were faults in the DS for the jth channel of observations on the time interval T.

The procedures (35) and (36) complements the test (29) in order to increase the reliability of the diagnosis.

Advertisement

5. Fault diagnosis by the χ2 and the ϑ2 combined criterion

When diagnosing DS, the problem of detecting short-term outliers against a background of failures arises. If short-term outliers are recognized, then they can be parried without affecting the operation of the DS. The problem of recognizing the type of violations can be solved by combining the χ2 and the ϑ2 goodness-of-fit tests. Indeed, DS diagnosis based on the AA criterion makes it possible to immediately detect all violations, both random short-term outliers and failures. Indeed, DS diagnosis based on the χ2 criterion makes it possible to immediately detect all violations, both random short-term outliers and failures. At the same time, DS diagnosis based on the ϑ2 criterion is implemented using a set of residuals on a sliding time interval. Random short-term outliers that may appear on a sliding time interval when calculating the Fj parameter are averaged and, as a rule, do not affect to the diagnosis results of the ϑ2 test. On the other hand, the appearance of gradual and sudden failures, which are characterized by constant deviations of residuals from their nominal values, lead to the Fj parameter going beyond the tolerance limits. Thus, if a violation in the jth observation was detected according to both criteria, then most likely a permanent failure occurred in the DS, and if only according to the χ2 criterion, then it was a random short-term outlier.

From the considered properties of the combined criterion of consent, the following algorithm for detecting and parrying violations in the observed DS is as follows:

  • If there is no discrepancy between the diagnostic parameters relative to the tolerances, the residuals are processed using EKF.

  • Anomalous observation signals detected by the χ2 criterion are excluded from processing or processed with certain robust confidence coefficients.

  • Violations detected by both criteria χ2 and ϑ2 are countered by connecting a redundant channel.

The procedures presented above allow for the diagnosis of DS with a depth of up to the element of the observations vector. At the same time, in practice, it becomes necessary to perform diagnosis with a depth up to the element of the DS errors vector.

Advertisement

6. Fault diagnosis based on joint filtering and smoothing procedures

The diagnosis of DS with a depth up to the elements of the errors vector can be implemented on the basis of joint procedures for filtering observations in “forward” time and smoothing estimates in “backward” time. By analogy with expression (10), the diagnostic parameter, which reflects the deviations of error vector estimates from their predicted root-mean-square values, can be represented by a quadratic form:

Ji=νi/NTΔPi1νi/N,E37

where νi/N=δfiδsi=Φi+11x̂i+1/Nx̂i/i;

δfi=xix̂i/i; δsi=Φi+11xi+1x̂i+1/N;

ΔPi=Pi/i+Φi1Pi+1/NΦiT;

x̂i/i and x̂i/N are estimates of the DS error vector at the ith moment of time, obtained, respectively, at the filtering and smoothing stage; Pi/i,Pi/N are the covariance matrices of the above estimates; ФТ=Ф1Т. Φi1 is the inverse transition matrix, which is determined from the solution of the differential equation:

dΦ1tti1/dt=Ф1tti1At;Φ1ti1ti1=I.E38

Stable smoothing (∣δ∣ < 3σ), which reflects the absence of anomalous observations signals, is characterized by the following distribution of the residual νi/N and the quadratic form Jsi:

νi/NN0ΔPi;Jsiχ2n2n,E39

where n is the dimension of the DS error vector.

Taking into account the properties of the χ2 distribution and the 3σ rule, it is possible to formulate the necessary conditions for the absence of violations for the DS in general and for the jth element of the error vector in particular

Jsin+32n;E40
Jsi/j=Jsi/j1+ν˜i/Nj2/ΔDijγj2=j+32j,E41

where ν˜i/N=ΔUi1νi/N; j=1,n¯; ΔUi1;ΔDi1 are an upper triangular matrix with identity diagonal and a diagonal matrix, respectively. These matrices are obtained by the following orthogonal transformation:

ΔPi1=ΔUiTΔDi1ΔUi1;E42

ΔDij1 is the jth element of the diagonal matrix ΔDi1.

Taking into account decomposition (42) and the statistical properties of the Fisher distribution [4]:

Fsj=α̂i/j/ΔDi/jϑ2bc,E43

it is possible to formulate the necessary condition for the operable DS state (“absence” of failures) for the jth element of the error vector, that is,

Fsjηj2=b+3c,E44

where α̂i/j is an estimate of the variance of the residual ν˜i/Nj on a sliding time interval; b and c are tabulated values (33) of the mathematical expectation and variance for the parameter Fj.

Diagnosis based on joint procedures for optimal filtering and smoothing assumes the relationship between prediction and interpolation models of the DS error vector:

xi+1=Φi+1xi+Γi+1ξiprediction one step ahead ati=i0,if¯;E45
xi=Φ1i+1xi+1Γi+1ξiinterpolation one step back ati=if,i0¯.E46

It should be noted, however, that the traditional RTS (Rauch-Tung-Striebel) algorithm [9] smooths out the predicted estimates x̂i/i1, the reliability of which is significantly lower than the updated estimates x̂i/i.

The value x̂i/iN for the smoothed estimate of the xi error vector can be found by solving the following optimization problem using the least squares method

x̂i/iN=argminxiJi/s,E47
whereJi/s=0.5{xix̂i/i+1NTPi/i+1N1xix̂i/i+1Na++xix̂i/iTPi/i1xix̂i/i}b;E48

x̂i/i+1N; Pi/i+1N is an estimate of the error vector interpolated at the ith moment of time and its covariance matrix.

The quadratic form (48) combines components “a” and “b”, which characterize interpolation and filtering errors at the ith moment of time. It should be noted that component “b” can also be formed for predicted estimates. However, in this case, smoothing errors will significantly depend on the frequency of registration these estimates.

Applying the rule of differentiation with respect to the parameter for the quadratic form (48), we can obtain the following solution to problem (47)

JixiTxi=x̂i/iN=Pi/i+1N1x̂i/iNx̂i/i+1N+Pi/i1x̂i/iNx̂i/i=0.E49

From (Eq. (49)), an expression can be obtained for a smoothed estimate of the error vector

x̂i/iN=x̂i/i+1NPi/iNPi/i1x̂i/i+1Nx̂i/i,E50
wherePi/iN=Pi/i+1N1+Pi/i11;E51
Ki/N=Pi/iNPi/i1is the optimal amplification coefficient.E52

Applying the matrix inversion lemma [10] to expression (51), it can be represented in expanded form:

Pi/iN=Pi/i+1NPi/i+1NPi/i+1N+Pi/i1Pi/i+1N.E53

Substituting expression (53) into relation (52), we obtain the following procedure for calculating the optimal amplification coefficient

Ki/N=Pi/i+1NPi/i+1N+Pi/i1.E54

The expression for the interpolated estimate follows from the ratio (46)

x̂i/i+1N=Фi1x̂i+1/i+1N.E55

The difference between relations (46) and (55) determines the equation for interpolation errors

ei/i+1N=Фi1ei+1/i+1NГiξi,E56

where ei/i+1N=xix̂i/i+1N; ei+1/i+1N=xi+1x̂i+1/i+1N.

The following covariance matrix corresponds to the interpolation error vector

Pi/i+1N=Eei/i+1Nei/i+1NT=Фi1Pi+1/i+1N+ГiQiГiTФiT.E57

Combining (Eqs. (50), (53)(55), (57)) into a single structure, we obtain an algorithm for optimal smoothing of error vector estimates

Interpolation:

x̂i/i+1N=Фi+11x̂i+1/i+1N;E58
P˜i/i+1N=Pi+1/i+1N+Гi+1QiГi+1T.E59

Updating:

νi/N=x̂i/i+1Nx̂i/i;E60
Pi/iN1=Фi+1TP˜i/i+1N1Фi+1+Pi/i1;E61
Ki/N=Pi/iNPi/i1;E62
x̂i/iN=x̂i/i+1NKi/Nνi/N,E63

To implement the algorithm (58)(63), it is necessary to register estimates x̂i/iof the DS error vector and their covariance matrices Pi/i1 obtained by the EKF in the process of functional monitoring of the DS. The matrix Pi/i1 is determined from the solution in “forward” time according to the following equations:

Pi/i1=Pi/i11+HiTRi1Hi;E64
Pi/i11=ΦiPi1/i1ΦiT+ΓiQi1ΓiT1.E65

The implementation of (Eqs. (58)(65)) is associated with computational difficulties due to the need to determine the inverse covariance matrices Pi/i1 and Pi/iN1. These difficulties can be significantly reduced with the sequential implementation of smoothing algorithms. With this implementation, the inversion of covariance matrices of n×n dimension is replaced by a sequence of division operations. Such a replacement is based on the transformation of (Eq. (61), (64), (65)) according to the matrix inversion lemma [10]. For (Eq. (65)) it will have the form:

Pi/i11=P˜i/i11P˜i/i11ΓiΓiTP˜i/i11Γi+Qi1/i111ΓiTP˜i/i11=P˜i/i1+j=irΓjΓjTQjj1,E66

where Γj is the jth column of matrix Γi, which has n×r dimension; Qjj is the jth element of the diagonal matrix Qi, which has r×r dimension.

Hence, the following procedure for the sequential implementation of Eq. (66) is valid

M01=ΦiTPi1/i11Φi1;E67
Kj=Mj11Γj/ΓjTMj11Γj+Qjj1;E68
Mj1=KjΓjTIMj11KjΓjTIT+KjQjj1KjT;E69
Pi/i11=Mr1;j=1,r¯.E70

Transformations similar to (66)(70) will also be valid for Eq. (61), if the covariance matrix of filtering errors is presented in the following form

Pi/i1=Ui/iTDi/i1Ui/i1,E71

where Ui/i is an upper triangular matrix with unit diagonal elements; Di/i is a diagonal matrix. These matrices form the basis of the U-D filter [11].

Taking into account decomposition (71) and preliminary decomposition of residuals ν˜i/N=Ui1νi/N, a sequential modification of the smoothing and diagnostic algorithm can be formed.

Interpolation:

m0=x˜i/i+1N=Φi+11x̂i+1/i+1N;E72
νi/N=x˜i/i+1Nx̂i/i;E73
MWGSW¯0=Φi+11Ui+1/NΦi+11Γi+1D¯0=diagDi+1/NQiU˜0D˜0;E74

Smoothing:

fj=U¯j1U˜j1;Vj=D˜jfjT;E75
αj2=fjVj+Djj;E76
K˜j=U˜j1Vj/αj2;E77
MWGSW¯j=K˜jfjU˜j1K˜jD¯j=diagD˜j1DjjU˜jD˜j;E78
Kj=U˜jD˜jU˜jTU¯jTDj1;E79
ν˜i/Nj=U¯j1νi/N;E80
mj=mj1Kjν˜j;j=1,n¯;E81
x̂i/N=mn;Ui/iN=U˜n;Di/iN=D˜n,E82

where U¯j1 is the jth row of the matrix Ui/i1; Dj1 is the jth element of the diagonal matrix Di/i1; MWGS is the modified weighted Gram-Schmidt procedure [11], intended to transform the aggregate of matrices W¯j and D¯j, which are n×n+r matrix and n+r×n+r matrix, respectively, into the aggregate of the n×n an upper triangular matrix U˜j with identity diagonal and a diagonal matrix D˜j. The algorithm for implementing the MWGS procedure has the form

dm=j=1n+rW¯mjD¯jW¯mj,m=n,2¯;E83
ukm=j=1n+rW¯mjD¯jW¯mj/dm,W¯kjW¯kjukmW¯mj,j=1,n+r¯k=1,m1¯;E84
d1=j=1n+rW¯1jD¯jW¯1j;E85

ukm and dj are the elements of the required matrix U and D.

To implement the U-D smoothing algorithm, it is necessary to determine the Ui/i1, Di/i1 components of the a posteriori covariance matrix of estimation errors Pi/i1. The traditional U-D algorithm [11] for observation filtering, which includes additional procedures necessary to smoothing estimates of the DS error vector, is presented below.

Prediction for filtering:

m0=xi/i1=Φix̂i1/i1;E86
MWGSW¯i=ΦiUi1/i1ΓiD¯i=diagDi1/i1Qi1U0;D0.E87

Prediction for smoothing:

U0T=ΦiTUi1/i1T;D01Di1/i11;E88
fj=ΓjUj1T;Vj=Dj11fjT;E89
Kj=Uj1TVj/fjVj+Qjj1;E90
MWGSLW˜j=KjfjUjTKjD˜j=diagDj1Qjj1UjT;Dj1;j=1,r¯;E91
U0TUi/i1T=UrT;D01Di/i11=Dr1.E92

Filtering:

fj=HjUj1;Vj=Dj1fjT;E93
αj2=fjVj+Rj;E94
Kj=Uj1Vj/αj2;E95
mj=mj1+Kjνj;E96
MWGSW¯j=KjfjUj1KjD¯j=diagDj1RjUj;Dj.j=1,l¯.E97

Updating for smoothing:

MWGSLW˜j=Uj1THjTD˜j=diagDj11Rj1UjT;Dj1;j=1,l¯;E98
Ui/iT=UlT;Di/i1=Dl1,E99

where MWGSL is a similar procedure to MWGS for forming matrices: the lower triangular with unit diagonal elements UjT and diagonal Dj1. Eq. (88)(92), (98), (99) are redundant with respect to the basic U-D filtering algorithm [11].

Diagnosis in “backward time” is performed when violations detected in “forward time”.

Advertisement

7. Analysis of the results of mathematical modeling of diagnosis procedures based on filtering and smoothing of observations

The Schuler pendulum [12], which includes a gyro-accelerometer (G-A) system moving in a gravitational field along a sphere of R radius, is considered as an object of mathematical modeling. The invariance of the modeled vertical (radius vector) to the movement of the base of the G-A system along the sphere when changing the velocity V and angular position φ is ensured by precession of the gyros. For this purpose, a φ̇=V/R signal that is proportional to angular velocity of G-A system motion relative to the sphere is fed to the gyros moment sensor. When the vector xt=ΔVδΔaΔωT of errors of such a system is observed from the velocity signals ZVt=VGAtVSEIt the parameters of Eqs. (3) and (5) will have the following form

At=0g101/R001001/τa00001/τω;E100
Gt=0000000000σa2/τa0000σω2/τω;E101
Ht=1000T,E102

where g is the acceleration of gravity; δ is an angular error of determining the vertical; ΔV is an error of velocity reckoning; Δa is an accelerometer error; Δω is the gyro drift; τa,τω are the correlation time for the accelerometer error and gyro drift, respectively; σa,σω are mean-square values of errors of the accelerometer and gyro, respectively; Δ stands for an error.

In Figures 2 and 3, characteristic results of the studies of an algorithm for G-A system diagnosis from the recorded data are presented. At the 500th second, the accelerometer failure corresponding to the imitation bias was simulated. Such a failure indirectly makes itself evident during filtering for the velocity observation channel when the diagnostic parameter βV2 is above upper tolerance.

Figure 2.

Estimates of the accelerometer error.

Figure 3.

Estimates of the gyro drift.

In post processing the recorded estimates and during the diagnosis by rules (41) and (44), it is possible to determine which of the sensors – an accelerometer or a gyro – has most probably caused a fault.

Figures 2 and 3 show the dynamics of variation of the estimates of both the bias of the ax accelerometer output signal and the ωx gyro drift when velocity observations are processed in “forward time” and when the estimates mentioned are refined in “backward time”. During the diagnosis from the recorded data, a failed accelerometer is isolated if the generalized parameters JSaj (the χ2 test) and FSaj (the ϑ2 test) are above upper tolerances (see Figure 2). It can also be seen (see Figure 3) that the accelerometer failure has had an insignificant influence on the variation of the generalized parameters JSωj and FSωj, which characterize the status of the ωx gyro.

Thus, combined processing of observations in “forward” and “backward” time enables us to solve diagnosis problems with a depth of the component of the state vector of a dynamical system.

Advertisement

8. Analysis and parrying of violations by robust filtering

The analysis and parrying of anomalous observations can be performed based on a robust modification of the U-D filter. This modification relies on the application of the influence function ψ, which defines the level of confidence in incoming observations. Such a function can be formed for the normalized residual βj=νj/αj, where αj is a scaling parameter which may be the mean square deviation Rj. For the residual βj, one can not only form the robust-likelihood function ρ(β) but also perform the optimization of the estimates

x̂i=argminxii=i0+1ifρβi,E103

where ρβ=lnfβ; fβ is a probability density function (PDF) [3].

The solution of the problem (116), in view of the constraint xiФixi1Гiξi1=0, is an algorithm for robust estimation.

Prediction: Eq. (97), (98)

Tuning:

βj=νj/αj;ψj=ψβj;ψj=ψβj;E104

Updating:

fj=HjUj1;Vj=Dj1fjT;E105
α˜j2=fjVjψ+αj2;E106
Kj=Uj1Vj/α˜j2;E107
mj=mj1+Kjαjψj;E108
MWGSW¯j=KjfjψjUj1KjD¯j=diagDj1αj2ψjUj;Dj;E109
Ui/i=Ul;Di/i=Dl;x̂i/i=ml;j=1,l¯,E110

where ψj=ψβj=∂ρβ∂ββ=βj and ψj=ψβj=2ρββ2β=βj.

The influence function [13, 14] ψj and its derivative ψj can be formed taking into account “a priori” assumptions about the laws of distribution of the useful signal and noise. The choice of the values of above functions is based on the necessary conditions [15] for the filter to be divergence-free, namely, the diagnostic parameter must have a Gaussian distribution, that is, βN01; in addition, the rule of 3σ is fulfilled [3] for the probability P that a random variable having the Gaussian distribution will be on the interval 3σ3σ, that is,

PβEβ3σ=0.0027.E111

For the random variable β, the rule (111) can be written as Pβ<3=0.9973.

Thus, the correct functioning of the DS can be matched by the condition β<3 and also the following values of the functions:

fgβ=2π12exp0.5β2;ρgβ=0.5ln2π+0.5β2;ψgβj=βj;ψgβj=1.E112

The following functions can be made to correspond to such a distribution and to off-design conditions of the filter operation.

flβ=0.5expβ;ρlβ=ln2+β;ψlβj=1;ψlβj=0.E113

The vagueness of boundaries between anomalous and conditioned signals can be taken into account based on the convolution of the Gauss and Laplace PDF. Such a convolution can be performed using the following moment generating functions (MGFs) [16]:

MlgT=MlTMgT=1T21expT2/2,E114

where M(T) is an MGF; T is generally a complex number; and KT=lnMT is a cumulant generating function.

Relying on the results of [14], the following relations can be shown to hold for the normalized residual βj:

ψβj=Т0+K3T/2T=T0;E115
ψβj=1+K3T/TT∂βT=T0,E116

where Т0 is the value of the argument at a saddle point for which the following equality is valid:

КТТβj=0.E117

In view of the approximation ln1T2T2 and relations (112)(117), the parameters ψjand ψj take the form ψlgβj=β/3; ψlgβj=1/3.

The influence function which reflect the considered assumptions is shown in Figure 4.

Figure 4.

Diagram for the control of an estimating filter with an influence function.

In this figure, a typical influence function is denoted by a dotted line, too. However, the interrelation between such a function and the parameter ψj is not obvious.

The need to parry anomalous observations can be illustrated by the results of mathematical modeling of a G-A system, which is described by Eq. (100)(102). Such results are presented in Figures 5 and 6, where the true Δδ=хδх̂δ and root mean square (RMS) error σδ=Pδ of vertical determination are shown. An anomalous observation signal ZVt was simulated. In the case of a sequential EKF (20)(26), estimation errors are shown in Figure 5, and for a robust filter (97)(98) and (100)(117), such errors are presented in Figure 6.

Figure 5.

Estimates obtained by the EKF.

Figure 6.

Estimates obtained by the robust filter.

It can be seen that in the absence of protection from anomalous observations, the divergence of the EKF is possible when the actual estimation errors Δδ considerably differ from their predicted mean square values σδ.

Advertisement

9. Conclusions

In this chapter, some approaches to monitoring and diagnosis of dynamic systems using the extended Kalman filter were presented. The considered approaches are based on the application of the DS error model, the formation and processing of residuals between the observed parameters and their predicted values. The statistical properties of residuals make it possible to form diagnostic parameters and the corresponding criteria of agreement. For the considered criteria, formalized tolerances can be generated that make it possible to detect failures and failures in the DS.

The conducted studies have confirmed the following possibilities of the proposed approaches.

  • Detection and exclusion or processing with certain confidence coefficients of anomalous observations.

  • Selection of outliers against the background of gradual biases of observations.

  • Improving the reliability of monitoring based on combined goodness-of-fit tests.

  • Detection of faults in observable dynamical systems with a depth of the component of the state vector on a basis of the joint procedures of optimal filtering and smoothing of experimental data.

References

  1. 1. Maybeck P. Stochastic Models, Estimation and Control. Vol. 2. New York: Academic Press; 1982
  2. 2. Gertler J. Fault Detection and Diagnosis in Engineering Systems. New York: Marcel Dekker; 1998
  3. 3. Skorokhod A, Hoppensteadt F, Salehi H. Random Perturbation Methods with Applications in Science and Engineering. New York: Springer-Verlag; 2002
  4. 4. Korn G, Korn T. Mathematical Handbook. New York: Mc Graw-Hill; 1968
  5. 5. Tanil C, Khanafseh S, Joerger M, et al. An INS monitor to detect GNSS spoofers capable of tracking vehicle position. IEEE Transactions on Aerospace and Electronic Systems. 2018;54(1):131-143
  6. 6. Yu Z, Zhang Q, Yu K, Zheng N. A state-domain robust Chi-Square test method for GNSS/INS integrated navigation. Journal of Sensors. 2021;2021:1745383, 8 pages. DOI: 10.1155/2021/1745383SINS
  7. 7. Wilkinson J, Reinsch C. Handbook for Automatic Computation. Vol. II, Linear Algebra. Berlin: Springer-Verlag; 1971
  8. 8. Bierman G. Factorization Methods for Discrete Sequential Estimation. New York: Academic Press; 1977
  9. 9. Rauch H, Tung F, Striebel C. Maximum likelihood estimates of linear dynamic systems. AIAA Journal. 1965;3(8):1445-1450
  10. 10. Sage A, White C. Optimum Systems Control. New Jersey: Prentice – Hall; 1977
  11. 11. Thornton C, Bierman G. UDUT covariance factorization for Kalman filtering. In: Control and Dynamic Systems. New York: Academic Press; 1980. pp. 117-247
  12. 12. Rogers R. Applied Mathematics in Integrated Navigation Systems. 2nd ed. Reston: AIAA Education Series; 2003
  13. 13. Abutaleb A, Papaiouannou M. New results in Sridhar filtering theory: The discrete case. Journal of Optimization Theory and Applications. 1990;64(1):5-14
  14. 14. Hampel F, Ronchetti E, Rousseeuw P. Robust Statistics: The Approach Based on Influence Functions. New York: Wiley; 1986
  15. 15. Fitzgerald R. Divergence of the Kalman filter. IEEE Transactions on Automatic Control. 1971;16(6):736-747
  16. 16. Wu W. Target tracking with glint noise. IEEE Transactions on Aerospace and Electronic Systems. 1993;29(1):174-185

Written By

Alexander Chernodarov

Submitted: 28 December 2023 Reviewed: 21 January 2024 Published: 19 June 2024