ORIGINAL RESEARCH article

Front. Appl. Math. Stat., 27 November 2020

Sec. Dynamical Systems

Volume 6 - 2020 | https://doi.org/10.3389/fams.2020.585439

On the Relation Between Smooth Variable Structure and Adaptive Kalman Filter

  • Chair of Dynamics and Control, University of Duisburg-Essen, Duisburg, Germany

Abstract

This article is addressed to the topic of robust state estimation of uncertain nonlinear systems. In particular, the smooth variable structure filter (SVSF) and its relation to the Kalman filter is studied. An adaptive Kalman filter is obtained from the SVSF approach by replacing the gain of the original filter. Boundedness of the estimation error of the adaptive filter is proven. The SVSF approach and the adaptive Kalman filter achieve improved robustness against model uncertainties if filter parameters are suitably optimized. Therefore, a parameter optimization process is developed and the estimation performance is studied.

1. Introduction

State estimation plays an important role in the field of control. System states are required for the calculation of state feedback controllers, exact input-/output linearizations, equivalent control, backstepping control, etc. Noise reduction of signals is desirable to improve performance of sliding mode approaches under real conditions. Combined input-state estimation is useful for the estimation and rejection of unknown exogenuous inputs. Additionally, robust model-based fault detection and localization approaches can be designed based on filters.

Related to linear systems minimum variance unbiased [1, 2] and augmented state filters [3, 4] can be used for combined input-state estimation. In case of known uncertainty bounds robust Kalman filters [5] can improve state estimation of uncertain linear systems. In the field of filtering robustness of the state estimation may be improved by minimizing the effect of worst possible energy-bounded disturbances on the estimation error [6]. Multiple-model approaches [7] are a powerful tool for state estimation of uncertain systems. Combining them with the particle filter allows state estimation of nonlinear systems [8].

The smooth variable structure filter (SVSF) introduced in Ref. 9 is an approach for state estimation of uncertain nonlinear systems. Several applications of SVSF can be found in the literature. The filter has been applied to estimate the states and parameters of an uncertain linear hydraulic system in Ref. 9. A multiple-model approach has been formulated for fault detection e.g., leakage of the hydraulic system [10]. The state of charge and state of health of batteries is estimated in Refs. 11 and 12. A multiple-model approach has been applied for target tracking in Ref. 13 and a SVSF based probabilistic data association (PDA) approach has been proposed for tracking in cluttered environment [14]. For multiple object tracking a SVSF based joint-PDA approach has been developed [15]. Online multiple vehicle tracking on real road scenarios has been investigated in Ref. 16. Several SVSF based simultaneous localization and mapping algorithms have been proposed e.g., Refs. 1719. Training of neural networks based on SVSF and classification of engine faults has been studied in Ref. 20. Dual estimation of states and model parameters has been considered in Ref. 21. The estimation strategy works as follows. The bi-section method, the shooting method, and SVSF are combined. The bi-section and shooting method are applied to determine best-fitting model parameter combinations. The obtained model is used by SVSF to estimate the system states. To apply the bi-section method the measurement signals are divided into segments in which the model parameters remain nearly constant. In comparison to the Kalman filter the SVSF approach facilitates detection of these segments based on an evaluation of chattering process.

As described in Ref. 9 the SVSF approach uses a switching gain to drive the estimated state trajectory into a region around the true state trajectory called existence subspace. Due to measurement noise chattering occurs in the existence subspace. By introducing a boundary layer, similar to the saturation function approach of sliding mode control, the high frequency switching in the existence subspace can be attenuated. The claimed advantage of SVSF approach is that if the boundary layer is not used, the filter guarantees to reach the existence subspace for sure although an imprecise system model may be used. It is shown in Ref. 22 that in case of not using the boundary layer the estimations of the filter converge to the measurements, which guarantees the estimation error to be bounded. However, this is a trivial result because if the estimations are equal to the measurements and every state is measured, than the filter is not required. If the boundary layer is used, the estimations diverge from the noisy measurements and estimation performance improves. Unfortunately, it has never been proven that the SVSF with boundary layer has bounded estimation error in case of using an imprecise model.

As already mentioned a serious limitation of the SVSF approach is that all system states have to be measured. Additionally, the measurement model is required to be linear. However, related to tracking a linear measurement model may be achievd by applying a measurement conversion [23] and measurements of the vehicle velocities could also be derived from measured positions.

Another problem of SVSF results from the dependency of the estimation performance on the width of the introduced smoothing boundary layer. In Ref. 24 an estimation error model for the SVSF is proposed and in Ref. 25 the estimation error is minimized according to the smoothing boundary layer width. A maximum a posteriori estimation of the noise statistics of the error model is discused in Ref. 26. However, the derived estimation error model and the related approaches require the system to be linear and precisely known which contradicts the idea of robustness.

In our previous publication [22] a new tuning parameter for the SVSF approach was introduced to achieve online optimization of the estimation performance. In this paper the relation between the SVSF approach and the Kalman filter is studied. An adaptive Kalman filter is obtained from the SVSF approach by replacing the original filter gain. The estimation performance of SVSF and the adaptive filter variant is compared with one another. Therefore, a parameter optimization scheme is proposed. In the simulation results the adaptive Kalman filter shows superior performance compared to the original SVSF approach.

The paper is organized as follows. In Section 2 the preliminaries and the original SVSF approach are discussed. In Section 3 the relation of SVSF and Kalman filter is studied. Parameter optimization of the Kalman filter leading to an adaptive filtering approach is considered in Section 4. The stability of the adaptive filter is studied in Section 5. A performance evaluation of SVSF and adaptive Kalman filter is provided in Section 6.

Notations. An overview of the notations used within the paper is given in Table 1.

TABLE 1

SymbolDescription
Estimated quantity
Real system model
Nominal system model (imprecise)
Expectation value
Transpose operator
Trace of matrix
Element i of vector
Diagonal element i of square matrix
Positive definite matrix
Positive semi-definite matrix
Absolute value applied element-wize on vector
Euclidean vector norm
Weighted Euclidean vector norm
Frobenius matrix norm
Schur product of vectors and
Kronecker product of matrices and
Vectorization by concatenating matrix columns of
Inverse vector operator (dimension of assumed as known)
Signum function applied element-wize on vector
Build diagonal matrix with
Minimal eigenvalue of matrix

Nomenclature.

2. Problem Formulation and Previous Work

Consider the dynamics of a nonlinear system to be exactly described by the discrete-time modelwith states , inputs , and outputs . The process is assumed to be a white noise process with independent samples described by the covariances , and , for , and the mean . The measurement model Eq. 2 can be obtained from any model of the form as the considered SVSF approach requires to be invertible [9]. Consider to be a nominal description of system , which may differ from the true behavior. According to Ref. 9 an estimation of the system states can be obtained using the SVSF algorithmwhere the operator “” denotes the Schur product, and are diagonal matrices with , denoting the ith diagonal element, and the ith element of vector is defined aswhere . According to Ref. 22 the SVSF algorithm (Eqs 37) reduces toin case of .

Theorem 1 of Ref. 9 proves that the output estimation error of algorithm (Eqs 37) approaches zero if instead of the gain the gain is used. As a consequence the estimations of the filter equal the measurements, i.e., . This is a trivial result as all states are required to be measured. So one might just use the measurements instead of the filter estimates. The estimations of the filter diverge from the measurements and estimation performance improves if a boundary layer is introduced and gain is replaced by . However, it has neither been proven that algorithm (Eqs 37) with gain has a bounded estimation error, nor it has been shown that the introduced boundary layer minimizes the squared estimation error or some other performance criterion.

3. Relation Between Smooth Variable Structure and Kalman Filter

In this section the stochastic gain of SVSF approach is replaced by a deterministic but yet undefined gain . Using the deterministic gain the estimation error covariance matrix of the filter is determined. By minimizing the mean squared estimation error (MSE) the deterministic gain becomes the specific optimal one . Replacing the original gain of SVSF by the optimal one gives a direct link to the Kalman filter.

From Eqs 1, 2, and 912 it follows that the state estimation error and the output estimation error can be determined asandwhere is a deterministic but yet undefined gain. To derive the optimal gain that minimizes the MSE an expression of the a posteriori error covariance dependent on is required to be derived. However, first of all the calculation of the output error covariance is considered. Inserting Eq. 14 into the definition of the output error covariance givesExpanding Eq. 15 and considering the definition of the a priori estimation error covariance and the stationary measurement noise covariance yieldsThe value of the remaining expectations in Eq. 16 is studied as follows. First of all the a priori estimation error is known to directly depend on the noise realizations with but not on the realization . In addition with and are independent of each other due to the independent white noise assumption. Finally, can not have any effect on and both random variables are stochastically independent. It follows and . From the zero-mean assumption of the noise i.e., it followsNext the expression for the a posteriori estimation error covariance is considered. Inserting Eq. 13 into the definition of the a posteriori estimation error covariance leads toExpanding Eq. 18 and considering the definition of the output error covariance yieldsBased on Eq. 14 the two remaining expectations in Eq. 19 can be written aswhere and again vanish due to the stochastic independency of and . Finally, the a posteriori error covariance is achieved asIn the following minimization of MSE is considered. Based on Eq. 21 the error covariance can also be written asBy adding expression Eq. 22 can be rewritten asThe trace of gives the MSE of the a posteriori estimation. The last term in Eq. 23 is at least a positive semidefinite matrix with a trace greaterequal zero as . All other terms do not depend on . Consequently, the minimum of MSE is achieved byas the trace of the last term is zero for . The connection between algorithm (Eqs 37) and the Kalman filter is studied as follows.

Theorem 1. The state prediction and correction of algorithm (Eqs 912) equals the one of the extended Kalman filter if is replaced by as stated in Eq. 24.

Proof. 1The state prediction (Eq. 9) obviously equals the one of the extended Kalman filter. Regarding the correct step (Eq. 11) it follows thatholds true if is replaced by of (Eq. 24). As the introduced of (Eq. 26) equals the Kalman filter gainstep (Eq. 26) and thus (Eq. 11) is identical to the correction step of the Kalman filter. ∎

4. Parameter Optimization of the Kalman Filter

The robustness of SVSF against model uncertainties is achieved by tuning the parameters of the smoothing boundary layer width [25]. However, also the Kalman filter gain can be made adaptive to achieve improved robustness [27]. In this section an adaptation law for the unknown a priori estimation error covariance of the optimal gain (Eq. 24) is derived. The error covariance should not be propagated based on the system model (like it is done usually in the field of Kalman filtering) as this would lead to prediction errors due to the imprecize model description. Instead itself is required to be estimated.

As according to Eq. 17 the state error covariance is related to the output error covariance the expressionwithmight be considered to gain information about . Estimating based on the innovation process is common in the field of adaptive Kalman filtering, e.g., Refs. 27 and 28. From Eq. 31 it can be seen that if is constant over N time steps, and is ergodic in the sense of , and then equals the a priori error covariance. Additional information about can be obtained by considering the suboptimal gain . Replacing of Eq. 23 by the suboptimal gain gives the error covariance of the suboptimal filter. As the filter with the optimal gain minimizes the MSE the upper boundcan be established. If it is assumed that , which is the case for sufficient small sampling time thenmight be considered to gain information about . Due to the recursive nature of the filtering approach the relation of to the previous value might be considered. For small sampling time this can be roughly described byIn order to find an estimation that fits best to the established Eqs 29, 33, and 34 a weighted least squares (WLS) estimation problem is formulated. The importance of the individual equations is expressed by the scalar weights which will be determined in a parameter optimization process. Using the vector operator “vec” the WLS problemsubject tois considered. The solution of this WLS problem iswhich is a weighted sum of the solutions of the individual equations. The obtained solution (Eq. 36) is not guaranteed to be positive semidefinite. However, as is a covariance matrix it must be at least positive semidefinite. In practice it will even be positive definite as otherwise the MSE would be zero. Incorporating the constraint into the WLS problem (Eq. 35) would lead to a nonlinear optimization problem. In order to keep the problem simple it is suggested to replace by some if has non-positive eigenvalues. The modified error covariance iswhere diagonalizes asand matrix is a diagonal matrix with the ith diagonal element defined bywhere . The scaling factor will be part of the parameter optimization process. The diagonalization (Eq. 38) can always be achieved as is symmetric if a symmetric initial matrix is chosen. Finally, estimating the optimal Kalman filter gain based on leads to the adaptive Kalman filtering (A-KF) approachwithwhere diagonalizes so that diagonalization is achieved asand the diagonal matrix is obtained fromwith and factor .

The adaptive filtering approach (Eqs 3946) depends on the setof tuning parameters. In order to optimize these parameters a training model is introduced. The filter is applied to the training model instead of the real system. The training model simulates the effects of model uncertainty occurring from the real unknown system by varying the system parameters of the known nominal model . Let be the number of system parameters and let be the nominal value of parameter i. Assume that a priori knowledge about the amount of model uncertainty is available meaning that parameter is required to be varied by percent to account for the severity of model uncertainty. It is suggested to obtain the ith parameter of the training model by drawing a sample from an uniform distribution. Repeating the procedure for all system parameters forms one set of training parameters denoted as a training model. The optimization of the filter parameters based on the training model can be achieved as follows. Let be the states and be the measurements of the training model and let be the estimations of the filter using the nominal model but receiving the measurements from the training model then the optimized parameters are given aswhere J denotes the costfunction. The optimization is a mixed integer optimization problem which can be solved by e.g., genetic approaches. In order to make the optimization process more reliable it is recommended to build several training models and optimize the filter parameters for each of them separately. The final optimized filter parameters can be obtained by taking the mean or median of the optimized parameters of the individual training models. The process of filter parameter optimization is illustrated in Figure 1.

FIGURE 1

5. Stability Analysis of the Adaptive Kalman Filter

As mentioned previously boundedness of the estimation error of SVSF approach (Eqs 37) using smoothing boundary layer has never been proven. However, if the adaptive Kalman filter is used instead ( is replaced by ) the boundedness of the estimation error can be proven as follows.

Theorem 2. The estimation error of the a posteriori estimationof the adaptive Kalman filtering approach (Eqs 3946) is bounded byif the measurement noise is bounded by and the following conditions are fulfilledwhere , are sufficient small positive constants, , are sufficient large positive constants.

Proof. From Eq. 51 it can be concluded that can be written aswith . As known from Eq. 13 the a posteriori estimation error isWith the definition the estimated optimal filter gain can be written asApplying the Sherman-Morrison-Woodbury formular (Fact 2.16.3 [29]) in combination with assumption (Eq. 52) givesUsing the submultiplicative Frobenius norm (Proposition 9.3.5 [29]) and Eq. 58, Eq. 57 can be written asBased on the minimum eigenvalue of the inequalitycan be established (Lemma 8.4.1 [29]). Using Eq. 60 and assumption Eq. 53 an upper bound of Eq. 59 is obtained asUnfortunately, implicitly depends on so that assumption Eq. 54 has to be taken into consideration leading to

Based on the derivative of Eq. 62 with respect to it can be shown that the maximum of the upper bound is achieved for which leads toThen (Eq. 50) is proven by applying the triangle inequality on Eq. 56.

6. Numerical Example

In this section state estimation and control of a chemical plant is considered in order to evaluate the performance of original SVSF and the adaptive Kalman filtering variant. According to Ref. 30, Ref. 31 a species A reacts in a continuous stirred tank reactor. The dynamics of the effluent flow concentration of species A and the reactor temperature can be described by the time-continuous modelwith measurements , and control variable z defined as

The parameters of the system are shown in Table 2. The input of the system is the change of the coolant stream temperature related to the nominal value . An input saturation of is considered. The system is simulated based on Euler method using a sampling time of 0.1 s. The resulting discrete-time measurement model iswhere the zero-mean, white noise is Gaussian with covarianceFor the initial state values , and are considered.

TABLE 2

DescriptionParameterUnit
Tank volumeVl
Feed flow rateq
Feed concentration
Feed temperatureK
Densityρ
Enthalpy
Exponential factorK
Frequency factor
Heat transfer characteristic
Specific heat
Coolant flow temperatureK

System parameter description [30].

6.1. Filter Training Process

In the following the optimization of the tuning parameters of A-KF and SVSF approach is considered. As explained in Section 4 variation of the nominal system parameters is required to build the training model. To account for the model uncertainty a variation of 20% of the nominal parameters is considered which is assumed as a priori known. Consequently, the ith parameter of the training model is obtained from the uniform distribution . Repeating the step for all parameters forms one training model. The parameter is not varied as it is required to be precisely known for control. In Table 3 an overview of the true and nominal system parameters and the parameters used to build the training model is given. In order to account for the different combination and variation of the system parameters three training models are build denoted as Training I, Training II, and Training III.

TABLE 3

ParameterNominalRealTraining
V10092[80, 120]
Q10094[80, 120]
10.98[0.8, 1.2]
350354[280, 420]
ρ1,000953[800, 1,200]
5e44.91e4[4e4, 6e4]
8,7508,893[7,000, 10,500]
7.2e107.68e10[5.76e10, 8.64e10]
5e44.68e4[4e4, 6e4]
0.2390.299[0.191, 0.287]
300300300

System parameters of nominal, real, and training model.

Interval [a,b] denotes: Parameter for training equals a sample u drawn from uniformly distribution , i.e., .

For each set the tuning parameters of the filters are optimized separately. The parameters of A-KF approach required to be optimized are given by Eq. 47. For the SVSF approach the boundary layer widths , and convergence rates , of the first and second state are optimized. The optimization is achieved using “genetic algorithm” of MATLAB with default settings. During optimization only one realization of measurement noise is considered so that the costfunction does not vary due to the noise.

The results of optimization are shown in Table 4. The A-KF approach requires at least 60 times more computational time for the optimization than SVSF. The SVSF algorithm is computationally more efficient and in addition it requires less parameters to be optimized. However, the achieved minimal value of the costfunction is always lower in case of A-KF.

TABLE 4

Training ITraining IITraining III
A-KFSVSFA-KFSVSFA-KFSVSF
GA iterations4,7954,3306,0493,8136,1064,095
GA convergenceAchievedAchievedAchievedAchievedAchievedAchieved
Min. value of costfunction0.268770.499760.412470.742740.379310.56349
Total number of FLOPS74867e6779e694446e6686e695337e6737e6
Training time [min]173.242.48215.232.37208.602.65

Computational effort of training process.

GA: Genetic Algorithm of MATLAB, GA iterations: Calls of costfunction, GA convergence: Change of costfunction falls below tolerance, Min. value of costfunction equals of Eq. 48, FLOPS: floating point operations, Hardware: 4xCPU@3.7GHz with 8 GB memory.

6.2. Open Loop Case

In the open loop case the step functions and are applied to the system and the resulting step responses are considered for state estimation. The system behavior is simulated over a time horizon of 6 min.

The performance values of A-KF and SVSF approach are shown in Table 5. The filters are applied based on four sets of optimized parameters. Three sets result from Training I, Training II, and Training III. The fourth set denoted as Training I-III considers the mean values of the optimized parameters of the other three sets. As all states are measured the MSE of the measurements is considered also. In order to account for the effect of different noise realizations the results are obtained by simulating the step response 100 times and taking the mean value and variance of the squared estimation error. The filters are applied to the same measurements with same noise realizations. From the results it can be seen that both A-KF and SVSF estimations are more precise than the measurements. In comparison to SVSF the A-KF approach achieves better estimation performance for all considered sets of optimized parameters and both step responses. For one specific run the step responses and the corresponding state estimations are visualized in Figures 2 and 3. The computational time required to generate Table 5 (to simulate both step responses 100 times and apply 8 filters each time) is 18.90 min on a 4xCPU@3.7Ghz with 8 GB memory. The computational time required to apply only one filter on one specific step response is 0.11 s in case of SVSF and 1.49 s in case of A-KF. Both values are far less than the 6 min of simulated system behavior.

TABLE 5

Training ITraining IITraining IIITraining I-IIIMeas
A-KFSVSFA-KFSVSFA-KFSVSFA-KFSVSF
K0.1940.2970.1980.5090.2000.2950.1970.3290.889
7.63e−57.10e−57.43e−52.44e−48.82e−51.00e−48.27e−51.21e2.66e−4
K0.1140.2760.1170.5030.1170.2790.1170.3140.892
7.65−58.50−57.12e−52.59−48.86e−51.20−48.45−51.37−42.67−4

Estimation performance in open loop case (100 realizations).

Meas., Measurements, : Mean of (squared estimation error), : Variance of (squared estimation error).

FIGURE 2

FIGURE 3

6.3. Closed Loop Case

In the closed loop case the effluent flow concentration of the chemical plant is controlled. The controller relies on the filter estimates. Consequently, the estimation performance as well as the control performance dependent on the filters is studied.

In order to achieve reference tracking of the control variable the super-twisting sliding mode approach [32] is applied. By introducing the estimated tracking errorwith reference value the sliding variable can be calculated asThen reference tracking can be achieved by applying the super-twisting approach [32]The controller parameters are chosen as and based on trial and error. For the sliding dynamics is considered. The simulated time horizon is 20 min with reference values defined bywhere . Related to the control performance the following four cases are studied. The sliding variable (Eq. 68) is calculated using the estimations of A-KF approach or the estimations of SVSF approach or the measurements denoted as “W/O filters” or the true states denoted as “Optimum”. For the filters the optimized filter parameters of the different training models are considered. The control of the plant over the time horizon of 20 min is simulated 100 times to account for the noise realizations. The measurements used by the filters have same noise realizations. The performance values are shown in Table 6. The best control performance can be achieved based on the true states which are noise-free. Control based on estimations of A-KF or SVSF achieves better performance than applying the controller directly on the noisy measurements. In comparison to SVSF approach A-KF method achieves better reference tracking and more precise state estimation for all considered training models. The control variable and the input values of one specific realization are visualized in Figures 4 and 5. The computational time required to generate Table 6 (to simulate 100 runs of 10 closed loop systems in parallel) is 34.54 min on a 4xCPU@3.7Ghz with 8 GB memory. The computational time required to simulate the closed loop behavior only once based on one controller is 0.45 s if the controller is fed by the estimations of SVSF and 5.24 s if the controller is fed by the estimations of A-KF. This is in both cases far less than the time horizon of 20 min over which the system behavior is simulated.

TABLE 6

Training ITraining IITraining IIITraining I-IIIW/OOpt
A-KFSVSFA-KFSVSFA-KFSVSFA-KFSVSFFilter
4.999e25.420e25.006e25.884e25.006e25.518e25.005e25.562e26.229e24.711e2
31.51246.29231.44162.39131.62558.46031.78760.6231.939e0
0.2210.3270.2220.5200.2250.3220.2220.3520.8900
2.38e−53.55e−52.44e−58.70e−52.43e−54.36e−52.37e−55.12e−59.07e−50
1.557e21.495e21.551e21.369e21.559e21.460e21.554e21.458e21.086e21.588e2
14.3834.796415.0227.30115.3975.92814.9686.1008.7580

Control and estimation performance in closed loop case (100 realizations).

Opt., Optimum;: Mean of (squared tracking error), : Variance of (squared tracking error), : Mean of (squared estimation error), : Variance of (squared estimation error), : Mean of (input energy), : Variance of (input energy).

FIGURE 4

FIGURE 5

7. Conclusion

In this paper the relation of smooth variable structure filter (SVSF) and Kalman filter has been studied. An adaptive Kalman filter (A-KF) was derived from the SVSF approach. Sufficient conditions for the boundedness of the estimation error of A-KF approach have been formulated. The simulation results show that beside the SVSF approach also an adaptive Kalman filter can be used to achieve robust estimation performance.

Statements

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

The development and theoretical description of the adaptive Kalman filter approach was undertaken by MS and DS. The numerical simulation was designed and realized by MS. The manuscript was drafted by MS and reviewed by DS.

Acknowledgments

The authors want to thank the reviewers for their valueable comments and suggestions.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Footnotes

1.^The authors thank the anonymous reviewers of European Control Conference 2020 for the insightful comments and suggestions related to the proof of the theorem.

References

  • 1.

    KitanidisPK. Unbiased minimum-variance linear state estimation. Automatica (1987). 23:7758. 10.1016/0005-1098(87)90037-9.

  • 2.

    GillijnsSDe MoorB. Unbiased minimum-variance input and state estimation for linear discrete-time systems. Automatica (2007). 43:1116. 10.1016/j.automatica.2006.08.002.

  • 3.

    AndersonBDMooreJB. Optimal filtering.Englewood Cliffs, NJ: Prentice-Hall (1979). 256 p.

  • 4.

    HmidaFBKhémiriKRagotJGossaM. Three-stage Kalman filter for state and fault estimation of linear stochastic systems with unknown inputs. J Franklin Inst (2012). 349:236988. 10.1016/j.jfranklin.2012.05.004.

  • 5.

    DongZYouZ. Finite-horizon robust Kalman filtering for uncertain discrete time-varying systems with uncertain-covariance white noises. IEEE Signal Process Lett (2006). 13:4936. 10.1109/LSP.2006.873148.

  • 6.

    HassibiBSayedAHKailathT. Indefinite quadratic estimation and control: a unified approach to H2 and H heoriesSociety for Industrial and Applied Mathematics. Philadelphia, PA (1999). p. 572.

  • 7.

    BlomHAPBar-ShalomY. The interacting multiple model algorithm for systems with Markovian switching coefficients. IEEE Trans Automat Contr. (1988). 33:7803. 10.1109/9.1299.

  • 8.

    MartinoLReadJElviraVLouzadaF. Cooperative parallel particle filters for online model selection and applications to urban mobility. Digit Signal Process (2017). 60:17285. 10.1016/j.dsp.2016.09.011.

  • 9.

    HabibiS. The smooth variable structure filter. Proc IEEE (2007). 95:102659. 10.1109/JPROC.2007.893255.

  • 10.

    GadsdenSASongYHabibiSR. Novel model-based estimators for the purposes of fault detection and diagnosis. IEEE ASME Trans Mechatron (2013). 18:123749. 10.1109/TMECH.2013.2253616.

  • 11.

    KimTWangYFangHSahinogluZWadaTHaraSet alModel-based condition monitoring for lithium-ion batteries. J Power Sources (2015). 295:1627. 10.1016/j.jpowsour.2015.03.184.

  • 12.

    QiaoHHAttariMAhmedRDelbariAHabibiSShoaT. Reliable state of charge and state of health estimation using the smooth variable structure filter. Contr Eng Pract (2018). 77:114. 10.1016/j.conengprac.2018.04.015.

  • 13.

    GadsdenSAHabibiSRKirubarajanT. A novel interacting multiple model method for nonlinear target tracking. In: FUSION 2010 : 13th international conference on information fusion; 2010 Jul 26 - 29;Edinburgh, Scotland. Piscataway, NJ: IEEE.

  • 14.

    AttariMGadsdenSAHabibiSR. Target tracking formulation of the SVSF as a probabilistic data association algorithm. In: The 2013 American control conference; 2013 June 17 - 19;Washington, DC;IEEE, Piscataway, NJ. p. 632832. 10.1109/ACC.2013.6580830.

  • 15.

    AttariMLuoZHabibiS. An SVSF-based generalized robust strategy for target tracking in clutter. IEEE Trans Intell Transport Syst (2016). 17:138192. 10.1109/TITS.2015.2504331.

  • 16.

    LuoZAttariMHabibiSMohrenschildtMV. Online multiple maneuvering vehicle tracking system based on multi-model smooth variable structure filter. IEEE Trans Intell Transport Syst (2020). 21:60316. 10.1109/TITS.2019.2899051.

  • 17.

    DemimFNemraALouadjK. Robust SVSF-SLAM for unmanned vehicle in unknown environment. IFAC-PapersOnLine (2016). 49:38694. 10.1016/j.ifacol.2016.10.585.

  • 18.

    AllamATadjineMNemraAKobziliE. Stereo vision as a sensor for SLAM based smooth variable structure filter with an adaptive boundary layer width. In: ICSC 2017, 6th International conference on systems and control. 2017 May 7–9;Batna, Algeria; Piscataway, NJ: IEEE. p. 1420. 10.1109/ICoSC.2017.7958700.

  • 19.

    LiuYWangC. A FastSLAM based on the smooth variable structure filter for UAVs. In: 15th international conference on ubiquitous robots (UR); June 26-30, 2018;Hawaii Convention Center; Piscataway, NJIEEE (2018). p. 5916. 10.1109/URAI.2018.8441876.

  • 20.

    AhmedRMSayedMAEGadsdenSAHabibiSR. Fault detection of an engine using a neural network trained by the smooth variable structure filter. In: IEEE international conference on control applications (CCA); Denver, CO; 2011 Sep 28-30;Piscataway, NJIEEE (2011). p. 11906. 10.1109/CCA.2011.6044515.

  • 21.

    Al-ShabiMHabibiS. Iterative smooth variable structure filter for parameter estimation. ISRN Signal Processing (2011). 2011:1. 10.5402/2011/725108.

  • 22.

    SpillerMBakhshandeFSöffkerD. The uncertainty learning filter: a revised smooth variable structure filter. Signal Process (2018). 152:21726. 10.1016/j.sigpro.2018.05.025.

  • 23.

    LongbinMXiaoquanSYiyuZKangSZBar-ShalomY. Unbiased converted measurements for tracking. IEEE Trans Aero Electron Syst (1998). 34:10237. 10.1109/7.705921.

  • 24.

    GadsdenSAHabibiSR. A new form of the smooth variable structure filter with a covariance derivation. In: CDC 2010 : 49th IEEE conference on decision and control 2010 Dec 15 - 17;Atlanta, GA; Piscataway, NJ:IEEE (2010). p. 738994. 10.1109/CDC.2010.5717397.

  • 25.

    GadsdenSAEl SayedMHabibiSR. Derivation of an optimal boundary layer width for the smooth variable structure filter. In: 2011 American Control Conference - ACC 2011; San Francisco, CA; 2011 Jun 29 - Jul 1;Piscataway, NJ:IEEE (2011). p. 49227. 10.1109/ACC.2011.5990970.

  • 26.

    TianYSuwoyoHWangWLiL. An ASVSF-SLAM algorithm with time-varying noise statistics based on MAP creation and weighted exponent. Math Probl Eng. (2019). 2019:1. 10.1155/2019/2765731.

  • 27.

    HideCMooreTSmithM. Adaptive Kalman filtering for low-cost INS/GPS. J Navig. (2003). 56:14352. 10.1017/S0373463302002151.

  • 28.

    YangYGaoW. An optimal adaptive Kalman filter. J Geodes. (2006). 80:17783. 10.1007/s00190-006-0041-0.

  • 29.

    BernsteinDS. Matrix mathematics: theory, facts, and formulasPrinceton, NJ: Princeton University Press (2009). 1184 p.

  • 30.

    MagniLNicolaoGDMagnaniLScattoliniR. A stabilizing model-based predictive control algorithm for nonlinear systems. Automatica (2001). 37:135162. 10.1016/S0005-1098(01)00083-8.

  • 31.

    SeborgDEMellichampDAEdgarTFDoyleFJIII. Process dynamics and controlNew York, NY: John Wiley & Sons (2010). p. 10856.

  • 32.

    ShtesselYEdwardsCFridmanLLevantA. Sliding mode control and observation.New York: Springer (2014). 356 p.

Summary

Keywords

state estimation, nonlinear systems, stochastic systems, Kalman filtering, control

Citation

Spiller M and Söffker D (2020) On the Relation Between Smooth Variable Structure and Adaptive Kalman Filter. Front. Appl. Math. Stat. 6:585439. doi: 10.3389/fams.2020.585439

Received

20 July 2020

Accepted

16 October 2020

Published

27 November 2020

Volume

6 - 2020

Edited by

Lili Lei, Nanjing University, China

Reviewed by

Xanthi Pedeli, Athens University of Economics and Business, Greece

Yong Xu, Northwestern Polytechnical University, China

Updates

Copyright

*Correspondence: Mark Spiller,

This article was submitted to Dynamical Systems, a section of the journal Frontiers in Applied Mathematics and Statistics

Disclaimer

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Outline

Figures

Cite article

Copy to clipboard


Export citation file


Share article

Article metrics