Next Article in Journal
Effect of Positive Bias and Pressure on Plasma Flow Characteristics in a Chemical Vapor Deposition Chamber
Previous Article in Journal
Control Efficiency of Biochar Loaded with Bacillus subtilis Tpb55 against Tobacco Black Shank
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Robust Hammerstein-Wiener Model Identification Method for Highly Nonlinear Systems

1
School of Electronics and Information Engineering, Taizhou University, Taizhou 318000, China
2
College of Automation, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
3
College of Computer Science and Technology, Heilongjiang Institute of Technology, Harbin 150050, China
*
Authors to whom correspondence should be addressed.
Processes 2022, 10(12), 2664; https://doi.org/10.3390/pr10122664
Submission received: 21 November 2022 / Revised: 5 December 2022 / Accepted: 8 December 2022 / Published: 11 December 2022

Abstract

:
The existing results show the applicability of the Over-Parameterized Model based Hammerstein-Wiener model identification methods. However, it requires to estimate extra parameters and performer a low rank approximation step. Therefore, it may give rise to unnecessarily high variance in parameter estimates for highly nonlinear systems, especially using a small and noisy data set. To overcome this corruptive phenomenon. To overcome this corruptive phenomenon, in this paper, a robust Hammerstein-Wiener model identification method is developed for highly nonlinear systems when using a small and noisy data set, where two parsimonious parametrization models with fewer parameters are used, and an iteration method is then used to retrieve the true system parameters from the parametrization models. Such modification can improve the parameter estimation performance in terms of accuracy and variance compared with the over-parametrization model based identification methods. All the above-mentioned developments are analyzed with variance analysis, along with a simulation example to confirm the effectiveness.

1. Introduction

The Hammerstein–Wiener (H-W) models, also called N-L-N models, are known as a particular class of block-oriented nonlinear models where a Linear (L) subsystem is embedded in two static Nonlinear subsystems (N). Due to the use of two nonlinear elements instead of one, H-W models offer convenient higher modeling capabilities for nonlinear systems, especially for nonlinear systems with both actuator and sensor nonlinearities [1]. H-W models identification has attracted considerable attention in the past years and a large amount of works have been explored in different frameworks, see, for example, iterative methods [2,3], and Least-Squares (LS) based Two-Step Algorithms (TSA) [4].
In this paper, the emphasis is put on the LS issue for H-W models with over-parametrization methods, which was perhaps first studied by Bai [4] where first characterizes the original H-W model by an over-parametrization model and then estimates the parameters of this over-parametrization model (OPM) by using a LS method. In consequence, the original system parameters are extracted from the estimated over parametrization model parameters by performing low-rank approximation using Singular Value Decompositions (SVDs). The TSA is attractive because of its numerical simplicity. Thus, following from this pioneering work [4], a number of TSA-based methods have been proposed to identify H-W models under different conditions: For instance, in [5], a auxiliary model multi-innovation stochastic gradient (AM-MISG) based TSA was proposed for H-W models online identification, which has fast convergence rate with robust characteristics and its convergence was proved by using the stochastic process theory. In [6], an instrumental variable based TSA was proposed for H-W continuous-time models in the presence of colored noise. In [7], a modified Bias-Eliminating Least-Squares (BELS) based TSA was proposed for H-W errors-in-variables models.The TSA has also been extensively used for the identification of other types of block-oriented nonlinear systems, e.g., Hammerstein models [8,9], and Wiener models [10].
In these contributions, all existing TSAs provide asymptotically consistent unbiased estimations for H-W models when the data length tends to infinity. However, it is well known that the existing TSAs related to the use of over-parametrization model and low-rank approximation step may result in a degradation of the accuracy and variance of the estimated quantities for highly nonlinear systems, especially for short noisy data length sequences. One reason is that the over-parametrization model has much more parameters than the actual Hammerstein-Wiener model, the extra parameters in TSAs may lead to larger estimation errors and variances, and the number of the extra parameters increase with the degree of nonlinearity. The other is that the low-rank approximation step in TSAs inevitably brings some truncation errors, which will result in some loss of model quality.
In this paper, inspired by the parsimonious models based identfication methods [11,12,13,14] for block-oriented nonlinear Hammerstein models can improve the variance properties of the over-parametrization model based methods, a robust consistent identification formulation for H-W models is thus proposed to overcome the above-mentioned drawbacks for highly nonlinear systems when using short noisy data length sequences, without the necessity of estimating extra parameters and performing low-rank approximation.The main contributions of the paper are stated below:
  • Two parsimonious models with fewer parameters constructed to characterize the H-W system instead of an over-parametrization model, and a projection based LS iterative method is proposed to estimate all system parameters in a parallel fashion. The proposed method avoids the estimating extra parameters and a low rank approximation step in classical over-parametrization model based methods, which leads to improved results compared to existing over-parametrization model based methods, because that estimating extra parameters and performing low rank approximation may result into a degradation of the accuracy of the estimated quantities.
  • The variance analysis is given to demonstrate that the new method generally gives smaller variance compared to conventional TSA method. Note that the results of variance analysis explains why the proposed identfication methods can lead to improved results compared to existing over-parametrization model based methods for H-W systems.
Note that Hammerstein models and Wiener models are two special cases of Hammerstein–Wiener models, that is, when n c = 1 and f 1 ( u ( t ) ) = u ( t ) in following (1), the Hammerstein–Wiener models become Wiener models, and when n d = 1 and g 1 ( y ( t ) ) = y ( t ) in following (1), the Hammerstein–Wiener models become Hammerstein models. The presented methodology in this paper is general for N-L-N models, L-N models and N-L models.
The paper is organized as follows. In Section 2, the identification problem is presented. In Section 3, the proposed method is presented, followed by the Variance analysis in Section 4. Section 5 presents a simulation example to validate and evaluate the performance of the new algorithm. Finally, some conclusions are drawn in Section 6.

2. Problem Statement and Analysis

Consider a particular class of nonlinear H-W model with polynomial nonlinearities, see also [4]
y ( t ) = j = 1 n b b j m = 1 n c c m f m [ u ( t j ) ] + i = 1 n a a i l = 1 n d d l g l [ y ( t i ) ] + e ( t )
where system parameters are a = [ a 1 , , a n a ] , b = [ b 1 , , b n b ] , c = [ c 1 , , c n c ] , and d = [ d 1 , , d n d ] , { u ( t ) } and { y ( t ) } are the system input and output sequences, respectively, { e ( t ) } is the noise sequence, and f m ( ) and g l ( ) are nonlinear functions. The following general assumptions are made in this paper:
A1
The first nonzero elements of a and b are positive and a 2 = 1 and b 2 = 1 .
A2
The functions { f m ( ) } m = 1 n c and g l ( ) l = 1 n d and degrees n a , n b , n c and n d are known.
A3
{ e ( i ) } is a zero-mean white noise sequence with finite variance δ e 2 uncorrelated with { u ( j ) } for all i , j , and uncorrelated with { y ( j ) } for i j .
Assumption 1 ensures uniqueness of the N-L-N nonlinear system model, as the introduction of a factor q ( q 0 ) to the pair ( q a , d / q ), the input-output relationship will not change. The terms a 2 and b 2 can be set as any other positive number, the proposed method can be directly used without any change.
Assumption 2 are related to a priori knowledge about the structure of the true system. In this context the static nonlinearity of the N-L-N model is approximated using a linear expansion in terms of basis functions. In generally, the basis functions and system orders lie in the fact that tuning them are more directly related to a priori knowledge about the true system. However, since the unknown coefficients of parameter vectors a , b , c and d can give the degrees of freedom to accurately modeling the nonlinear systems, the basis functions and system orders can be assumed to known simply for having not directly related knowledge about the nonlinear function in practice. More specifically, the system orders can be set high enough to provide enough degrees of freedom to accurately describing the nonlinear N-L-N systems. The basis functions can be chosen as polynomial functions simply, which is realistic in many applications where the nonlinearities characterizing the systems are smooth enough [9]. Note that the identification method developed under those assumptions in which the nonlinear basis functions as well as the system orders in polynomial model supposed to be known can be regarded as a grey-box method compared with the black-box model approach without any assumption about the structure of the N-L-N model.
Assumption 3 requires the noises to be white, which is realistic in many applications when the main objective is on the plant model.
Based on (1), an over-parametrization model can be formulated as
y ( t ) = θ Φ ( t ) + e ( t )
where
θ = [ θ 1 , θ 2 ] 1 × ( n b n c + n a n d )
θ 1 = [ b 1 c 1 , , b 1 c n c , , b n b c 1 , , b n b c n c ] 1 × n b n c
θ 2 = [ a 1 d 1 , , a 1 d n d , , a n a d 1 , , a n a d n d ] 1 × n a n d
Φ ( t ) = [ f ( t ) , g ( t ) ] ( n b n c + n a n d ) × 1
f ( t ) = [ f 1 [ u ( t 1 ) ] , , f n c [ u ( t 1 ) ] , , f 1 [ u ( t n b ) , , f n c [ u ( t n b ) ] ] n b n c × 1
g ( t ) = [ g 1 [ y ( t 1 ) ] , , g n d [ y ( t 1 ) ] , , g 1 [ y ( t n b ) , , g n d [ y ( t n a ) ] ] n a n d × 1 .
Define
Y = [ y ( 1 ) , y ( 2 ) , , y ( N ) ]
E = [ e ( 1 ) , e ( 2 ) , , e ( N ) ]
F = [ f ( 1 ) , f ( 2 ) , , f ( N ) ]
G = [ g ( 1 ) , g ( 2 ) , , g ( N ) ]
Φ = [ F , G ] .
It follows from (1)–(4) that
Y = θ 1 F + θ 2 G + E = θ Φ + E .
For the estimation of the unknown system parameters in (5), the following criterion is considered in the TSA:
min a , b , c , d 1 2 ( Y θ Φ ) W ( Y θ Φ )
s . t . rank ( b c ) = 1 , rank ( a d ) = 1
where W is a symmetric weighting matrix, e.g., an identity matrix of appropriate dimension I.
The TSA is a relaxation method by solving the unconstrained LS problem (6a) without considering the rank constraint (6b), where the system parameters are extracted via SVDs, see [4] for more details. Clearly, the number of extra parameters in the TSA method [4] is n a n d + n b n c n a n d n b n c , and two SVDs have to be used to extract the system parameters from θ . These may deteriorate the accuracy and variance of estimated quantities. Note that the number of the extra parameters in the TSA will increase as the degree of nonlinearity of H-W models increases. The performance of the TSA method in terms of the accuracy and variance of estimated quantities will be deteriorated for identifying highly nonlinear systems from a limited number of noisy data.
This paper aims to seek a method to estimate unknown parameter vectors a , b , c and d from the known observed input-output data { u ( t ) , y ( t ) } collected at instants t = 1 , , N without estimating extra parameters and performing SVDs.

3. The Proposed Method

3.1. Parsimonious Parametrization Model Formulation

We decouple θ into two parts θ 1 and θ 2 . The over-parametrization model in (2) can be then represented as following two parsimonious parametrization models (PPMs):
y ( t ) = a Φ 1 ( t ) + b Φ 2 ( k ) + e ( t ) = d Φ 3 ( t ) + c Φ 4 ( k ) + e ( t )
where
Φ 1 ( t ) = ( d 1 g 1 [ y ( t 1 ) ] + + d n d g n d [ y ( t 1 ) ] , , d 1 g 1 [ y ( t n a ) + + d n d g n d [ y ( t n a ) ] )
Φ 2 ( t ) = ( c 1 f 1 [ u ( t 1 ) ] + + c n c f n c [ u ( t 1 ) ] , , c 1 f 1 [ u ( t n b ) + + c n c f n c [ u ( t n b ) ] )
Φ 3 ( t ) = i = 1 n a ( a i g 1 [ y ( t i ) ] ) , , i = 1 n a ( a i g n d [ y ( t i ) ] )
Φ 4 ( t ) = i = 1 n b ( b i f 1 [ u ( t i ) ] ) , , i = 1 n b ( b i f n c [ u ( t i ) ] ) .
Define
Φ 1 = [ Φ 1 ( 1 ) , Φ 1 ( 2 ) , , Φ 1 ( N ) ] = d 0 0 0 0 d 0 0 : : : : : 0 0 0 d η 1 G
Φ 2 = [ Φ 2 ( 1 ) , Φ 2 ( 2 ) , , Φ 2 ( N ) ]
Φ 3 = [ Φ 3 ( 1 ) , Φ 3 ( 2 ) , , Φ 3 ( N ) ]
Φ 4 = [ Φ 4 ( 1 ) , Φ 4 ( 2 ) , , Φ 4 ( N ) ] .
It follows from (7) that
Y = a Φ 1 + b Φ 2 + E = d Φ 3 + c Φ 4 + E .

3.2. Parsimonious Identification Method

We consider an iterative estimation procedure to estimate unknown parameters from (10). Projecting the row space of Y in (10) onto the orthogonal complement of the row space of the matrix F, we have
Y Π F = ( a Φ 1 + b Φ 2 + E ) Π F = ( d Φ 3 + c Φ 4 + E ) Π F
where Π F = I N F ( F F ) 1 F . Due to b Φ 2 = θ 1 F and c Φ 4 = θ 1 F , we have
Y Π F = a Φ 1 Π F + E Π F = d Φ 3 Π F + E Π F .
Once d is assumed to be known, the parameter vector a can be estimated without knowing c and b , while the parameter vector d can be estimated with only knowing a , which allows an iterative method to obtain a and d based (12). The details of estimation of a and d are listed as below.
  • An LS estimate a ^ ( k ) can be obtained by substituting d ^ ( k 1 ) into (12) as
    a ^ ( k ) = Y Π F Φ ^ 1 ( k 1 ) ( Φ ^ 1 ( k 1 ) Π F Φ ^ 1 ( k 1 ) ) 1
    where the estimate Φ ^ 1 ( k 1 ) of Φ 1 is a function of d ^ ( k 1 ) .
  • The unique solution a ^ ( k ) can be obtained by performing a normalization operation as follows:
    a ^ ( k ) = a ^ ( k ) / a ^ ( k ) 2 .
  • An LS estimate d ^ ( k ) can be obtained by substituting a ^ ( k ) into (12) as
    d ^ ( k ) = Y Π F Φ ^ 3 ( k ) ( Φ ^ 3 ( k ) Π F Φ ^ 3 ( k ) ) 1
    where the estimate Φ ^ 3 ( k ) of Φ 3 is a function of a ^ ( k ) .
In similar arguments to estimate a and d , by projecting the row space of Y in (10) onto the orthogonal complement of the row space of the matrix G, we get
Y Π G = b Φ 2 Π G + E Π G = c Φ 4 Π G + E Π G .
  • An LS estimate b ^ ( k ) can be estimated by substituting c ^ ( k 1 ) into (16) as
    b ^ ( k ) = Y Π G Φ ^ 2 ( k 1 ) ( Φ ^ 2 ( k 1 ) Π G Φ ^ 2 ( k 1 ) ) 1
    where the estimate Φ ^ 2 ( k 1 ) of Φ 2 is a function of c ^ ( k 1 ) .
  • The unique solution b ^ ( k ) can be obtained by performing a normalization operation as follows:
    b ^ ( k ) = b ^ ( k ) / b ^ ( k ) 2 .
  • An LS estimate c ^ ( k ) can be estimated by substituting b ^ ( k ) into (16) as
    c ^ ( k ) = Y Π G Φ ^ 4 ( k ) ( Φ ^ 4 ( k ) Π G Φ ^ 4 ( k ) ) 1
    where the estimate Φ ^ 4 ( k ) of Φ 4 is a function of b ^ ( k ) .
The following general assumption A4 about persistently exciting (PE) conditions is made for system identifiability.
A4
Matrices F and G are full row rank.
The assumption A4 is a general assumption for all identification methods, which can be easily guaranteed by constructing F and G based on random signals. Based on which, once the initial values d ^ ( 0 ) and c ^ ( 0 ) are nonzero, it follows from (10) that matrices ( Φ ^ 1 ( 0 ) , Φ ^ 2 ( 0 ) ) will be full row rank, due to that the rank of matrices η ^ 1 and η ^ 2 with nonzero parameters d ^ ( 0 ) and c ^ ( 0 ) are full row rank, such that the estimates a ^ ( 1 ) in (13) and b ^ ( 1 ) in (17) are ensured nonzero. Once the nonzero estimated parameters a ^ ( 1 ) and b ^ ( 1 ) are obtained, and then c ^ ( 1 ) and d ^ ( 1 ) are ensured nonzero, due to that the rank of matrices η ^ 3 and η ^ 4 with nonzero parameters a ^ ( 1 ) and b ^ ( 1 ) are also full row rank. Following above arguments, we conclude that once the PE matrices and the nonzero initial values are given, the proposed algorithm can provide consistent estimated parameters.
The parameters are estimated asynchronously from the PPMs by two independent projection based iterative methods, thus the proposed method, referred to as Asynchronous Parsimonious Method (APM), is summarized as Algorithm 1.
Algorithm 1:
1.
Initialization: d ^ ( 0 ) and c ^ ( 0 ) : arbitrary nonzero values.
2.
Iteration:
for k = 0 : convergence
a1.
Estimate a ^ ( k ) as (13) and (14).
a2.
Estimate d ^ ( k ) as (15).
a3.
If the stopping criterion is satisfied, and break, else go to step (a1).
end
3.
Iteration:
for k = 0 : convergence
b1.
Estimate b ^ ( k ) as (17) and (18).
b2.
Estimate c ^ ( k ) as (19).
b3.
If the stopping criterion is satisfied, and break, else go to step (b1).
end
4.
Let a ^ = a ^ ( k ) sgn ( a ^ 1 ( k ) ) , b ^ = b ^ ( k ) sgn ( b ^ 1 ( k ) ) , d ^ 2 = d ^ 2 ( k + 1 ) sgn ( a ^ 1 ( k ) ) , and c ^ 2 = c ^ 2 ( k + 1 ) sgn ( b ^ 1 ( k ) ) , where sgn ( ) is a sign function and a ^ 1 ( k ) and b ^ 1 ( k ) are the first element of a ^ ( k ) and b ^ ( k ) , respectively.

4. Variance Analysis

Following the Theorem 1 in [11], one can eaily prove that the proposed method with arbitrary nonzero initial values d ^ ( 0 ) and c ^ ( 0 ) provides consistent estimates for H-W models. This section, we analyze the variance of the AMP estimates relative to that of conventional TSA algorithm. Due to all paramters are estimated in a similar way, we only analy the variances of estimated paramter a ^ . The LS estimation error variance of θ ^ in TSA [4] can be computed as,
c o v ( θ ^ 1 , θ ^ 2 ) = δ e 2 ( Φ Φ ) 1
Substituting (4e) into (20), we have
c o v ( θ ^ 1 , θ ^ 2 ) = δ e 2 F F F G G F G G 1
Based the block matrix inversion relation in [15], we have
c o v ( θ ^ 2 ) = δ e 2 ( G G ) 1 + δ e 2 ( G G ) 1 G F Q 1 ( G F ) ( G G ) 1
where Q = F F F G ( G G ) 1 G F .
Since the solutions of a ^ , b ^ , c ^ and d ^ are provided by SVD, decompositions of θ ^ 1 , and θ ^ 2 . Omitting the error terms more than first order which do not contribute to the estimation error variance, deriving based on first-order Taylor expansion of SVD in [16], the variance of the parameter estimation errors of a ^ can be computed in approximately as following,
c o v ( a ^ T S A ) = δ e 2 ( η 1 F F η 1 ) 1 + δ e 2 ( η 1 F F ) 1 F G Q 1 ( F G ) ( F F η 1 ) 1
Note that the second terms at the right-hand side of (52) are positive definite, we have
c o v ( a ^ T S A ) δ e 2 ( η 1 F F η 1 ) 1
In a similar way as (20), the estimation error variances of a ^ , b ^ , c ^ and d ^ in APM are computed as following
c o v ( a ^ A P M ) = δ e 2 ( η ^ 1 F F η ^ 1 ) 1
Since estimates from AMP are consistent, neglecting the high-order terms in above equations for asymptotic variance, and only considering the noise terms for estimation error variances, we have
c o v ( a ^ A P M ) = δ e 2 ( η 1 F F η 1 ) 1
Compared (24) and (26), we conclude that the APM estimates generally have a smaller variance than those of the TSA estimates by neglecting high-order terms in estimation error.

5. Case Studies

In this section, we use a benchmark Hammerstein–Wiener model with highly nonlinearities modified from [4] to evaluate the proposed APM and compare it with the TSA. The system model of the benchmark example is presented in the following:
a = [ 0.1569 , 0.0934 , 0.5477 , 0 , 0.7303 , 0.3651 ] b = [ 0.1569 , 0.0934 , 0.5477 , 0 , 0.7303 , 0.3651 ] c = [ 1 , 4 , 3 , 7 , 1 , 3.4 ] d = [ 0.4 , 0.25 ] f ( t ) = [ u ( t ) , u 2 ( t ) , u 3 ( t ) , u 4 ( t ) , u 5 ( t ) , u 6 ( t ) ] g ( t ) = [ y ( t ) , sin ( y ( t ) ) ] .
The input { u ( t ) } and the measurement noise { e ( t ) } are chosen as two zero-mean Gaussian random sequences. The variance of { u ( t ) } is fixed as one, while the variance of { e ( t ) } is adjusted to achieve a given Signal-to-Noise Ratio (SNR). In this example, three SNR values will be considered, i.e., 15, 20, and 25 dB. For each experimental condition, we conduct a Monte-Carlo (MC) simulation of 100 runs, each with N = 500 data points. The performance of the methods is evaluated with respect to the accuracy and standard deviations in tracking the system parameters θ = ( a , b , c , d ) and output.
More specifically, the quality of the estimates is evaluated by investigating the Parameter Estimation Errors (PEE), Variance Accounted For (VAF) on a validation data set which is different from the one used for model identification. The PEE and VAF are defined as follows:
PEE = θ θ ^ 2 θ 2 100 % VAF = max 0 , 1 var ( y ¯ k y ^ ) var ( y ¯ k ) 100 %
where the operator var ( · ) denotes the variance of its argument, y ¯ k denotes the noise-free simulated output of the system, and y ^ k denotes the the noise-free simulated model output. The VAF is allowed to reach 100 % , and PPE is allowed to reach 0 % when the model is equal to the true system.
Theoretically, the computational complexity of the proposed APM per iteration is O ( N 2 ) , which is obviously more computationally demanding than O ( N ) in the TSA, because of the projection operations [17]. Fortunately, from the averaged estimation results of Euclidean norm θ for 100 MC simulations at each iteration in APM under different SNRs shown in Figure 1, we observe that APM converges very quickly within 4 iterations for high SNRs (e.g., 20 and 25 dB), and also provides accurate estimation results even for a low SNR of 15 dB. It is concluded that the APM can provide accurate estimation results for highly nonlinear systems from a limited number of noisy samples. The estimated parameters in θ along with the mean values and standard deviations and the PEE are listed in Table 1. From the results shown in Table 1, we can conclude that
  • Under the same noise levels, the proposed APM provides better estimation results in terms of accuracy and standard deviations than TSA.
  • The two methods yield similar results for SNR = 25 dB (the corresponding PEEs are 10.8245   ( ± 5.8194 ) and 11.8245   ( ± 7.0916 ) , respectively). The performance of APM is obviously better than the performance of TSA for c 2 , c 4 , c 6 , PEE for SNR = 20 dB, and c 1 , c 2 , c 3 , c 4 , c 5 , c 6 , VAF , PPE for SNR = 15 dB. The attenuation of the signal levels does not obviously affect the APM while it deteriorates the performance of TSA significantly, showing that proposed APM is more robust to noise and data size.
The calculated VAFs for the proposed APM and TSA are given in Figure 2, while the corresponding histograms of the VAF values are shown in Figure 3. In Table 2 the mean values and standard deviations of VAF are given. The results show that the VAFs of APM are better than that of TSA under the same noise levels. The results verify again our argument: as the avoidance of introducing extra parameters and performing SVD to estimate the parameters of ( a , b , c , d ) from the products a d and b c , the proposed APM based on PPMs with fewer parameters can enhance the accuracy of the estimated quantities and decrease the variance of estimates, for the identification of highly nonlinear Hammerstein-Wiener systems given short noisy data length sequences.
To evaluate the sensitivity of the proposed APM under different sample lengths, the mean values of PEE and VAF for 100 Monte-Carlo simulations with respect to different N and SNRs using proposed APM and TSA are shown in Figure 4 and Figure 5, respectively. It can be seen that the proposed method provides quite good performance when N > 100 . The performance of the proposed APM in terms of PEE and VAF is better than the TSA especially for identifying N-L-N systems from a limited number of noisy data with a low SNR, see PEE and VAF when data point N < 1000 and SNR = 25 dB. These results show that proposed APM is more robust to noise and data size.

6. Conclusions

A new identification formulation for H-W models, referred to as APM, is presented. The avoidance of using over-parametrization models leads to a number of improvements, i.e., reducing the number of unknown parameters to be estimated and avoiding a truncated SVD step, so improving the estimated quantities in terms of accuracy and variance compared to TSA. These improvements especially hold for highly nonlinear H-W model identification from a limited number of noisy samples. The reasons why the new method results in an improvement model over the previous LS method have been demonstrated by variance analysis. The application to an illustrative example has well demonstrated the effectiveness and merit of the proposed method in comparison with TSA. Future work should focus on identifying other nonlinear systems by taking into account some additional prior knowledge about the systems.

Author Contributions

Conceptualization, L.S. and J.H.; methodology, L.S. and J.H.; software, L.S. and J.H.; validation, L.S. and J.H.; formal analysis, L.S. and J.H.; investigation, L.S. and J.H.; writing—original draft preparation, L.S. and J.H.; writing—review and editing, J.H. and C.X.; visualization, J.H., Z.F. and C.X.; supervision, J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Chongqing Natural Science Foundation under Grant CSTB2022NSCQ-MSX1225; in part by the China Postdoctoral Science Foundation under Grant 2022MD713688; in part by the Chongqing Postdoctoral Science Foundation under Grant 2021XM3079; in part by the Science and Technology Research Program of Chongqing Municipal Education Commission under Grant KJQN202000602.

Institutional Review Board Statement

This study does not involve any institutional review issues.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data is given by The system model of the benchmark example is presented in Case Studies in Section 3 and simulated by matlab, which can be obtained by emailing jiehou.phd@hotmail.com.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
H-WHammerstein–Wiener
N-L-NNonlinear-Linear-Nonlinear
BELSBias-Eliminating Least-Squares
TSATwo-Step Algorithms
OPMover-parametrization model
SVDsSingular Value Decompositions
AM-MISG   auxiliary model multi-innovation stochastic gradient
APMAsynchronous Parsimonious Method
TSATwo-Step Algorithms
SNRSignal-to-Noise Ratio
PEEParameter Estimation Errors
VAFVariance Accounted For

References

  1. Schoukens, M.; Tiels, K. Identification of block-oriented nonlinear systems starting from linear approximations: A survey. Automatica 2017, 85, 272–292. [Google Scholar] [CrossRef] [Green Version]
  2. Wang, D.; Ding, F. Hierarchical least squares estimation algorithm for Hammerstein-Wiener systems. IEEE Signal Process. Lett. 2012, 19, 825–828. [Google Scholar] [CrossRef]
  3. Li, F.; Jia, L. Parameter estimation of Hammerstein-Wiener nonlinear system with noise using special test signals. Neurocomputing 2019, 344, 37–48. [Google Scholar] [CrossRef]
  4. Bai, E. An optimal two-stage identification algorithm for Hammerstein-Wiener nonlinear systems. Automatica 1998, 34, 333–338. [Google Scholar] [CrossRef]
  5. Xu, L.; Ding, F.; Yang, E.F. Auxiliary model multiinnovation stochastic gradient parameter estimation methods for nonlinear sandwich systems. Int. J. Robust Nonlinear Control 2020, 31, 148–165. [Google Scholar] [CrossRef]
  6. Ni, B.; Gilson, M.; Garnier, H. Refined instrumental variable method for Hammerstein-Wiener continuous-time model identification. IET Control Theory Appl. 2013, 7, 1276–1286. [Google Scholar] [CrossRef]
  7. Wang, Z.; Wang, Y.; Ji, Z. A novel two-stage estimation algorithm for nonlinear Hammerstein-Wiener systems from noisy input and output data. J. Frankl. Inst. 2017, 354, 1937–1944. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, J.; Zhang, Q.; Ljung, L. Revisiting Hammerstein system identifification through the two-stage algorithm for bilinear parameter estimation. Automatica 2009, 45, 2627–2633. [Google Scholar] [CrossRef]
  9. Hou, J.; Su, H.; Yu, C.; Chen, F.; Li, P. Bias-Correction Errors-in-Variables Hammerstein Model Identification. IEEE Trans. Ind. Electron. 2022. [Google Scholar] [CrossRef]
  10. Gomez, J.; Baeyens, E. Identifification of block-oriented nonlinear systems using or thonormal bases. J. Process Control 2004, 14, 685–697. [Google Scholar] [CrossRef]
  11. Hou, J.; Chen, F.; Li, P.; Zhu, Z. Fixed point iteration-based subspace identification of Hammerstein state-space models. IET Control Theory Appl. 2019, 13, 1173–1181. [Google Scholar] [CrossRef]
  12. Hou, J.; Chen, F.; Li, P.; Shu, J.; Zhao, F. Recursive Parsimonious Subspace Identification for Closed-Loop Hammerstein Nonlinear Systems. IEEE Access 2019, 7, 173515–173523. [Google Scholar] [CrossRef]
  13. Hou, J.; Chen, F.; Li, P.; Zhu, Z. Gray-Box Parsimonious Subspace Identification of Hammerstein-Type Systems. IEEE Trans. Ind. Electron. 2021, 68, 9941–9951. [Google Scholar] [CrossRef]
  14. Hou, J. Parsimonious model based consistent subspace identification of Hammerstein systems under periodic disturbances. Int. J. Control. Autom. Syst. 2022. [Google Scholar] [CrossRef]
  15. Ding, F. Decomposition based fast least squares algorithm for output error systems. Signal Process. 2013, 93, 1235–1242. [Google Scholar] [CrossRef]
  16. Hou, J.; Su, H.; Yu, C.P.; Chen, F.; Li, P.; Xie, H.; Li, T. Consistent Subspace Identification of Errors-In-Variables Hammerstein Systems. IEEE Trans. Syst. Man Cybern. Syst. 2022. [Google Scholar] [CrossRef]
  17. Ding, F. Computational efficiency of the identification methods. PartB: Iterative algorithm. J. Nanjing Univ. Inf. Sci. Technol. Nat. Sci. Ed. 2021, 4, 385–401. [Google Scholar]
Figure 1. Averaged estimation results of θ 2 at each iteration for APM under different SNRs.
Figure 1. Averaged estimation results of θ 2 at each iteration for APM under different SNRs.
Processes 10 02664 g001
Figure 2. VAF of APM (left, red line) and TSA (right, blue line) under different SNRs.
Figure 2. VAF of APM (left, red line) and TSA (right, blue line) under different SNRs.
Processes 10 02664 g002
Figure 3. Histogram of VAF values of APM and TSA under different SNRs.
Figure 3. Histogram of VAF values of APM and TSA under different SNRs.
Processes 10 02664 g003
Figure 4. Averaged results of PEE for APM (left, red line) and TSA (right, blue line) under different N and SNRs, where N [ 100 , 4000 ] .
Figure 4. Averaged results of PEE for APM (left, red line) and TSA (right, blue line) under different N and SNRs, where N [ 100 , 4000 ] .
Processes 10 02664 g004
Figure 5. Averaged results of VAF for APM (left, red line) and TSA (right, blue line) under different N and SNRs, where N [ 100 , 4000 ] .
Figure 5. Averaged results of VAF for APM (left, red line) and TSA (right, blue line) under different N and SNRs, where N [ 100 , 4000 ] .
Processes 10 02664 g005
Table 1. Estimated results of θ and PEE.
Table 1. Estimated results of θ and PEE.
TrueAPM (SNR = 15)APM (SNR = 20)APM (SNR = 25)TSA (SNR = 15)TSA (SNR = 20)TSA (SNR = 25)
a 1 = 0.1569 0.1604 (±0.0793)0.1654 (±0.0617)0.1477 (±0.0423)0.1643 (±0.0840)0.1669 (±0.0597)0.1470 (±0.0467)
a 2 = 0.0934 0.0912 (±0.0632)0.0906 (±0.0443)0.0979 (±0.0326)0.0855 (±0.0836)0.0907 (±0.0485)0.0928 (±0.0367)
a 3 = 0.5477 0.5442 (±0.1472)0.5480 (±0.1187)0.5465 (±0.0277)0.5162 (±0.1961)0.5407 (±0.0413)0.5474 (±0.0280)
a 4 = 0 0.0009 (±0.0500)0.0032 (±0.0357)0.0002 (±0.0267)0.0014 (±0.00826)0.0041 (±0.0413)0.0026 (±0.0333)
a 5 = 0.7303 −0.7218 (±0.0481)−0.7178 (±0.1473)−0.7271 (±0.0252)−0.6758 (±0.2514)−0.7317 (±0.0419)−0.7249 (±0.0257)
a 6 = 0.3651 −0.3597 (±0.07317)−0.3478 (±0.0839)−0.3690 (±0.0228)−0.3164 (±0.1460)−0.3510 (±0.0398)−0.3713 (±0.0267)
b 1 = 0.1569 0.1572 (±0.0128)0.1577 (±0.0068)0.1558 (±0.0040)0.2757 (±0.2053)0.1515 (±0.0770)0.1620 (±0.0377)
b 2 = 0.0934 0.0930 (±0.0121)0.0931 (±0.0060)0.0939 (±0.0038)0.0039 (±0.3265)0.0692 (±0.1037)0.0913 (±0.0367)
b 3 = 0.5477 0.5465 (±0.0155)0.5478 (±0.0112)0.5468 (±0.0079)0.1405 (±0.4996)0.4736 (±0.2627)0.5469 (±0.0313)
b 4 = 0.7303 0.7293 (±0.0114)0.7306 (±0.0074)0.7299 (±0.0050)0.1432 (±0.5629)0.6178 (±0.3892)0.7220 (±0.0511)
b 5 = 0.3651 0.3676 (±0.0176)0.3634 (±0.0156)0.3675 (±0.0093)0.1299 (±0.4004)0.3028 (±0.2003)0.3720 (±0.1045)
c 1 = 1 1.0774 (±0.2146)0.9999 (±0.1701)1.0022 (±0.0967)0.2651 (±0.8143)0.8495 (±0.5122)0.9914 (±0.3376)
c 2 = 4 4.0237 (±0.8490)3.9649 (±0.4590)4.0077 (±0.2340)1.1182 (±3.0569)3.3149 (±1.7629)3.9646 (±0.4040)
c 3 = 3 2.6377 (±1.3845)2.9785 (±0.6425)3.0432 (±0.3804)0.9981 (±2.5809)2.5448 (±1.6301)3.0505 (±0.3848)
c 4 = 7 7.1772 (±1.6055)7.0295 (±1.4038)7.0416 (±0.7487)1.9601 (±10.3455)6.1131 (±4.6544)7.1346 (±0.8611)
c 5 = 1 1.0199 (±1.3840)1.0168 (±0.5661)0.9645 (±0.3401)0.0885 (±1.6670)0.8553 (±0.7665)0.9455 (±0.3612)
c 6 = 3.4 3.5277 (±1.3443)3.4013 (±1.0777)3.3914 (±0.5844)0.7933 (±5.1063)2.7925 (±1.8014)3.2967 (±0.6655)
d 1 = 0.4 0.5043 (±0.1304)0.4927 (±0.1017)0.5014 (±0.0040)0.4731 (±0.1723)0.5027 (±0.0071)0.5064 (±0.0045)
d 2 = 0.25 0.2551 (±0.0675)0.2481 (±0.0595)0.2500 (±0.0211)0.2518 (±0.1244)0.2549 (±0.0330)0.2501 (±0.0216)
PPE (%)38.7442 (±21.5544)18.7887 (±10.3743)10.8245 (±5.8194)130.4982 (±71.5442)36.6713 (±48.9617)11.8245 (±7.0916)
Table 2. Estimated Results of VAF under Different SNRs.
Table 2. Estimated Results of VAF under Different SNRs.
MethodsVAF
APM (SNR = 15 dB)99.79 (±0.0977)
APM (SNR = 20 dB)99.94 (±0.0270)
APM (SNR = 25 dB)99.97 (±0.0214)
TSA (SNR = 15 dB)56.81 (±31.1700)
TSA (SNR = 20 dB)97.03 (±3.5800)
TSA (SNR = 25 dB)99.40 (±0.5200)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sun, L.; Hou, J.; Xing, C.; Fang, Z. A Robust Hammerstein-Wiener Model Identification Method for Highly Nonlinear Systems. Processes 2022, 10, 2664. https://doi.org/10.3390/pr10122664

AMA Style

Sun L, Hou J, Xing C, Fang Z. A Robust Hammerstein-Wiener Model Identification Method for Highly Nonlinear Systems. Processes. 2022; 10(12):2664. https://doi.org/10.3390/pr10122664

Chicago/Turabian Style

Sun, Lijie, Jie Hou, Chuanjun Xing, and Zhewei Fang. 2022. "A Robust Hammerstein-Wiener Model Identification Method for Highly Nonlinear Systems" Processes 10, no. 12: 2664. https://doi.org/10.3390/pr10122664

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop