Next Article in Journal
The Log-Bimodal Asymmetric Generalized Gaussian Model with Application to Positive Data
Next Article in Special Issue
Generalized Pandemic Model with COVID-19 for Early-Stage Infection Forecasting
Previous Article in Journal
Modified and Improved Algorithm for Finding a Median Path with a Specific Length () for a Tree Network
Previous Article in Special Issue
Forecasting Day-Ahead Brent Crude Oil Prices Using Hybrid Combinations of Time Series Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Two-Threshold-Variable Integer-Valued Autoregressive Model

1
School of Mathematics, Jilin University, Changchun 130012, China
2
School of Mathematics and Statistics, Henan University, Kaifeng 475004, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(16), 3586; https://doi.org/10.3390/math11163586
Submission received: 21 July 2023 / Revised: 7 August 2023 / Accepted: 17 August 2023 / Published: 18 August 2023
(This article belongs to the Special Issue Time Series Analysis)

Abstract

:
In the past, most threshold models considered a single threshold variable. However, for some practical applications, models with two threshold variables may be needed. In this paper, we propose a two-threshold-variable integer-valued autoregressive model based on the binomial thinning operator and discuss some of its basic properties, including the mean, variance, strict stationarity, and ergodicity. We consider the conditional least squares (CLS) estimation and discuss the asymptotic normality of the CLS estimator under the known and unknown threshold values. The performances of the CLS estimator are compared via simulation studies. In addition, two real data sets are considered to underline the superior performance of the proposed model.

1. Introduction

Integer-valued time series data often occur in real applications, such as in the number of births at a hospital for several consecutive months; in the number of workers that are fired from a factory each month; in the number of claims and in information transmission times of insurance companies every month; and, particularly, in health studies as the daily number of infected patients or deaths due to a virus. The binomial thinning operator proposed by Steutel and van Harn [1] has been widely used to construct the autoregressive model, i.e., the integer-valued autoregressive (INAR) model (Al-Osh and Alzaid [2], McKenzie [3]), which is a popular method to analyze the integer-valued time series data and is defined as follows:
X t = α X t 1 + ϵ t , t = 1 , 2 ,
where { ϵ t } is a sequence of independently and identified distributed (i.i.d.) random variables and is independent of X s , s > t , “ α " is the binomial thinning operator with
α X : = i = 1 X B i , X N ,
and B i is independent of X and is a sequence of i.i.d. Bernoulli random variables with P ( B i = 1 ) = α ( 0 , 1 ) = 1 P ( B i = 0 ) . See Du and Li [4], Silva and Oliveira [5], Silva and Silva [6], and Zhang et al. [7] for more extensions of the INAR model, among others.
However, the above INAR model and its extensions aim to analyze the integer-valued time series with a linear structure and are unavailable to analyze the integer-valued time series with a nonlinear structure. Scotto et al. [8] proposed a discrete counterpart of the conventional max- autoregressive process of order one. It is based on the binomial thinning operator and driven by a sequence of i.i.d. non-negative integer-valued random variables with either a regularly varying right tail or an exponential-type right tail. Aleksić and Ristić [9] introduced a new minification INAR model of the first-order to solve the problem, which can arise when the binomial thinning operator or the negative binomial thinning operator is used. Namely, if one of these thinning operators is used in the construction of the minification model then it is possible that the model becomes zero constantly over time.
The threshold autoregressive (TAR) model proposed by Tong [10] provides an efficient method with which to handle continuous-valued time series data with a nonlinear structure. Boero and Marrocu [11] and Potter [12] applied threshold models to economy and finance. Dueker et al [13] proposed a contemporaneous TAR model for the bond market. See Tong [14] for more discussion on continuous-valued threshold models. Li and Tong [15] proposed a faster approach (called the nested sub-sample search algorithm) to fit a threshold model using the least squares method. In an analogy to the TAR model, Monteiro et al. [16] introduced an integer-valued self-exciting threshold autoregressive process (SETINAR(2,1)), which is driven by an independent Poisson-distributed random variable. Wang et al. [17] proposed a self-excited threshold Poisson autoregressive model, which assumes a two-regime structure of the conditional mean process according to the magnitude of the lagged observations. Yang et al. [18] proposed an integer-valued threshold autoregressive process driven by an independent negative-binomial distributed random variable and the negative binomial thinning operator. To explore the relationship between stock return autocorrelation and trading volume, Zhang et al. [19] proposed a multiple-threshold-variable autoregressive model and applied it to analyze quarterly U.S. real GNP data. But the multiple-threshold-variable autoregressive model is restricted to a continuous-valued time series, and few studies have discussed a similar model for the nonnegative integer-valued time series. To fill this gap, we propose a new two-threshold-variable INAR (2-TINAR) model, which provides an alternative way to analyze nonnegative integer-valued time series with a nonlinear structure.
The paper is organized as follows. Section 2 defines the 2-TINAR model and establishes its stability properties. Section 3 considers conditional least squares (CLS) estimation. Section 4 gives a simulation study. Section 5 considers two real data applications to illustrate the effectiveness of the proposed model. Section 6 concludes.

2. Two-Threshold-Variable Integer-Valued Autoregressive Model

In this paper, we first give the definition of the 2-TINAR model and then discuss some properties of the model.
Definition 1.
The 2-TINAR process { X t } is defined as
X t = j = 1 4 ( α j 1 X t 1 + α j 2 X t 2 + ϵ j t ) I j t ( r , s ) , t = 1 , 2 , ,
where
(1) 
( r , s ) are the threshold parameters and
I 1 t ( r , s ) = I ( X t 1 > r , X t 2 > s ) = I ( X t 1 > r ) I ( X t 2 > s ) , I 2 t ( r , s ) = I ( X t 1 r , X t 2 > s ) = I ( X t 1 r ) I ( X t 2 > s ) , I 3 t ( r , s ) = I ( X t 1 r , X t 2 s ) = I ( X t 1 r ) I ( X t 2 s ) , I 4 t ( r , s ) = I ( X t 1 > r , X t 2 s ) = I ( X t 1 > r ) I ( X t 2 s ) ;
(2) 
α j i ( 0 , 1 ) , 0 < i = 1 2 α j i < 1 , i = 1 , 2 , j = 1 , 2 , 3 , 4 , is the binomial thinning operator, and the operators in α j 1 X t 1 and α j 2 X t 2 operate independently;
(3) 
t , ϵ j t P o i s ( λ j ) and for fixed j, ϵ j t is i.i.d. and independent of α j i X t i and X t i .
In the following, we consider the properties of the 2-TINAR model, including stationarity, ergodicity, mean and variance, which will be given in the next three propositions, whose proofs are delegated to Appendix A.
Proposition 1.
Let Y t = ( X t , X t 1 ) ; then,
(1) 
{ Y t } is an irreducible, aperiodic, and positive recurrent Markov chain;
(2) 
{ Y t } is an ergodic sequence, and a strictly stationary process satisfying (1) exists.
Proposition 2.
Assume that { X t } is generated from (1); then, E ( X t k ) < , k = 1 , 2 , 3 .
Proposition 3.
Assume that { X t } is generated from (1) and F = σ { X t i , i 1 } . Then, the mean and variance are
(1) 
E ( X t | F t 1 ) = j = 1 4 ( α j 1 X t 1 + α j 2 X t 2 + λ j ) I j t ( r , s ) ;
(2) 
E ( X t ) = j = 1 4 p j ( α j 1 u j + α j 2 u j * + λ j ) ;
(3) 
V a r ( X t | F t 1 ) = j = 1 4 ( α j 1 ( 1 α j 1 ) X t 1 + α j 2 ( 1 α j 2 ) X t 2 + λ j ) I j t ( r , s ) ;
(4) 
V a r ( X t ) = j = 1 4 ( α j 1 2 ( p j ( v j + u j 2 ) p j 2 u j 2 ) + α j 2 2 ( p j ( v j * + ( u j * ) 2 ) p j 2 ( u j * ) 2 ) + 2 ( α j 1 α j 2 w j p j α j 1 α j 2 p j 2 u j u j * ) + 2 ( α j 1 λ j p j u j α j 1 λ j p j 2 u j ) + 2 ( α j 2 λ j p j u j * α j 2 λ j p j 2 u j * ) + α j 1 ( 1 α j 1 ) p j u j + α j 2 ( 1 α j 2 ) p j u j * + λ j p j ) + m = 1 6 C m , m = 1 , 2 , , 6 , where p j , u j , v j , u j * , v j * , w j and C m are given in Appendix A.

3. Conditional Least Squares Estimation

In this section, we use the CLS method to estimate the parameters involved in the 2-TINAR model. Here, we consider the following two cases: the first one is that the threshold values r and s are known and the second one is that the threshold values r and s are unknown.

3.1. Known Case of ( r , s )

In this part, we assume that { X t , t = 1 , . . . , n } comprises observations generated by (1), ψ j = ( α j 1 , α j 2 , λ j ) is the vector of regression parameters, j = 1 , 2 , 3 , 4 , ϕ = ( ψ 1 , ψ 2 , ψ 3 , ψ 4 ) = ( ϕ 1 , . . . , ϕ 12 ) , and g ( ϕ , X t 1 , X t 2 ) = E ϕ ( X t 1 | F t 1 ) = j = 1 4 ( α j 1 X t 1 + α j 2 X t 2 + λ j ) I j t ( r , s ) . Then, the CLS estimate ψ ^ j , C L S = ( α ^ j 1 , C L S , α ^ j 2 , C L S , λ ^ j , C L S ) is obtained by minimizing the function
Q ( ϕ ) = t = 3 n ( X t g ( ϕ , X t 1 , X t 2 ) ) 2 = t = 3 n q t 2 ( ϕ ) ,
where q t ( ϕ ) = X t j = 1 4 ( α j 1 X t 1 + α j 2 X t 2 + λ j ) I j t ( r , s ) . Then, the closed form of ψ j , C L S is obtained by the following equations:
Q ( ϕ ) α j 1 = 2 t = 3 n X t j = 1 4 α j 1 X t 1 + α j 2 X t 2 + λ j I j t ( r , s ) X t 1 = 0 , Q ( ϕ ) α j 2 = 2 t = 3 n X t j = 1 4 ( α j 1 X t 1 + α j 2 X t 2 + λ j ) I j t ( r , s ) X t 2 = 0 , Q ( ϕ ) λ j = 2 t = 3 n X t j = 1 4 ( α j 1 X t 1 + α j 2 X t 2 + λ j ) I j t ( r , s ) = 0 .
Denote
A j = t = 3 n I j t ( r , s ) X t 1 2 t = 3 n I j t ( r , s ) X t 1 X t 2 t = 3 n I j t ( r , s ) X t 1 t = 3 n I j t ( r , s ) X t 1 X t 2 t = 3 n I j t ( r , s ) X t 2 2 t = 3 n I j t ( r , s ) X t 2 t = 3 n I j t ( r , s ) X t 1 t = 3 n I j t ( r , s ) X t 2 t = 3 n I j t ( r , s ) ,
D j = t = 3 n I j t ( r , s ) X t X t 1 , t = 3 n I j t ( r , s ) X t X t 2 , t = 3 n I j t ( r , s ) X t .
Then, ψ j , C L S = A j 1 D j .
To study the asymptotic behaviour of the estimators, we make the following assumptions about the underlying process and the parameter space.
Assumption 1.
If { X t } is generated from (1), then the parameter space ϕ is a compact subset of D × R + 4 , and D = ( 0 , 1 ) × ( 0 , 1 ) × ( 0 , 1 ) × ( 0 , 1 ) × ( 0 , 1 ) × ( 0 , 1 ) × ( 0 , 1 ) × ( 0 , 1 ) is a compact subset of R + 8 .
Assumption 2.
The model (1) is identifiable, i.e., p ϕ p ϕ 0 , if ϕ ϕ 0 , where p ϕ denotes the marginal distribution of { X t } with parameter ϕ.
For Assumption 1, assume the parameter space is compact so that the asymptotic properties of the CLS estimator can be guaranteed, which is common in INAR models. Parameter identifiability in Assumption 2 is a property that concerns whether the model parameters can be uniquely determined, which is the foundation for parameter estimation.
The following theorem establishes the asymptotic porperties of the CLS estimator, whose proof will be given in Appendix A.
Theorem 1.
Under the Assumptions 1 and 2, ϕ ^ C L S is strongly consistent and asymptotically normally distributed with
T ( ϕ ^ C L S ϕ 0 ) d N ( 0 , V 1 W V 1 ) ,
where V = E E ( X t | F t 1 ) ϕ E ( X t | F t 1 ) ϕ and W = E ( q t ( ϕ ) ) 2 E ( X t | F t 1 ) ϕ E ( X t | F t 1 ) ϕ .

3.2. Unknown Case of ( r , s )

Under the unknown case of ( r , s ) , we first estimate ( r , s ) , which is obtained by minimizing (2) by the following steps:
(1)
For each candidate for ( r , s ) in CR, we estimate ϕ ^ by minimizing Q ( ϕ , r , s ) , i.e.,
ϕ ^ = arg min ϕ Q ( ϕ , r , s ) .
(2)
The estimator for thresholds ( r , s ) is obtained by searching over all of the candidates for ( r , s ) in CR, i.e.,
( r ^ , s ^ ) = arg min ( r , s ) C R Q ( ϕ ^ , r , s ) ,
where CR is the set of candidates for estimators for ( r , s ) with CR = { X { i } , X { i + 1 } } , i = ( 0.2 , 0.25 , 0.3 , 0.35 , 0.4 , 0.45 , 0 , 5 , 0.55 , 0.6 , 0.65 , 0.7 , 0.75 , 0.8 ) . To select a proper CR, we propose a validated method, which can put less pressure on the capacity of computing and guarantee the accuracy of estimates; { X { i } , X { i + 1 } } are thirteen candidates for ( r ^ , s ^ ) . Actually, the estimators for ( r , s ) are searched for from X { 0.20 } to X { 0.85 } , and this is a sufficient search range for the threshold. Hence, this method guarantees reasonable and sufficient search ranges without too much pressure on computing.
Based on the initial setting of the parameter space, both of the thresholds r and s are integers. Therefore, the consistency of r ^ means that r ^ = r utterly, and so does s ^ . It is feasible that we estimate the other parameters by assuming that thresholds ( r , s ) are known, which is similar to the discussion in Wang et al. [17] because the validity of the estimates with unknown thresholds, as for the other parameters, is asymptotically identical to that obtained with known ( r , s ) . Hence, in this subsection, we treat the thresholds r and s as known parameters and consider the consistency for ϕ = ( α j 1 , α j 2 , λ j ) , j = 1 , 2 , 3 , 4 .

4. Simulation Study

In this section, we illustrate the finite sample property of the CLS estimates under the known and unknown cases of ( r , s ) . In the simulation, we use m = 10,000 replications and set the sample size is T = 500 , 1000 , 2000 and 10,000.

4.1. Known Case of ( r , s )

In the simulation study, we consider the following parameter combinations:
( C 1 ) ( 0.3 , 0.2 , 7 , 0.2 , 0.25 , 6 , 0.2 , 0.3 , 8 , 0.3 , 0.2 , 6 ) with ( r , s ) = ( 13 , 11 ) , ( C 2 ) ( 0.3 , 0.35 , 15 , 0.3 , 0.35 , 20 , 0.3 , 0.4 , 25 , 0.35 , 0.25 , 15 ) with ( r , s ) = ( 55 , 53 ) .
As discussed in Li and Tong [15], if the proportion of observations in one regime to the whole is less than 5 % , the estimation results may not be reliable. To illustrate the reasonableness of ( r , s ) given in (C1) and (C2), we give two sample paths of the sample generated by the 2-TINAR model with (C1) and (C2) in Figure 1, which shows that the proportion of the observations in each range is no less than 20 % . In Figure 1, circle means sample point, red dot line means the value of r, and blue dot line means the value of s.
The mean and standard deviation (SD) of the estimates are summarized in Table 1, from which we obtain that the CLS method performs reasonably well when ( r , s ) is known because the mean gradually approaches the true value of the parameter and SD decreases gradually, when the sample size is increasing.
To further account for the reasonableness of the CLS estimates, we present the boxplots of the parameter combinations (C1) in Figure 2 (the boxplots of (C2) are similar, and we omit them), and the QQ-plots of the parameter combinations (C1) and (C2) indicate the asymptotic normality of the CLS estimator. For saving space, we omit the QQ-plots, which are available upon request. All of them highlight that the good performances of the CLS estimate under the case of ( r , s ) are known.

4.2. Unknown Case of ( r , s )

In this part, we first let (C3) ( 0.2 , 0.25 , 3 , 0.25 , 0.35 , 5 , 0.3 , 0.3 , 4 , 0.3 , 0.25 , 6 , 13 , 14 ) and (C4) ( 0.30 , 0.25 , 6 , 0.25 , 0.35 , 6 , 0.4 , 0.3 , 9 , 0.3 , 0.35 , 8 , 30 , 31 ) . Then, we use the approach given in Section 3.2 to obtain ( r ^ , s ^ ) .
To illustrate the reasonableness of ( r ^ , s ^ ) given in (C3) and (C4), we give two sample paths of the sample generated by 2-TINAR model with (C3) and (C4) in Figure 3, which shows that the proportion of the observations in each range is no less than 20 % . In Figure 3, circle, red and blue dot line in figure are the same as Figure 1.
The mean and SD of the estimates are summarized in Table 2, from which we can see that with the increase in sample size, the mean gradually approaches the true value of the parameter and SD decreases gradually. The boxplots (C3) and (C4) are similar to (C1) and (C2). We present the boxplots of the parameter combinations (C3) in Figure 4(the boxplots of (C4) are similar and we omit them).
From Figure 4, the median of the estimator is closer to the true value and the quartile range and overall range of the estimated values become narrower, both of which indicate the consistency of the estimators. The QQ−plots of the parameter combinations (C3) and (C4) indicate the asymptotic normality of the CLS estimator. For the same reason, we omit the QQ-plots. All of them highlight that the good performances of the CLS estimate under the case of ( r , s ) are unknown.

5. Two Real Examples

In this section, we use 2-TINAR(2) models to study two stock datasets listed in the New York Stock Exchange (NYSE).

5.1. Siparex Croissance Stock

In this subsection, we consider the daily number of trades of a stock listed in the NYSE (Siparex Croissance). By computation, the mean is 10.0190 and the variance is 129.7295, which shows that this dataset is over-dispersed and implies that it may be better suited to the piecewise structure. Figure 5 shows the path of the data, whose autocorrelation (ACF) and partial autocorrelation functions (PACF) are presented in Figure 6.
We compare the proposed model with the max-INAR(1) model with geometric innovations (Scotto et al. [8]), the min-INAR(1) model (Aleksić and Ristić [9]), the Poisson INAR(2) (P-INAR) model (Du and Li [4]), and the SETINAR(2,1) model (Monteiro et al. [16]) with Z t P o i s ( λ j ) to fit the data set by the CLS method and compare their mean squared error (MSE) and mean absolute deviation error (MADE), where
M S E = 1 T 3 t = 3 T ( x t x ^ t ) 2 , M A D E = 1 T 3 t = 3 T x t x ^ t .
For each model, we obtain the values of the CLS estimates of the parameters, include the SD of the estimates, and include the in-sample and out-of-sample MSE and MADE values.
For the in-sample values, all of the observations are used to estimate the parameters, while for the out-of-sample values, the first 3618 observations are used to estimate parameters; we predict that the last m = 15 observations, the r for SETINAR(2,1) model, and ( r , s ) for the 2-TINAR(2) model are obtained by the method described in Section 3.2.
The results of the CLS estimates, MSE and MADE, are summarized in Table 3, from which we can see that the max-INAR(1) model and the min-INAR(1) model are not well fitted, so these two models are not suitable for the kind of datum applied in this paper. The 2-TINAR(2) model takes the smallest MSE and MADE values; hence, 2-TINAR(2) is more appropriate for this data set.

5.2. Westar Energy Stock

In this subsection, we consider the number of trades in 5-min intervals between 9:45 a.m. and 4:00 p.m. of a stock listed in the NYSE (Westar Energy, Inc. (WR)), which belongs to the industry subsector conventional electricity. The time period covered is the first quarter of 2005 (3 January 2005–31 March 2005) with 61 trading days; the sample size is T = 4575 . Data are taken from the Trades and Quotes (TAQ) dataset. By computation, the mean is 9.6070 and the variance is 34.8908, which shows it is overdispersed and implies that the piecewise structure may be more suitable for this set of data. Figure 7 shows the path of the data, whose autocorrelation (ACF) and partial autocorrelation functions (PACF) are presented in Figure 8.
Like in Section 5.1, we compare the proposed model with the max-INAR(1) model with geometric innovations (Scotto et al. [8]), the min-INAR(1) model (Aleksić and Ristić [9]), the Poisson INAR(2) (P-INAR) model (Du and Li [4]), and the SETINAR(2,1) model (Monteiro et al. [16]) with Z t P o i s ( λ j ) to fit the data set by the CLS method and compare their MSE and MADE. For each model, we obtain the values of the CLS estimates of the parameters, include the SD of the estimates, and include the in-sample and out-of-sample MSE and MADE values.
For the in-sample values, all observations are used to estimate parameters, while for the out-of-sample values, the first 4560 observations are used to estimate the parameters; we predict that the last m = 15 observations, r for the SETINAR(2,1) model and ( r , s ) for the 2-TINAR(2) model, are obtained by the method described in Section 3.2. The results of the CLS estimates, MSE and MADE, are summarized in Table 4, from which we can see that the max-INAR(1) model and the min-INAR(1) model are not well fitted, so these two models are not suitable for this kind of data again. The 2-TINAR(2) model takes the smallest MSE and MADE values; hence, 2-TINAR(2) is more appropriate for this data set.
Obviously, we can see that as one of the main novelties of the proposed model, it means that the regime of r and s is determined by considering lots of past information, which makes it more practical. Compared with other models, the 2-TINAR(2) model distinguishes innovations in different regions, which makes our model more flexible, but at the same time it also increases the parameters of the model and makes the model more complex. Therefore, we can find that the 2-TINAR(2) model is more suitable for the analysis of the scattered dataset, such as two real examples mentioned in this paper, but it is not suitable for data with small variances and variations.

6. Conclusions

In this paper, we propose a new two-threshold-variable INAR(2) model, which is a generalization of existing INAR models. We consider the CLS estimate with known ( r , s ) and unknown ( r , s ) , respectively. To verify the asymptotic behaviour of the estimators, we give the results of the simulation in each case. A superior performance of the proposed model in real example is demonstrated.
In model (1), we use X t 1 and X t 2 as threshold variables, while other variables can also be used as threshold variables, i.e., the method considered here can be easily extended to other INAR models, such as INAR models with explanatory variable or covariate defined in Enciso-Mora et al. [20],
X t = α X t 1 + Z t , Z t P o i s ( exp ( w t γ ) ) ,
or
X t = α t X t 1 + Z t , α t = [ 1 + exp ( w t δ ) ] 1 , Z t P o i s ( exp ( w t γ ) ) ,
where w t is explanatory variable or covariate; then, we can use X t 1 and w t as threshold variables.
Furthermore, there should be more efficient methods of determining the search range for the threshold estimates, and the possibility of extending this model to the high-dimensional situation is worthy of attention. Moreover, we can extend the results to the three-threshold-variable case. These remains topics of future study. Extensions to the models in Chen et al. [21], Qian and Zhu [22], Su and Zhu [23] and Zhang et al. [7] are similar. Details will be discussed in a future project.

Author Contributions

Conceptualization, F.Z.; methodology, F.Z. and J.Z.; software, J.Z. and H.C.; validation, J.Z. and H.C.; formal analysis, J.Z., F.Z. and H.C.; investigation, J.Z. and H.C.; resources, F.Z.; data curation, J.Z. and H.C.; writing—original draft preparation, J.Z.; writing—review and editing, F.Z. and H.C.; visualization, J.Z. and H.C.; supervision, F.Z. All authors have read and agreed to the published version of the manuscript.

Funding

Zhu’s work is supported by National Natural Science Foundation of China (No. 12271206), the Natural Science Foundation of Jilin Province (No. 20210101143JC), and the Science and Technology Research Planning Project of Jilin Provincial Department of Education (No. JJKH20231122KJ). Chen’s work is supported by Natural Science Foundation of Henan Province (No. 222300420127) and Postdoctoral Research in Henan Province (No. 202103051).

Data Availability Statement

No new data were created in this review.

Acknowledgments

We thank three reviewers for their insightful and constructive comments, which greatly improved the overall presentation.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Proof of Proposition 1.
(1) It is easy to see that { Y t } is a Markov chain with state space N 0 , and the transition probability of the 2-TINAR(2) process is
P ( Y t = ( j , l ) | Y t 1 = ( l , m ) ) = k = 0 min ( m , j ) m k α j 2 k ( 1 α j 2 ) m k P ( R = j k ) ,
note that P ( R = j k ) = b = 0 min ( l , j k ) j k b α j 1 b ( 1 α j 1 ) l b P ( ϵ j t = j k b ) ,
P ( Y t = ( j , l ) | Y t 1 = ( l , m ) ) > 0 ; thus, we can see that { Y t } is an irreducible, aperiodic Markov chain. To prove that Y t is positive recurrent, let Y t = A Y t 1 + ( ϵ j t , 0 ) , ρ ( A j ) < 1 , where A j = α j 1 α j 2 1 0 . The definition of * can be expressed as α 1 α 2 α 3 α 4 X 1 X 2 : = α 1 X 1 + α 2 X 2 α 3 X 3 + α 4 X 4 ; then, based on Proposition 2.1 in Yang et al. [18], we can know that { Y t } is a positive recurrent Markov chain.
(2) It is the direct conclusion of (1); thus, we can see the existence of a strictly stationary distribution of (1). □
Proof of Proposition 2.
Note that three of the four values I 1 t ( r , s ) , I 2 t ( r , s ) , I 3 t ( r , s ) , I 4 t ( r , s ) must be equal to 0; thus, from Silva and Oliveira [5], let max ρ ( A j ) < 1 , max ρ ( A j A l ) < 1 , and max ρ ( A j A l A k ) < 1 . We can know that { X t } is a 3-order stationarity process. The 3-order joint moment of X t , X t + s 1 , X t + s 2 , for s 1 , s 2 R is a function of 2 variables defined by μ X ( s 1 , s 2 ) = E ( X t , X t + s 1 , X t + s 2 ) with μ X = E ( X t ) . Similar to Silva and Silva [6], we let μ j i = E ( B i ) = α j i , σ j i 2 = Var ( B i ) = α j i ( 1 α j i ) , γ j i = E ( B i 3 ) = α j i , γ ϵ j t = E ( ϵ j t 3 ) , C j i = i = 1 2 ( γ j i 3 α j i σ j i 2 α j i 3 ) . Then, for k > 0 ,
μ X ( 0 , 0 ) i = 1 2 l = 1 2 k = 1 2 max ( α j i ) max ( α j l ) max ( α j k ) μ X ( j l , j k ) + 3 i = 1 2 l = 1 2 max ( α j l ) max ( σ j i 2 ) μ X ( i l ) + 3 λ j i = 1 2 l = 1 2 max ( α j i ) max ( α j l ) μ X ( i l ) + 3 λ j i = 1 2 max ( σ j i 2 ) μ X + i = 1 2 max ( C j i ) μ X + γ ϵ j t , μ X ( 0 , k ) i = 1 2 max ( α j i ) μ X ( 0 , k i ) + λ j μ X ( 0 ) , μ X ( k , k ) i = 1 2 l = 1 2 max ( α j i ) max ( α j l ) μ X ( k i , k l ) + i = 1 2 max ( σ j i 2 ) μ X ( k i ) + 2 λ j μ X ( k ) , μ X ( k , m ) i = 1 2 max ( α j i ) μ X ( k , m i ) + λ j μ X ( k ) , m > k ,
the second-order moment of { X t } is μ X ( 0 ) i = 1 2 max ( α j i ) μ X ( i ) + λ j μ X + max ( V j p ) , where V j p = λ j + μ X i = 1 2 σ j i 2 . □
Proof of Proposition 3.
(1) and (3) are obvious, so we just present the proofs of (2) and (4), which are obtained by similar arguments after some tedious calculations.
(2). Let
p 1 = P ( X t 1 > r , X t 2 > s ) , p 2 = P ( X t 1 r , X t 2 > s ) , p 3 = P ( X t 1 r , X t 2 s ) , p 4 = P ( X t 1 > r , X t 2 s ) , u 1 = E ( X t 1 | X t 1 > r , X t 2 > s ) , u 2 = E ( X t 1 | X t 1 r , X t 2 > s ) , u 3 = E ( X t 1 | X t 1 r , X t 2 s ) , u 4 = E ( X t 1 | X t 1 > r , X t 2 s ) , u 1 * = E ( X t 2 | X t 1 > r , X t 2 > s ) , u 2 * = E ( X t 2 | X t 1 r , X t 2 > s ) , u 3 * = E ( X t 2 | X t 1 r , X t 2 s ) , u 4 * = E ( X t 2 | X t 1 > r , X t 2 s ) , v 1 = V a r ( X t 1 | X t 1 > r , X t 2 > s ) , v 2 = V a r ( X t 1 | X t 1 r , X t 2 > s ) , v 3 = V a r ( X t 1 | X t 1 r , X t 2 s ) , v 4 = V a r ( X t 1 | X t 1 > r , X t 2 s ) , v 1 * = V a r ( X t 2 | X t 1 > r , X t 2 > s ) , v 2 * = V a r ( X t 2 | X t 1 r , X t 2 > s ) , v 3 * = V a r ( X t 2 | X t 1 r , X t 2 s ) , v 4 * = V a r ( X t 2 | X t 1 > r , X t 2 s ) , w 1 = E ( X t 1 X t 2 | X t 1 > r , X t 2 > s ) , w 2 = E ( X t 1 X t 2 | X t 1 r , X t 2 > s ) , w 3 = E ( X t 1 X t 2 | X t 1 r , X t 2 s ) , w 4 = E ( X t 1 X t 2 | X t 1 > r , X t 2 s ) .
Hence,
E ( X t ) = E [ E ( X t | F t 1 ) ] = E j = 1 4 ( α j 1 X t 1 + α j 2 X t 2 + λ j ) I j t ( r , s ) = j = 1 4 I j t ( r , s ) ( α j 1 E X t 1 + α j 2 E X t 2 + λ j ) = j = 1 4 p j ( α j 1 u j + α j 2 u j * + λ j ) .
(4). According to the variance formula, the variance of X t is
V a r ( X t ) = V a r ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) + ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s )   + ( α 31 X t 1 + α 32 X t 2 + λ 3 ) I 3 t ( r , s ) + ( α 41 X t 1 + α 42 X t 2 + λ 4 ) I 4 t ( r , s ) )   = V a r ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) ) + V a r ( ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s ) )   + V a r ( ( α 31 X t 1 + α 32 X t 2 + λ 3 ) I 3 t ( r , s ) ) + V a r ( ( α 41 X t 1 + α 42 X t 2 + λ 4 ) I 4 t ( r , s ) )   + 2 C o v ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) , ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s ) )   + 2 C o v ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) , ( α 31 X t 1 + α 32 X t 2 + λ 3 ) I 3 t ( r , s ) )   + 2 C o v ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) , ( α 41 X t 1 + α 42 X t 2 + λ 4 ) I 4 t ( r , s ) )   + 2 C o v ( ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s ) , ( α 31 X t 1 + α 32 X t 2 + λ 3 ) I 3 t ( r , s ) )   + 2 C o v ( ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s ) , ( α 41 X t 1 + α 42 X t 2 + λ 4 ) I 4 t ( r , s ) )   + 2 C o v ( ( α 31 X t 1 + α 32 X t 2 + λ 3 ) I 3 t ( r , s ) , ( α 41 X t 1 + α 42 X t 2 + λ 4 ) I 4 t ( r , s ) )   : = V 1 + V 2 + V 3 + V 4 + C 1 + C 2 + C 3 + C 4 + C 5 + C 6 .
In the following, we compute these quantities in (A1). First, we have
V 1 = V a r ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) ) = V a r ( E ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) | F ) ) + E ( V a r ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) | F ) ) = V a r ( α 11 X t 1 I 1 t + α 12 X t 2 I 1 t + λ 1 I 1 t ) + E ( α 11 ( 1 α 11 ) X t 1 I 1 t + α 12 ( 1 α 12 ) X t 2 I 1 t + λ 1 I 1 t ) = α 11 2 V a r ( I 1 t X t 1 ) + α 12 2 V a r ( I 2 t X t 2 ) + 2 C o v ( α 11 X t 1 I 1 t , α 12 X t 2 I 1 t ) + 2 C o v ( α 11 X t 1 I 1 t , λ 1 I 1 t ) + 2 C o v ( α 12 X t 2 I 1 t , λ 1 I 1 t ) + α 11 ( 1 α 11 ) p 1 u 1 + α 12 ( 1 α 12 ) p 1 u 1 * + λ 1 p 1 = α 11 2 ( p 1 ( v 1 + u 1 2 ) p 1 2 u 1 2 ) + α 12 2 ( p 1 ( v 1 * + ( u 1 * ) 2 ) p 1 2 ( u 2 * ) 2 ) + 2 ( α 11 α 12 w 1 p 1 α 11 α 12 p 1 2 u 1 u 1 * ) + 2 ( α 11 λ 1 p 1 u 1 α 11 λ 1 p 1 2 u 1 ) + 2 ( α 12 λ 1 p 1 u 1 * α 12 λ 1 p 1 2 u 1 * ) + α 11 ( 1 α 11 ) p 1 u 1 + α 12 ( 1 α 12 ) p 1 u 1 * + λ 1 p 1 .
By the same arguments as above, it follows that
V j = V a r ( ( α j 1 X t 1 + α j 2 X t 2 + λ j ) I j t ( r , s ) ) = α j 1 2 ( p j ( v j + u j 2 ) p j 2 u j 2 ) + α j 2 2 ( p j ( v j * + ( u j * ) 2 ) p j 2 ( u j * ) 2 ) + 2 ( α j 1 α j 2 w j p j α j 1 α j 2 p j 2 u j u j * ) + 2 ( α j 1 λ j p j u j α j 1 λ j p j 2 u j ) + 2 ( α j 2 λ j p j u j * α j 2 λ j p j 2 u j * ) + α j 1 ( 1 α j 1 ) p j u j + α j 2 ( 1 α j 2 ) p j u j * + λ j p j , j = 2 , 3 , 4 .
We can see that C j takes the form
C 1 = 2 C o v ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) , ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s ) ) = 2 E ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) ) E ( ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s ) ) = 2 ( α 11 u 1 p 1 + α 12 u 1 * p 1 + λ 1 p 1 ) ( α 21 u 2 p 2 + α 22 u 2 * p 2 + λ 2 p 2 ) ,
C 2 = 2 C o v ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) , ( α 31 X t 1 + α 32 X t 2 + λ 3 ) I 3 t ( r , s ) )   = 2 ( α 11 u 1 p 1 + α 12 u 1 * p 1 + λ 1 p 1 ) ( α 31 u 3 p 3 + α 32 u 3 * p 3 + λ 3 p 3 ) ,
C 3 = 2 C o v ( ( α 11 X t 1 + α 12 X t 2 + λ 1 ) I 1 t ( r , s ) , ( α 41 X t 1 + α 42 X t 2 + λ 4 ) I 4 t ( r , s ) )   = 2 ( α 11 u 1 p 1 + α 12 u 1 * p 1 + λ 1 p 1 ) ( α 41 u 4 p 4 + α 42 u 4 * p 4 + λ 4 p 4 ) ,
C 4 = 2 C o v ( ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s ) , ( α 31 X t 1 + α 32 X t 2 + λ 3 ) I 3 t ( r , s ) )   = 2 ( α 21 u 2 p 2 + α 22 u 2 * p 2 + λ 2 p 2 ) ( α 31 u 3 p 3 + α 32 u 3 * p 3 + λ 3 p 3 ) ,
C 5 = 2 C o v ( ( α 21 X t 1 + α 22 X t 2 + λ 2 ) I 2 t ( r , s ) , ( α 41 X t 1 + α 42 X t 2 + λ 4 ) I 4 t ( r , s ) )   = 2 ( α 21 u 2 p 2 + α 22 u 2 * p 2 + λ 2 p 2 ) ( α 41 u 4 p 4 + α 42 u 4 * p 4 + λ 4 p 4 ) ,
C 6 = 2 C o v ( ( α 31 X t 1 + α 32 X t 2 + λ 3 ) I 3 t ( r , s ) , ( α 41 X t 1 + α 42 X t 2 + λ 4 ) I 4 t ( r , s ) )   = 2 ( α 31 u 3 p 3 + α 32 u 3 * p 3 + λ 3 p 3 ) ( α 41 u 4 p 4 + α 42 u 4 * p 4 + λ 4 p 4 ) ,
then, (4) follows by replacing (A2)–(A9) in (A1). □
Proof of Theorem 2
It can easily be seen that the conditions in Klimko and Nelson [24] are verified; we can see that
g ( ϕ , X t 1 , X t 2 ) , g ( ϕ , X t 1 , X t 2 ) ϕ i , g 2 ( ϕ , X t 1 , X t 2 ) ϕ i ϕ j , g 3 ( ϕ , X t 1 , X t 2 ) ϕ i ϕ j ϕ k
satisfy all the regularity conditions for i , j , k = 1 , , 12 . Thus, the CLS estimator is strongly consistent. Moreover, when proving asymptotic normality we first have to check the following conditions:
(1)
E ( X t | X t 1 , , X 0 ) = E ( X t | X t 1 , X t 2 ) , t 3 , a . s . ;
(2)
E ( q t 2 ( ϕ ) | g ( ϕ , X t 1 , X t 2 ) ϕ i g ( ϕ , X t 1 , X t 2 ) ϕ j | ) < , i , j = 1 , , 12 ;
(3)
V is non singular,
then, we know from Klimko and Nelson [24] that the CLS estimation is asymptotically normal with the asymptotic variance V 1 W V 1 . □

References

  1. Steutel, F.W.; van Harn, K. Discrete analogues of self-decomposability and stability. Ann. Probab. 1979, 7, 893–899. [Google Scholar] [CrossRef]
  2. Al-Osh, M.A.; Alzaid, A.A. First-order integer-valued autoregressive (INAR(1)) process. J. Time Ser. Anal. 1987, 8, 261–275. [Google Scholar] [CrossRef]
  3. McKenzie, E. Some simple models for discrete variate time series. Water Resour. Bull. 1985, 21, 645–650. [Google Scholar] [CrossRef]
  4. Du, J.; Li, Y. The integer-valued autoregressive (INAR(p)) model. J. Time Ser. Anal. 1991, 12, 129–142. [Google Scholar]
  5. Silva, I.; Oliveira, V.L. Difference equations for the higher-order moments and cumulants of the INAR(p) model. J. Time Ser. Anal. 2005, 26, 17–36. [Google Scholar] [CrossRef]
  6. Silva, I.; Silva, M.E. Parameter estimation for INAR processes based on high-order statistics. REVSTAT Stat. J. 2009, 7, 105–117. [Google Scholar]
  7. Zhang, J.; Zhu, F.; Mamode Khan, N. A new INAR model based on Poisson-BE2 innovations. Commun. Stat.-Theory Methods 2023, 52, 6063–6067. [Google Scholar] [CrossRef]
  8. Scotto, M.G.; Weiß, C.H.; Möller, T.A.; Gouveia, S. The max-INAR(1) model for count processes. TEST 2018, 27, 850–870. [Google Scholar] [CrossRef]
  9. Aleksić, M.S.; Ristić, R.M. A geometric minification integer-valued autoregressive model. Appl. Math. Model. 2021, 90, 265–280. [Google Scholar] [CrossRef]
  10. Tong, H. On a threshold model. In Pattern Recognition and Signal Processing; Sijthoff and Noordhoff: Amsterdam, The Netherlands, 1978. [Google Scholar]
  11. Boero, G.; Marrocu, E. The performance of SETAR models: A regime conditional evaluation of point, interval and density forecasts. Int. J. Forecast. 2004, 20, 305–320. [Google Scholar] [CrossRef]
  12. Potter, S.M. A nonlinear approach to U.S. GNP. J. Appl. Econom. 1995, 10, 109–125. [Google Scholar] [CrossRef]
  13. Dueker, M.; Martin, S.; Spagnolo, F. Contemporaneous threshold autoregressive models: Estimation, testing and forecasting. J. Econom. 2007, 141, 517–547. [Google Scholar] [CrossRef]
  14. Tong, H. Threshold models in time series analysis 30 years on. Stat. Its Interface 2011, 4, 107–118. [Google Scholar] [CrossRef]
  15. Li, D.; Tong, H. Nested sub-sample search algorithm for estimation of threshold models. Stat. Sin. 2016, 26, 1543–1554. [Google Scholar] [CrossRef]
  16. Monteiro, M.; Scotto, M.G.; Pereira, I. Integer-valued self-exciting threshold autoregressive processes. Commun. Stat.-Theory Methods 2012, 41, 2717–2737. [Google Scholar] [CrossRef]
  17. Wang, C.; Liu, H.; Yao, J.; Davis, R.A.; Li, W.K. Self-excited threshold Poisson autoregression. J. Am. Stat. Assoc. 2014, 109, 776–787. [Google Scholar] [CrossRef]
  18. Yang, K.; Wang, D.; Jia, B.; Li, H. An integer-valued threshold autoregressive process based on negative binomial thinning. Stat. Pap. 2018, 59, 1131–1160. [Google Scholar] [CrossRef]
  19. Zhang, X.; Li, D.; Tong, H. On the least squares estimation of multiple-threshold-variable autoregressive models. J. Bus. Econ. Stat. 2023, 1–14. [Google Scholar] [CrossRef]
  20. Enciso-Mora, V.; Neal, P.; Subba Rao, T. Integer valued AR processes with explanatory variables. Sankhyā 2009, 71, 248–263. [Google Scholar]
  21. Chen, H.; Li, Q.; Zhu, F. Two classes of dynamic binomial integer-valued ARCH models. Braz. J. Probab. Stat. 2020, 34, 685–711. [Google Scholar] [CrossRef]
  22. Qian, L.; Zhu, F. A new minification integer-valued autoregressive process driven by explanatory variables. Aust. N. Z. J. Stat. 2022, 64, 478–494. [Google Scholar] [CrossRef]
  23. Su, B.; Zhu, F. Comparison of BINAR(1) models with bivariate negative binomial innovations and explanatory variables. J. Stat. Comput. Simul. 2021, 91, 1616–1634. [Google Scholar] [CrossRef]
  24. Klimko, L.A.; Nelson, P.I. On conditional least squares estimation for stochastic processes. Ann. Stat. 1978, 6, 629–642. [Google Scholar] [CrossRef]
Figure 1. Sample paths for combinations (C1) and (C2).
Figure 1. Sample paths for combinations (C1) and (C2).
Mathematics 11 03586 g001
Figure 2. Boxplots of (C1).
Figure 2. Boxplots of (C1).
Mathematics 11 03586 g002
Figure 3. Sample paths of (C3) and (C4).
Figure 3. Sample paths of (C3) and (C4).
Mathematics 11 03586 g003
Figure 4. Boxplots of (C3).
Figure 4. Boxplots of (C3).
Mathematics 11 03586 g004
Figure 5. Path of the stock data.
Figure 5. Path of the stock data.
Mathematics 11 03586 g005
Figure 6. Daily number of trades of a stock data: (a). ACF (b). PACF.
Figure 6. Daily number of trades of a stock data: (a). ACF (b). PACF.
Mathematics 11 03586 g006
Figure 7. Path of the stock data.
Figure 7. Path of the stock data.
Mathematics 11 03586 g007
Figure 8. The number of trades of a stock data: (a). ACF (b). PACF.
Figure 8. The number of trades of a stock data: (a). ACF (b). PACF.
Mathematics 11 03586 g008
Table 1. The mean and standard deviation (in brackets) of the estimates of 2-TINAR model with known ( r , s ) .
Table 1. The mean and standard deviation (in brackets) of the estimates of 2-TINAR model with known ( r , s ) .
T α 11 α 12 λ 1 α 21 α 22 λ 2 α 31 α 32 λ 3 α 41 α 42 λ 4
(C1) = (0.3, 0.2, 7, 0.2, 0.25, 6, 0.2, 0.3, 8, 0.3, 0.2, 6)
5000.29130.18847.34240.20870.24956.00190.20900.30348.13200.33010.29216.5232
(0.1106)(0.0908)(2.1188)(0.1225)(0.1116)(2.0209)(0.1285)(0.1692)(2.1968)(0.2155)(0.2147)(4.1374)
10000.29510.19367.17250.20220.24935.99520.20000.29748.07020.30600.23736.1198
(0.0781)(0.0658)(1.4782)(0.0918)(0.0802)(1.4289)(0.0990)(0.1280)(1.5499)(0.1634)(0.1635)(3.1374)
20000.29780.19657.08620.20040.24975.99800.19910.29818.03620.30140.21106.0076
(0.0548)(0.0467)(1.0395)(0.0657)(0.0565)(0.9988)(0.0728)(0.0921)(1.0918)(0.1208)(0.1281)(2.3074)
10,0000.29950.19947.01700.20010.24995.99980.19990.29978.00480.30010.19986.0013
(0.0244)(0.0209)(0.4650)(0.0294)(0.0252)(0.4464)(0.0328)(0.0413)(0.4897)(0.0537)(0.0645)(1.0270)
(C2) = (0.3, 0.35, 15, 0.3, 0.35, 20, 0.3, 0.4, 25, 0.35, 0.25, 15)
5000.29550.339517.39320.30140.346020.21520.29780.398525.31530.35060.249315.7572
(0.1487)(0.1278)(11.5915)(0.1017)(0.1130)(8.1557)(0.1371)(0.1181)(6.2198)(0.1245)(0.1334)(8.8644)
10000.29480.344815.87560.30070.348220.07210.29740.399625.17090.35030.247915.1737
(0.1088)(0.0898)(8.7210)(0.0713)(0.0793)(5.7577)(0.1053)(0.0891)(5.2321)(0.0870)(0.0994)(6.5817)
20000.29690.347515.34610.30010.348920.06420.29860.400225.06910.35020.248315.0749
(0.0775)(0.0629)(6.4147)(0.0502)(0.0558)(4.0647)(0.0774)(0.0661)(4.1621)(0.0611)(0.0708)(4.7424)
10,0000.29950.349515.05940.30010.349920.00170.29990.400125.00100.35000.249715.0124
(0.0343)(0.0279)(2.8665)(0.0223)(0.0250)(1.8115)(0.0363)(0.0311)(2.1124)(0.0272)(0.0317)(2.1158)
Table 2. The mean and standard deviation (in brackets) of the estimates of 2-TINAR model with unknown ( r , s ) .
Table 2. The mean and standard deviation (in brackets) of the estimates of 2-TINAR model with unknown ( r , s ) .
T α 11 α 12 λ 1 α 21 α 22 λ 2 α 31 α 32 λ 3 α 41 α 42 λ 4 rs
(C3) = (0.2, 0.25, 3, 0.25, 0.35, 5, 0.3, 0.3, 4, 0.3, 0.25, 6, 13, 14)
5000.31650.38829.63830.28970.43086.75830.30920.30003.95580.39080.25476.68199.985110.5800
(0.2458)(0.3007)(7.8787)(0.2093)(0.2878)(5.0400)(0.0655)(0.0590)(0.6470)(0.2612)(0.1715)(4.5692)(2.0175)(2.1507)
10000.24820.29666.60490.25440.38605.61380.30410.30123.97270.33700.24046.222311.310312.0423
(0.1764)(0.2082)(5.6639)(0.1584)(0.2204)(3.8261)(0.0471)(0.0419)(0.4652)(0.1945)(0.1337)(3.4623)(1.9055)(2.0830)
20000.21560.26344.57000.24860.35845.14600.30140.30073.98910.31270.24276.017612.340013.2323
(0.1327)(0.1540)(3.9383)(0.1235)(0.1626)(2.8763)(0.0325)(0.0291)(0.3225)(0.1352)(0.0971)(2.3564)(1.3604)(1.5242)
10,0000.20030.24863.04700.25030.34915.00880.29960.29984.00570.30040.25015.992312.996213.9960
(0.0627)(0.0725)(1.4592)(0.0569)(0.0717)(1.3053)(0.0133)(0.0127)(0.1368)(0.0537)(0.0411)(0.8949)(0.1104)(0.1166)
(C4) = (0.30, 0.25, 6, 0.25, 0.35, 6, 0.4, 0.3, 9, 0.3, 0.35, 8, 30, 31)
5000.51520.489322.86060.26210.391912.31900.39230.28499.51500.31380.333711.072426.593427.4885
(0.4299)(0.4219)(19.7688)(0.1613)(0.2612)(10.0337)(0.0666)(0.0647)(1.8098)(0.2115)(0.1714)(7.6731)(3.9233)(4.0751)
10000.37940.343714.34850.25020.35648.88320.39660.29339.23360.30060.33959.109228.573529.5242
(0.2716)(0.2547)(11.4391)(0.1216)(0.1978)(7.1729)(0.0465)(0.0458)(1.3281)(0.1619)(0.1253)(5.7439)(3.0484)(3.1956)
20000.32160.281110.07190.24910.35146.82340.39830.29839.08340.30090.34878.145429.678830.6630
(0.2052)(0.1911)(7.6392)(0.0876)(0.1446)(4.7774)(0.0321)(0.0307)(0.9129)(0.1163)(0.0873)(4.1778)(1.5502)(1.6476)
10,0000.30070.25026.44170.24960.34986.02270.39960.29999.01310.30040.34928.006530.000031.0000
(0.1036)(0.0985)(4.0326)(0.0390)(0.0638)(2.2480)(0.0143)(0.0135)(0.4048)(0.0513)(0.0382)(1.8861)(0.0000)(0.0000)
Table 3. Fitting results of the Siparex Croissance data.
Table 3. Fitting results of the Siparex Croissance data.
Model in SampleEstimate MSEMADE
max-INAR(1) α ^ 1 q ^
0.49670.1208
(1.8621)(0.0627) 21.15913.9308
min-INAR(1) α ^ 1 μ ^
1.50748.7789
(50.3267)(414.5962) 37.87805.2950
P-INAR(2) α ^ 1 α ^ 2 λ ^
0.37140.22284.0663
(0.0316)(0.0279)(0.2790) 21.13863.8917
SETINAR(2)
r = 9
α ^ 11 α ^ 12 λ 1 α ^ 21 α ^ 22 λ 2
0.40150.22953.76340.34970.21555.0422
(0.0373)0.0312)(0.2895)(0.0485)(0.0392)(1.1459) 21.35043.8975
2-TINAR(2)
r = 14 , s = 17
α ^ 11 α ^ 12 λ 1 α ^ 21 α ^ 22 λ 2 α ^ 31 α ^ 32 λ 3 α ^ 41 α ^ 42 λ 4
0.35440.14139.09320.73530.14642.89760.36660.26993.72580.34900.20224.2198
(0.0739)(0.0583)(2.4291)(0.1802)(0.0796)(2.5865)(0.0379)( 0.0314)(0.2639)(0.0630)(0.1057)(1.5101)20.55773.7951
out of sample
max-INAR(1) α ^ 1 q ^
0.49670.1208
(1.8660)(0.0629) 25.33174.0658
min-INAR(1) α ^ 1 μ ^
1.50748.7789
(49.6399)(408.9382) 40.68004.4990
P-INAR(2) α ^ 1 α ^ 2 λ ^
0.37170.22244.0767
(0.0316)(0.0279)(0.2796) 22.39613.1670
SETINAR(2)
r = 9
α ^ 11 α ^ 12 λ 1 α ^ 21 α ^ 22 λ 2
0.40580.22823.75720.34970.21555.0422
(0.0374)(0.0313)(0.2898)(0.0485)(0.0392)(1.1459) 22.91523.2409
2-TINAR(2)
r = 14 , s = 17
α ^ 11 α ^ 12 λ 1 α ^ 21 α ^ 22 λ 2 α ^ 31 α ^ 32 λ 3 α ^ 41 α ^ 42 λ 4
0.35440.14139.09320.73530.14642.89760.37100.26813.72130.34900.20224.2198
(0.0737)(0.0582)(2.4241)(0.1798)(0.0795)(2.5812)(0.0380)(0.0315)(0.2636)(0.0629)(0.1055)(1.5070)17.34522.9123
Table 4. Fitting results of the WR data.
Table 4. Fitting results of the WR data.
Model in SampleEstimate MSEMADE
max-INAR(1) α ^ 1 q ^
0.24610.3982
(1.2089)(2.4770) 150.019611.5280
min-INAR(1) α ^ 1 μ ^
1.76251.6307
(50.1349)(27.6055) 175.161712.6286
P-INAR(2) α ^ 1 α ^ 2 λ ^
0.33620.21874.2795
(0.0192)(0.0172)(0.1804) 25.15774.4766
SETINAR(2)
r = 14
α ^ 11 α ^ 12 λ 1 α ^ 21 α ^ 22 λ 2
0.33250.24244.05950.25860.14996.9125
(0.0227)(0.0186)(0.1909)(0.0634)(0.0396)(1.1693) 23.97464.4016
2-TINAR(2)
r = 11 , s = 12
α ^ 11 α ^ 12 λ 1 α ^ 21 α ^ 22 λ 2 α ^ 31 α ^ 32 λ 3 α ^ 41 α ^ 42 λ 4
0.18550.045810.84600.46190.09855.39520.31160.30093.80490.36600.19524.3789
(0.0848)(0.0919)(2.2813)(0.0962)(0.0585)(1.4106)(0.0233)(0.0225)(0.1971)(0.0884)(0.0942)(1.6380)23.53364.4005
out of sample
max-INAR(1) α ^ 1 q ^
0.24610.3982
(1.2109)(2.4811) 179.540112.6882
min-INAR(1) α ^ 1 μ ^
1.76251.6307
(50.2126)(27.6483) 178.716612.7448
P-INAR(2) α ^ 1 α ^ 2 λ ^
0.33590.21724.2877
(0.0192)(0.0172)(0.1808) 73.55377.7076
SETINAR(2)
r = 14
α ^ 11 α ^ 12 λ 1 α ^ 21 α ^ 22 λ 2
0.33100.24124.07590.26300.14796.8355
(0.0227)(0.0186)(0.1909)(0.0637)(0.0397)(1.1741) 73.84547.7045
2-TINAR(2)
r = 11 , s = 12
α ^ 11 α ^ 12 λ 1 α ^ 21 α ^ 22 λ 2 α ^ 31 α ^ 32 λ 3 α ^ 41 α ^ 42 λ 4
0.18850.045810.76830.45500.09845.42780.31050.30043.81350.37060.18924.3368
(0.0849)(0.0922)(2.3067)(0.0961)(0.0583)(1.4070)(0.0233)(0.0224)(0.1969)(0.0887)(0.0948)(1.6435)71.61807.5889
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, J.; Zhu, F.; Chen, H. Two-Threshold-Variable Integer-Valued Autoregressive Model. Mathematics 2023, 11, 3586. https://doi.org/10.3390/math11163586

AMA Style

Zhang J, Zhu F, Chen H. Two-Threshold-Variable Integer-Valued Autoregressive Model. Mathematics. 2023; 11(16):3586. https://doi.org/10.3390/math11163586

Chicago/Turabian Style

Zhang, Jiayue, Fukang Zhu, and Huaping Chen. 2023. "Two-Threshold-Variable Integer-Valued Autoregressive Model" Mathematics 11, no. 16: 3586. https://doi.org/10.3390/math11163586

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop