Next Article in Journal
Quantitative Study of Non-Linear Convection Diffusion Equations for a Rotating-Disc Electrode
Next Article in Special Issue
Forecast of Winter Precipitation Type Based on Machine Learning Method
Previous Article in Journal
A Comparative Study of Rank Aggregation Methods in Recommendation Systems
Previous Article in Special Issue
Comparing Several Gamma Means: An Improved Log-Likelihood Ratio Test
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Essay

Detecting Structural Change Point in ARMA Models via Neural Network Regression and LSCUSUM Methods

1
School of Mathematics and Statistic, Qinghai Normal University, Xining 810008, China
2
The State Key Laboratory of Tibetan Intelligent Information Processing and Application, Xining 810008, China
3
School of Computer Science and Technology, Qinghai Normal University, Xining 810008, China
4
School of Preparatory Education for Nationalities, Ningxia University, Yinchuan 750002, China
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(1), 133; https://doi.org/10.3390/e25010133
Submission received: 1 December 2022 / Revised: 2 January 2023 / Accepted: 6 January 2023 / Published: 9 January 2023
(This article belongs to the Special Issue Recent Advances in Statistical Theory and Applications)

Abstract

:
This study considers the change point testing problem in autoregressive moving average (ARMA) ( p , q ) models through the location and scale-based cumulative sum (LSCUSUM) method combined with neural network regression (NNR). We estimated the model parameters via the NNR method based on the training sample, where a long AR model was fitted to obtain the residuals. Then, we selected the optimal model orders p and q of the ARMA models using the Akaike information criterion based on a validation set. Finally, we used the forecasting errors obtained from the selected model to construct the LSCUSUM test. Extensive simulations and their application to three real datasets show that the proposed NNR-based LSCUSUM test performs well.

1. Introduction

The change point problem has been extensively studied in time series analysis because time series often suffer from structural changes due to issues such as changes of governmental policy, health care quality, and critical social events. This problem has also been crucial in engineering, medicine, and economics, and numerous related articles have been published. We refer the reader to Csörgö and Horváth [1], Hušková et al. [2], Franke et al. [3], Kang and Lee [4], Lee and Chen [5], and Huh et al. [6] for a general review.
The cumulative sum (CUSUM) test proposed by Page [7] is one of the most popular methods for change point tests due to its convenience of usage. Lee and Na [8] proposed the conventional estimate-based CUSUM test, which generally performs well, but suffers from severe size distortions and produces low powers on some occasions. Na et al. [9] and Lee [10] suggested the residual-based CUSUM test for time series models as a remedy. This test is more stable, but it suffers serious power loss when location parameters change because it can only detect the change in scale parameters. To improve upon this shortcoming, Oh [11,12] proposed a score vector-based CUSUM test and a modified residual-based CUSUM method to enhance the power performance. Lee [13] introduced a location- and scale-based CUSUM (LSCUSUM) test, and simulations confirmed its improved performance in terms of stability and power over any estimate-, residual- and score vector-based CUSUM tests developed thus far.
A key step in constructing the LSCUSUM test is to accurately estimate the residuals, i.e., fitting a correct model is important if we want to test change points in time series. The most popular time series fitting method is the classical ARMA method, which uses maximum likelihood estimation (QMLE) or least squares estimation (LSE) to fit an ARMA model. Classical linear ARMA models produce an accurate prediction when a time series truly follows them, but the prediction is not correct if the time series has obvious nonlinear characteristics. Because many time series carry nonlinear characteristics, new analytical tools such as neural network regression (NNR), support vector regression (SVR), and least absolute shrinkage and selection operator (LASSO) have been applied to fit them, and have been demonstrated to outperform classical time series models, especially when a time series has certain nonlinear and nonstationary characteristics. For surveys, we refer the reader to Hwarng and Ang [14], Hossain et al. [15], Zafar et al. [16], and Giacomazzo [17]. Harchaoui and Lévy-Leduc [18] considered the change point estimation problem as a variable selection problem and used the LASSO method to estimate multiple change points in one-dimensional piecewise constant signals. Chan et al. [19] applied a group LASSO method to estimate multiple change points in a time series. Jin et al. [20] proposed a fast algorithm to estimate multiple change points in nonstationary time series models. Qian and Su [21] studied a multiple change points estimation problem for panel data models via an adaptive group fused LASSO. For more studies about change point estimation via a penalized model selection method, please refer to Lee et al. [22], Jin et al. [23], and Ciuperca and Maciak [24], among many others.
Lee et al. [25,26] proposed an SVR with a hybridization of the CUSUM method to detect change points in ARMA and generalized autoregressive conditional heteroscedastic (GARCH) models. As is well known, the NNR can approximate the nonlinearity of a time series without knowing the underlying dynamic structure by minimizing the empirical risk. Motivated by this, we adopted NNR to obtain ARMA residuals, and then constructed the LSCUSUM statistic to test change points. Our simulations show that the proposed method outperforms methods based on SVR, the adaptive least absolute shrinkage and selection operator (ALASSO), and classical ARMA.
The rest of this paper is organized as follows. Section 2 presents the principle of the LSCUSUM test for ARMA models. Section 3 introduces the NNR method in a general framework. Section 4 proposes a forecasting method based on the NNR fitting ARMA model, herein the NNR-ARMA model, and then explains how to apply the NNR-ARMA model to the construction of the LSCUSUM test. Section 5 performs Monte Carlo simulations to evaluate the validity of the proposed method under various linear and nonlinear ARMA models. Section 6 applies the method to three real datasets for illustration. Finally, Section 7 provides concluding remarks.

2. LSCUSUM Test for ARMA Models

The model discussed in this paper is
y t = f ( Θ ; y t 1 , , y t p , ε t 1 , , ε t q ) + ε t ,
where f is an unknown function to be estimated, Θ = ( ϕ 1 , , ϕ p , θ 1 , , θ q , σ 2 ) is an unknown parameter vector, p and q are real numbers, and ε i are IID random variables with mean zero, covariance σ 2 > 0 , and E ε t 4 < . For given observations y 1 , , y n , we are interested in the null and alternative hypotheses
H 0 : Θ t = Θ ,   t = 1 , , n   ,
H 1 : Θ t = Θ , t = 1 , , k Θ + Δ , t = k + 1 , , n ,
where Θ is the unknown parameter vector, the nonzero constant vector Δ represents the parameter skip, and k = [ n τ ] is the location of the change point, where τ 0 , 1 and [ ] denotes the integral function. Model (1) includes AR, ARMA, threshold ARMA (TARMA) and time-varying AR (TVAR) models as special cases.
To test the above null and alternative hypotheses, we applied the following two LSCUSUM tests proposed by Lee [13]:
T ^ n L S = max 1 k n 1 n γ ^ 1 , n 2 t = 1 k ( y t ε ^ t ) ε ^ t k n t = 1 n ( y t ε ^ t ) ε ^ t 2 + 1 n γ ^ 2 , n 2 t = 1 k ε ^ t 2 k n t = 1 n ε ^ t 2 2 ,
T ^ n max = max 1 k n max 1 n γ ^ 1 , n t = 1 k ( y t ε ^ t ) ε ^ t k n t = 1 n ( y t ε ^ t ) ε ^ t , 1 n γ ^ 2 , n t = 1 k ε ^ t 2 k n t = 1 n ε ^ t 2 ,
where ε ^ t = y t f ( Θ ^ ; y t 1 , , y t p , ε t 1 , , ε t q ) denotes the residuals with Θ ^ as the estimator of Θ , and
γ ^ 1 , n 2 = 1 n t = 1 n ( y t ε ^ t ) 2 ε ^ t 2 1 n t = 1 n ( y t ε ^ t ) ε ^ t 2 ,
γ ^ 2 , n 2 = 1 n t = 1 n ε ^ t 4 1 n t = 1 n ε ^ t 2 2 .
We rejected H 0 if T ^ n L S > 2.4503 or T ^ n max > 1.4596 at the nominal 0.05 level. The critical values of these two tests were obtained though Monte Carlo simulations using two-dimensional standard Brownian motion.
Next, we introduce the NNR method in a general framework, propose a fitting method based on NNR to the ARMA model, and explain how to apply the NNR-ARMA method when constructing the LSCUSUM test.

3. Neural Network Regression

A neural network (NN) is a simplified model obtained by abstracting and modeling the human brain, which is a parallel connection of a set of nodes called neurons. It can effectively solve complex regression and classification problems with a large number of correlated variables and make accurate predictions for time series.
In this study, we define an NN with an input layer, a hidden layer, and an output layer. We consider a feedforward net with n input nodes, one layer of H hidden nodes, and one output node. As shown in Figure 1, n nodes representing independent variables on the left form the input layer, H nodes in the middle form the hidden layer, and the rightmost node belongs to the output layer representing dependent variables. These nodes are connected according to the arrows, and the number next to an arrow marks its weight.
Considering an input vector x = ( x 1 , , x n ) n , the output z h ( x ; θ ) of the h - th hidden node is
z h ( x ; θ ) = ψ i = 1 n ω i h x i + ω 0 h .
The output of the net is
F ( x ; θ ) = ψ h = 1 H υ h z h ( x ; θ ) + υ 0 = ψ h = 1 H υ h ψ i = 1 n ω i h x i + ω 0 h + υ 0 ,
where ω i h ( i = 0 , , n ; h = 1 , , H ) is the weight connecting the i - th input layer node and h - th hidden layer node; υ h ( h = 0 , , H ) is the weight of the h - th hidden layer node to the j - th dependent variable; and θ stands for all parameters υ 0 , , υ H and ω i h ( i = 0 , , n ; h = 1 , , H ) of the network. We also write υ = ( υ 0 , , υ H ) T and ω = ( ω i h , i = 0 , , n ; h = 1 , , H ) . Here, ψ ( ) and ψ ( ) are activation functions of the NN, which are usually defined as S-shaped logistic functions,
ψ ( x ) = 1 1 + exp ( a x ) ,
and
lim ψ ( x ) = 1 , x 0 , x ,
where a is a parameter that controls the inclination of the function. Different function types can be selected as required, such as the threshold, piecewise, Gaussian, and hyperbolic tangent functions.
The NN connection weights were adjusted through training, as follows:
1.
Assume that the input vector and the associated output vector are x = x 1 , , x m n and y = y 1 , , y m , respectively.
2.
Initialize the weights in the network:
ω ( 0 )   ,   υ ( 0 ) .
3.
Use forward propagation to obtain the predicted value and calculate the loss function,
S ( y , x ; θ ) = 1 2 j = 1 m ( y j F ( x j ; θ ) ) 2 .
4.
Update the connection weights:
ω k = ω k 1 + Δ ω ;
υ k = υ k 1 + Δ υ
Taking individual weights, the k - th iteration weights are
ω i h ( k ) = ω i h ( k 1 ) η S ( y , x ; θ ( k ) ) ω i h , f o r   i = 0 , , n , h = 1 , , H and j = 1 , , m .
Similarly,
υ h ( k ) = υ h ( k 1 ) η S ( y , x ; θ ( k ) ) υ h , f o r   h = 1 , , H and j = 1 , , m .
where η ( 0 < η < 1 ) is the learning rate; k is the number of updates or iterations; and S ( y , x ; θ ( k ) ) ω h j and S ( y , x ; θ ( k ) ) υ h are respective error terms of the hidden unit and output unit, i.e., the gradient values or partial derivatives of the loss function with respect to them. According to the chain rule, they are:
S ( y , x ; θ ( k ) ) ω h j = S ω h j = S F F z z ω h j ,
S ( y , x ; θ ( k ) ) υ h = S υ h = S F F z z υ h
5.
Repeat steps 3 and 4 until the loss function is less than the preset threshold or the number of iterations is exhausted and the output parameter is the best parameter at present.
An NN can have multiple hidden layers, but one is generally enough. The number of nodes in the hidden layer can be large or small. Too many nodes may lead to overfitting, and too few to poor fitting. The number of neurodes H in the hidden layer can be determined as in Looney [27] by rule of thumb as
H = 1.7 log 2 ( n ) + 1 .
Alternatively, one can use the Bayesian information criterion (BIC), as proposed by Swanson [28], to sequentially determine H . In this paper, the optimal number of nodes H is selected according to root mean square error (RMSE) recommendations through cross-validation.

4. Prediction Based on NNR-ARMA Model

Throughout this section, we assume that { y t : t = 1 , , n , n + 1 , , n + l } is generated according to the ARMA model (1). If the training sample is known to follow an NNR-ARMA ( p , q ) model with specific orders p and q , we can use this model. Otherwise, we determine the orders from the training sample as follows:
Step 1 Let y = ( y m , , y n ) , and design the matrix
X = ( X m , X m + 1 , , X n ) T = y m 1 y m p ε m 1 ε m q y m y m p + 1 ε m ε m q + 1 y n 1 y n p ε n 1 ε n q
as the NN input. The design matrix X involves the potential innovation terms ε t ( t = m q , , n 1 ) . Thus, the long A R ( p ) model is first fitted to y 1 , , y n to obtain the residual ε ^ t , and then the residual ε ^ t is used to approximate the unobserved ε t . See Hannan and Rissanen [29] for more details. The design matrix with ε t replaced is X ^ , where m = p + max ( p , q ) + 1 . When fitting the long A R ( p ) model, the order p can be determined using the Akaike information criterion (AIC), with the maximum order set to be 10 log 10 ( n ) .
Step 2 Fix p * > 0   ,   q * > 0 , and the underlying true ARMA orders p p * and q q * , where p * ,   q * are known upper bounds of the true orders. For each p ,   q , we assume that y 1 , , y n approximately satisfy
y t = f ( y t 1 , , y t p , ε ^ t 1 , ε ^ t q ) + ε ^ t ,
and the NNR-ARMA ( p , q ) model is fitted with y = ( y m , , y n ) and X ^ , as in Section 3. Then, after applying the obtained NNR-ARMA ( p , q ) model to y n + 1 ,   , y n + l , calculate the prediction errors ε ˜ t ( t = 1 , , l ) and the corresponding AIC, namely, AIC = 2 ( p + q )   + l ln ( RSS / l ) , where l denotes the number of effective observations and RSS denotes the sum of squared residuals. We then select the optimal order ( p 0 , q 0 ) by minimizing the AICs.
Step 3 Train the NNR-ARMA ( p 0 , q 0 ) model for the whole training sample y 1 , , y n + l and predict the testing samples with the estimated NNR-ARM ( p 0 , q 0 ) model. The obtained prediction errors are then used in the construction of the NNR-based LSCUSUM test T ^ n L S and T ^ n max .

5. Monte Carlo Simulation

We conducted Monte Carlo simulations to evaluate the finite sample performance of the NNR-based LSCUSUM test for linear and nonlinear ARMA models. For this, we used the AR, ARMA, TARMA, and TVAR models to generate data, and focused on the comparison of NNR-, ALASSO-, SVR-, and classical ARMA-based LSCUSUM tests. For each simulation, the empirical sizes and powers were calculated as the number of rejections of the null hypothesis of no changes out of 1000 repetitions at the 5% significance level. The sample sizes under consideration were n = 300 , 500, and 1000. The change point was assumed to occur at [ n / 2 ] under the alternatives. The simulations were conducted with R version 4.1.3, and we used the R packages nnet and caret for NNR, glmnet for ALASSO, and e1071 for SVR.
(1) 
Linear ARMA model
We generated time series from the ARMA (1,1) model:
y t = ϕ y t 1 + θ ε t 1 + ε t ,
where ϕ < 1 and ε t are IID normal random variables with mean zero and variance σ 2 . We considered the following setups for the null hypothesis:
  • Model 1: AR(1) with ϕ = 0.3 ,   θ = 0 , and σ 2 = 1 .
  • Model 2: ARMA(1,1) with ϕ = 0.3 ,   θ = 0.3 , and σ 2 = 1 .
Under the alternative, we assumed a single parameter changed while the others remained constant. Table 1 and Table 2 report the empirical sizes and powers for the AR(1) and ARMA(1,1) models, respectively, from which it can be seen that there was no size distortion, and reasonably good powers were produced using all four methods. In particular, the NNR-based LSCUSUM test outperformed the ALASSO-, SVR-, and classical ARMA-based LSCUSUM test, and for the LSCUSUM statistic, T ^ n max was preferable to T ^ n L S in terms of stability and power. In addition, we conducted simulations with a sample size of 500 and different change point locations; see Table 3. It can be seen that the empirical power is greatly affected by the position of the change point, the NNR-based LSCUSUM test has the highest power when a change occurs in the middle, and the power decreases as the change point comes nearer to the beginning and ending points of time series, indicating that the change point is more easily detected when in the middle of the sequence.
We also found that our method is obviously superior to the classical method when the change point occurs at the front of the sequence. Although not reported here, we carried out simulations for various settings of model parameters and sample sizes, and a similar pattern was seen.
(2) 
Threshold ARMA model
We generated time series from the TARMA(1,1) model:
y t = ϕ 1 y t 1 + θ 1 ε t 1 + ε t . y t 1 1 ϕ 2 y t 1 + θ 2 ε t 1 + ε t . y t 1 > 1
where ε t are IID normal random variables with mean zero and variance σ 2 . We considered the following setup for the null hypothesis:
  • Model 3: TARMA(1,1) with ϕ 1 = 0.1 ,   ϕ 2 = 0.5 ,   θ 1 = θ 2 = 0 . 5 , and σ 2 = 1 .
  • Model 4: TARMA(1,1) with ϕ 1 = 0.3 ,   ϕ 2 = 0.5 ,   θ 1 = θ 2 = 0 . 5 , and σ 2 = 1 .
Under the alternative, we assumed a single parameter changed while the others remained constant. Table 4 and Table 5 list some empirical sizes and powers for models 3 and 4, respectively. The results show that the classical ARMA- and ALASSO-based LSCUSUM tests had severe size distortion, which became more severe as the sample size increased. The NNR- and SVR-based LSCUSUM did not have such a problem, and the NNR-based LSCUSUM test produced better powers than the SVR-based LSCUSUM test. As anticipated, the power increased as the sample size increased or the deviance between the two parameter sets before and after the change point became larger. The simulation results considering different change point positions were the same as before, i.e., the LSCUSUM test had the highest power when a change point occurred in the middle of the time series (Table 6).
(3) 
Time-varying AR model
We generated a time series from the period AR(1) (PAR) model:
y t = ϕ t y t 1 + ε t
where ϕ t is periodic with period t = 1 , 2 and ε t ~ N ( 0 , σ 2 ) . We considered the following setup for the null hypothesis:
  • Model 5: PAR (1) with ϕ 1 = 0.3 ,   ϕ 2 = 0.5 , and σ 2 = 1 .
  • Model 6: PAR (1) with ϕ 1 = 0.3 ,   ϕ 2 = 0.5 , and σ 2 = 1 .
Under the alternative, we assumed a single parameter changed while the others remained constant. Table 7 and Table 8 summarize the empirical sizes and powers, respectively, for the PAR model. The results show that no tests had size distortions, and the NNR-based LSCUSUM test outperformed the ALASSO-, SVR- and classical ARMA-based LSCUSUM tests of stability and power. As in cases 1 and 2, NNR-based T ^ n max was superior to the other tests. Table 9 shows a similar pattern to cases 1 and 2.

6. Empirical Applications

We illustrate our proposed NNR-based LSCUSUM test scheme by analyzing three sets of real data: the annual volume of discharge from the Nile River at Aswan in 10 8 m 3 from 1871 to 1970, the weekly average oil prices of the USA from 1 September 2014 to 27 August 2018, and the weekly Hang Seng Index (HSI) of Hong Kong from 4 September 2011 to 26 Jun 2022, whose respective time series were obtained from the R package TSA and the websites www.macrotrends.net (accessed on 27 June 2022) and investing.com. We used 70% samples of each dataset for training and the remaining 30% for testing. Prior to fitting the NNR-ARMA model, we inspected the autocorrelation function (ACF) and partial autocorrelation function (PACF) of each time series, looking for irregular patterns of autocorrelations, as seen in Figure 2, where plot (a) shows that the ACFs and PACFs for the Nile River dataset support stationarity to a great extent, indicating that estimation via the NNR-ARMA model with this time series would not undermine the outcomes. In contrast, plots (b) and (c) show that oil prices and HSI are nonstationary time series, and we use the log-returns of the oil prices and HIS.
We applied the NNR-based LSCUSUM test to the Nile River data and obtained T ^ L S = 9 . 589 and T ^ max = 2 . 631 , which are larger than the respective critical values 2.4503 and 1.4596 at the 0.05 level. Thus, we rejected the null hypothesis of no changes. We calculated that both statistics detected the change point at 1899, which is the conclusion of Cobb [30], MacNeill et al. [31], and Wu and Zhao [32]. They believed that there is a change (decrease) in year 1899, which may be due to the construction of a new dam at Aswan. See Figure 3, where the red vertical line indicates a change point. For comparison, we also applied the classical ARMA-based LSCUSUM test to the Nile River data. Here, we expect this method to misbehave due to the nonlinearity of the datasets (Keenan’s linearity test [33] is quite against the linearity assumption). We obtained T ^ L S = 2 . 271 < 2 . 4503 , which does not reject the null hypothesis, whereas T ^ max = 1 . 468 > 1 . 4596 rejects it, indicating that the change point occurs in 1917. The blue vertical line in Figure 3 shows that the change point for the dataset is quite away from the previously obtained one. This result demonstrates that the classical ARMA-based LSCUSUM can be severely damaged when a time series sample has a strong nonlinear feature.
Furthermore, we obtained the NNR-based T ^ L S = 4.665 and T ^ max = 2 . 009 for log-returns of oil prices, which indicates the detection of change in both cases at the 0.05 level. Both statistics identified the location of a change point at 14 February 2016, whereas we found that T ^ L S = 7 . 004 and T ^ max = 2 . 153 for log-returns of HSI, and both statistics indicate that the change point occurs on 10 December 2017. Similar to the case of oil prices, a significant change in trend was observed in the original datasets, as shown in Figure 4 and Figure 5.

7. Concluding Remarks

We proposed a prediction method based on the NNR-ARMA model and described how to determine an optimal model. This model was used to obtain residuals for a training time series sample, which was divided into two subseries. A long AR model was fitted to the first subseries to obtain the residuals, which were used as the error terms in the NNR-ARMA ( p , q ) model, and for each estimated NNR-ARMA ( p , q ) model, we calculated AICs based on the second subseries and selected an optimal ARMA order with the smallest AIC. Prediction errors or residuals were obtained using the determined NNR-ARMA ( p , q ) model and used to construct the LSCUSUM test. Monte Carlo simulations were conducted using the linear and nonlinear ARMA models with various parameter settings, including ARMA, TARMA, and TVAR. The results show that the NNR-based LSCUSUM test outperformed the classical ARMA- and ALASSO-based LSCUSUM tests, especially when the underlying model was nonlinear. Finally, our method was applied to the analysis of real datasets, namely of Nile River data, oil prices, and HSI prices, and detected one change point in all cases. Our findings support the validity of our method and its practicality in various real-world circumstances.

Author Contributions

Conceptualization, Z.C. and X.-h.R.; methodology, X.-h.R.; software, X.-h.R.; validation, X.-h.R. and Y.L.; formal analysis, Z.C.; writing—original draft preparation, X.-h.R.; writing—review and editing, Z.C. and Y.L.; visualization, X.-h.R. and Y.L.; supervision, Z.C.; project administration, Z.C.; funding acquisition, Z.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No.12161072), Natural Science Foundation of Qinghai Province (No.2019-ZJ-920).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data is contained within the article.

Acknowledgments

The authors would like to thank the editor and the referees whose comments have improved the quality of this paper.

Conflicts of Interest

The authors declare no conflict interest.

References

  1. Csörgö, M.; Horváth, L. Limit Theorems in Change-Point Analysis; John Wiley & Sons, Inc.: New York, NY, USA, 1997. [Google Scholar]
  2. Hušková, M.; Prášková, Z.; Steinebach, J. On the detection of changes in autoregressive time series Asymptotic. J. Stat. Plan. Inference 2007, 137, 1243–1259. [Google Scholar] [CrossRef]
  3. Franke, J.; Kirch, C.; Kamgaing, J.T. Changepoints in times series of counts. J. Time Ser. Anal. 2012, 33, 757–770. [Google Scholar] [CrossRef]
  4. Kang, J.; Lee, S. Parameter change test for Poisson autoregressive models. Scand. J. Stat. 2014, 41, 1136–1152. [Google Scholar] [CrossRef]
  5. Lee, S.; Lee, Y.; Chen, C.W.S. Parameter change test for zero-inflated generalized Poisson autoregressive models. Statistics 2016, 50, 540–557. [Google Scholar] [CrossRef]
  6. Huh, J.; Kim, H.; Lee, S. Monitoring parameter shift with Poisson integer-valued GARCH models. J. Stat. Comput. Simul. 2017, 87, 1754–1766. [Google Scholar] [CrossRef]
  7. Page, E.S. A test for a change in a parameter occurring at an unknown point. Biometrika 1955, 42, 523–527. [Google Scholar] [CrossRef]
  8. Lee, S.; Na, O.; Na, S. On the cusum of squares test for variance change in nonstationary and nonparametric time series models. Ann. Inst. Stat. Math. 2003, 55, 467–485. [Google Scholar] [CrossRef]
  9. Na, O.; Lee, J.; Lee, S. Change point detection in SCOMDY models. AStA Adv. Stat. Anal. 2013, 97, 215–238. [Google Scholar] [CrossRef]
  10. Lee, J.; Lee, S. Parameter change test for nonlinear time series models with GARCH type errors. J. Korean Math. Soc. 2015, 52, 503–522. [Google Scholar] [CrossRef] [Green Version]
  11. Oh, H.; Lee, S. On score vector-and residual-based CUSUM tests in ARMA–GARCH models. Stat. Methods Appl. 2018, 27, 385–406. [Google Scholar] [CrossRef]
  12. Oh, H.; Lee, S. Modified residual CUSUM test for location-scale time series models with heteroscedasticity. Ann. Inst. Stat. Math. 2019, 71, 1059–1091. [Google Scholar] [CrossRef]
  13. Lee, S. Location and scale-based CUSUM test with application to autoregressive models. J. Stat. Comput. Simul. 2020, 90, 2309–2328. [Google Scholar] [CrossRef]
  14. Hwarng, H.B.; Ang, H.T. A simple neural network for ARMA (p, q) time series. Omega 2001, 29, 319–333. [Google Scholar] [CrossRef]
  15. Hossain, M.B.; Moon, J.; Chon, K.H. Estimation of ARMA model order via artificial neural network for modeling physiological systems. IEEE Access 2020, 8, 186813–186820. [Google Scholar] [CrossRef]
  16. Zafar, T.; Kamal, K.; Mathavan, S. A Hybrid Approach for Noise Reduction in Acoustic Signal of Machining Process Using Neural Networks and ARMA Model. Sensors 2021, 21, 8023. [Google Scholar] [CrossRef]
  17. Giacomazzo, M. Three Essays on Shrinkage Estimation and Model Selection of Linear and Nonlinear Time Series Models. Ph.D. Thesis, Arizona State University, Tempe, AZ, USA, 2018. [Google Scholar]
  18. Harchaoui, Z.; Lévy-Leduc, C. Multiple change-point estimation with a total variation penalty. J. Am. Stat. Assoc. 2010, 105, 1480–1493. [Google Scholar] [CrossRef] [Green Version]
  19. Chan, N.H.; Yau, C.Y.; Zhang, R.M. Group LASSO for structural break time series. J. Am. Stat. Assoc. 2014, 109, 590–599. [Google Scholar] [CrossRef]
  20. Jin, B.; Shi, X.; Wu, Y. A novel and fast methodology for simultaneous multiple structural break estimation and variable selection for nonstationary time series models. Stat. Comput. 2013, 23, 221–231. [Google Scholar] [CrossRef]
  21. Qian, J.; Su, L. Shrinkage estimation of common breaks in panel data models via adaptive group fused lasso. J. Econom. 2016, 191, 86–109. [Google Scholar] [CrossRef]
  22. Lee, S.; Seo, M.H.; Shin, Y. The lasso for high dimensional regression with a possible change point. J. R. Stat. Soc. Ser. B Stat. Methodol. 2016, 78, 193. [Google Scholar] [CrossRef]
  23. Jin, B.; Wu, Y.; Shi, X. Consistent two-stage multiple change-point detection in linear models. Can. J. Stat. 2016, 44, 161–179. [Google Scholar] [CrossRef]
  24. Ciuperca, G.; Maciak, M. Change-point detection in a linear model by adaptive fused quantile method. Scand. J. Stat. 2020, 47, 425–463. [Google Scholar] [CrossRef] [Green Version]
  25. Lee, S.; Lee, S.; Moon, M. Hybrid change point detection for time series via support vector regression and CUSUM method. Appl. Soft Comput. 2020, 89, 106101. [Google Scholar] [CrossRef]
  26. Lee, S.; Lee, S. Change Point Test for the Conditional Mean of Time Series of Counts Based on Support Vector Regression. Entropy 2021, 23, 433. [Google Scholar] [CrossRef] [PubMed]
  27. Looney, C.G. Pattern Recognition Using Neural Networks: Theory and Algorithms for Engineers and Scientist; Oxford University Press Inc.: Oxford, UK, 1997. [Google Scholar]
  28. Swanson, N.R.; White, H. A model-selection approach to assessing the information in the term structure using linear models and artificial neural networks. J. Bus. Econ. Stat. 1995, 13, 265–275. [Google Scholar]
  29. Hannan, E.J.; Rissanen, J. Recursive estimation of mixed autoregressive-moving average order. Biometrika 1982, 69, 81–94. [Google Scholar] [CrossRef]
  30. Cobb, G.W. The problem of the Nile: Conditional solution to a changepoint problem. Biometrika 1978, 65, 243–251. [Google Scholar] [CrossRef]
  31. MacNeill, I.B.; Tang, M.; Jandhyala, V.K. A search for the source of the Nile’s change-points. Environmetrics 1991, 2, 341–375. [Google Scholar] [CrossRef]
  32. Wu, W.B.; Zhao, Z. Inference of trends in time series. J. R. Stat. Soc. 2007, 69, 391–410. [Google Scholar] [CrossRef]
  33. Li, W.K. Diagnostic Checks in Time Series; Chapman and Hall/CRC: Boca Raton, FL, USA, 2003. [Google Scholar]
Figure 1. Proposed neural network topology.
Figure 1. Proposed neural network topology.
Entropy 25 00133 g001
Figure 2. Plot of ACF and partial autocorrelation function (PACF) up to lag 30 of: (a) Nile River data; (b) oil prices; and (c) HSI.
Figure 2. Plot of ACF and partial autocorrelation function (PACF) up to lag 30 of: (a) Nile River data; (b) oil prices; and (c) HSI.
Entropy 25 00133 g002aEntropy 25 00133 g002b
Figure 3. Annual volume of discharge from Nile River at Aswan, 1871–1970.
Figure 3. Annual volume of discharge from Nile River at Aswan, 1871–1970.
Entropy 25 00133 g003
Figure 4. (a) Weekly averages of oil prices; (b) log-returns with detected change point.
Figure 4. (a) Weekly averages of oil prices; (b) log-returns with detected change point.
Entropy 25 00133 g004
Figure 5. (a) Weekly price of HIS; (b) log-returns with detected change point.
Figure 5. (a) Weekly price of HIS; (b) log-returns with detected change point.
Entropy 25 00133 g005
Table 1. Empirical sizes and powers for AR(1) model with ϕ = 0.3 ,   σ 2 = 1 under the null and each alternative.
Table 1. Empirical sizes and powers for AR(1) model with ϕ = 0.3 ,   σ 2 = 1 under the null and each alternative.
ϕ = 0.3 , σ 2 = 1 n = 300 n = 500 n = 1000
NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
Size T ^ n L S 0.0560.0440.0380.0430.0460.0450.0470.0460.0610.0440.0440.045
T ^ n max 0.0500.0420.0350.0410.0450.0400.0460.0420.0570.0390.0430.039
ϕ = 0.5 T ^ n L S 0.2830.2670.2260.2680.4580.4570.4190.4490.8140.8090.7880.810
T ^ n max 0.2870.2570.2380.2640.4890.4630.4350.4580.8330.8160.7970.816
ϕ = 0.7 T ^ n L S 0.9290.8760.8180.8720.9990.9920.9830.9921.0001.0001.0001.000
T ^ n max 0.9370.8970.8130.8880.9970.9940.9840.9931.0001.0001.0001.000
σ 2 = 2 T ^ n L S 0.9080.9030.8850.9161.0000.9990.9970.9991.0001.0001.0001.000
T ^ n max 0.9300.9230.9900.9341.0001.0000.9981.0001.0001.0001.0001.000
Table 2. Empirical sizes and powers for ARMA(1,1) model with ϕ = 0.3 , θ = 0.3 , σ 2 = 1 under the null and each alternative.
Table 2. Empirical sizes and powers for ARMA(1,1) model with ϕ = 0.3 , θ = 0.3 , σ 2 = 1 under the null and each alternative.
ϕ = 0.3 , θ = 0.3 n = 300 n = 500 n = 1000
σ 2 = 1 NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
Size T ^ n L S 0.0480.0470.0530.0420.0520.0440.0490.0420.0640.0480.0430.043
T ^ n max 0.0450.0460.0470.0380.0410.0480.0460.0420.0630.0460.0430.042
ϕ = 0.5 T ^ n L S 0.3160.2810.2540.2520.5120.4730.4200.4390.8630.8300.7900.805
T ^ n max 0.3160.2790.2510.2420.5220.4740.4080.4400.8700.8500.7940.824
ϕ = 0.7 T ^ n L S 0.9490.8990.7840.8590.9970.9950.9750.9931.0001.0001.0001.000
T ^ n max 0.9540.9170.8020.8690.9980.9980.9810.9961.0001.0001.0001.000
θ = 0.7 T ^ n L S 0.4880.5150.4510.4920.7980.8290.7640.8130.9920.9940.9840.988
T ^ n max 0.5190.5270.4510.5120.8100.8370.7740.8210.9930.9930.9840.992
σ 2 = 2 T ^ n L S 0.9380.9230.8800.9220.9990.9980.9920.9981.0001.0001.0001.000
T ^ n max 0.9310.9250.8840.9270.9970.9970.9960.9971.0001.0001.0001.000
Table 3. Empirical powers when change point occurs at [ n τ ] with n = 500 for ARMA(1,1) model.
Table 3. Empirical powers when change point occurs at [ n τ ] with n = 500 for ARMA(1,1) model.
ϕ = 0.3 , θ = 0.3 τ = 0.25 τ = 0.50 τ = 0.75
σ 2 = 1 NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
Size T ^ n L S 0.2780.2280.2120.1070.5120.4730.4200.4390.3680.2810.2570.248
T ^ n max 0.2870.2260.2040.1000.5220.4740.4080.4400.3590.2980.2600.263
ϕ = 0.5 T ^ n L S 0.8990.8270.6570.3890.9970.9950.9750.9930.9470.9450.9100.935
T ^ n max 0.9140.8570.6910.3770.9980.9980.9810.9960.9520.9500.9310.952
ϕ = 0.7 T ^ n L S 0.6000.6760.5090.3750.7980.8290.7640.8130.5510.4940.4510.458
T ^ n max 0.6280.6840.5270.3890.8100.8370.7740.8210.5410.4970.4500.452
θ = 0.7 T ^ n L S 0.8340.8090.7570.4390.9990.9980.9920.9980.9640.9610.9400.962
T ^ n max 0.8610.8710.7960.5100.9970.9970.9960.9970.9690.9690.9531.969
Table 4. Empirical sizes and powers for threshold ARMA(1,1) model with ϕ 1 = 0.1 , ϕ 2 = 0.5 ,   θ = θ 1 = θ 2 = 0.5 ,   σ 2 = 1 under the null and each alternative.
Table 4. Empirical sizes and powers for threshold ARMA(1,1) model with ϕ 1 = 0.1 , ϕ 2 = 0.5 ,   θ = θ 1 = θ 2 = 0.5 ,   σ 2 = 1 under the null and each alternative.
ϕ 1 = 0.1 , ϕ 2 = 0.5 n = 300 n = 500 n = 1000
θ = 0.5 , σ 2 = 1 NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
Size T ^ n L S 0.0570.0700.0390.0730.0400.0860.0380.0960.0410.0830.0370.097
T ^ n max 0.0450.0660.0400.0680.0400.0920.0410.0950.0410.0810.0430.095
ϕ 1 = 0.5 T ^ n L S 0.4640.3100.2130.4450.8880.8170.8200.8251.0001.0000.9941.000
T ^ n max 0.5100.3270.3210.4750.9090.7690.7480.7561.0001.0000.9971.000
ϕ 1 = 0.7 T ^ n L S 0.9980.9750.6840.9771.0001.0000.9251.0001.0001.0001.0001.000
T ^ n max 0.9980.9510.7480.9591.0001.0000.9511.0001.0001.0001.0001.000
θ = 0.5 T ^ n L S 0.5240.3610.3230.7120.7470.3910.4810.8100.8880.6300.7530.946
T ^ n max 0.5310.3900.3520.7370.7560.3980.5130.8180.8990.6350.7760.946
σ 2 = 2 T ^ n L S 0.9040.9130.8680.9130.9950.9890.9970.9911.0001.0001.0001.000
T ^ n max 0.9090.9200.8800.9230.9940.9930.9910.9911.0001.0001.0001.000
Table 5. Empirical sizes and powers for threshold ARMA(1,1) model with ϕ 1 = 0.3 , ϕ 2 = 0.5 ,   θ = θ 1 = θ 2 = 0.5 ,   σ 2 = 1 under the null and each alternative.
Table 5. Empirical sizes and powers for threshold ARMA(1,1) model with ϕ 1 = 0.3 , ϕ 2 = 0.5 ,   θ = θ 1 = θ 2 = 0.5 ,   σ 2 = 1 under the null and each alternative.
ϕ 1 = 0.3 , ϕ 2 = 0.5 n = 300 n = 500 n = 1000
θ = 0.5 , σ 2 = 1 NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
Size T ^ n L S 0.0650.0940.0340.1000.0690.1370.0500.1370.0450.1340.0440.134
T ^ n max 0.0670.0940.0350.0960.0680.1400.0390.1340.0480.1360.0420.135
ϕ 1 = 0.7 T ^ n L S 0.8800.7990.4330.8360.9940.9990.7951.0001.0001.0000.9961.000
T ^ n max 0.8760.7810.5280.8410.9930.9950.8570.9951.0001.0000.9981.000
ϕ 1 = 0.9 T ^ n L S 1.0001.0000.4880.9961.0001.0000.7011.0001.0001.0001.0001.000
T ^ n max 1.0001.0000.5430.9971.0001.0000.7691.0001.0001.0001.0001.000
θ = 0.5 T ^ n L S 0.2250.1130.0900.4300.2640.1400.1180.5490.2840.2180.1600.678
T ^ n max 0.2590.1100.0950.4720.2780.1350.1260.5850.2950.2120.1830.701
σ 2 = 2 T ^ n L S 0.9040.8960.8520.9030.9930.8980.9870.9911.0001.0001.0001.000
T ^ n max 0.9190.9100.8650.9110.9960.9880.9880.9881.0001.0001.0001.000
Table 6. Empirical power of the change point occurs at [ n τ ] and n = 500 for threshold ARMA(1,1).
Table 6. Empirical power of the change point occurs at [ n τ ] and n = 500 for threshold ARMA(1,1).
ϕ 1 = 0.1 , ϕ 2 = 0.5 τ = 0.25 τ = 0.50 τ = 0.75
θ = 0.5 , σ 2 = 1 NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
ϕ 1 = 0.5 T ^ n L S 0.3650.3780.2980.4790.8880.8170.4780.8250.6290.5760.5100.626
T ^ n max 0.3580.3570.3020.4960.9090.7690.4890.7560.6350.5590.5160.616
ϕ 1 = 0.7 T ^ n L S 0.7000.7690.5750.8261.0001.0000.9251.0000.9890.9520.9220.960
T ^ n max 0.7260.7560.6080.8361.0001.0000.9511.0000.9930.9400.9320.951
θ = 0.5 T ^ n L S 0.9860.9750.2981.0000.7470.3910.4810.8100.7560.3540.5440.891
T ^ n max 0.9880.9750.2710.9990.7560.3980.5130.8180.7200.3510.4890.868
σ 2 = 2 T ^ n L S 0.8100.8260.7780.8360.9950.9890.9970.9910.9620.9490.9380.949
T ^ n max 0.8140.8640.7870.9620.9940.9930.9910.9910.9610.9540.9430.955
Table 7. Empirical sizes and powers for PAR(1) model with ϕ 1 = 0.3 , ϕ 2 = 0.5 ,   T = 2 under the null and each alternative.
Table 7. Empirical sizes and powers for PAR(1) model with ϕ 1 = 0.3 , ϕ 2 = 0.5 ,   T = 2 under the null and each alternative.
ϕ 1 = 0.3 , ϕ 2 = 0.5 n = 300 n = 500 n = 1000
T = 2 NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
Size T ^ n L S 0.0530.0490.0560.0390.0550.0630.0580.0510.0570.0490.0480.047
T ^ n max 0.0520.0470.0460.0380.0600.0530.0550.0490.0520.0500.0460.056
ϕ 1 = 0.5 T ^ n L S 0.1400.1130.1120.0950.1850.1630.1550.1360.3370.3330.2810.290
T ^ n max 0.1340.1130.1040.0970.1750.1620.1530.1340.3430.3400.3050.299
ϕ 1 = 0.7 T ^ n L S 0.3470.2910.2910.2470.5670.5020.4680.4450.8750.8700.8310.834
T ^ n max 0.3440.2890.2930.2400.5700.5060.4820.4520.8880.8690.8470.848
ϕ 2 = 0.7 T ^ n L S 0.0940.0790.0860.0690.1370.0910.1170.0820.2020.1990.2120.170
T ^ n max 0.0860.0740.0790.0620.1160.0920.1110.0750.1860.1860.1990.162
ϕ 2 = 0.3 T ^ n L S 0.7040.5760.4300.7120.9520.8980.7750.9501.0001.0000.9951.000
T ^ n max 0.6930.5800.4200.7140.9540.8990.7470.9501.0001.0000.9961.000
Table 8. Empirical sizes and powers for threshold PAR(1) model with ϕ 1 = 0.3 , ϕ 2 = 0.5 ,   T = 2 under the null and each alternative.
Table 8. Empirical sizes and powers for threshold PAR(1) model with ϕ 1 = 0.3 , ϕ 2 = 0.5 ,   T = 2 under the null and each alternative.
ϕ 1 = 0.3 , ϕ 2 = 0.5 n = 300 n = 500 n = 1000
T = 2 NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
Size T ^ n L S 0.0540.0480.0390.0440.0470.0500.0480.0390.0590.0410.0320.031
T ^ n max 0.0560.0390.0360.0390.0460.0530.0440.0460.0590.0390.0410.034
ϕ 1 = 0.5 T ^ n L S 0.1250.1070.0950.0900.1900.1540.1360.1380.3290.3270.2760.289
T ^ n max 0.1190.1060.0970.0940.1760.1530.1370.1390.3270.3170.2780.287
ϕ 1 = 0.7 T ^ n L S 0.3460.3150.2690.2620.5640.5060.4750.4430.8620.8590.8300.831
T ^ n max 0.3330.2990.2680.2440.5830.5110.4760.4570.8780.8720.8410.845
ϕ 2 = 0.7 T ^ n L S 0.1110.0810.0870.0650.1400.0860.1000.0830.2370.1780.1960.140
T ^ n max 0.0980.0740.0750.0620.1250.0860.0970.0700.2300.1600.1820.128
ϕ 2 = 0.3 T ^ n L S 0.7170.5890.4170.7540.9530.9210.7770.9611.0001.0000.9871.000
T ^ n max 0.7220.5800.4130.7380.9500.9240.7670.9641.0001.0000.9871.000
Table 9. Empirical power of the change point occurs at [ n τ ] for PAR(1).
Table 9. Empirical power of the change point occurs at [ n τ ] for PAR(1).
ϕ 1 = 0.3 , ϕ 2 = 0.5 τ = 0.25 τ = 0.50 τ = 0.75
T = 2 NNRALASSOSVRClassicalNNRALASSOSVRClassicalNNRALASSOSVRClassical
ϕ 1 = 0.5 T ^ n L S 0.1230.0950.0960.0880.1850.1630.1550.1360.1280.1060.1120.096
T ^ n max 0.1180.0950.0930.0830.1750.1620.1530.1340.1250.1210.1190.105
ϕ 1 = 0.7 T ^ n L S 0.3010.2140.2180.1810.5670.5020.4680.4450.3920.3460.3180.308
T ^ n max 0.2880.2250.2190.1700.5700.5060.4820.4520.3970.3440.3320.307
ϕ 2 = 0.7 T ^ n L S 0.0840.0640.0770.0540.1370.0910.1170.0820.1040.0830.1050.071
T ^ n max 0.0820.0620.0840.0530.1160.0920.1110.0750.0910.0860.0840.074
ϕ 2 = 0.3 T ^ n L S 0.4960.3180.2700.6080.9520.8980.7750.9500.7920.7510.6130.746
T ^ n max 0.5120.3150.2700.6210.9540.8990.7470.9500.7980.7630.6000.744
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ri, X.-h.; Chen, Z.; Liang, Y. Detecting Structural Change Point in ARMA Models via Neural Network Regression and LSCUSUM Methods. Entropy 2023, 25, 133. https://doi.org/10.3390/e25010133

AMA Style

Ri X-h, Chen Z, Liang Y. Detecting Structural Change Point in ARMA Models via Neural Network Regression and LSCUSUM Methods. Entropy. 2023; 25(1):133. https://doi.org/10.3390/e25010133

Chicago/Turabian Style

Ri, Xi-hame, Zhanshou Chen, and Yan Liang. 2023. "Detecting Structural Change Point in ARMA Models via Neural Network Regression and LSCUSUM Methods" Entropy 25, no. 1: 133. https://doi.org/10.3390/e25010133

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop