Next Article in Journal
MRI Brain Classification Using the Quantum Entropy LBP and Deep-Learning-Based Features
Next Article in Special Issue
An Appraisal of Incremental Learning Methods
Previous Article in Journal
TMEA: A Thermodynamically Motivated Framework for Functional Characterization of Biological Responses to System Acclimation
Previous Article in Special Issue
De-Biased Graphical Lasso for High-Frequency Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Inference for Convolutionally Observed Diffusion Processes

by
Shogo H Nakakita
1,2,3,* and
Masayuki Uchida
1,4
1
Graduate School of Engineering Science, Osaka University, 1-3 Machikaneyamacho, Toyonaka, Osaka 560-0043, Japan
2
Japanese Society for the Promotion of Science, 5 Chome-3-1 Kōjimachi, Chiyoda City, Tokyo 102-0083, Japan
3
The Ronin Institute for Independent Scholarship, 127 Haddon Place, Montclair, Montclair, NJ 07043, USA
4
Center for Mathematical Modeling and Data Science, Osaka University, 1-3 Machikaneyamacho, Toyonaka, Osaka 560-0043, Japan
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(9), 1031; https://doi.org/10.3390/e22091031
Submission received: 21 August 2020 / Revised: 10 September 2020 / Accepted: 11 September 2020 / Published: 15 September 2020

Abstract

:
We propose a new statistical observation scheme of diffusion processes named convolutional observation, where it is possible to deal with smoother observation than ordinary diffusion processes by considering convolution of diffusion processes and some kernel functions with respect to time parameter. We discuss the estimation and test theories for the parameter determining the smoothness of the observation, as well as the least-square-type estimation for the parameters in the diffusion coefficient and the drift one of the latent diffusion process. In addition to the theoretical discussion, we also examine the performance of the estimation and the test with computational simulation, and show an example of real data analysis for one EEG data whose observation can be regarded as smoother one than ordinary diffusion processes with statistical significance.

1. Introduction

We consider a d-dimensional diffusion process defined by the following stochastic differential equation,
d X t = b X t , β d t + a X t , α d w t , X λ = x λ ,
where λ > 0 , w t t λ is a standard r-dimensional Wiener process, x λ is an R d -valued random variable independent of w t t λ , α Θ 1 and β Θ 2 are unknown parameters, Θ 1 R m 1 and Θ 2 R m 2 are compact and convex parameter spaces, a : R d × Θ 1 R d R r and b : R d × Θ 2 R d are known functions. Our concern is statistical estimation for α and β from observation. θ = α , β denotes the true value of θ : = α , β .
We denote the observation as the discretised process X ¯ i h n , n : i = 0 , , n with discretisation step h n > 0 such that h n 0 and T n : = n h n , where the convoluted process X ¯ t , n t 0 is defined as
X ¯ t , n : = t ρ ¯ h n t V h n t s X s d s = R V h n t s X s d s = V h n X t ,
where V h n is an R d R d -valued kernel function whose support is a subset of 0 , ρ ¯ h n , and ρ ¯ > 0 such that sup n ρ ¯ h n λ . In this paper, we specify V h n = V ρ , h n which is a parametric kernel function whose support is a subset of 0 , ρ ¯ h n defined as
V ρ , h n i , j t : = ρ i h n 1 1 0 , ρ i h n t if i = j and ρ i > 0 , δ t if i = j and ρ i = 0 , 0 if i j ,
δ t is the Dirac-delta function, ρ = ρ 1 , , ρ d T Θ ρ : = 0 , ρ ¯ d is the smoothing parameter determining the smoothness of observation. That is to say, the observed process is defined as follows:
X ¯ i h n , n = ρ h n 1 i ρ h n i h n X s d s if ρ > 0 , X i h n if ρ = 0 ,
for all = 1 , , d . Let us consider both the problems that (i) ρ is a known parameter, and (ii) ρ is an unknown one whose true value is denoted by ρ and this is estimated by observation X ¯ i h n , n , and the parameter space is denoted as Ξ : = Θ ρ × Θ , where Θ : = Θ 1 × Θ 2 .
When assuming ρ as a known parameter, we can find literature for parametric estimation for α and/or β based on observation schemes which can be represented as special cases for some specific ρ . If ρ = 0 , our scheme is simply equivalent to parametric inference based on discretely observed diffusion processes X i h n : i = 0 , , n studied in [1,2,3,4,5,6,7,8] and references therein. If ρ = 1 , , 1 T , we can regard the problem as parametric estimation for integrated diffusion processes discussed in Gloter [9], Ditlevsen and Sørensen [10], Gloter [11], Gloter and Gobet [12], Sørensen [13]. Even for the case ρ = 0 , , 0 , 1 , , 1 T where some axes correspond to direct observation and the others do to integrated observation, we give consistent estimators for α and β by considering the scheme of convolutionally observed diffusion processes and this is one of the contributions of our study.
What is more, our contribution is to consider the scheme where ρ is unknown and succeed in representation of the microstructure noise which makes the observation smoother than the latent diffusion process itself, which can be seen in some biomedical time series data. Statistical modelling of biomedical time series data with stochastic differential equations has been one of the topics eagerly studied e.g., [14,15,16,17]. As [18] states, the existence of microstructure noise in financial data affects realised volatilities to increase as the subsampling frequency gets higher, for instance, see Figure 7.1 on p. 217 in [19]. However, realised volatilities of some biomedical data such as EEG decrease as subsampling frequency increases. For instance, some time series data for the 2nd participant in the dataset named Two class motor imagery (002-2014) of BNCI Horizon 2020 [20] show clear tendency of decreasing realised volatilities as subsampling frequency increases. Figure 1 shows the path of the 2nd axis of the data S02E.mat BNCI Horizon 2020 [20] for all 222 seconds (the observation frequency is 512 Hz, and hence the entire data size is 113664) and that for the first one second; it seems to perturb like a diffusion process.
We define realised volatilities with subsampling as for a sequence of real-valued observation Y i i = 0 , , n ,
R V k = 1 i n / k Y i k Y i 1 k 2 ,
where k = 1 , , 100 is the subsampling frequency parameter, and provide a plot of the realised volatilities the 2nd axis of the data S02E.mat in Figure 2:
The altitudes of the graph represented in the y-axis correspond to the values of the realised volatilities R V k with subsampling at every k observation represented in the x-axis. It is observable that the increasing subsampling frequency results in decreasing realised volatilities, which cannot be explained by the existent major microstructure noises, e.g., see [21,22,23,24,25]. To explain this phenomenon, we consider the smoother process than the latent one, though ordinarily microstructure noises make the observation rougher than the latent process, because quadratic variation of a sufficiently smooth function is zero. One way to deal with smoother observation than the latent state is convolutional observation. As a concrete example, we show a convolutionally observed diffusion process and its characteristics in realised volatilities: let us consider the following 1-dimensional stochastic differential equation defining an Ornstein–Uhlenbeck (OU) process:
d X t = 20 X t d t + 10 d w t , X λ = 0 ,
where λ = 10 2 / 5 . We simulate the stochastic differential equation by Euler–Maruyama method see [26] with parameters n = 10 7 , h n = 10 5 , and T n = 10 2 and its convolution approximated by summation with the smoothing parameter ρ = 10 (for details, see Section 5). Figure 3 shows the latent diffusion process and the convoluted observation on 0 , 1 , and we can see that the observation is indeed smoothed compared to the latent state.
In Figure 4, we also give the plot of realised volatilities of the convolutionally observed process with subsampling as Figure 2.
It is seen that the convolutional observation of a diffusion process also has the characteristics of decreasing realised volatilities as subsampling frequency increases, which can bee seen in some biological data such as BNCI Horizon 2020 [20]. Of course, graphically comparing characteristics of simulation and real data is insufficient to verify the convolutional observation with smoothing parameter ρ > 0 in 1-dimensional case; therefore, we propose statistical estimation method for unknown ρ and hypothesis test with the null hypothesis H 0 : ρ = 0 and the alternative one H 1 : ρ 0 from convolutional observation in Section 3. Moreover, in Section 6, we examine the real EEG data plotted in Figure 2 by the statistical hypothesis testing we propose, and see it is more appropriate to consider the data as a convolutional observation of a latent diffusion process with ρ 0 rather than direct observation of the latent process, which indicates the validity to deal with the problem of the convolutional observation scheme with unknown ρ .
The paper is composed of the following sections: Section 2 gives the notations and assumptions used in this paper; Section 3 discusses the estimation and test for smoothing parameter ρ , and the discussion provides us with the tools to examine whether we should consider the convolutional observation scheme; Section 4 proposes the quasi-likelihood functions for the parameter of diffusion processes α and β , and corresponding estimators with consistency; Section 5 is for the computational simulation to examine the theoretical results in the previous sections; and Section 6 shows an application of the methods we propose in real data analysis.

2. Notations and Assumptions

2.1. Notations

First of all, we set A x , α : = a x , α 2 , a x : = a x , α , A x : = A x , α and b x : = b x , β . We also give the notation for a matrix-valued function G x , α | ρ such that G i , j x , α | ρ : = A i , j x , α f G ρ i , ρ j , where
f G ρ i , ρ j : = 1 if ρ i = ρ j = 0 , 1 ρ j 2 if ρ i = 0 , ρ j 0 , 1 , 1 2 ρ j if ρ i = 0 , ρ j 1 , ρ ¯ , 1 ρ i 2 if ρ i 0 , 1 , ρ j = 0 , 1 2 ρ i if ρ i 1 , ρ ¯ , ρ j = 0 , 3 ρ i 2 ρ j + 3 ρ i ρ j 2 + 6 ρ i ρ j 2 ρ j 3 6 ρ i ρ j if ρ i , ρ j 0 , 1 , ρ i > ρ j , 3 ρ i 2 ρ j 3 ρ i ρ j 2 + 6 ρ i ρ j 2 ρ i 3 6 ρ i ρ j if ρ i , ρ j 0 , 1 , ρ i ρ j , 3 ρ j 2 + 3 ρ j ρ j 3 6 ρ i ρ j if ρ i 1 , ρ ¯ , ρ j 0 , 1 , ρ i > ρ j + 1 , 6 ρ j 1 6 ρ i ρ j if ρ i , ρ j 1 , ρ ¯ , ρ i > ρ j + 1 , ρ i ρ j 3 3 ρ i 2 + 6 ρ i ρ j + 3 ρ i 1 ρ j 3 6 ρ i ρ j if ρ i 1 , ρ ¯ , ρ j 0 , 1 , ρ i ρ j + 1 , ρ i ρ j 3 3 ρ i 2 + 6 ρ i ρ j + 3 ρ i 3 ρ j 2 + 3 ρ j 2 6 ρ i ρ j if ρ i , ρ j 1 , ρ ¯ , ρ j < ρ i ρ j + 1 , ρ i ρ j 3 + 6 ρ i ρ j ρ i 3 3 ρ j 2 + 3 ρ j 1 6 ρ i ρ j if ρ i 0 , 1 , ρ j 1 , ρ ¯ , ρ j ρ i + 1 , ρ i ρ j 3 3 ρ i 2 3 ρ j 2 + 6 ρ i ρ j + 3 ρ i + 3 ρ j 2 6 ρ i ρ j if ρ i , ρ j 1 , ρ ¯ , ρ i ρ j ρ i + 1 , 3 ρ i 2 + 3 ρ i ρ i 3 6 ρ i ρ j if ρ i 0 , 1 , ρ j 1 , ρ ¯ , ρ j > ρ i + 1 , 6 ρ i 1 6 ρ i ρ j if ρ i , ρ j 1 , ρ ¯ , ρ j > ρ i + 1 .
The continuity of the function f G is shown in the supplementary material. For the detailed discussion of the necessity of f G , see Remark 4.
In addition, we also give the notation used throughout this paper.
  • For every matrix A, A T is the transpose of A, and A 2 : = A A T .
  • For every set of matrices A and B of the same size, A B : = tr A B T . Moreover, for any m N , A R m R m and u , v R m , A u , v : = v T A u .
  • v and A 1 , 2 denote the -th element of a vector v and the 1 , 2 -th one of a matrix A, respectively.
  • For any vector v and any matrix A, v : = tr v T v and A : = tr A T A .
  • Ω , P , F , F t denotes the stochastic basis, where F t : = σ x λ , w s : s t .
  • μ f · denotes the integral of a μ -integrable function f where μ is a measure.

2.2. Assumptions

With respect to X t , we assume the following conditions.
[A1]
(i)
For a constant C, for all x 1 , x 2 R d ,
sup α Θ 1 a x 1 , α a x 2 , α + sup β Θ 2 b x 1 , β b x 2 , β C x 1 x 2 .
(ii)
For all p 0 , sup t λ E θ X t p < .
(iii)
There exists a unique invariant measure ν 0 on R d , B R d and for all p 1 and f L p ν 0 with polynomial growth,
1 T λ T f X t d t P R d f x ν 0 d x .
Remark 1.
For sufficient conditions of these regularity ones, please see [A1] and Remark 1 in Uchida and Yoshida [7].
[A2]
There exists C > 0 such that a : R d × Θ 1 R d R r and b : R d × Θ 2 R d have continuous derivatives satisfying
sup α Θ 1 x j α i a x , α C 1 + x C , 0 i 2 , 0 j 2 , sup β Θ 2 x j β i b x , β C 1 + x C , 0 i 2 , 0 j 2 .
With the invariant measure ν 0 , ξ : = ρ T , θ T T , and ξ denoting the true value of ξ , we define
V 1 α | ξ : = R d G x , α | ρ G x , α | ρ 2 ν 0 d x , V 2 β | ξ : = R d b x , β b x , β 2 ν 0 d x .
For these functions, let us assume the following identifiability conditions hold.
[A3]
There exist χ 1 α > 0 and χ 1 β > 0 such that for all α Θ 1 and β Θ 2 , V 1 α | ξ χ 1 α α α 2 and V 2 β | ξ χ 1 β β β 2 .

3. Estimation and Test of the Smoothing Parameter

In this section, we discuss the case where the smoothing parameter ρ of the kernel function V ρ , h n is unknown. The estimation is significant for estimation of α and β since we utilise the estimate for ρ in quasi-likelihood functions of α and β . The test problem for hypotheses H 0 : ρ = 0 and H 1 : ρ 0 is also important to examine whether our framework of convolutional observation is meaningful.

3.1. Estimation of the Smoothing Parameter

For simplicity of notation, let us consider the case ρ ¯ > 2 ; otherwise the discussion is quite parallel. We should note that for all i = 1 , , d ,
G i , i x | ρ = A i , i x if ρ i = 0 , A i , i x 1 ρ i 3 if ρ i 0 , 1 , A i , i x 1 ρ i 1 3 ρ i 2 if ρ i 1 , ρ ¯ .
Let us consider the estimation of ρ i with using the next statistics: the full quadratic variation
1 n h n k = 1 n X ¯ k h n , n i X ¯ k 1 h n , n i 2 P ν 0 A i , i · if ρ i = 0 , ν 0 A i , i · 1 ρ i 3 if ρ i 0 , 1 , ν 0 A i , i · 1 ρ i 1 3 ρ i 2 if ρ i 1 , ρ ¯ ,
because of Proposition 3 in supplementary material Appendix A, and the reduced quadratic variation defined as 1 n h n 2 2 k n X ¯ 2 k h n , n i X ¯ 2 k 2 h n , n i 2 converges in probability as follows.
Lemma 1.
Under [A1], we have the convergence in probability such that
1 n h n 2 2 k n X ¯ 2 k h n , n i X ¯ 2 k 2 h n , n i 2 P ν 0 A i , i · if ρ i = 0 , ν 0 A i , i · 1 ρ i 6 if ρ i 0 , 2 , ν 0 A i , i · 2 ρ i 4 3 ρ i 2 if ρ i 2 , ρ ¯ .
Then we define the ratio of those statistics such that
R n i : = 1 n h n k = 1 n X ¯ k h n , n i X ¯ k 1 h n , n i 2 1 n h n 2 2 k n X ¯ 2 k h n , n i X ¯ 2 k 2 h n , n i 2 1 P 1 if ρ i = 0 , 1 ρ i 3 1 ρ i 6 1 if ρ i 0 , 1 , 1 ρ i 1 3 ρ i 2 1 ρ i 6 1 if ρ i 1 , 2 , 1 ρ i 1 3 ρ i 2 2 ρ i 4 3 ρ i 2 1 if ρ i 2 , ρ ¯ = 1 if ρ i = 0 , 6 2 ρ i 6 ρ i 1 if ρ i 0 , 1 , 6 ρ i 2 6 ρ i 2 ρ i 3 1 if ρ i 1 , 2 , 3 ρ i 1 6 ρ i 4 1 if ρ i 2 , ρ ¯ , = : R ρ i ,
where R has the next property.
Lemma 2.
R is a 3 ρ ¯ 1 6 ρ ¯ 4 1 , 1 -valued monotonically decreasing continuous function, and has a continuous inverse R 1 : 3 ρ ¯ 1 6 ρ ¯ 4 1 , 1 0 , ρ ¯ .
We define ρ ^ n such that
ρ ^ n i : = 0 if R n i > 1 , R 1 R n i if R n i 3 ρ ¯ 1 6 ρ ¯ 4 1 , 1 , ρ ¯ if R n i < 3 ρ ¯ 1 6 ρ ¯ 4 1 ,
and then continuous mapping theorem for convergence in probability verifies the next result.
Theorem 1.
Under [A1], ρ ^ n has consistency, i.e., ρ ^ n P ρ .
Remark 2.
We can compute y = R 1 x , x 3 ρ ¯ 1 6 ρ ¯ 4 1 , 1 by solving the following equations:
( i ) x = 6 2 y 6 y 1 if x 4 / 5 , 1 , ( i i ) x = 6 2 y 6 y 2 y 3 1 if x 5 / 8 , 4 / 5 , ( i i i ) x = 3 y 1 6 y 4 1 if x 3 ρ ¯ 1 6 ρ ¯ 4 1 , 5 / 8 .

3.2. Test for Smoothed Observation

For all i = 1 , , d , we consider the next hypothesis testing:
H 0 : ρ i = 0 , H 1 : ρ i > 0 .
Let us consider the following test statistic:
T i , n : = n 2 3 n h n 2 k = 1 n X ¯ k h n , n i X ¯ k 1 h n , n i 4 × 1 n h n k = 1 n X ¯ k h n , n i X ¯ k 1 h n , n i 2 1 n h n 2 2 k n X ¯ 2 k h n , n i X ¯ 2 k 2 h n , n i 2 = 3 / 2 k = 1 n X ¯ k h n , n i X ¯ k 1 h n , n i 4 × k = 1 n X ¯ k , n i X ¯ k 1 , n i 2 2 2 k n X ¯ 2 k h n , n i X ¯ 2 k 2 h n , n i 2 ,
and we abbreviate T i , n to T n if d = 1 . Under H 0 , we have the next result.
Theorem 2.
Under H 0 and [A1], we have the convergence in law such that
T i , n L N 0 , 1 .
We also obtain the result to support the consistency of the test.
Theorem 3.
Under H 1 and [A1], we have convergence such that for any c R ,
P T i , n < c 1 .
Remark 3.
These results are intuitive since the quadratic variation and the reduced one with some appropriate scaling should converge to the same value if H 0 holds and the quadratic variation with some appropriate scaling should converge to the value which is smaller than the value which the reduced one with scaling converge to.
Hence when we set the significance level α sig 0 , 1 , then we have the rejection region
T i , n < Φ 1 α sig
where Φ is the distribution function of the standard Gaussian distribution. Theorem 3 supports the consistency of the test.
This test is essential in terms of examining the validity to consider the scheme of convolutional observation: if ρ = 0 , then the ordinary observation scheme can be applied, but if ρ 0 , then we have the motivation to consider the convolutional observation scheme.

4. Least Square Estimation of the Diffusion and Drift Parameters

Let us set the least-square quasi-likelihood functions such that
H 1 , n α | ρ : = k = 1 n 1 h n X ¯ k h n , n X ¯ k 1 h n , n 2 G X ¯ k 1 h n , n , α | ρ 2 , H 2 , n β | ρ : = k = max i ρ i + 2 n 1 h n X ¯ k h n , n X ¯ k 1 h n , n h n b X ¯ k 2 max i ρ i h n , n , β 2 ,
and the least-square estimators α ^ n and β ^ n satisfying
H 1 , n α ^ n | ρ = sup α Θ 1 H 1 , n α | ρ , H 2 , n β ^ n | ρ = sup β Θ 2 H 2 , n β | ρ .
when ρ is known, and
H 1 , n α ^ n | ρ ^ n = sup α Θ 1 H 1 , n α | ρ ^ n , H 2 , n β ^ n | ρ ^ n = sup β Θ 2 H 2 , n β | ρ ^ n .
when ρ is unknown.
Theorem 4.
Under [A1]–[A3], α ^ n and β ^ n are consistent, i.e., α ^ n P α and β ^ n P β .
Remark 4.
The function G and f G are indeed complex and confusing; hence, we can consider some alternative estimation methods with subsampling or pre-averaging. However, these methods also have the problem what size of subsampling or pre-averaging is proper and the result of the estimation can be dependent on tuning the subsampling size or pre-averaging one. Therefore, our work proposes the estimation method which uses the observation without subsampling or pre-averaging.

5. Simulations

In this simulation section, we only consider the case where ρ is unknown and should be estimated by data with the method proposed in Section 3.

5.1. 1-Dimensional Simulation

We examine the following 1-dimensional stochastic differential equation whose solution is a 1-dimensional Ornstein–Uhlenbeck (OU) process:
d X t = β 1 X t + β 2 d t + α d w t , X λ = 0 ,
α Θ 1 : = 0.01 , 10 , β Θ 2 : = 10 , 0.01 × 10 , 10 , and λ = 10 7 / 3 . The procedure of the simulation is as follows: in the first place we iterate an approximated OU process by Euler–Maruyama scheme, for example, see [26] with simulation parameters n sim = 10 5 + m , h sim = 10 10 / 3 m , T sim = 10 5 / 3 where m N is a parameter to determine the precision of approximation; secondly, we give the approximation of convolution by summation such that
X ¯ i h n , n 1 10 m ρ k = 0 10 m 1 X i h n k h sim if 10 m ρ 1 , X i h n if 10 m ρ < 0 ,
where i = 0 , , n , the sampling frequency h n = 10 10 / 3 and n = 10 5 / 3 . In this Section 5.1, we fix the true value of α and β as α = 3 and β = 2 , 1 T , and change the true value of ρ Θ ρ : = 0 , 100 to see the corresponding changes of performance of estimation for ξ , and test for ρ in comparison to estimation by an existent method called local Gaussian approximation (LGA) for parametric estimation of discretely observed diffusion processes, e.g., see [4] which does not concern convolutional observation. All the numbers of iterations for different ρ ’s are 1000.
In the first place, we see the estimation and test with small values of ρ such that ρ = 0 , 0.1 , 0.2 , , 1 to observe how the performance of statistics changes by difference in ρ . Table 1 summarises the results of simulation of ρ ^ n for ρ ’s with respective empirical means and root mean square error (RMSE).
We can see the proposed estimator ρ ^ n works well for small ρ . With respect to the performance of the test statistic T n proposed in Section 3.2, Table 2 shows the empirical ratio of the number of iterations whose T n is lower than some typical critical values where Φ indicates the distribution function of 1-dimensional standard Gaussian distribution as well as the maximum value of T n in 1000 iterations.
Even for ρ = 0.1 , the simulation result supports the theoretical discussion of the test with consistency. Because Φ 10 16 = 8.222 , all the iterations with ρ 0.3 result in rejection of H 0 with substantially significance level 10 16 . Let us see the estimation for α and β by our proposal method and that by the LGA in Table 3.
Note that the biases of the estimation by LGA increase as the true value of ρ gets larger, while the estimation by our proposal method is not influenced by the true value of ρ . This result of the simulation supports the theoretical discussion in Section 4 stating the consistency of θ ^ n , and necessity to consider the convolutional observation scheme where the LGA method does not work properly.
Secondly, we consider the estimation and test with large ρ such that ρ = 10, 15, 20 to see if our proposal methods work even for large ρ . We note that the maximum values of T n for ρ = 10, 15, 20 in 1000 iterations are 55.091 , 68.462 and 79.105 , and hence we can detect the smoothed observation easily. Table 4 shows the empirical means and RMSEs of ρ ^ n for ρ = 10, 15, 20 and we can see that the RMSEs increase as ρ ’s increase; it indicates the difficulty to estimate ρ accurately when ρ is large.
Table 5 summarises the estimation for θ by means and RMSE, and tells us that although the large RMSE of ρ ^ n results in increase of RMSE of α ^ n , estimation by our method is substantially better than that by LGA.

5.2. 2-Dimensional Simulation

We consider the following 2-dimensional stochastic differential equation whose solution is a 2-dimensional OU process:
d X t 1 X t 2 = β 1 β 2 β 4 β 5 X t 1 X t 2 + β 3 β 6 d t + α 1 α 2 α 2 α 3 d w t , X λ = 0 0 ,
λ = 10 7 / 3 . The simulation is conducted with the settings as follows: firstly, we iterate the OU process by Euler–Maruyama scheme with the simulation sample size n sim = 10 5 + m , T sim = 10 5 / 3 and discretisation step h sim = 10 10 / 3 m , where m = 2 is the precision parameter for approximation of convolution; in the second place, we approximate the convoluted process with summation such that
X ¯ i h n , n j 1 10 m ρ j k = 0 10 m 1 X i h n k h sim j if 10 m ρ j 1 , X i h n j if 10 m ρ j < 0 ,
where i = 0 , , n , j = 1 , 2 , the sampling scheme for inference is defined as n = 10 5 and h n = 10 10 / 3 ; the true value of ρ , α and β are set as ρ = 2 , 4 T , α = 2 , 0 , 3 T , β = 2 , 0.4 , 0 , 0.1 , 3 , 5 T ; the parameter spaces are defined as Θ ρ = 0 , 10 2 , Θ 1 = 1 + 10 8 , 10 × 1 + 10 8 , 1 10 8 × 1 + 10 8 , 10 , and Θ 2 = 10 , 10 6 ; the total iteration number is set to 1000.
Table 6 summarises the estimation for ρ with the method proposed in Section 3 (the inverse of r is computationally obtained) with empirical means and empirical RMSEs of ρ ^ n in 1000 iterations. We can see that ρ ^ n is sufficiently precise to estimate the true value of ρ indeed in this result, which is significant to estimate the other parameters α and β .
We also note that the maximum values of the test statistics for smoothed observation proposed in Section 3.2 in 1000 iterations are 17.947 and 33.159 for each axis. The p-value for them are smaller than 10 16 ; therefore, we can conclude that it is possible to detect the smoothed observation with the proposed test statistic in the case ρ i = 2.0 if d = 2 from this result.
With respect to the estimation for α and β , we compare the estimates by our proposal method with that by LGA which does not concern convolutional observation. Table 7 is the summary for α estimate by both the methods:
We can see that the estimation precision for α by our proposal outperforms those by LGA. This results support validity of our estimation method if we have convolutional observation for diffusion processes. Regarding β , the simulation result is summarised in Table 8:
Though the estimation for β 3 by our method has the smaller bias in comparison to that by LGA, the RMSE of our method is larger than that of LGA; in the estimation for other parameters, our proposal method outperforms the method by LGA. We can conclude that our proposal for estimation of α and β concerning convolutional observation performs better than that not considering this observation scheme.

6. Real Data Analysis

In this section, we analyse the EEG dataset named S02E.mat provided in “2. Two class motor imagery (002-2014)” of BNCI Horizon 2020 [20]. The datasets including S02E.mat are also studied by Steryl et al. [27].

6.1. Estimation and Test for the Smoothing Parameters

In the first place, we pick up the first 15 axes of the dataset and compute ρ ^ n and T n proposed in Section 3.1 and Section 3.2 respectively. The results are shown in Table 9.
We can observe that all the 15 time series data have the smoothing parameter ρ > 0 with statistical significance when we assume ordinary significance levels. These results motivate us to use our methods for parametric estimation proposed in Section 4 when we fit stochastic differential equations for these data.

6.2. Parametric Estimation for a Diffusion Process

We fit a 1-dimensional OU process for the time series data in the 2nd column of the data file S02E.mat with 512 Hz observation for 222 s (the plot of the path can be seen in Figure 1), whose ρ ^ n = 1.037 is the largest among those for the 15 axes and it is larger than 0 with statistical significance. According to the simulation result shown in Section 5.1, this size of the smoothing parameter gives critical biases when we estimate α and β with LGA method not concerning convolutional observation scheme.
The stochastic differential equation with parameters α Θ 1 : = 0.01 , 200 and β Θ 2 : = 100 , 0.01 × 100 , 100 is defined as follows:
d X t = β 1 X t + β 2 d t + α d w t , X λ = x λ .
We set 5 s as the time unit: hence n = 113,664 and h n = 1 / 5 × 512 . If we fit the OU process with the LGA method, i.e., we do not concern convolutional observation scheme, we obtain the fitting result such that
d X t = 17.378 X t + 1.091 d t + 122.892 d w t , X λ = x λ .
In the next place, we fit α and β with the least square method proposed in Section 4, and then we have the next fitting result:
d X t = 2.146 X t + 0.552 d t + 151.919 d w t , X λ = x λ .
It is worth noting that this fitting result is substantially different to that by LGA as shown above: hence these results indicate the significance to examine if the observation is convoluted with the smoothing parameter ρ > 0 and otherwise the estimation is strongly biased.

7. Summary

We have discussed the convolutional observation scheme which deals with the smoothness of observation in comparison to ordinary diffusion processes. The first contribution is to propose this new observation scheme with the statistical test to confirm whether this scheme is valid in real data. The second one is to prove consistency of the estimator ρ ^ n for the smoothing parameter ρ , those for parameters in diffusion and drift coefficient, i.e., α and β , of the latent diffusion process X t . Thirdly, we have examined the performance of those estimators and the test statistics in computational simulation, and verified these statistics work well in realistic settings. In the fourth place, we have shown a real example of observation where ρ 0 holds with statistical significance.
If we combine the test for noise detection by Nakakita and Uchida [28] and that for smoothed observation proposed in this paper, we can test if the observed process is diffusion or not in terms that the observation is noisy or smoothed. Note that the realised volatilities of the noisy observation of diffusion processes increase as observation frequency increases while those of the smoothed observation decreases as the frequency grows. On that point, the noisy observation in ordinary meaning and the smoothed one are ‘opposite’ to each other.
These contributions, especially the third one, will cultivate the motivation to study statistical approaches for convolutionally observed diffusion processes furthermore, such as estimation of kernel function V appearing in the convoluted diffusion X ¯ t : = V X t , test theory for parameters α and β as likelihood-ratio-type test statistics, for example, see [29,30], large deviation inequalities for quasi-likelihood functions and associated discussion of Bayes-type estimators, e.g., [6,31,32,33]. By these future works, it is expected that the applicability of stochastic differential equations in real data analysis and contributions to the areas with high frequency observation of phenomena such as EEG will be enhanced.

8. Proofs

8.1. the Results for Some Laws of Large Numbers

In this subsection, we give the notations and statements of propositions without proofs except for Proposition 3: the detailed proofs are given in supplementary material. We assume Δ λ , k N , M > 0 , and consider a class of R ¯ k R ¯ d -valued kernel functions on R denoted as K Δ , k , M such that for all Φ Δ K Δ , k , M , it holds:
( i ) supp Φ Δ 0 , Δ , ( ii ) for all f : 0 , Δ × Ω R k , ω Ω , 0 Δ Φ Δ Δ s f s , ω d s M sup s 0 , Δ f s , ω ( iii ) for all t 0 λ , f : R d R which is continuous and at most polynomial growth , E t Δ t Φ Δ t s f X s d s F t 0 = t Δ t Φ Δ t s E f X s F t 0 d s .
Remark 5.
Note one sufficient condition for Φ Δ K Δ , k , M is (i) Φ Δ : R R k R d , (ii) supp Φ Δ 0 , Δ , (iii) 0 Δ Φ Δ Δ s d s M and (iv) B 0 , Δ -measurable since
0 Δ Φ Δ Δ s f s , ω d s 0 Δ Φ Δ Δ s f s , ω d s M sup s 0 , Δ f s , ω
for the Cauchy–Schwarz inequality, and Fubini’s theorem.
It is easily checked that V ρ , h n K max i = 1 , , d ρ i h n , d , d .
Let p denote an integer such that sup n N p h n λ , Δ n : = p h n . We set the sequence of the kernels Φ Δ n , n n N such that Φ Δ n , n K Δ n , d , M for some M > 0 , 0 Δ n Φ Δ n , n d s = I d and there exist a matrix B R d R d such that
0 Δ n + h n Φ Δ n , n Δ n + h n s Φ Δ n , n Δ n s s d s h n B C h n 2 1 + x C ,
a set L 0 , , p such that there exist functions D : R d R d R d for L such that
E 0 Δ n + 1 + h n Φ Δ n , n Δ n s 1 0 s 1 a x d w s 2 d s 1 0 Δ n + 1 + h n Φ Δ n , n Δ n + 1 + h n s 1 Φ Δ n , n Δ n + h n s 1 × 0 s 1 a x d w s 2 d s 1 T h n D x C h n 2 1 + x C ,
a function G : R d R d R d such that
E 0 Δ n + h n Φ Δ n , n Δ n + h n s 1 Φ Δ n , n Δ n s 1 0 s 1 a x d w s 2 d s 1 0 Δ n + h n Φ Δ n , n Δ n + h n s 1 Φ Δ n , n Δ n s 1 0 s 1 a x d w s 2 d s 1 T h n G x C h n 2 1 + x C .
We define
X ¯ t , n = t Δ n t Φ Δ n , n t s X s d s ,
and the following random quantities such that
ν ¯ n f · , ξ : = 1 n i = 1 n f X ¯ i h n , n , ξ , I ¯ , n v · , ξ : = 1 n h n i = 1 + n v X ¯ i 1 h n , n , ξ X ¯ i h n , n X ¯ i 1 h n , n h n B b X ¯ i 1 h n , n , Q ¯ n M · , ξ : = 1 n h n i = 1 n M X ¯ i 1 h n , n , ξ X ¯ i h n , n X ¯ i 1 h n , n 2 ,
where f : R d × Ξ R , v : R d × Ξ R d , M : R d × Ξ R d R d are in C 2 -class, and their first and second derivatives and themselves are at most polynomial growth uniformly in ξ Ξ .
Proposition 1.
Under [A1], ν ¯ n f · , ξ P ν 0 f · , ξ uniformly in ξ Ξ .
Proposition 2.
If L and [A1] hold, I ¯ , n v · , ξ P ν 0 x v D T · , ξ uniformly in ξ Ξ .
Proposition 3.
Under [A1], Q ¯ n M · , ξ P ν 0 M G · , ξ uniformly in ξ Ξ .
We set p = ρ ¯ + 1 , Δ n = p h n and see the evaluation of B, D and G when setting our kernel Φ Δ , n = V ρ , h n as follows (for the derivation of the evaluation, see the supplementary material): we have Δ n = p h n , B = I d , D 0 x = D 0 x | ρ ρ = ρ , where D 0 i , j x | ρ = A i , j x f D 0 ρ i , ρ j ,
f D 0 ρ i , ρ j : = 0 if ρ j = 0 , ρ j 2 if ρ i = 0 , ρ j 0 , 1 , 2 ρ j 1 2 ρ j if ρ i = 0 , ρ j 1 , ρ ¯ , 6 ρ i ρ j 3 ρ i 2 3 ρ i 6 ρ i ρ j if ρ i > 0 , ρ i + 1 < ρ j , ρ i ρ j 3 + 3 ρ j 2 3 ρ j + 1 6 ρ i ρ j if ρ j > 1 , ρ i < ρ j ρ i + 1 , 3 ρ j 2 3 ρ j + 1 6 ρ i ρ j if ρ j > 1 , ρ i ρ j , ρ i ρ j 3 + ρ j 3 6 ρ i ρ j if ρ j 0 , 1 , ρ i 0 , ρ j , ρ j 3 6 ρ i ρ j if ρ j 0 , 1 , ρ i ρ j ,
D = O for max i = 1 , , d ρ d + 1 because of independent increments of the Wiener process, and G x = G x | ρ ρ = ρ where G x | ρ = G x , α | ρ α = α .

8.2. Proof of the Results in Section 3.1

Proof of Lemma 1.
By following the proof of the Proposition 3, it is sufficient to evaluate
0 p + 2 h n 0 p + 2 h n A i , i x min s , s V ρ , h n i , i p + 2 h n s V ρ , h n i , i p h n s V ρ , h n i , i p + 2 h n s V ρ , h n i , i p h n s d s d s
for the asymptotic behaviour of the reduced quadratic variation by choosing
M 1 , 2 = 1 if 1 = 2 = i , 0 otherwise .
If ρ i = 0 ,
0 p + 2 h n 0 p + 2 h n A i , i x min s , s V ρ , h n i , i p + 2 h n s V ρ , h n i , i p h n s V ρ , h n i , i p + 2 h n s V ρ , h n i , i p h n s d s d s = A i , i x 0 p + 2 h n 0 p + 2 h n min s , s δ p + 2 h n s δ p h n s δ p + 2 h n s δ p h n s d s d s = A i , i x p + 2 h n 2 p h n + p h n = 2 h n A i , i x ,
and if ρ i 0 , ρ ¯ ,
0 p + 2 h n 0 p + 2 h n A i , i x min s , s V ρ , h n i , i p + 2 h n s V ρ , h n i , i p h n s V ρ , h n i , i p + 2 h n s V ρ , h n i , i p h n s d s d s = A i , i x ρ i h n 2 0 p + 2 h n 0 p + 2 h n min s , s × 1 0 , ρ i h n p + 2 h n s 1 0 , ρ i h n p h n s × 1 0 , ρ i h n p + 2 h n s 1 0 , ρ i h n p h n s d s d s = A i , i x ρ i h n 2 p + 2 ρ i h n p + 2 h n p + 2 ρ i h n p + 2 h n min s , s d s d s 2 A i , i x ρ i h n 2 p + 2 ρ i h n p + 2 h n p ρ i h n p h n min s , s d s d s + A i , i x ρ i h n 2 p ρ i h n p h n p ρ i h n p h n min s , s d s d s = A i , i x ρ i h n 2 p + 2 ρ i h n p + 2 h n s p + 2 h n s d s + p + 2 ρ i h n s s d s d s 2 A i , i x ρ i h n 2 1 2 , ρ ¯ ρ i p h n p + 2 h n p ρ i h n p h n s d s d s 2 A i , i x ρ i h n 2 1 2 , ρ ¯ ρ i p + 2 ρ i h n p h n s p h n s d s + p ρ i h n s s d s d s 2 A i , i x ρ i h n 2 1 0 , 2 ρ i p + 2 ρ i h n p + 2 h n p ρ i h n p h n s d s d s + A i , i x ρ i h n 2 p ρ i h n p h n s p h n s d s + p ρ i h n s s d s d s = 2 h n A i , i x 1 ρ i 6 if ρ i 0 , 2 , 2 h n A i , i x 2 ρ i 4 3 ρ i 2 if ρ i 2 , ρ ¯ .
Hence, we obtain the proof (for details, see the supplementary material). □
Proof of Lemma 2.
Continuity is obvious, and monotonicity is obtained as follows: if ρ i 0 , 1 ,
d d ρ i 6 2 ρ i 6 ρ i 1 = 2 6 ρ i 6 2 ρ i 1 6 ρ i 2 = 12 6 ρ i 2 < 0 ,
and if ρ i 1 , 2 ,
d d ρ i 6 ρ i 2 6 ρ i 2 ρ i 3 1 = 6 6 ρ i 2 ρ i 3 6 ρ i 2 12 ρ i 3 ρ i 2 6 ρ i 2 ρ i 3 2 = 6 ρ i 7 ρ i + 4 + 2 ρ i 2 6 ρ i 2 ρ i 3 2 < 0 ,
and if ρ i 2 , ρ ¯ ,
d d ρ i 3 ρ i 1 6 ρ i 4 1 = 18 6 ρ i 4 2 < 0 .
The inverse can be obtained directly. □
Proof of Theorem 1.
It follows from Lemma 1, 2 and continuous mapping theorem. □

8.3. Proofs of the Results in Section 3.2

Proof of Theorem 2.
We can clearly prove the result by using Lemma 7 in Kessler [4], Proposition 7 in Nakakita and Uchida [28], and Slutsky’s theorem. □
Proof of Theorem 3.
By Lemma 1, there exists a number < 0 such that
1 n h n k = 1 n X ¯ k h n , n i X ¯ k 1 h n , n i 2 1 n h n 2 2 k n X ¯ 2 k h n , n i X ¯ 2 k 2 h n , n i 2 P < 0 ,
and hence it is sufficient to show that
sup n N E 2 3 n h n 2 k = 1 n X ¯ k h n , n i X ¯ k 1 h n , n i 4 < ;
and it is obvious that
X ¯ k h n , n i X ¯ k 1 h n , n i 4 C X ¯ k h n , n i X k ρ ¯ 1 h n 4 + C X ¯ k 1 h n , n i X k ρ ¯ 1 h n 4
and
E X ¯ k h n , n i X k ρ ¯ 1 h n 4 F k ρ ¯ 1 h n = E 1 ρ i h n k ρ i h n k h n X s i X k ρ ¯ 1 h n d s 4 F k ρ ¯ 1 h n E 1 ρ i h n k ρ i h n k h n X s i X k ρ ¯ 1 h n d s 4 F k ρ ¯ 1 h n E 1 ρ i h n k ρ i h n k h n sup s k ρ i h n , k h n X s i X k ρ ¯ 1 h n d s 4 F k ρ ¯ 1 h n = E sup s k ρ i h n , k h n X s i X k ρ ¯ 1 h n 4 F k ρ ¯ 1 h n C h n 2 1 + X k ρ ¯ 1 h n C
by Proposition A in Gloter [9], and a parallel result holds for X ¯ k 1 h n , n i X k ρ ¯ 1 h n 4 . Hence we obtain the result. □

8.4. Proof of the Results in Section 4

Proof of Theorem 4.
We only deal with the case where ρ is unknown because the discussion for the case where ρ is known is parallel. First of all, we prove the consistency of α ^ n . We obtain that
1 n H 1 , n α | ρ ^ n 1 n H 1 , n α | ρ = 1 n k = 1 n 1 h n X ¯ k h n , n X ¯ k 1 h n , n 2 G X ¯ k 1 h n , n , α | ρ ^ n 2 + 1 n k = 1 n 1 h n X ¯ k h n , n X ¯ k 1 h n , n 2 G X ¯ k 1 h n , n , α | ρ 2 2 n h n k = 1 n G X ¯ k 1 h n , n , α | ρ ^ n G X ¯ k 1 h n , n , α | ρ X ¯ k h n , n X ¯ k 1 h n , n 2 + 1 n k = 1 n G X ¯ k 1 h n , n , α | ρ ^ n 2 G X ¯ k 1 h n , n , α | ρ 2 2 n h n k = 1 n X ¯ k h n , n X ¯ k 1 h n , n 2 × i = 1 d j = 1 d A i , j X ¯ k 1 h n , n , α f G ρ ^ n i , ρ ^ n j f G ρ i , ρ j + 1 n k = 1 n i = 1 d j = 1 d A i , j X ¯ k 1 h n , n , α 2 f G 2 ρ ^ n i , ρ ^ n j f G 2 ρ i , ρ j C n h n k = 1 n 1 + X ¯ k 1 h n , n C X ¯ k h n , n X ¯ k 1 h n , n 2 × i = 1 d j = 1 d f G ρ ^ n i , ρ ^ n j f G ρ i , ρ j + C n k = 1 n 1 + X ¯ k 1 h n , n C i = 1 d j = 1 d f G 2 ρ ^ n i , ρ ^ n j f G 2 ρ i , ρ j P 0 uniformly in α ,
because continuous mapping theorem holds. Therefore, it follows from Proposition 1 and Proposition 3 that
1 n H 1 , n α | ρ ^ n 1 n H 1 , n α | ρ = 2 n h n k = 1 n G X ¯ k 1 h n , n , α | ρ X ¯ k h n , n X ¯ k 1 h n , n 2 1 n k = 1 n G X ¯ k 1 h n , n , α | ρ 2 2 n h n k = 1 n G X ¯ k 1 h n , n , α | ρ X ¯ k h n , n X ¯ k 1 h n , n 2 + 1 n k = 1 n G X ¯ k 1 h n , n , α | ρ 2 + o P 1 P V 1 α | ξ uniformly in α
where o P 1 indicates the term converging in probability to zero uniformly in θ . Then we obtain that α ^ n α in the same way as Kessler [4] with Assumption [A3].
In the next place, we consider the consistency of β ^ n . Firstly, we consider the case max i ρ i 1 , for an integer 1 , , ρ ¯ + 1 . Then it is sufficient to show
1 n h n H 2 , n β | ρ ^ n 1 n h n H 2 , n β | ρ P V 2 β | ξ uniformly in β
due to Assumption [A3]. Because the evaluation D j x = O where j max i = 1 , , n ρ i + 1 using independent increments of the Wiener process, Proposition 1 and Proposition 2 verify
F β 1 n h n H 2 , n β | ρ P V 2 β | ξ uniformly in β ,
where
F j β : = 1 n h n 2 k = 1 + j n X ¯ k h n , n X ¯ k 1 h n , n h n b X ¯ k 1 j h n , n , β 2 .
In addition, the exact convergences such that
P 1 max i ρ ^ n i + 1 = 1 1 , P 1 j max i ρ ^ n i + 1 = 1 0
hold for all j , since for all j = 1 , , ρ ¯ + 1 ,
P 1 j max i ρ ^ n i + 1 = 1 = P max i ρ ^ n i j 1 , j .
Therefore, for any ϵ > 0 ,
P sup β Θ 2 1 n h n H 2 , n β | ρ ^ n 1 n h n H 2 , n β | ρ V 2 β | ξ > ϵ j P 1 j max i ρ ^ n i + 1 = 1 + P 1 max i ρ ^ n i + 1 = 1 sup β Θ 2 F β 1 n h n H 2 , n β | ρ V 2 β | ξ > ϵ 0 .
For the case max i ρ i = for an integer = 0 , , ρ ¯ + 1 , we similarly obtain
1 n h n H 2 , n β | ρ ^ n 1 n h n H 2 , n β | ρ P V 2 β | ξ uniformly in β
because we have
F β 1 n h n H 2 , n β | ρ P V 2 β | ξ uniformly in β , F + 1 β 1 n h n H 2 , n β | ρ P V 2 β | ξ uniformly in β ,
and
P 1 max i ρ ^ n i + 1 + 1 + 1 max i ρ ^ n i + 1 = 1 1 , P 1 j max i ρ ^ n i + 1 = 0 1 , for all j , + 1 ,
and it holds that for any ϵ > 0 ,
P sup β Θ 2 1 n h n H 2 , n β | ρ ^ n 1 n h n H 2 , n β | ρ V 2 β | ξ > ϵ j , + 1 P 1 j max i ρ ^ n i + 1 = 1 + P 1 max i ρ ^ n i + 1 = 1 sup β Θ 2 F β 1 n h n H 2 , n β | ρ V 2 β | ξ > ϵ + P 1 + 1 max i ρ ^ n i + 1 = 1 sup β Θ 2 F + 1 β 1 n h n H 2 , n β | ρ V 2 β | ξ > ϵ 0 .
Hence it is shown that β ^ n P β with Assumption [A3]. □

Supplementary Materials

The following are available online at https://www.mdpi.com/1099-4300/22/9/1031/s1.

Author Contributions

Conceptualization, S.H.N. and M.U.; Methodology, S.H.N. and M.U.; Software, S.H.N.; Validation, S.N.; Formal Analysis, S.H.N.; Investigation, S.H.N.; Resources, S.H.N. and M.U.; Data Curation, S.H.N.; Writing—Original Draft Preparation, S.H.N.; Writing—Review & Editing, M.U.; Visualization, S.H.N.; Supervision, M.U.; Project Administration, M.U.; Funding Acquisition, S.H.N. and M.U. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially funded by JST CREST, JSPS KAKENHI Grant Number JP17H01100 and JP20J10058.

Acknowledgments

This work was partially supported by Cooperative Research Program of the Institute of Statistical Mathematics.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Florens-Zmirou, D. Approximate discrete time schemes for statistics of diffusion processes. Statistics 1989, 20, 547–557. [Google Scholar] [CrossRef]
  2. Yoshida, N. Estimation for diffusion processes from discrete observation. J. Multivar. Anal. 1992, 41, 220–242. [Google Scholar] [CrossRef] [Green Version]
  3. Bibby, B.M.; Sørensen, M. Martingale estimating functions for discretely observed diffusion processes. Bernoulli 1995, 1, 17–39. [Google Scholar] [CrossRef]
  4. Kessler, M. Estimation of an ergodic diffusion from discrete observations. Scand. J. Stat. 1997, 24, 211–229. [Google Scholar] [CrossRef]
  5. Kessler, M.; Sørensen, M. Estimating equations based on eigenfunctions for a discretely observed diffusion process. Bernoulli 1999, 5, 299–314. [Google Scholar] [CrossRef]
  6. Yoshida, N. Polynomial type large deviation inequalities and quasi-likelihood analysis for stochastic differential equations. Ann. Inst. Stat. Math. 2011, 63, 431–479. [Google Scholar] [CrossRef]
  7. Uchida, M.; Yoshida, N. Adaptive estimation of an ergodic diffusion process based on sampled data. Stoch. Process. Their Appl. 2012, 122, 2885–2924. [Google Scholar] [CrossRef] [Green Version]
  8. Uchida, M.; Yoshida, N. Adaptive Bayes type estimators of ergodic diffusion processes from discrete observations. Stat. Inference Stoch. Process. 2014, 17, 181–219. [Google Scholar] [CrossRef]
  9. Gloter, A. Discrete sampling of an integrated diffusion process and parameter estimation of the diffusion coefficient. ESAIM Probab. Stat. 2000, 4, 205–227. [Google Scholar] [CrossRef]
  10. Ditlevsen, S.; Sørensen, M. Inference for observations of integrated diffusion processes. Scand. J. Stat. 2004, 31, 417–429. [Google Scholar] [CrossRef]
  11. Gloter, A. Parameter estimation for a discretely observed integrated diffusion process. Scand. J. Stat. 2006, 33, 83–104. [Google Scholar] [CrossRef]
  12. Gloter, A.; Gobet, E. LAMN property for hidden processes: The case of integrated diffusions. Ann. l’Institut Henri Poincaré Probab. Stat. 2008, 44, 104–128. [Google Scholar] [CrossRef]
  13. Sørensen, M. Prediction-based estimating functions: Review and new developments. Braz. J. Probab. Stat. 2011, 25, 362–391. [Google Scholar] [CrossRef]
  14. Donnet, S.; Samson, A. A review on estimation of stochastic differential equations for pharmacokinetic/pharmacodynamic models. Adv. Drug Deliv. Rev. 2013, 65, 929–939. [Google Scholar] [CrossRef] [PubMed]
  15. Delattre, M.; Genon-Catalot, V.; Samson, A. Mixtures of stochastic differential equations with random effects: Application to data clustering. J. Stat. Plan. Inference 2016, 173, 109–124. [Google Scholar] [CrossRef]
  16. Picchini, U.; Forman, J.L. Bayesian inference for stochastic differential equation mixed effects models of a tumour xenography study. J. R. Stat. Soc. Ser. C Appl. Stat. 2019, 68, 887–913. [Google Scholar] [CrossRef]
  17. Ruse, M.G.; Samson, A.; Ditlevsen, S. Inference for biomedical data by using diffusion models with covariates and mixed effects. J. R. Stat. Soc. Ser. C Appl. Stat. 2020, 69, 167–193. [Google Scholar] [CrossRef]
  18. Zhang, L.; Mykland, P.A.; Aït-Sahalia, Y. A Tale of Two Time Scales: Determining Integrated Volatility with Noisy High-Frequency Data. J. Am. Stat. Assoc. 2005, 100, 1394–1411. [Google Scholar] [CrossRef]
  19. Aït-Sahalia, Y.; Jacod, J. High-Frequency Financial Econometrics; Princeton University Press: Princeton, NJ, USA, 2014. [Google Scholar]
  20. BNCI Horizon 2020. Two Class Motor Imagery. 2014. Available online: http://bnci-horizon-2020.eu/database/data-sets (accessed on 20 April 2019).
  21. Jacod, J.; Li, Y.; Mykland, P.A.; Podolskij, M.; Vetter, M. Microstructure noise in the continuous case: The pre-averaging approach. Stoch. Process. Their Appl. 2009, 119, 2249–2276. [Google Scholar] [CrossRef] [Green Version]
  22. Jacod, J.; Podolskij, M.; Vetter, M. Limit theorems for moving averages of discretized processes plus noise. Ann. Stat. 2010, 38, 1478–1545. [Google Scholar] [CrossRef]
  23. Bibinger, M.; Hautsch, N.; Malec, P.; Reiß, M. Estimating the quadratic covariation matrix from noisy observations: Local method of moments and efficiency. Ann. Stat. 2014, 42, 1312–1346. [Google Scholar] [CrossRef]
  24. Koike, Y. Quadratic covariation estimation of an irregularly observed semimartingale with jumps and noise. Bernoulli 2016, 22, 1894–1936. [Google Scholar] [CrossRef] [Green Version]
  25. Ogihara, T. Parametric inference for nonsynchronously observed diffusion processes in the presence of market microstructure noise. Bernoulli 2018, 24, 3318–3383. [Google Scholar] [CrossRef] [Green Version]
  26. Iacus, S.M. Simulation and Inference for Stochastic Differential Equations: With R Examples; Springer: New York, NY, USA, 2008. [Google Scholar]
  27. Steyrl, D.; Scherer, R.; Faller, J.; Müller-Putz, G.R. Random forests in non-invasive sensorimotor rhythm brain-computer interfaces: A practical and convenient non-linear classifier. Biomed. Tech. 2016, 61, 77–86. [Google Scholar] [CrossRef]
  28. Nakakita, S.H.; Uchida, M. Inference for ergodic diffusions plus noise. Scand. J. Stat. 2019, 46, 470–516. [Google Scholar] [CrossRef] [Green Version]
  29. Kitagawa, H.; Uchida, M. Adaptive test statistics for ergodic diffusion processes sampled at discrete times. J. Stat. Plan. Inference 2014, 150, 84–110. [Google Scholar] [CrossRef]
  30. Nakakita, S.H.; Uchida, M. Adaptive test for ergodic diffusions plus noise. J. Stat. Plan. Inference 2019, 203, 131–150. [Google Scholar] [CrossRef] [Green Version]
  31. Ogihara, T.; Yoshida, N. Quasi-likelihood analysis for the stochastic differential equation with jumps. Stat. Inference Stoch. Process. 2011, 14, 189–229. [Google Scholar] [CrossRef]
  32. Clinet, S.; Yoshida, N. Statistical inference for ergodic point processes and application to Limit Order Book. Stoch. Process. Their Appl. 2017, 127, 1800–1839. [Google Scholar] [CrossRef] [Green Version]
  33. Nakakita, S.H.; Uchida, M. Quasi-likelihood analysis of an ergodic diffusion plus noise. arXiv 2018, arXiv:1806.09401. [Google Scholar]
Figure 1. The path of the second column of S02E.mat of BNCI Horizon 2020 [20] for all 222 seconds (left) and the first one second (right).
Figure 1. The path of the second column of S02E.mat of BNCI Horizon 2020 [20] for all 222 seconds (left) and the first one second (right).
Entropy 22 01031 g001
Figure 2. Realised volatilities with subsampling of the 2nd axis of data S02E.mat in two class motor imagery (002-2014) [20].
Figure 2. Realised volatilities with subsampling of the 2nd axis of data S02E.mat in two class motor imagery (002-2014) [20].
Entropy 22 01031 g002
Figure 3. The left figure is the plot of the latent diffusion process, and the right one is that of the convolutionally observed process on 0 , 1 respectively.
Figure 3. The left figure is the plot of the latent diffusion process, and the right one is that of the convolutionally observed process on 0 , 1 respectively.
Entropy 22 01031 g003
Figure 4. The realised volatilities of the convolutionally observed diffusion process with subsampling.
Figure 4. The realised volatilities of the convolutionally observed diffusion process with subsampling.
Entropy 22 01031 g004
Table 1. Estimation performance of ρ with small ρ .
Table 1. Estimation performance of ρ with small ρ .
ρ 0.00.10.20.30.40.50.60.70.80.91.0
mean 0.00990 0.0971 0.198 0.298 0.398 0.498 0.598 0.699 0.799 0.899 0.999
RMSE ( 0.0182 ) ( 0.0256 ) ( 0.0235 ) ( 0.0215 ) ( 0.0197 ) ( 0.0180 ) ( 0.0164 ) ( 0.0150 ) ( 0.0135 ) ( 0.0123 ) ( 0.0110 )
Table 2. Empirical ratio of test statistic T n less than some critical values, and the maximum value of T n in 1000 iterations.
Table 2. Empirical ratio of test statistic T n less than some critical values, and the maximum value of T n in 1000 iterations.
Empirical Ratio of T n Less Than…Max. of T n
Φ 1 0.10 Φ 1 0.05 Φ 1 0.025 Φ 1 0.01 Φ 1 0.001
ρ = 0.0 0.101 0.053 0.025 0.005 0.000 3.060
ρ = 0.1 0.989 0.980 0.966 0.914 0.759 0.710
ρ = 0.2 1.000 1.000 1.000 1.000 1.000 4.593
ρ = 0.3 1.000 1.000 1.000 1.000 1.000 9.341
ρ = 0.4 1.000 1.000 1.000 1.000 1.000 13.985
ρ = 0.5 1.000 1.000 1.000 1.000 1.000 19.152
ρ = 0.6 1.000 1.000 1.000 1.000 1.000 24.816
ρ = 0.7 1.000 1.000 1.000 1.000 1.000 30.848
ρ = 0.8 1.000 1.000 1.000 1.000 1.000 37.557
ρ = 0.9 1.000 1.000 1.000 1.000 1.000 44.829
ρ = 1.0 1.000 1.000 1.000 1.000 1.000 52.759
Table 3. Estimation of θ by the proposed method and LGA with small ρ .
Table 3. Estimation of θ by the proposed method and LGA with small ρ .
The Proposed MethodLGA
α β 1 β 2 α β 1 β 2
True Value 3.0 2.0 1.0 3.0 2.0 1.0
ρ = 0.0 mean 3.004 2.091 1.036 2.999 2.095 1.037
RMSE ( 0.0109 ) ( 0.318 ) ( 0.497 ) ( 0.00679 ) ( 0.320 ) ( 0.497 )
ρ = 0.1 mean 2.999 2.091 1.035 2.949 2.026 1.003
RMSE ( 0.0146 ) ( 0.319 ) ( 0.496 ) ( 0.0509 ) ( 0.297 ) ( 0.480 )
ρ = 0.2 mean 2.998 2.091 1.035 2.898 1.955 0.967
RMSE ( 0.0142 ) ( 0.319 ) ( 0.496 ) ( 0.102 ) ( 0.290 ) ( 0.464 )
ρ = 0.3 mean 2.998 2.092 1.036 2.846 1.885 0.932
RMSE ( 0.0139 ) ( 0.319 ) ( 0.497 ) ( 0.155 ) ( 0.299 ) ( 0.452 )
ρ = 0.4 mean 2.998 2.091 1.036 2.792 1.815 0.897
RMSE ( 0.0135 ) ( 0.319 ) ( 0.497 ) ( 0.208 ) ( 0.324 ) ( 0.442 )
ρ = 0.5 mean 2.998 2.092 1.036 2.738 1.744 0.862
RMSE ( 0.0132 ) ( 0.319 ) ( 0.497 ) ( 0.262 ) ( 0.361 ) ( 0.436 )
ρ = 0.6 mean 2.998 2.091 1.036 2.683 1.674 0.827
RMSE ( 0.0129 ) ( 0.319 ) ( 0.497 ) ( 0.317 ) ( 0.408 ) ( 0.434 )
ρ = 0.7 mean 2.998 2.092 1.036 2.626 1.604 0.792
RMSE ( 0.0126 ) ( 0.319 ) ( 0.497 ) ( 0.374 ) ( 0.460 ) ( 0.434 )
ρ = 0.8 mean 2.998 2.092 1.036 2.568 1.534 0.757
RMSE ( 0.0124 ) ( 0.319 ) ( 0.496 ) ( 0.432 ) ( 0.517 ) ( 0.439 )
ρ = 0.9 mean 2.998 2.092 1.036 2.509 1.464 0.722
RMSE ( 0.0121 ) ( 0.319 ) ( 0.497 ) ( 0.491 ) ( 0.578 ) ( 0.445 )
ρ = 1.0 mean 2.998 2.091 1.036 2.449 1.394 0.687
RMSE ( 0.0119 ) ( 0.319 ) ( 0.497 ) ( 0.551 ) ( 0.640 ) ( 0.456 )
Table 4. The performance of ρ ^ n for ρ = 10 , 15 , 20 in 1000 iterations.
Table 4. The performance of ρ ^ n for ρ = 10 , 15 , 20 in 1000 iterations.
ρ = 10 ρ = 15 ρ = 20
mean of ρ ^ n 9.919 14.980 19.751
RMSE of ρ ^ n ( 0.145 ) ( 0.240 ) ( 0.409 )
Table 5. Estimation of θ by the proposed method with large ρ .
Table 5. Estimation of θ by the proposed method with large ρ .
The Proposed MethodLGA
α β 1 β 2 α β 1 β 2
True Value 3 . 0 2.0 1.0 3.0 2.0 1.0
ρ = 10 mean 2.989 2.101 1.030 0.933 0.204 0.0811
RMSE ( 0.0347 ) ( 0.323 ) ( 0.496 ) ( 2.067 ) ( 1.796 ) ( 0.920 )
ρ = 15 mean 2.996 2.095 1.027 0.765 0.138 0.0473
RMSE ( 0.0475 ) ( 0.321 ) ( 0.495 ) ( 2.235 ) ( 1.862 ) ( 0.953 )
ρ = 20 mean 2.977 2.090 1.024 0.664 0.104 0.0302
RMSE ( 0.0526 ) ( 0.319 ) ( 0.493 ) ( 2.336 ) ( 1.896 ) ( 0.970 )
Table 6. Summary for ρ estimate.
Table 6. Summary for ρ estimate.
ρ 1 ρ 2
true value 2.0 4.0
empirical mean 1.988 3.966
empirical RMSE ( 0.0207 ) ( 0.0514 )
Table 7. Summary for α estimate.
Table 7. Summary for α estimate.
α 1 α 2 α 3
True Value 2.0 0.0 3.0
Our proposalmean 1.993 0.000256 2.992
RMSE ( 0.0115 ) ( 0.00739 ) ( 0.0213 )
LGAmean 1.295 0.00320 1.442
RMSE ( 0.705 ) ( 0.0154 ) ( 1.558 )
Table 8. Summary for β estimate.
Table 8. Summary for β estimate.
β 1 β 2 β 3 β 4 β 5 β 6
True Value 2.0 0.4 0.0 0.1 3.0 5.0
Our proposalmean 2.137 0.408 0.0439 0.0788 3.103 5.091
RMSE ( 0.362 ) ( 0.252 ) ( 0.540 ) ( 0.473 ) ( 0.399 ) ( 0.777 )
LGAmean 0.917 0.340 0.326 0.696 0.221 1.243
RMSE ( 1.093 ) ( 0.802 ) ( 0.386 ) ( 0.804 ) ( 3.242 ) ( 3.765 )
Table 9. The values of ρ ^ n and T n for the first 15 axes of S02.mat by BNCI Horizon 2020 [20].
Table 9. The values of ρ ^ n and T n for the first 15 axes of S02.mat by BNCI Horizon 2020 [20].
1st axis2nd axis3rd axis4th axis5th axis
ρ ^ n 0.449 1.037 0.894 0.736 0.937
T n 20.398 58.631 46.649 35.201 49.741
6th axis7th axis8th axis9th axis10th axis
ρ ^ n 0.951 0.971 1.017 0.958 0.967
T n 51.392 52.607 55.455 51.221 51.797
11th axis12th axis13th axis14th axis15th axis
ρ ^ n 0.949 0.649 0.952 0.977 0.932
T n 50.457 30.094 50.633 50.978 48.842

Share and Cite

MDPI and ACS Style

Nakakita, S.H.; Uchida, M. Inference for Convolutionally Observed Diffusion Processes. Entropy 2020, 22, 1031. https://doi.org/10.3390/e22091031

AMA Style

Nakakita SH, Uchida M. Inference for Convolutionally Observed Diffusion Processes. Entropy. 2020; 22(9):1031. https://doi.org/10.3390/e22091031

Chicago/Turabian Style

Nakakita, Shogo H, and Masayuki Uchida. 2020. "Inference for Convolutionally Observed Diffusion Processes" Entropy 22, no. 9: 1031. https://doi.org/10.3390/e22091031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop