Next Article in Journal
Compressed Sensing Techniques Applied to Medical Images Obtained with Magnetic Resonance
Previous Article in Journal
Framed Natural Mates of Framed Curves in Euclidean 3-Space
Previous Article in Special Issue
Optimal Designs for Antoine’s Equation: Compound Criteria and Multi-Objective Designs via Genetic Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Nested Maximum Entropy Designs for Computer Experiments

1
School of Science, Beijing University of Civil Engineering and Architecture, Beijing 100044, China
2
NCMIS, KLSC, Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing 100190, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(16), 3572; https://doi.org/10.3390/math11163572
Submission received: 16 July 2023 / Revised: 3 August 2023 / Accepted: 9 August 2023 / Published: 18 August 2023
(This article belongs to the Special Issue Optimal Experimental Design and Statistical Modeling)

Abstract

:
Presently, computer experiments with multiple levels of accuracy are widely applied in science and engineering. This paper introduces a class of nested maximum entropy designs for such computer experiments. A multi-layer DETMAX algorithm is proposed to construct nested maximum entropy designs. Based on nested maximum entropy designs, we also propose an integer-programming procedure to specify the sample sizes in multi-fidelity computer experiments. Simulated annealing techniques are used to tackle complex optimization problems in the proposed methods. Illustrative examples show that the proposed nested entropy designs can yield better prediction results than nested Latin hypercube designs in the literature and that the proposed sample-size determination method is effective.

1. Introduction

With the rapid development of computer simulation technology, computer experiments have been widely used in the manufacturing industry, system engineering, natural science, and other fields [1,2]. Statistical designs for computer experiments have received considerable attention [3,4,5,6]. In many real applications, multi-fidelity computer experiments with different levels of accuracy are often encountered. More accurate experiments need longer computational time, while faster experiments have relatively low accuracy. It seems inefficient to study them separately. Therefore, many authors have studied statistical analysis for integrating multi-fidelity computer experiments with different levels of accuracy [7,8,9,10]. The experimental design issue for such computer experiments has been investigated by [11,12,13,14,15], and many others.
Shannon entropy is a basic concept of information theory [16]. Ref. [17] introduced the use of Shannon entropy as a measure of experimental information for spatial design. He argued that the experiment that minimizes the expected entropy, which is the entropy of the posterior distribution, can provide the largest amount of information for prediction. Ref. [18] proved that minimizing the posterior entropy is equivalent to maximizing the entropy of the prior distribution. The maximum entropy criterion was subsequently adopted as one of the major approaches for computer experiments [1]. Ref. [19] proposed the DETMAX algorithm [20] to efficiently construct the maximum entropy design. Ref. [21] proposed a sequential framework for conducting computer experiments with the maximum entropy criterion. However, to the best of our knowledge, there is no research on the maximum entropy design for multi-fidelity computer experiments.
In this paper, we introduce a class of nested maximum entropy designs with multi-layer structures for multi-fidelity computer experiments. Unlike [14]’s nested Latin hypercube designs, a nested maximum entropy design allows for flexibility in sample sizes as the sample size in each larger design does not need to be a multiple of that in a smaller one. Since computer experiments with higher accuracy are more important, we first consider the optimization of lower layers in the nested maximum entropy designs. Based on a layer-by-layer optimization strategy [11,22], a multi-layer DETMAX algorithm is proposed to construct such nested maximum entropy designs. The algorithm begins to generate a maximum entropy design for the lowest layer. Subsequently, we fix the design points optimized in lower layers, and optimize the current layer according to the maximum entropy criterion step by step, until the whole design is completely optimized. Based on nested maximum entropy designs, we also propose an integer-programming procedure to specify the sample sizes in multi-accuracy computer experiments under the budget constraint. Simulated annealing techniques [23] are adopted to tackle complex optimization problems in the proposed approaches. Illustrative examples are presented to show the effectiveness of our methods.
The contributions of this paper are summarized as follows. First, we introduce a new type of model-based design for multi-fidelity computer experiments based on information entropy. Second, our methods are flexible in sample sizes of multi-fidelity computer experiments. Third, this paper first provides an entropy-based strategy to determine sample sizes of multi-fidelity computer experiments.
The rest of this paper is organized as follows. Section 2 reviews the concept of maximum entropy designs for a single level of accuracy. In Section 3, the DETMAX algorithm is extended to construct nested maximum entropy designs. Section 4 deals with the sample-size determination of multi-accuracy computer experiments. Section 5 provides numerical examples. We end this paper with some concluding remarks in Section 6.

2. Review of Maximum Entropy Designs

In this section, we give a review of maximum entropy designs. Consider the following Kriging model,
y ( x ) = f ( x ) β + Z ( x ) ,
where x R p , f ( x ) = ( f 1 ( x ) , , f m ( x ) ) is a prespecified vector of regression functions, β = ( β 1 , , β m ) presents a vector of unknown regression coefficients, Z ( x ) is a stationary Gaussian process with zero mean, variance σ 2 , and the correlation function
R ( x 1 x 2 | ϕ ) = exp i = 1 p ϕ i ( x 1 i x 2 i ) 2
for x 1 = ( x 11 , , x 1 p ) and x 2 = ( x 21 , , x 2 p ) and a vector of correlation parameters ϕ = ( ϕ 1 , , ϕ p ) with ϕ i > 0 for i = 1 , , p , denoted by Z GP ( 0 , σ 2 , ϕ ) .
Let D = { x 1 , , x n } and y = ( y ( x 1 ) , , y ( x n ) ) represent a design with n runs and the corresponding vector of response values, respectively. Ref. [18] used the expected change in information to evaluate the design D. Since entropy is the negative of information, maximizing the expected change in information is equivalent to maximizing the entropy of the responses at the points in the design, denoted by H ( Y D ) . In the context of Gaussian process models, the design relevant part of H ( Y D ) is log ( det ( σ 2 R ) ) / 2 , where R is the n × n correlation matrix whose ( i , j ) th entry is R ( x i x j | ϕ ) . Therefore, a maximum entropy design D maximizes the determinant of the covariance matrix of the set of responses y at the points in the design [2],
max D det ( σ 2 R ) .
Because of the independence between σ 2 and the design D, (2) is equivalent to
max D det ( R ) .
Here the vector of correlation parameters, ϕ , in R are assumed to be known.

3. Construction of Nested Maximum Entropy Designs

In this section, we extend the maximum entropy design to the case of multiple layers and propose the corresponding construction algorithms.

3.1. Nested Maximum Entropy Designs

Nested designs with multiple layers are usually used for multi-fidelity computer experiments [11,12,22,24,25]. Assume that we have a computer experiment with K levels, and the accuracy declines gradually from level 1 to level K. For each k = 1 , , K , the Kriging model for computer experiments at the kth level of accuracy is
y k ( x i ) = f ( x i ) β k + Z k ( x i ) , i = 1 , , n k ,
where f ( x i ) and β k are straightforward extensions to those in (1), and Z k GP ( 0 , σ k 2 , ϕ k ) . R k is the n k × n k matrix whose ( i , j ) th entry is R ( x i x j | ϕ k ) with the known vector ϕ k .
Let a nested design D ( K ) = { x 1 , , x n K } with K layers D ( 1 ) D ( K ) , where D ( k ) = { x 1 , , x n k } denotes the kth layer of the nested design for each k = 1 , , K and n 1 < < n K . The vector s = ( n 1 , , n K ) represents the structure of D ( K ) . Please note that D ( k ) with smaller k is used for computer experiments with higher accuracy, which are more important. Similar to some definitions of optimal nested designs [11,22], we call D ( K ) = { x 1 , , x n K } a nested maximum entropy design if the following conditions hold: the first layer D ( 1 ) is a maximum entropy design that maximizes det ( R 1 ) ; for each k = 2 , , K , D ( k ) = { x 1 , , x n k } maximizes det ( R k ) with fixed optimized D ( k 1 ) .
By the above definition, a nested maximum entropy design can be constructed by a sequential algorithm; see Algorithm 1.
Algorithm 1 Construction of a nested maximum entropy design with K layers
  • Initialization:
    k = 1 , randomly construct the first layer D ( 1 ) . Optimize D ( 1 ) using the maximum criteria to obtain D best ( 1 ) .
    Recursive step:
  • for  k = 2 , , K  do
  •    Enlarge D best ( k 1 ) to D ( k ) , where D ( k ) is the kth layer of the design with the structure s . Maximize the entropy det ( R k ) corresponding to D ( k ) by optimizing D ( k ) D best ( k 1 ) .
  • Output D best ( k ) .
  • end for

3.2. Multi-Layer DETMAX Algorithm

This subsection presents optimization algorithms for constructing each layer of a nested maximum entropy design.
The maximum entropy design can be obtained by a DETMAX-based algorithm [19]. It is optimized through a series of “excursions” to improve the det( R ) corresponding to the current design by adding or removing appropriate points from the current design until the det( R ) for the resulting n-point design cannot be increased. Except for the initial and final designs which have exactly n points, the number of chosen points of any designs constructed on this excursion can be greater or less than n. We extend this algorithm to the multi-layer case.
The flow chart in Figure 1 describes the procedure of the multi-layer DETMAX algorithm. The layer-by-layer optimization strategy [11,22] is adopted here. First, the initial design of the first layer, denoted as D 0 ( 1 ) , is randomly generated and then optimized. If D best ( 1 ) obtained through a series of excursions meets the stopping condition, then the first layer has been completed. Subsequently, optimize the second layer with the first layer fixed. Each layer is optimized in turn until the last layer is optimized, and then output D best ( K ) .
We give the details for optimizing each layer in the above algorithm. Suppose we now optimize the kth layer of the design D ( k ) . One excursion starts with the n k -point design and ends when the number of points in D ( k ) reaches exactly n k again. The procedure for making excursions is described as follows. Let ϵ denote a prespecified threshold.
Step 1. Add a point at which the variance function σ 0 | D ( k ) 2 is largest, or subtract a point corresponding to the maximum element of the diagonal of R k 1 .
Step 2. The current design D ( k ) has n k points.
If n k > n k , remove a point if D ( k ) is not in F k and add a point otherwise.
If n k < n k , add a point if D ( k ) is not in F k and remove a point otherwise.
The new design updated by this step has n k new points and correlation matrix R k new .
Step 3. If n k new n k , back to Step 2. Otherwise, if det ( R k new ) det ( R k ) < ϵ , stop the procedure and output D best ( k ) ; if det ( R k new ) det ( R k ) ϵ , do the following:
If det( R k new ) < det( R k ), let R k = R k new , place all the designs generated on this excursion into F k . Go to Step 1.
If det ( R k new ) det ( R k ) , let R k = R k n e w , F k = , D best ( k ) = D ( k ) and R k best = R k . Go to Step 1.
In Step 1, the selection of whether to add or subtract a point is made randomly. The best point x 0 is obtained by maximizing σ 0 | D ( k ) 2 , which is given by
σ 0 | D ( k ) 2 ( x 0 ) = σ k 2 ( 1 r k R k 1 r k ) ,
where r k = ( R ( x 0 x 1 | ϕ k ) , , R ( x 0 x n k | ϕ k ) ) . To determine the best site x 0 to add to the current design, we adopt a grid search procedure [26] for p = 2 , and the simulated annealing algorithm [23] for p 3 (see Algorithm 2). The point we subtract in Step 1 is always in D ( k ) D best ( k 1 ) .
Algorithm 2 Simulated annealing in excursions
  • Step 0: Input the starting temperature T = T 0 > 0 , the ending temperature T end > 0 , the length of Markov chain L, search step size λ , Boltzmann’s constant k 0 = 1 , reduction factor α ( 0 < α < 1 ) and an initial solution x = x ( 0 ) , which is randomly generated in [ 0 , 1 ] p .
  • while  T > T end  do
  •    for  k = 1 , , L , do
  •      Step 1:  x new = x + λ u , where u is generated by sampling random values from N p ( 0 , 1 ) .
         Step 2:
  •      if  σ 0 | D ( k ) 2 ( x new ) > σ 0 | D ( k ) 2 ( x ) , then
  •          x = x new , λ = 0.99 λ .
  •      else
  •         Randomly generate a real number r in [ 0 , 1 ] .
  •         if  r < exp σ 0 | D ( k ) 2 ( x new ) σ 0 | D ( k ) 2 ( x ) k 0 T  then
  •            x = x new , λ = 0.99 λ .
  •         else
  •           Go back to Step 1.
  •         end if
  •      end if
  •    end for
  •    Step 3: T = α T.
  • end while
  • Step 4: Output the best solution x .
Since the design points are bounded, the determinant of the corresponding covariance matrix is bounded. According to the monotone convergence theorem, our algorithm can converge to a limit after considerable iteration times. However, because of the nonconvex feature of the problem like other design construction problems [27], the limit may not be the global solution. To better approximate the global solution, the above algorithm can be conducted repeatedly with several random initial designs, and the final output is the best one among the corresponding results. Several two-dimensional nested maximum entropy designs constructed by the proposed algorithm can be seen in Figure 2.

4. Sample-Size Determination of Multi-Accuracy Computer Experiments

A related issue to experimental design is sample-size determination. Both sample-size determination and experimental design should be implemented before we obtain data. Sample-size determination should be considered earlier than experimental design, since the latter is usually conducted with given sample sizes. The problem of sample-size determination in computer experiments has attracted much attention in the literature; see [28,29,30], among others. However, these studies focused on computer experiments with one level of accuracy, and there is little work for the case of more than one level. In this section, we propose a method to determine sample sizes of multi-accuracy computer experiments based on the entropy criterion.
There is no data available when we implement sample-size determination. For multi-fidelity computer experiments, we consider the maximum entropy of possible nested designs with different sample sizes. We first introduce a concept of integrative entropy, which is an extension to (3). For a K-layer nested maximum entropy design D with structure s = ( n 1 , , n K ) , the integrative entropy of D is defined by
E n = k = 1 K w k det ( R k ) ,
where w k is a non-negative weight and R k is the correlation matrix of layer k for k = 1 , , K . The choice of weights, w k = n k 1 p , in (4) can be found in [15,27]. Please note that the integrative entropy in (4) is a function that depends on n 1 , , n K , i.e., E n = E n ( D ) = E n ( n 1 , , n K ) .
For a computer experiment with K levels of accuracy, let b k denote the cost at the kth level, k = 1 , , K . Assume that the total budget is B. We specify the sample sizes n 1 , , n K through maximizing the integrative entropy under the budget constraint, i.e., solving the optimization problem,
max E n ( n 1 , , n K ) , s . t . k = 1 K n k b k B , n 1 , , n K N ,
where N denotes the set of non-negative integers.
The above optimization problem is a nonlinear knapsack problem [31]. There are many techniques for this problem such as the branch-and-bound algorithm, dynamic programming, and the decomposition method. Please note that the objective function in (5) is very complicated. We adopt the simulated annealing algorithm to solve this integer-programming problem since the algorithm possesses the features of avoiding local optimum, high flexibility, and good convergence properties. Due to the complexity of this problem, we select multiple initial points to run Algorithm 3, and output the solution corresponding to the greatest objective function value. Let randint ( 1 , m ) denote an integer randomly chosen from { 1 , , m } , m N .
Algorithm 3 Simulated annealing in sample-size determination
  • Step 0: Initialize T = T 0 = 100 , T end = 1 , L = 100 , k 0 = 1 , α = 0.7 . Randomly generate an initial solution n = n ( 0 ) = ( n 1 , , n K ) .
  • Step 1: Repeat Step 0 until k = 1 K n k b k B is satisfied.
  • while  T > T end  do
  •    for  k = 1 , , L , do
  •      Step 2:  n s = n s + ( 1 ) t , where s = randint ( 1 , K ) and t = randint ( 1 , 2 ) . n new = ( n 1 , , n s , , n K ) .
         Step 3:
  •      if  E n ( n new ) > E n ( n ) , then
  •          n = n new .
  •      else
  •         Randomly generate a real number r in [ 0 , 1 ] .
  •         if  r < exp E n ( n new ) E n ( n ) k 0 T  then
  •            n = n new .
  •         else
  •           Go back to Step 2.
  •         end if
  •      end if
  •    end for
  •    Step 4: T= α T.
  • end while
  • Step 5: Output the best solution n .

5. Illustrations

In this section, we provide several examples to illustrate the effectiveness of the proposed methods. Examples 1, 2, and 3 demonstrate the prediction performance of the proposed nested maximum entropy designs, which are constructed by applying the algorithm in Section 3. Example 4 presents an application of the proposed sample-size determination method in Section 4.
Here we consider the case of K = 2 . In Examples 1–3, let D h = { x 1 h , , x n 1 h } and D l = { x 1 l , , x n 2 l } represent the design sets of the high-accuracy experiment (HE) with n 1 runs and the design sets of the low-accuracy experiment (LE) with n 2 runs, respectively. We denote the HE response associated with D h as y h , and the LE response associated with D l as y l . The prediction model in [32] is used, and the corresponding predictor of y h is denoted by y ^ h . The FNLHD [11] is compared with the proposed entropy design. Prediction performance is evaluated with the empirical mean squared prediction error (MSPE),
MSPE = 1 10000 i = 1 10000 [ y h ( x i ) y ^ h ( x i ) ] 2 ,
where { x 1 , , x 10000 } is generated by Latin hypercube sampling ([33]).

5.1. Example 1

In this example, we use the following function from [19],
y h ( x ) = [ 1 exp ( 1 / ( 2 x 2 ) ) ] ( 2300 x 1 3 + 1900 x 1 2 + 2092 x 1 + 60 ) / ( 100 x 1 3 + 500 x 1 2 + 4 x 1 + 20 ) , x [ 0 , 1 ] 2 ,
as the HE function. In addition, the LE function is:
y l = [ 1 + y h ( x 1 + 1 / 20 , x 2 + 1 / 20 ) + y h ( x 1 + 1 / 20 , max ( 0 , x 2 1 / 20 ) ) + y h ( x 1 1 / 20 , x 2 1 / 20 ) + y h ( x 1 1 / 20 , max ( 0 , x 2 1 / 20 ) ) ] / 4
We implement the two design methods with s = ( 16 , 25 ) and s = ( 25 , 32 ) , respectively. In our method, ϕ 1 = ( 200 , 200 ) , ϕ 2 = ( 50 , 50 ) , ϵ = 10 15 , and σ 1 2 = σ 2 2 = 1 . MSPEs over 100 repetitions are shown in Figure 3. It can be seen that the proposed method outperforms FNLHD.

5.2. Example 2

In this example, the following function from [34] is exploited,
y h ( x ) = 2 / 3 exp ( x 1 + x 2 ) x 4 sin ( x 3 ) + x 3 , x [ 0 , 1 ] 4 ,
as the HE function. In addition, the LE function is:
y l = 1.2 y h 1 .
In Example 2, we use Algorithm 2 to optimize the nested maximum entropy designs with s = ( 10 , 15 ) and s = ( 20 , 24 ) for p = 4 . Take ϕ 1 = ( 200 , 200 , 200 , 200 ) , ϕ 2 = ( 10 , 10 , 10 , 10 ) , σ 1 2 = σ 2 2 = 1 , ϵ = 10 15 , L = 100 , λ = 0.3 , k 0 = 1 , T 0 = 100 , and T end = 1 in the proposed method. Table 1 displays the MSPEs over 100 replicates for the proposed method with four values of reduction factors α , 0.99, 0.9, 0.8, 0.7 in Algorithm 2. The results from FNLHD are also shown. We can see that our designs can yield better prediction results compared with FNLHDs. Moreover, the proposed simulated annealing algorithm is insensitive to the selection of parameter α .

5.3. Example 3

In this example, let the following function from [34],
y h ( x ) = x 1 2 [ 1 + ( x 2 + x 3 2 ) x 4 / x 1 2 1 ] + ( x 1 + 3 x 4 ) × exp [ 1 + sin ( x 3 ) ] , x [ 0 , 1 ] 4 ,
act as the HE function. In addition, the LE function is:
y l = [ 1 + sin ( x 1 ) / 10 ] y h ( x 1 , x 2 , x 3 , x 4 ) 2 x 1 + x 2 2 + x 3 2 + 0.5 .
In Example 3, s = ( 10 , 25 ) and s = ( 15 , 22 ) , respectively. Algorithm 2 is used again in the proposed method for p = 4 . We set ϕ 1 = ( 200 , 200 , 200 , 200 ) , ϕ 2 = ( 10 , 10 , 10 , 10 ) , σ 1 2 = σ 2 2 = 1 , ϵ = 10 15 , L = 100 , λ = 0.3 , k 0 = 1 , T 0 = 100 , T end = 1 , and four values of α , 0.99, 0.9, 0.8, 0.7, as in Section 5.2. Table 2 presents the MSPEs over 100 replicates. Similar to the conclusions in Section 5.2, the proposed approach is better than FNLHD, and our algorithm is insensitive to the selection of parameter α .

5.4. Example 4

Consider the experiments with two levels of accuracy and four input variables. Suppose that the costs of running an HE and LE are 40 thousand RMB and 25 thousand RMB, respectively. Let the total budget be one million RMB. According to the method in Section 4, we solve the optimization problem,
max E n ( n 1 , n 2 ) ,   s . t . 40 n 1 + 25 n 2 1000 , n 1 , n 2 N ,
where E n ( n 1 , n 2 ) is defined by (4). We run Algorithm 3 with ϕ 1 = ( 200 , 200 , 200 , 200 ) , ϕ 2 = ( 10 , 10 , 10 , 10 ) , σ 1 2 = σ 2 2 = 1 , L = 100 , λ = 0.3 , k 0 = 1 , T 0 = 100 , T end = 1 , and five initial points. The optimal sample sizes are given by ( n 1 , n 2 ) = ( 10 , 24 ) , corresponding to the integrative entropy value 3.67. The corresponding nested design is shown in Table 3, where the first ten rows display the HE design.

6. Concluding Remarks

In this paper, we have introduced a new class of nested designs, nested maximum entropy designs, for multi-fidelity computer experiments. Such designs possess flexible run numbers in each layer and can provide a considerable amount of information for prediction. A multi-layer DETMAX algorithm has been proposed to construct nested maximum entropy designs. The related maximum entropy criterion has been used to determine the sample sizes for each level of accuracy in multi-fidelity computer experiments.
There are some limitations of our work. Due to the complexity of the optimization problem with the entropy criterion, the proposed algorithms can only handle relatively simple cases, such as relatively low dimensions and relatively small run sizes. In addition, extensions of the proposed approaches can be made in several directions. First, the corresponding designs for finite design regions [35] can be studied in the future. Second, our methods can be modified to accommodate both qualitative and quantitative factors [9,36,37]. Third, sequential frameworks [21,32,38,39] for multi-fidelity computer experiments can be developed by the proposed entropy criterion.

Author Contributions

Conceptualization, W.M. and S.X.; methodology, W.M.; software, C.L.; validation, W.M.; formal analysis, W.M. and C.L.; investigation, W.M.; resources S.X.; data curation, C.L.; writing—original draft preparation, W.M. and C.L.; writing—review and editing, W.M., C.L. and S.X.; visualization, C.L.; supervision, W.M.; project administration, S.X.; funding acquisition, S.X. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is supported by the National Key R&D Program of China (Grant nos. 2021YFA1000300, 2021YFA1000301, and 2021YFA1000303) and the National Natural Science Foundation of China (Grant no. 12171462).

Data Availability Statement

Not applicable.

Acknowledgments

The authors gratefully acknowledge the editors and reviewers for their professional comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sacks, J.; Welch, W.J.; Mitchell, T.J.; Wynn, H.P. Design and analysis of computer experiments. Stat. Sci. 1989, 4, 409–423. [Google Scholar]
  2. Santner, T.J.; Williams, B.J.; Notz, W.I. The Design and Analysis of Computer Experiments; Springer: New York, NY, USA, 2018; Volume 2. [Google Scholar]
  3. Gryder, R.W.; Wilson, S.R.; Swieringa, K.A.; Edwards, D.J. Space-filling designs for multi-layer nested factors. Qual. Eng. 2019, 31, 269–278. [Google Scholar]
  4. Mu, W.; Wei, Q.; Cui, D.; Xiong, S. Best Linear Unbiased Prediction for Multifidelity Computer Experiments. Math. Probl. Eng. 2018, 2018, 1–7. [Google Scholar]
  5. Shang, B.; Apley, D.W. Fully-sequential space-filling design algorithms for computer experiments. J. Qual. Technol. 2021, 53, 173–196. [Google Scholar]
  6. Wang, Y.; Sun, F.; Xu, H. On design orthogonality, maximin distance, and projection uniformity for computer experiments. J. Am. Stat. Assoc. 2022, 117, 375–385. [Google Scholar]
  7. Kennedy, M.C.; O’Hagan, A. Predicting the output from a complex compute code when fast approximations are available. Biometrika 2000, 87, 1–13. [Google Scholar]
  8. Mu, W.; Xiong, S. A class of space-filling designs and their projection properties. Stat. Probab. Lett. 2018, 141, 129–134. [Google Scholar]
  9. Qian, P.Z.G.; Wu, H.; Wu, C.F.J. Gaussian Process Models for Computer Experiments with Qualitative and Quantitative Factors. Technometrics 2008, 50, 383–396. [Google Scholar]
  10. Wei, Y.; Xiong, S. Bayesian integrative analysis for multi-fidelity computer experiments. J. Appl. Stat. 2019, 46, 1973–1987. [Google Scholar]
  11. Chen, D.; Xiong, S. Flexible nested Latin hypercube designs for computer experiments. J. Qual. Technol. 2017, 49, 337–353. [Google Scholar]
  12. Dash, S.; Mandal, B.N.; Parsad, R. On the construction of nested orthogonal Latin hypercube designs. Metrika 2020, 83, 347–353. [Google Scholar]
  13. Guo, B.; Chen, X.P.; Liu, M.Q. Construction of Latin hypercube designs with nested and sliced structures. Stat. Pap. 2020, 61, 727–740. [Google Scholar]
  14. Qian, P.Z. Nested Latin hypercube designs. Biometrika 2009, 96, 957–970. [Google Scholar]
  15. Rennen, G.; Husslage, B.; Van Dam, E.R.; Den Hertog, D. Nested maximin Latin hypercube designs. Struct. Multidiscip. Optim. 2010, 41, 371–395. [Google Scholar]
  16. Cover, T.M.; Thomas, J.A. Elements of Information Theory, 2nd ed.; John-Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  17. Lindley, D.V. On a measure of the information provided by an experiment. Ann. Math. Stat. 1956, 27, 986–1005. [Google Scholar] [CrossRef]
  18. Shewry, M.C.; Wynn, H.P. Maximum entropy sampling. J. Appl. Stat. 1987, 14, 165–170. [Google Scholar]
  19. Currin, C.; Mitchell, T.; Morris, M.; Ylvisaker, D. Bayesian prediction of deterministic functions, with applications to the design and analysis of computer experiments. J. Am. Stat. Assoc. 1991, 86, 953–963. [Google Scholar]
  20. Mitchell, T.J. An Algorithm for the Construction of “D. Optimal” Experimental Designs. Technometrics 1974, 16, 203–210. [Google Scholar]
  21. Jiang, Z.; Zhang, W.; Zhang, L. Sequential Maximum Entropy Approach to Design of Virtual Experiment. J. Syst. Simul. 2007, 19, 3876–3879. [Google Scholar]
  22. Chen, D.; Xiong, S. Optimization of Nested Latin Hypercube Designs. J. Sys. Sci. Math. Scis 2017, 37, 53. [Google Scholar]
  23. Kirkpatrick, S.; Gelatt, C.D., Jr.; Vecchi, M.P. Optimization by simulated annealing. Readings Comput. Vis. Issues Probl. Princ. Paradig. 1987, 220, 606–615. [Google Scholar]
  24. Chen, H.; Zhang, Y.; Yang, X. niform projection nested Latin hypercube designs. Stat. Pap. 2021, 62, 2031–2045. [Google Scholar]
  25. Xu, J.; Duan, X.; Wang, Z.; Yan, L. A general construction for nested Latin hypercube designs. Stat. Probab. Lett. 2018, 134, 134–140. [Google Scholar]
  26. Hebble, T.L.; Mitchell, T.J. “Repairing” response surface designs. Technometrics 1972, 14, 767–779. [Google Scholar]
  27. Mu, W.; Xiong, S. On algorithmic construction of maximin distance designs. Commun. Stat.-Simul. Comput. 2017, 46, 7972–7985. [Google Scholar]
  28. Harari, O.; Bingham, D.; Dean, A.; Higdon, D. Computer experiments: Prediction accuracy, sample size and model complexity revisited. Stat. Sin. 2009, 899–919. [Google Scholar]
  29. Loeppky, J.L.; Sacks, J.; Welch, W.J. Choosing the sample size of a computer experiment: A practical guide. Technometrics 2009, 51, 366–376. [Google Scholar]
  30. Sahama, T.R.; Diamond, N.T. Sample size considerations and augmentation of computer experiments. J. Stat. Comput. Simul. 2001, 68, 307–319. [Google Scholar]
  31. Ma, L.; Zhang, H.; Liu, Y.; Ning, A. Advanced Operations Research Tutorial; Shanghai People’s Publishing House: Shanghai, China, 2015. [Google Scholar]
  32. Xiong, S.; Qian, P.Z.; Wu, C.J. Sequential design and analysis of high-accuracy and low-accuracy computer codes. Technometrics 2013, 55, 37–46. [Google Scholar]
  33. McKay, M.D.; Beckman, R.J.; Conover, W.J. A Comparison of Three Methods for Selecting Values of Input Variables in The Analysis of Output from a Computer Code. Technometrics 1979, 21, 239–245. [Google Scholar]
  34. Cox, D.D.; Park, J.S.; Singer, C.E. A Statistical Method for Tuning a Computer Code to a Data Base. Comput. Stat. Data Anal. 2001, 37, 77–92. [Google Scholar]
  35. Tan, M.H. Minimax designs for finite design regions. Technometrics 2013, 55, 346–358. [Google Scholar]
  36. Han, G.; Santner, T.J.; Notz, W.I.; Bartel, D.L. Prediction for Computer Experiments Having Quantitative and Qualitative Input Variables. Technometrics 2009, 51, 278–288. [Google Scholar]
  37. Zhou, Q.; Qian, P.Z.G.; Zhou, S. A Simple Approach to Emulation for Computer Models With Qualitative and Quantitative Factors. Technometrics 2011, 53, 266–273. [Google Scholar] [CrossRef]
  38. Le Gratiet, L.; Cannamela, C. Cokriging-based sequential design strategies using fast cross-validation techniques for multi-fidelity computer codes. Technometrics 2020, 57, 418–427. [Google Scholar]
  39. Li, X.; Wang, X.; Xiong, S. A sequential design strategy for integrating low-accuracy and high-accuracy computer experiments. Commun. Stat.-Simul. Comput. 2020, 52, 817–824. [Google Scholar]
Figure 1. The procedure of multi-layer DETMAX algorithm.
Figure 1. The procedure of multi-layer DETMAX algorithm.
Mathematics 11 03572 g001
Figure 2. Some nested maximum entropy designs with p = 2 . The first four rows display nested maximum entropy designs with K = 2 and the last row shows five nested maximum entropy designs with K = 3 . The blue dots denote the points in the first layer. Adding the points depicted by red stars, we obtain the second layer. In the last row, the whole design comprises all the points including green plus signs. The correlation parameters are set as: for two-layer designs, (I) ϕ 1 = ( 80 , 80 ) , ϕ 2 = ( 30 , 30 ) , (II) ϕ 1 = ( 100 , 100 ) , ϕ 2 = ( 50 , 50 ) , (III) ϕ 1 = ( 100 , 100 ) , ϕ 2 = ( 100 , 100 ) , (IV) ϕ 1 = ( 150 , 150 ) , ϕ 2 = ( 80 , 80 ) , and (V) ϕ 1 = ( 200 , 200 ) , ϕ 2 = ( 50 , 50 ) ; for three-layer designs, ϕ 1 = ( 200 , 200 ) , ϕ 2 = ( 50 , 50 ) , ϕ 3 = ( 50 , 50 ) .
Figure 2. Some nested maximum entropy designs with p = 2 . The first four rows display nested maximum entropy designs with K = 2 and the last row shows five nested maximum entropy designs with K = 3 . The blue dots denote the points in the first layer. Adding the points depicted by red stars, we obtain the second layer. In the last row, the whole design comprises all the points including green plus signs. The correlation parameters are set as: for two-layer designs, (I) ϕ 1 = ( 80 , 80 ) , ϕ 2 = ( 30 , 30 ) , (II) ϕ 1 = ( 100 , 100 ) , ϕ 2 = ( 50 , 50 ) , (III) ϕ 1 = ( 100 , 100 ) , ϕ 2 = ( 100 , 100 ) , (IV) ϕ 1 = ( 150 , 150 ) , ϕ 2 = ( 80 , 80 ) , and (V) ϕ 1 = ( 200 , 200 ) , ϕ 2 = ( 50 , 50 ) ; for three-layer designs, ϕ 1 = ( 200 , 200 ) , ϕ 2 = ( 50 , 50 ) , ϕ 3 = ( 50 , 50 ) .
Mathematics 11 03572 g002aMathematics 11 03572 g002b
Figure 3. The boxplots of MSPEs in Example 1. Class 1 and 2 correspond to nested maximum entropy designs and FNLHDs, respectively. (a) s = ( 16 , 30 ) ; (b) s = ( 20 , 35 ) .
Figure 3. The boxplots of MSPEs in Example 1. Class 1 and 2 correspond to nested maximum entropy designs and FNLHDs, respectively. (a) s = ( 16 , 30 ) ; (b) s = ( 20 , 35 ) .
Mathematics 11 03572 g003
Table 1. MSPEs in Example 2 (standard deviations in parentheses).
Table 1. MSPEs in Example 2 (standard deviations in parentheses).
s = ( 10 , 15 ) s = ( 20 , 24 )
MSPEMSPE
Proposed ( α = 0.99 )0.3122 (0.0535)0.3321 (0.0345)
Proposed ( α = 0.9 )0.3100 (0.0580)0.3330 (0.0302)
Proposed ( α = 0.8 )0.3195 (0.0556)0.3295 (0.0354)
Proposed ( α = 0.7 )0.3149 (0.0596)0.3264 (0.0338)
FNLHD0.3932 (0.0620)0.3745 (0.0453)
Table 2. MSPEs in Example 3 (standard deviations in parentheses).
Table 2. MSPEs in Example 3 (standard deviations in parentheses).
s = ( 10 , 25 ) s = ( 15 , 22 )
MSPEMSPE
Proposed ( α = 0.99 )1.2816 (0.1726)1.3878 (0.2737)
Proposed ( α = 0.9 )1.3397 (0.2759)1.3736 (0.2785)
Proposed ( α = 0.8 )1.3026 (0.2108)1.4316 (0.2973)
Proposed ( α = 0.7 )1.3352 (0.2526)1.4249 (0.3299)
FNLHD1.4756 (0.3505)1.5116 (0.3773)
Table 3. The best design with structure ( 10 , 24 ) in Section 5.4.
Table 3. The best design with structure ( 10 , 24 ) in Section 5.4.
Run x 1 x 2 x 3 x 4
10.5280.9600.1130.525
20.3890.4040.2270.895
30.2510.8660.1100.738
40.2560.3430.6840.205
50.3150.5720.4040.783
60.7770.4330.4340.893
70.5630.2140.5660.196
80.1600.8730.7870.401
90.8070.7750.7120.271
100.3810.6880.7020.763
110.9990.9650.1250.904
120001
1310.58100
141111
15110.4970
1610.58710
1700.38111
1800.36700.396
190100.860
200.301111
210.758110.397
2210.15911
231000.397
2410.5820.5270.615
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mu, W.; Liu, C.; Xiong, S. Nested Maximum Entropy Designs for Computer Experiments. Mathematics 2023, 11, 3572. https://doi.org/10.3390/math11163572

AMA Style

Mu W, Liu C, Xiong S. Nested Maximum Entropy Designs for Computer Experiments. Mathematics. 2023; 11(16):3572. https://doi.org/10.3390/math11163572

Chicago/Turabian Style

Mu, Weiyan, Chengxin Liu, and Shifeng Xiong. 2023. "Nested Maximum Entropy Designs for Computer Experiments" Mathematics 11, no. 16: 3572. https://doi.org/10.3390/math11163572

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop