Next Article in Journal
Probabilistic Models and Deep Learning Models Assessed to Estimate Design and Operational Ocean Wave Statistics to Reduce Coastal Hazards
Previous Article in Journal
Numerical Modelling and Sensitivity Analysis of the Pitztal Valley Debris Flow Event
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Technical Note

Maximum Likelihood Instead of Least Squares in Fracture Analysis by Means of a Simple Excel Sheet with VBA Macro

by
Vincenzo Guerriero
Department of Civil, Construction-Architectural and Environmental Engineering (DICEAA), University of L’Aquila, 67100 L’Aquila, Italy
Geosciences 2023, 13(12), 379; https://doi.org/10.3390/geosciences13120379
Submission received: 28 August 2023 / Revised: 24 November 2023 / Accepted: 3 December 2023 / Published: 11 December 2023
(This article belongs to the Section Structural Geology and Tectonics)

Abstract

:
This technical note illustrates a linear regression algorithm based on the Maximum Likelihood Estimation (MLE), with a related Excel spreadsheet and VBA program, adapted to the case of fracture aperture data sets in which sampling of the smallest values is problematic. The method has been tested by means of Monte Carlo simulations and exhibits significantly better convergence against Least Squares criterion (LSM). As the method is conceptually simple and, following the indications illustrated here, the relative spreadsheet can be easily designed, it may be routinely used, instead of the Least Squares, in fracture analysis. Furthermore, the proposed method, with the appropriate modifications, might be potentially extended to other cases in geology and geophysics, in which significant biases at the lower limits of the sampling scale occur.

1. Introduction

Stratabound joints, as defined by Odling et al. [1] (also defined as perfect bed-bounded joints [2]), are fractures exhibiting in outcrop both terminations on the bed boundaries (Figure 1a). Let us consider the fracture sub-network composed only of this kind of joint. Such joint system, together with bedding joints bounding the mechanical layers, can form a well-connected hydraulic network [3], observable at the meter–decimeter scale, which divides rock into blocks that are parallelepiped shaped, often clearly observable in field.
In the last decades several studies (e.g., [4,5,6,7]) have highlighted how permeable structures are normally detectable on several scales of observation. Other studies have emphasized the advantages of using multiple-porosity models to simulate hydraulic behavior of naturally fractured rocks [8,9,10,11,12,13,14]. Multiple-porosity models allow to consider the different permeable structures in rock (e.g., pores, vugs, fracture networks, from micro-fractures to faults, observable at different scales) as hydraulic systems that are conventionally distinct, overlapping and interacting among them (e.g., [15] and references therein).
Figure 1. (a) Stratabound or perfect bed-bounded joint network; (b) example of field scanline; (c) example of field data, provided as paired values of joint aperture vs. bed thickness, highlighted by red arrows (from [3], modified); (d) comparator utilized for field measuring of joint aperture (from [6], modified); the threshold value of 0.265 mm is highlighted by a red circle.
Figure 1. (a) Stratabound or perfect bed-bounded joint network; (b) example of field scanline; (c) example of field data, provided as paired values of joint aperture vs. bed thickness, highlighted by red arrows (from [3], modified); (d) comparator utilized for field measuring of joint aperture (from [6], modified); the threshold value of 0.265 mm is highlighted by a red circle.
Geosciences 13 00379 g001
Guerriero et al. [4] proposed a hierarchical model for permeable structures in carbonate rocks, which identified four main kinds of interacting hydraulic structures, observable at different scales: (i) fault network, (ii) stratabound joints, (iii) non-stratabound joints (including micro- fractures) and (iv) a pore scale system. Stratabound joint networks assume a key role in the hydraulic behavior of fractured rock masses because they are an element of communication between the small-scale fractures (including micro-fractures) with the large-scale fault network.
For these reasons, the geometric characterization of stratabound joint networks in terms of fracture aperture and spacing may be of considerable interest. In fact, for a given succession, the knowledge of how aperture and spacing vary with bed thickness, would allow, for a given succession, a hydraulic characterization or modeling of stratabound joint networks, based on thickness data. Joint spacing and its dependence on rock layer thickness has been extensively analyzed [1,16,17,18,19,20,21,22,23]. Joint aperture and related statistical distribution have also been studied [2,4,6,24,25,26,27,28,29,30,31,32]. Nevertheless, in the specific case of perfect bed-bounded (i.e., stratabound) joints, aperture statistical distribution and its dependence on bed thickness and/or lithology has been insufficiently studied.
Therefore, Guerriero et al. [3] carried out a specific statistical investigation aimed at assessing whether and how stratabound joint aperture depends on bed thickness and/or lithology. Such analysis involved two Lower Cretaceous (Albian) carbonate successions, outcropping at Faito and Chianello Mts. (southern Italy). These successions, which had been selected as surface analogues of buried oil reservoirs in Val D’Agri (Basilicata, Italy), were previously well studied in terms of sedimentology and petrophysics [33,34], as well as of geological structural settings [3,30,31,35].
The statistical analysis carried out by Guerriero et al. [3] pointed out that stratabound joint aperture depends on bed thickness, according to a linear function, characterized by a non-zero y-intercept. This may have important consequences for the hydraulic and structural characterization of fractured rocks. As explained by Guerriero et al. [3], non-zero intercept implies that packages of thin beds may exhibit very high porosity and permeability values, whereas thicker beds are expected to exhibit low porosity but high permeability values.
In that analysis, fracture sampling was carried out by measuring aperture and bed thickness for each joint (and other attributes that are not relevant for the present study), along scanlines oriented parallel to the bedding along a vertical face through the center of each fracture (Figure 1b,c). To measure joint aperture in the field, the caliper proposed by Ortega et al. [6] was utilized (Figure 1d).
Guerriero et al. [3] explained the various drawbacks associated with field measurement of fracture aperture. Based on the visibility conditions at the studied outcrops, aperture measurements that were less than 0.265 mm (Figure 1d) were considered unreliable. For this reason, it was decided to set a minimum threshold value of 0.265 mm for them, thus including all aperture values detected equal to or less than this limit into this single category. This produces a data truncation that should not be confused with the truncation artifact (e.g., [6]) because the latter is an underestimation of the number of fractures detected, and small fractures (and micro-fractures) might not be visible. In the case of stratabound joints in our study area, we expect that they are practically always visible (as these do not include micro-fractures), but it is the measure of their aperture that could be wrong. By way of example, if in a certain layer six joints are measured with the following (true) aperture values (mm) [0.316, 0.224, 0.118, 0.081, 0.426, 0.277], the values of the recorded measurements according to the above-mentioned scale and threshold are [0.33, 0.265, 0.265, 0.265, 0.4, 0.265]. If the layer has, e.g., a thickness of 28 cm, the sample consists of the following pairs of values [(0.33, 28), (0.265, 28), (0.265, 28), (0.265, 28), (0.4, 28), (0.265, 28)]. Therefore, the number of fractures sampled is not altered (i.e., no truncation artifact, sensu Ortega et al., [6]), but there is a loss of information about the recorded measurement values.
The inclusion of all smaller aperture values into a single aperture class may induce a marked tendency of the regression line, in a diagram showing aperture vs. bed thickness, to intersect the ordinate axis near this threshold [3]. Because the intercept value identified by Guerriero et al. [3] was just close to this limit, the suspicion arose that it could have been affected by an artifact. Therefore, it has been decided to repeat the analysis of those data using a more effective method based on Maximum Likelihood Estimation (MLE). Such criterion has been successfully used by Rizzo et al. [32] to study fracture aperture and length statistics, which demonstrated the validity of the MLEs against linear regression. In the present work, the MLE has been used for different purposes, i.e., for the specific goal of correcting the above-mentioned aperture measurement bias within the framework of linear regression analysis.
The aim of this paper is to describe the adopted statistical method and the utilized algorithm and software, as well as to provide details about its validation and effectiveness. To this end, an example of application based on synthetic simulated fracture data is illustrated to clarify the differences between the two approaches and the added value of the Maximum Likelihood approach. Instead, illustrating the results of the analysis involving real field data from Faito and Chianello outcrops [3] goes beyond the scope of this work and these will be presented in a companion paper to be published later.

2. Recalls about Maximum Likelihood Estimation in Linear Regression

Let us consider a certain observed statistical sample and assume that it follows a known probability distribution. MLE is a parametric method in which the parameters (e.g., mean, variance, etc.) are estimated by adjusting them until the probability of observing that sample is maximized (e.g., [36]). As an example, suppose we observe an event that we know to be distributed according to a Poisson distribution, which occurs on average λ times in a time interval T. Let us imagine that we have the following outcomes of six independent observations (i.e., that we have observed the occurrences for six different time intervals T): [0, 3, 5, 6, 3, 5]. If we knew the λ parameter, we could calculate the probability L(λ) associated with this sample as the product of the probabilities of each observation:
L λ = λ 0 · e λ 0 ! · λ 3 · e λ 3 ! · λ 5 · e λ 5 ! · λ 6 · e λ 6 ! · λ 3 · e λ 3 ! · λ 5 · e λ 5 !   .
In case λ is not known and we want to estimate it, the basic idea of MLE consists of finding the one that maximizes L(λ) among all the possible values of λ, as seen in Equation (1). The function L(λ) is called Likelihood Function (LF). This function is also used for continuous probability distributions as the product of probability density functions. By way of example, if we have N observations of a Normal variable z, as [Z1, Z2, … ZN], then the LF is given by:
L μ , σ = 1 σ 2 π j = 1 N e Z j μ 2 2 σ 2 ;
where μ and σ denote the parameters mean and standard deviation, respectively. In the case where μ and σ are both unknown, LF must be minimized as a function of two variables. If one of the two parameters is known, then LF must be minimized with respect to the other unknown parameter. The cases above illustrated that the parameter values which maximize the LF can be calculated in closed form. If the involved equations are more complicated, maximization can be carried out numerically. Usually, the logarithm of the LF, denoted by Log-Likely Function, is utilized (e.g., [36,37]). In case of numerical maximization, its use has the advantage of avoiding overflow/underflow drawbacks.
The MLE can also be employed in (linear or nonlinear) regression analysis, as an alternative method to LSM [37,38]. In regression analysis between the two variables z and T, it is assumed to be:
z i = y T i + r i ,
where y(Ti) (i.e., the expected value) is a function of T, and r is a random residual with zero mean and constant standard deviation. In linear regression, the expected value y is defined as a linear function of T: y = m T + n, where m and n are constant parameters. If the analysis is carried out by means of MLE (which is a parametric method), a kind of distribution needs to be assumed for r. Usually, r is assumed normally distributed, with zero mean and constant standard deviation (here denoted by σ). Considering Equation (2), the LF involving r is:
L = 1 σ 2 π j = 1 N e r j 2 2 σ 2 = 1 σ 2 π j = 1 N e ( z j y ( T j ) ) 2 2 σ 2 = 1 σ 2 π j = 1 N e ( z j ( m · T j + n ) ) 2 2 σ 2 .
It should be noted that, in Equation (4), zj and Tj are known observed values, whereas m, n and σ need to be calculated by maximizing L(m,n,σ) or its logarithm.
It can be proven that such an approach to linear regression provides the same results as the classical LSM (e.g., [37]). Nevertheless, in the present study, the LF is defined in a different way (Section 3.1), and Section 3.2 explains why such MLE formulation provides different results than LSM.
Guerriero et al. [3] suggested that it is more appropriate to calculate residuals ri as the difference between the logarithms of observed values zi and predicted yi:
r i = ln z i ln y i .
Also in this case, residual r is assumed normally distributed, with zero mean and constant standard deviation denoted by d. Furthermore, the expected value is again: y = m T + n. It should be noted that this definition implies that, for each rock layer, i.e., fixed T, the result is ln z = r + ln y T ; therefore, ln(z) is the sum of a Normal variable with zero mean (i.e., r) and a constant (i.e., ln(y(T)), i.e., it is a Normal variable with a mean equal to such constant. In other words, within each layer, the observed aperture is well described by a Log Normal variable (i.e., whose logarithm is Normal), with a median value equal to y(T) = m T + n, and standard deviation d.
Once r, z and y are defined, an expression of the LF similar to that given in Equation (4) may be found. In the present study, a differently defined LF is used, whose details are illustrated in the next section.

3. Methods

3.1. Linear Regression by Means of Maximum Likelihood Estimation

Here, the LF involved in regression analysis, and utilized by the proposed algorithm and spreadsheet, is defined. Given the joint aperture classes illustrated in Figure 1d, let us denote the limit value by xi between contiguous classes si. In this instance, xi is an intermediate value between si-1 and si, opportunely chosen (Figure 2, at #1). Under the hypothesis that joint aperture values exhibit Log Normal distribution (Sect. 2), let us denote a probability distribution by Fmn.d(x) whose mean is ln(y) and the standard deviation is d. Here, y denotes the expected aperture value, which is a linear function of bed thickness T, whose parameters are coefficient m (mm/cm) and intercept n (mm). Then, the probability that a measured aperture value S falls within the class si denoted by pmn,d(si), for i > 1, is:
Probability ( x i 1 < S < x i ) = p m n , d s i = F m n , d x i 1 F m n , d x i ;             i > 1 ;
Whilst, for the first aperture class:
Probability S < x 1 = p m n , d s 1 = F m n , d x 1   ;             i = 1   .
Therefore, for a given sample [Sk, Tk], the Likelihood Function L(m,n,d) assumes the following form:
L m , n , d = k p m n , d S k   .
Searching for the maximum of this function on the space of the three parameters m, n and d (numerically), the MLE estimates of these three parameters are achieved.
Figure 2. Spreadsheet utilized for simulations. Yellow cells with red font denote variables to be adjusted in order to maximize the object function. Dark blue cells with yellow font denote parameter true values. #1: Aperture classes and related limits. #2: True m, n and d values of the model simulating fracture data. #3: Estimated m, n and d values, by MLE. #4: Estimated m, n and d values, by Excel linear regression functions. #5: Excel function LOGINV() to produce a single random aperture value, starting from a bed thickness value. #6: Simulated data set; from left: resampled bed thickness value, fracture aperture class and its upper and lower limits. #7: Field bed thickness data. #8: Likelihood data (see main text). #9: Object function; in sheet MLE, it is the Log Likelihood, and in sheet LSM, it is the sum of square of residuals. #10: Data to build up probability plots of residuals. #11: Monte Carlo simulation output data. Three columns on the left include each one 100 estimated values of m, n and d; on the right side, average and standard deviation of these columns are calculated.
Figure 2. Spreadsheet utilized for simulations. Yellow cells with red font denote variables to be adjusted in order to maximize the object function. Dark blue cells with yellow font denote parameter true values. #1: Aperture classes and related limits. #2: True m, n and d values of the model simulating fracture data. #3: Estimated m, n and d values, by MLE. #4: Estimated m, n and d values, by Excel linear regression functions. #5: Excel function LOGINV() to produce a single random aperture value, starting from a bed thickness value. #6: Simulated data set; from left: resampled bed thickness value, fracture aperture class and its upper and lower limits. #7: Field bed thickness data. #8: Likelihood data (see main text). #9: Object function; in sheet MLE, it is the Log Likelihood, and in sheet LSM, it is the sum of square of residuals. #10: Data to build up probability plots of residuals. #11: Monte Carlo simulation output data. Three columns on the left include each one 100 estimated values of m, n and d; on the right side, average and standard deviation of these columns are calculated.
Geosciences 13 00379 g002

3.2. Different Response to Data Truncation of the Proposed MLE and LSM, in Linear Regression

To understand the systematic error in the LSM analysis, caused by data truncation, let us recall the example illustrated in the Introduction, in which we imagine having acquired six joint aperture measurements from a 28-cm-thick layer. Imagine reporting these values in a diagram with bed thickness in abscissae and joint aperture in ordinate (e.g., Figure 1c). Of the six values above illustrated, three of them (with true apertures of 0.224, 0.118 and 0.081 mm) should have been placed in aperture classes lower than 0.265 mm. The grouping of these aperture values in the single class of 0.265 mm produces an upward migration of the points associated with the related measure. If we imagine inserting data recorded from several layers into this diagram, such migration of points will occur more frequently for smaller thicknesses (Figure 1c). This migration produces a lifting of the left part of the trendline, resulting in an increase in the value of the intercept, with consequent alteration of the estimated coefficient, due to a clockwise rotation of the trendline.
Whenever an experimental point has ordinates in the 0.265 mm class, the LSM introduces a bias due to the assumption that the residual around the trendline can take on the values +r and −r with equal probability. Due to data truncation, the measured aperture values can take on any exceeding values (according to the scale in Figure 1d) but cannot take on values lower than 0.265 mm. The MLE, on the other hand, does not introduce this bias, since the definition of the LF (Equation (8)) associates with values belonging to the 0.265 mm aperture class, and the probability that an aperture value is equal to or lower than this threshold (true statement). If a sample contains many aperture values in that class, LSM may provide highly biased estimates, while MLE is expected to be unaffected by that condition.

3.3. The Excel Sheet and VBA Program

The Excel folder utilized in this work includes three sheets: MLE, LSM and Results. Figure 2 illustrates the sheet MLE in detail; with respect to the latter, the sheet LSM is different only in column K cells, whose formula calculate the square of residual in the adjacent cell in column J. The first 30 rows, which are not depicted here, include a header illustrating some user instructions. The routines, written in Visual Basic for Applications (VBA), which utilize this folder to analyze data and carry out Monte Carlo simulations, are:
-
Sub Maximize() and Sub Minimize_LS(): analyze a data set by maximizing or minimizing an object function, which is Log likelihood for the former and residual standard deviation for the latter.
-
Sub Simul_Apert_Data(): based on thickness field data (#7 in Figure 2) and assigned (true) values of the parameters m, n and d (#2 in Figure 2), it produces a data set, composed of 35 paired values, by (i) resampling thickness data and (ii) producing a random aperture value for each thickness value (Section 3.3). Then, it identifies which class, and related limits, belongs to (#6).
-
Sub Simul_100(): for each triplet of true values m, n and d, it produces 100 simulated data sets and analyzes each one by means of Sub Maximize(), in sheet MLE, and Sub Minimize_LS(), in sheet LSM. Then, it saves the estimated values in columns S, T and U.
-
Sub Monte_Carlo(): varies the n true value in the range 0.05–4.75 mm, and for each value produces simulations by calling Sub Simul_100(), then saves the related results (#11) in sheet Results (Figure 3).
The core of the calculation method in the MLE sheet is in formulas in column K (#8 in Figure 2), in which for each aperture value (in column E, #6), the probability logarithm that it falls within the range to which it belongs is calculated as follows:
LN ( LOGNORMDIST G 60 ; LN I $ 55 × C 60 + I $ 57 ; J $ 48 LOGNORMDIST F 60 ; LN I $ 55 × C 60 + I $ 57 ; J $ 48
The formula LOGNORMDIST G 60 ; LN I $ 55 × C 60 + I $ 57 ; J $ 48 provides the probability that an aperture value is lesser or equal to the limit in cell G60 when its median value is a linear function of thickness (term LN(I$55 × C60 + I$57)) and its standard deviation is the value d in cell J48. The difference between distribution values in the formula above illustrated the probability that an aperture value lies within the range limited by cells G60–F60. The sum of logarithms of these probabilities returns the Log likelihood in cell J55. The routine Sub Maximize() starts calculations by assigning likely initial values to parameters m, n and d achieved by Excel least squares functions in cells J46:L46 (#4) by means of instructions such as:
Range(“I55”) = Range(“K46”)
Range(“I57”) = Range(“L46”)
Range(“J48”) = Range(“J46”)
then, it iteratively adjusts the values of parameters m, n and d in cells I55, I57 and J48, respectively (#3, in Figure 2), according to a simple steeper descent algorithm, until the maximum of the Log Likelihood (cell J55) is reached.
The spreadsheet also includes formulas to build up probability plots of residuals, indicated in panel #10 in Figure 2. Illustrating the use of these plots lies outside the scope of this paper. An explanation of their construction and usefulness is provided, such as in the work of Chambers et al. [39].

3.4. Validation by Means of Monte Carlo Simulation

A convenient way to evaluate the effectiveness of the MLE method compared to the LS consists of carrying out the analysis on simulated data sets whose parameters are already known, then comparing the known parameter values with the estimates achieved according to the two methodologies. To this end, a series of Monte Carlo simulations was carried out in which synthetic fracture aperture data sets were produced using likely values for the parameters m and d (equal to 0.03 and 0.2, respectively) and with parameter n (which was critical in our analysis) varying in the range 0.050–0.475 mm according to the following procedure:
  • Produce a data set, constituted by paired values (aperture, thickness), using the known parameters (m,n,d),
  • Simulate data truncation, including all joints belonging to the aperture class of 0.265 mm or lesser, in the 0.265 mm class,
  • Analyze by means of MLE,
  • Analyze by means of LS,
  • Compare known parameter values with estimates by MLE and LS,
  • Go back to step #1.
From sheet MLE, the Sub Monte_Carlo() is called, which assigns, as “true” values from which simulated data are produced, m = 0.003, d = 0.2 and varies n, starting from 0.050 to 0.475 with step of 0.025. For each n value, 35 thickness values are chosen from field data (in column A) by means of random numbers (Sub Simul_Apert_Data()). For each thickness value, an aperture one is produced by means of a random number and of the function in cell F52:
L O G I N V E 52 ; L N H 41 × G 52 + H 43 ; I $ 48
which returns the inverse Log Normal distribution of: (1) random number, (2) mean as logarithm of a linear function with m value from cell H41 and n from H43 and (3) standard deviation from cell I48. Then, the subroutine Sub Simul_Apert_Data() individuates the class and related limits to which this value belongs, and saves these in columns E, F and G. Hence, the Sub Maximize() is called, which maximizes the Log Likelihood. Then, the simulated data set is copied within the sheet LSM, and the best fit line is calculated by minimizing the sum of the square of residuals, calculated according to Equation (5) (Sub Minimize_LS()).
For each n value, this procedure is repeated 100 times; then, (m, n, d) estimates are saved in columns U, T and S, respectively. After 100 iterations, the average and standard deviation of these column values are saved into the sheet Results. Then, n is incremented by 0.025 mm.

4. Use of the Spreadsheet for Analysis of Field Data

The illustrated spreadsheet allows us to reproduce the mentioned Monte Carlo simulations. Nevertheless, it can be used to analyze our own field data by means of the routines Sub Initialize_Apert_Data() and Sub Maximize(). The fracture aperture data will be entered in column E and, for each item, the associated bed thickness value will be entered at the same row, in column C. Then, call the routine Sub Initialize_Apert_Data(), which will insert the appropriate values into columns F and G by individuating the class and related limits to which each aperture value belongs. Then, call the Sub Maximize() on the MLE sheet, which will estimate the parameters m, n and d, by maximizing the object function (i.e., the Log Likelihood function). This routine only works on the sheet denoted by MLE.
The spreadsheet, in its current form, analyzes the data set of 35 items. When analyzing a data set of a different size—for example, 50 items—the arrays in columns I, J, K, M, N, O and P need to be updated so that they have the same size (i.e., 50). Furthermore, the formulas in cells J55, J46, K46 and L46 must be updated so that they perform calculations on arrays of suitable length. By way of example, the formula in cell J55, which calculates the Log Likelihood, is:
SUM(K60:K94).
This calculates the Log Likelihood as sum of items in column K and rows from 60 to 94. In cases in which, for example, the analyzed data set includes 50 values, then the last one will fall at row 109. Therefore, such a formula will need to be updated as SUM(K60:K109).
This spreadsheet utilizes the aperture classes according to the logarithmic comparator proposed by Ortega et al. [6], which are stored in column F, at rows from 39 to 48. The class limits are in column G. By way of example, the limit between the aperture class of 0.33 mm (in cell F40) and 0.4 mm (in cell F41) is given by their average, equal to 0.365 mm (in cell G41). In case it is needed to use different aperture classes, in addition to modifying the values in that column, the routine Sub Initialize_Apert_Data() also needs to be updated in order to take into account both the varied limit values for each class and the varied number of aperture classes.

5. Results Discussion

Figure 3 shows the results of Monte Carlo simulations aimed at comparing the MLE and modified LSM results. The estimates of the parameters m, n and d, are plotted against the true n value, here denoted by ntrue. The MLE estimates of m and n exhibit a good convergence over the whole analyzed ntrue range. The estimates of the three parameters m, n and d, achieved by MLE, show an excellent convergence for ntrue > 0.1 mm. The LSM estimates of m, n and d show notably strong deviations from true values, for ntrue < 0.2 mm (i.e., LSM provides strongly biased estimators), and a good convergence only for ntrue > 0.3 mm. It is noteworthy that, for ntrue < 0.2 mm, the LS estimate of the parameter n assumes the same value (about 0.25 mm) and is independent of the statistical sample. This highlights the absolute ineffectiveness of such a statistical approach in estimating this parameter for small ntrue values.
Table 1 illustrates the results of Monte Carlo simulations for the MLE, which are partly visible in Figure 3. For each parameter analyzed, the estimate and, in the adjacent column, the relative standard deviation are reported. For ntrue > 0.1 mm, the mean values of m, n and d estimates converge, thus pointing out that MLE provides unbiased estimators. The standard deviation of the parameter n estimate increases as ntrue decreases. This highlights a random uncertainty that can be reduced by increasing the statistical sample size (as the estimator is unbiased).

6. Concluding Remarks

MLE is particularly effective in analyzing fracture data sets in which the field measurement of minor fracture aperture is problematic. Comparison of MLE against LSM has pointed out that the former is able to remove the bias due to data truncation, and then provide unbiased estimators, whereas the latter provides strongly biased estimators. The utilized algorithm, the related spreadsheet and routines have been illustrated in detail, thus allowing the reader to use the Excel folder proposed here and modify it or to create their own version. As this linear regression method can be easily performed with an Excel spreadsheet, it may be routinely used in fracture analysis. Furthermore, with the appropriate modifications, it might be potentially extended to other experimental situations where biases occur at the lower limit of the sampling scale.

Funding

Ministry of Economic Development (MiSE), Italy, Grant Id: C19C20000520004.

Data Availability Statement

Name of the code/library: MonteCarlo_MLE.xlsm. Contact: vincenzo.guerriero@univaq.it or vincenzo.guerriero@unina.it. Program language: Visual Basic, Applications Edition (VBA). Software required: MS Office or equivalent. Program size: 82 KB. The source codes are available for downloading at the link: https://github.com/vincenzo-guerriero/MonteCarlo_MLE.git (accessed on 2 December 2023).

Acknowledgments

The author thanks the editor and anonymous reviewers for their constructive comments and suggestions, which helped to significantly improve this manuscript. The research leading to these results has received funding from the Italian Ministry of Economic Development (MiSE) under the project “SICURA—CASA INTELLIGENTE DELLE TECNOLOGIE PER LA SICUREZZA CUP. C19C20000520004—Piano di investimenti per la diffusione della banda ultra larga FSC 2014-2020”.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Odling, N.E.; Gillespie, P.; Bourgine, B.; Castaing, C.; Chiles, J.P.; Christensen, N.P.; Fillion, E.; Genter, A.; Olsen, C.; Thrane, L.; et al. Variations in fracture system geometry and their implications for fluid flow in fractured hydrocarbon reservoirs. Pet. Geosci. 1999, 5, 373–384. [Google Scholar] [CrossRef]
  2. Hooker, J.N.; Laubach, S.E.; Marrett, R. Fracture-aperture size—Frequency, spatial distribution, and growth processes in strata-bounded and non-strata-bounded fractures, Cambrian Mesón Group, NW Argentina. J. Struct. Geol. 2013, 54, 54–71. [Google Scholar] [CrossRef]
  3. Guerriero, V.; Dati, F.; Giorgioni, M.; Iannace, A.; Mazzoli, S.; Vitale, S. The role of stratabound fractures for fluid migration pathways and storage in well-bedded carbonates. Ital. J. Geosci. 2015, 134, 383–395. [Google Scholar] [CrossRef]
  4. Guerriero, V.; Mazzoli, S.; Iannace, A.; Vitale, S.; Carravetta, A.; Strauss, C. A permeability model for naturally fractured carbonate reservoirs. Mar. Pet. Geol. 2013, 40, 115–134. [Google Scholar] [CrossRef]
  5. Camacho Velazquez, R.; Fuentes-Cruz, G.; Vasquez-Cruz, M.A. Decline curve analysis of fractured reservoirs with fractal geometry. SPE Reserv. Eval. Eng. 2008, 11, 606–619. [Google Scholar] [CrossRef]
  6. Ortega, O.J.; Marrett, R.A.; Laubach, S.E. Scale-independent approach to fracture intensity and average spacing measurement. AAPG Bull. 2006, 90, 193–208. [Google Scholar] [CrossRef]
  7. Massonnat, G.; Viszkok, J. Hierarchical organization of flow network in fractured carbonate reservoirs: Identification and characterization of key parameters. In Proceedings of the SPE Annual Technical Conference and Exhibition, San Antonio, TX, USA, 29 September–2 October 2002. [Google Scholar] [CrossRef]
  8. Ozkan, E.; Brown, M.; Raghavan, R.; Kazemi, H. Comparison of Fractured Horizontal-Well Performance in Conventional and Unconventional Reservoirs. In Proceedings of the SPE Western Regional Meeting, San Jose, CA, USA, 24–26 March 2009. [Google Scholar]
  9. Yan, B.; Wang, Y.; Killough, J.E. Beyond dual-porosity modeling for the simulation of complex flow mechanisms in shale reservoirs. Comput. Geosci. 2016, 20, 69–91. [Google Scholar] [CrossRef]
  10. Spagnuolo, A.M.; Wright, S. Analysis of a multiple-porosity model for single-phase flow through naturally fractured porous media. J. Appl. Math. 2003, 2003, 327–364. [Google Scholar] [CrossRef]
  11. Kim, J.; Sonnenthal, E.L.; Rutqvist, J. Formulation and sequential numerical algorithms of coupled fluid/heat flow and geomechanics for multiple porosity materials. Int. J. Numer. Methods Eng. 2012, 92, 425–456. [Google Scholar] [CrossRef]
  12. Dreier, J.; Ozkan, E.; Kazemi, H. New Analytical Pressure-Transient Models to Detect and Characterize Reservoirs with Multiple Fracture Systems. In Proceedings of the SPE International Petroleum Conference in Mexico, Puebla, Mexico, 8–9 November 2004. [Google Scholar]
  13. Al-Ahmadi, H.A.; Wattenbarger, R.A. Triple-porosity Models: One Further Step Towards Capturing Fractured Reservoirs Heterogeneity. In Proceedings of the SPE/DGS Saudi Arabia Section Technical Symposium and Exhibition, Al-Khobar, Saudi Arabia, 15–18 May 2011. [Google Scholar]
  14. Liu, C.-Q. Exact solution for the compressible flow equations through a medium with triple-porosity. Appl. Math. Mech. 1981, 2, 457–462. [Google Scholar]
  15. Guerriero, V. 1923–2023: One Century since Formulation of the Effective Stress Principle, the Consolidation Theory and Fluid–Porous-Solid Interaction Models. Geotechnics 2022, 2, 961–988. [Google Scholar] [CrossRef]
  16. Huang, Q.; Angelier, J. Fracture spacing and its relation to bed thickness. Geol. Mag. 1989, 126, 355–362. [Google Scholar] [CrossRef]
  17. Narr, W.; Suppe, J. Joint spacing in sedimentary rocks. J. Struct. Geol. 1991, 13, 1037–1048. [Google Scholar] [CrossRef]
  18. Mandal, N.; Deb, S.K.; Khan, D. Evidence for a nonlinear relationship between fracture spacing and layer thickness. J. Struct. Geol. 1994, 16, 1275–1281. [Google Scholar] [CrossRef]
  19. Narr, W. Estimating average fracture spacing in subsurface rock. AAPG Bull. 1996, 80, 1565–1586. [Google Scholar] [CrossRef]
  20. Pascal, C.; Angelier, J.; Cacas, M.-C.; Hancock, P.L. Distribution of joints: Probabilistic modeling and case study near Cardiff (Wales, U.K.). J. Struct. Geol. 1997, 19, 1273–1284. [Google Scholar] [CrossRef]
  21. Wu, H.; Pollard, D.D. An experimental study of the relationship between joint spacing and layer thickness. J. Struct. Geol. 1995, 17, 887–905. [Google Scholar] [CrossRef]
  22. Bai, T.; Pollard, D.D. Fracture spacing in layered rocks: A new explanation based on the stress transition. J. Struct. Geol. 2000, 22, 43–57. [Google Scholar] [CrossRef]
  23. Guerriero, V.; Mazzoli, S. Theory of Effective Stress in Soil and Rock and Implications for Fracturing Processes: A Review. Geosciences 2021, 11, 119. [Google Scholar] [CrossRef]
  24. Marrett, R.; Ortega, O.J.; Kelsey, C.M. Extent of power-law scaling for natural fractures in rock. Geology 1999, 27, 799–802. [Google Scholar] [CrossRef]
  25. Gale, J.F.W.; Laubach, S.E.; Marrett, R.A.; Olson, J.E.; Holder, J.; Reed, R.M. Predicting and characterizing fractures in dolostone reservoirs: Using the link between diagenesis and fracturing. Geol. Soc. Lond. Spec. Publ. 2004, 235, 177–192. [Google Scholar] [CrossRef]
  26. Ortega, O.J.; Gale, J.F.; Marrett, R.A. Quantifying diagenetic and stratigraphic controls on fracture intensity in platform carbonates: An example from the Sierra Madre Oriental, northeast Mexico. J. Struct. Geol. 2010, 32, 1943–1959. [Google Scholar] [CrossRef]
  27. Hooker, J.N.; Gale, J.F.W.; Gomez, L.A.; Laubach, S.E.; Marrett, R.; Reed, R.M. Aperture-size scaling variations in a low-strain opening-mode fracture set, Cozzette Sandstone, Colorado. J. Struct. Geol. 2009, 31, 707–718. [Google Scholar] [CrossRef]
  28. Hooker, J.N.; Gomez, L.A.; Laubach, S.E.; Gale, J.F.W.; Marrett, R. Effects of diagenesis (cement precipitation) during fracture opening on fracture aperture-size scaling in carbonate rocks. In Advances in Carbonate Exploration and Reservoir Analysis: Geological Society; Garland, J., Laubach, S., Neilson, J., Whidden, K.J., Eds.; Special Publication: London, UK, 2012; Volume 370, pp. 187–206. [Google Scholar]
  29. Hooker, J.N.; Laubach, S.E.; Marrett, R. A universal power-law scaling exponent for fracture apertures in sandstone. Geol. Soc. Am. Bull. 2014, 126, 1340–1362. [Google Scholar] [CrossRef]
  30. Guerriero, V.; Iannace, A.; Mazzoli, S.; Parente, M.; Vitale, S.; Giorgioni, M. Quantifying uncertainties in multi-scale studies of fractured reservoir analogues: Implemented statistical analysis of scan line data from carbonate rocks. J. Struct. Geol. 2010, 32, 1271–1278. [Google Scholar] [CrossRef]
  31. Guerriero, V.; Vitale, S.; Ciarcia, S.; Mazzoli, S. Improved statistical multi-scale analysis of fractured reservoir analogues. Tectonophysics 2011, 504, 14–24. [Google Scholar] [CrossRef]
  32. Rizzo, R.E.; Healy, D.; De Siena, L. Benefits of maximum likelihood estimators for fracture attribute analysis: Implications for permeability and up-scaling. J. Struct. Geol. 2017, 95, 17–31. [Google Scholar] [CrossRef]
  33. Giorgioni, M.; Iannace, A.; D’Amore, M.; Dati, F.; Galluccio, L.; Guerriero, V.; Mazzoli, S.; Parente, M.; Strauss, C.; Vitale, S. Impact of early dolomitization on multi-scale petrophysical heterogeneities and fracture intensity of low-porosity platform carbonates (Albian-Cenomanian, southern Apennines, Italy). Mar. Pet. Geol. 2016, 73, 462–478. [Google Scholar] [CrossRef]
  34. Iannace, A.; Galluccio, L.; Guerriero, V.; Mazzoli, S.; Parente, M.; Vitale, S. Dolomites within the Mesozoic carbonates of Southern Apennines (Italy): Genetic models and reservoir implications. Rend. Online Soc. Geol. Ital. 2008, 2, 109–114. [Google Scholar]
  35. Vitale, S.; Dati, F.; Mazzoli, S.; Ciarcia, S.; Guerriero, V.; Iannace, A. Modes and timing of fracture network development in poly-deformed carbonate reservoir analogues, Mt. Chianello, southern Italy. J. Struct. Geol. 2012, 34, 223–235. [Google Scholar] [CrossRef]
  36. Dekking, F.M.; Kraaikamp, C.; Lopuhaa, H.P.; Meester, L.E. A Modern Introduction to Probability and Statistics: Understanding Why and How; Springer: London, UK, 2005. [Google Scholar]
  37. Myung, I.J. Tutorial on maximum likelihood estimation. J. Math. Psychol. 2003, 47, 90–100. [Google Scholar] [CrossRef]
  38. Da Silva, S.L.E.F.; dos Santos Lima, G.Z.; Volpe, E.V.; de Araújo, J.M.; Corso, G. Robust approaches for inverse problems based on Tsallis and Kaniadakis generalised statistics. Eur. Phys. J. Plus 2021, 136, 518. [Google Scholar] [CrossRef]
  39. Chambers, J.; Cleveland, W.; Kleiner, B.; Tukey, P. Graphical Methods for Data Analysis; Chapter 6; Wadsworth: Belmont, CA, USA, 1983. [Google Scholar]
Figure 3. Spreadsheet “Results”. The diagrams on the left side point out that MLE estimators exhibit modest deviations over the whole analyzed range of true n values and excellent convergence for n > 0.1, whereas LSM ones converge only for n > 0.3, showing large deviations elsewhere.
Figure 3. Spreadsheet “Results”. The diagrams on the left side point out that MLE estimators exhibit modest deviations over the whole analyzed range of true n values and excellent convergence for n > 0.1, whereas LSM ones converge only for n > 0.3, showing large deviations elsewhere.
Geosciences 13 00379 g003
Table 1. Results of Monte Carlo simulations. For each parameter, the mean value of the sampling estimate is illustrated, as well as the related standard deviation, as a descriptor of uncertainty. Note that the 6th column shows the estimated value of the parameter ‘residual standard deviation’, whereas the last column shows the related standard deviation (i.e., the uncertainty).
Table 1. Results of Monte Carlo simulations. For each parameter, the mean value of the sampling estimate is illustrated, as well as the related standard deviation, as a descriptor of uncertainty. Note that the 6th column shows the estimated value of the parameter ‘residual standard deviation’, whereas the last column shows the related standard deviation (i.e., the uncertainty).
Method: Maximum Likelihood Estimation
True m = 0.003, True Residual std dev = 0.2, Sample Number = 35
True n ValueEstimated m Std dev.
m
Estimated nStd dev.
n
Estimated Residual std dev.Std dev. Residual std dev.
0.050.00240.001130.1020.0860.1250.105
0.0750.00280.001120.0980.0790.1320.089
0.10.00300.001060.0970.0720.1700.088
0.1250.00310.000860.1150.0600.1770.085
0.150.00290.000850.1490.0510.1710.071
0.1750.00310.000940.1670.0490.1720.053
0.20.00300.000690.2000.0260.1880.052
0.2250.00310.000570.2240.0220.1850.047
0.250.00300.000540.2490.0190.1930.043
0.2750.00300.000560.2740.0190.1960.039
0.30.00300.000580.3020.0180.1930.035
0.3250.00300.000690.3260.0200.1920.034
0.350.00310.000650.3480.0210.1900.033
0.3750.00310.000570.3720.0210.1900.032
0.40.00310.000740.3990.0230.1880.031
0.4250.00310.000690.4260.0240.1860.030
0.450.00320.000920.4500.0300.1940.029
0.4750.00300.000800.4760.0280.1890.027
0.050.00240.001130.1020.0860.1250.105
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Guerriero, V. Maximum Likelihood Instead of Least Squares in Fracture Analysis by Means of a Simple Excel Sheet with VBA Macro. Geosciences 2023, 13, 379. https://doi.org/10.3390/geosciences13120379

AMA Style

Guerriero V. Maximum Likelihood Instead of Least Squares in Fracture Analysis by Means of a Simple Excel Sheet with VBA Macro. Geosciences. 2023; 13(12):379. https://doi.org/10.3390/geosciences13120379

Chicago/Turabian Style

Guerriero, Vincenzo. 2023. "Maximum Likelihood Instead of Least Squares in Fracture Analysis by Means of a Simple Excel Sheet with VBA Macro" Geosciences 13, no. 12: 379. https://doi.org/10.3390/geosciences13120379

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop