Next Article in Journal
Existence and Uniqueness Results for a Pantograph Boundary Value Problem Involving a Variable-Order Hadamard Fractional Derivative
Next Article in Special Issue
Engineering Applications with Stress-Strength for a New Flexible Extension of Inverse Lomax Model: Bayesian and Non-Bayesian Inference
Previous Article in Journal
Properties of Various Entropies of Gaussian Distribution and Comparison of Entropies of Fractional Processes
Previous Article in Special Issue
Fuzzy vs. Traditional Reliability Model for Inverse Weibull Distribution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

New Partially Linear Regression and Machine Learning Models Applied to Agronomic Data

by
Gabriela M. Rodrigues
1,
Edwin M. M. Ortega
1,* and
Gauss M. Cordeiro
2
1
Department of Exact Sciences, University of São Paulo, Piracicaba 13418-900, Brazil
2
Department of Statistics, Federal University of Pernambuco, Recife 50670-901, Brazil
*
Author to whom correspondence should be addressed.
Axioms 2023, 12(11), 1027; https://doi.org/10.3390/axioms12111027
Submission received: 22 September 2023 / Revised: 24 October 2023 / Accepted: 30 October 2023 / Published: 31 October 2023

Abstract

:
Regression analysis can be appropriate to describe a nonlinear relationship between the response variable and the explanatory variables. This article describes the construction of a partially linear regression model with two systematic components based on the exponentiated odd log-logistic normal distribution. The parameters are estimated by the penalized maximum likelihood method. Simulations for some parameter settings and sample sizes empirically prove the accuracy of the estimators. The superiority of the proposed regression model over other regression models is shown by means of agronomic experimentation data. The predictive performance of the new model is compared with two machine learning techniques: decision trees and random forests. These methods achieved similar prediction performance, i.e., none stands out as a better predictor. In this sense, the objective of the research is to choose the best method. If the objective is only predictive, the decision tree can be used due to its simplicity. For inference purposes, the regression model is recommended, which can provide much more information regarding the relationship of the variables under study.

1. Introduction

Regression analysis is an important statistical tool to investigate the relationship between two or more variables. This relationship can be linear or nonlinear, which should be verified adequately. The incorrect assumption of linearity can compromise the reliability of the hypothesis tests and lead to poorly specified models. Semiparametric models are an interesting alternative for real data since they can access this relationship with more flexibility by means of nonparametric functions without imposing stringent conditions such as the commonly used nonlinear regression models. They enable the investigation of linear and nonlinear effects of the covariates simultaneously and can access this relationship. Because of these advantages, these models are receiving more attention in the literature.
Normal linear regression models are generally applied in different areas to model real data. However, it is known that several phenomena are not always in accordance with the normal model due to the lack of symmetry of the distribution or the presence of heavy tails. Another standard assumption in regression analysis is the homogeneity of error variability. Violation of this assumption may have adverse consequences for the efficiency of the estimators. Therefore, it becomes important to check heteroscedasticity. In this paper, we propose a partial linear regression model, where the mean and dispersion parameters vary across the observations through regression structures, assuming that the model response follows the exponentiated odd log-logistic normal (EOLLN) distribution. A special feature of this regression model is that it can model symmetric, asymmetric, and bimodal data.
Some models for nonparametric functions can be cited: a semiparametric model for the median and skewness [1], a skew-normal semiparametric model [2], an exponentiated sinh Cauchy model [3], and a log-sinh Cauchy model with cure fraction [4]. In the Bayesian context, the following can be cited: a semiparametric regression model in microbiology [5], a semiparametric model for tea productivity [6], an extended Maxwell semiparametric model [7], and an exponentiated power exponential regression [8].
Other alternatives that have been gaining popularity are machine learning models ([9,10,11,12]). Models based on decision trees (DTs) and random forests (RFs), which are nonlinear and nonparametric techniques, can be more flexible than the usual regression models. Since nonparametric methods are free of assumptions and do not require prior knowledge of the functional form of the relation between the response variable and the covariates, they are also robust to the presence of outliers and can be used with asymmetric data [13].
Articles related to plantain varieties are important for farmers, researchers, and professionals in the field. There are several variations of agronomic potential among varieties, which have been minimally explored ([14,15,16]). The knowledge of this potential can help to obtain more sustainable and profitable practices, which are the patterns of growth, a basic knowledge. Given the scarcity of these studies, we define a new regression model for the pseudostem height in the planting–flowering period of banana varieties.
Figure 1a displays the scatterplot between the pseudostem height and planting–flowering period, where these variables do not have a linear relationship. The second covariate related to the response variable is a variety (a factor with nine levels). The relationship of this variable with the response variables in Figure 1b shows that there is a heterogeneity of variances and several discrepant points (outliers). In view of these facts, two statistical tools may be appropriate.
First, a heterogeneous semiparametric partially linear regression model can be a suitable alternative. This model can adequately study the nonlinear relationship of the covariate period (continuous) with the response through nonparametric functions together with the linear effect of the covariate variety (factor). Further, we verify the effects of these covariates on the dispersion parameter.
Second, the DT and RF machine learning algorithms can be alternatives to predict the pseudostem height in terms of the covariates. The choice of these algorithms is based on their simplicity since they do not require prior knowledge of the functional form of the relation between these covariates and the response variable. Due to their easy interpretation and computational implementation, models based on decision trees are very attractive to researchers in areas other than statistics, although they are less precise than the usual regression models.
This article is summarized as follows. Section 2 defines the EOLLN distribution [17], which can model bimodal data and/or data with positive or negative asymmetry, and proposes a partially linear regression using P-splines based upon this distribution. The machine learning methodology is briefly addressed in Section 3. Some simulations are carried out in Section 4. The utility of the new regression model is illustrated through agronomic data in Section 5. Some conclusions are addressed in Section 6.

2. The EOLLN Model

The probability density function (pdf) of the EOLLN model is determined from [17] by taking the normal as baseline
f ( y ) = f ( y ; ν , τ , μ , σ ) = ν τ ϕ y μ σ Φ y μ σ ν τ 1 1 Φ y μ σ ν 1 σ Φ y μ σ ν + 1 Φ y μ σ ν τ + 1 ,
where y R , μ R is a location, σ > 0 is a scale, and Φ ( · ) and ϕ ( · ) are the cumulative distribution function (cdf) and pdf of the standard normal, respectively.
Hereafter, let Y EOLLN ( ν , τ , μ , σ ) be a random variable with density function (1). The three special EOLLN models are: OLLN [18] when τ = 1 , exponentiated normal (Exp-N) [19]) when ν = 1 , and clearly, normal when ν = τ = 1 .
Figure 2 shows the flexibility of the EOLLN density such as bimodality and positive and negative asymmetry.
The quantile function (qf) of Y is given by (for 0 < u < 1 )
Q ( u ) = Q N u 1 / ( ν τ ) u 1 / ( ν τ ) + ( 1 u 1 / τ ) 1 / ν ,
where Q N ( · ) = G 1 ( · ; μ , σ ) is the normal qf.

The EOLLN Partially Linear Regression Model

Consider independent observations Y 1 , , Y n , and the systematic components (for i = 1 , , n )
g 1 ( μ i ) = η i 1 = x i 1 β 1 , and g 2 ( σ i ) = η i 2 = x i 2 β 2 ,
where (for k = 1 , 2 ) x i k = ( 1 , x i k 1 , , x i k p k ) is the explanatory variable vector, β k = ( β k 0 , , β k p ) is a p k + 1 vector of unknown coefficients, and η i k is the linear predictor. The strictly monotone functions g 1 and g 2 are defined from R R and R R + , respectively.
The logarithm of the likelihood function for θ = ( ν , τ , β 1 , β 2 ) can be expressed as
l ( θ ) = n log ν τ σ + i = 1 n log ϕ y i μ i σ i + ( ν τ 1 ) i = 1 n log Φ y i μ i σ i + ( ν 1 ) i = 1 n log 1 Φ y i μ i σ i ( τ + 1 ) i = 1 n log Φ y i μ i σ i ν + 1 Φ y i μ i σ i ν .
We maximize (4) using R (gamlss) [20] or (optim), SAS (Proc NLMixed). The initial coefficients are the values from the fitted normal regression.
The EOLLN partially linear regression model is constructed using P-splines [21] for the nonlinear effects, and changing Equations (3) by
g 1 ( μ i ) = η i 1 = x i 1 β 1 + h ( t i ) , and g 2 ( σ i ) = η i 2 = x i 2 β 2 ,
where h ( t i ) is a smooth function of the explanatory variable t i in an interval [ a , b ] .
Let ω = ( θ , h ) be the full parameter vector, where h = ( h ( t 1 0 ) , , h ( t q 0 ) ) is the vector of nonlinear effects. The maximum likelihood estimates (MLEs) are calculated by maximizing the penalized log-likelihood
l p ( ω ) = l ( θ ) λ 2 h K h ,
where l ( θ ) is given by (4), λ > 0 is the unknown smoothing parameter, and K is a symmetric matrix depending on smoothing parameters (for details, see [22]).
The total degrees of freedom of the model is the sum of the degrees of freedom adopted to all additive and parametric terms as reported in [23]. We use the Penalized Quasi-Likelihood (PQL) method [24] to estimate the smoothing parameters and the degrees of freedom of the P-Spline smooth functions. We use the RS algorithm implemented in the gamlss package in R to maximize (6). The function pb(·) assigns additive terms to explanatory variables. The PQL method is described in [25].
The residuals can identify the discrepancy between the fitted model and the data, and the construction of an envelope to better interpret them is reported in [26]. The quantile residuals (qrs) for the proposed regression model are
q r i = Φ 1 Φ ν ^ τ ^ y i μ ^ i σ ^ i Φ ν ^ y i μ ^ i σ ^ i + 1 Φ y i μ ^ i σ ^ i ν ^ τ ^ ,
where μ i and σ i follow from Equation (5).

3. Machine Learning Methods

Machine learning is a sub-area of artificial intelligence. It involves supplying machines with data so that they learn and are able to predict new results when new data are presented. In this process, different models are fitted to a dataset (training dataset), applied to a validation dataset (or testing dataset), and compared according to a performance criterion with the objective of forecasting future units. In traditional regression analysis, the goal is also to study the relationship between the variables, while assuming that the available data belong to a random sample.
There are two classes of machine learning: supervised and unsupervised. In the first class, predictor variables are used to forecast one or more response variables (called labeled variables). For qualitative responses, this problem is known as classification, while for quantitative responses it is called prediction. In turn, in unsupervised learning, all of the variables of the dataset are considered without distinction between predictor and response variables, and the objective is to describe associations and patterns among them. This study is focused on regression problems using two supervised algorithms, as described below.

3.1. Decision Trees

The use of decision trees is a nonparametric method that leads to easily interpreted results. The tree is constructed by recursively subdividing the space generated by the predictor variables. Each partition is called a node and each final result is called a leaf (terminal node). This method is based on Classification and Regression Trees (CARTs) [27], according to the following strategy:
1.
Partition the space generated by the predictor variables ( X 1 , , X p ) into m regions R 1 , , R m .
2.
For each element belonging to R j , the predictor of Y (called Y ^ R j ) will be the mean between the points with values of X 1 , , X p in R j .
However, the division into many partitions is computationally unfeasible, and besides this, multiple divisions can rapidly fragment the data, making the dataset insufficient for the next levels and hampering the learning of the algorithm. In light of this situation, the decision tree algorithm utilizes a recursive binary splitting (top-down) approach. If the optimality criterion is to minimize the sum of the square errors (SSE), the steps follow as:
1.
Let the predictor variables X j and their possible cutoff points be t, and consider the regions, for all pairs ( j , t ) ,
R 1 ( j , t ) = { X : X j < t } , R 2 ( j , t ) = { X : X j t } .
2.
Select the pair ( j , t ) that gives the smallest SSE:
SSE = i : x i R 1 ( j , t ) ( y i y ^ i R 1 ) 2 + i : x i R 2 ( j , t ) ( y i y ^ i R 2 ) 2 .
3.
Repeat the procedure only with the data partition until some criterion is satisfied (e.g., obtaining a fixed minimum number of elements in each region).
One of the problems associated with decision trees is overfitting, which consists of obtaining the terminal nodes equal to the number of observations, i.e., each element of the dataset used is perfectly predicted. The pruning of the trees can resolve this problem, generating fewer nodes with lower variance and greater interpretability, although this can slightly increase the bias. A way to prune the tree is to examine the complexity parameter (CP), which serves to control the size of the tree and corresponds to the smallest increment in the cost of the model necessary to consider a new subdivision. In general, the goal is to find the level for which the CP is minimized.

3.2. Random Forests

Random forests (RFs) [28] are generalizations of models based on trees with the objective of reducing the variability and bias [29] by the insertion of two forms of randomness in their construction. Various trees are constructed with different bootstrap samples of the data and each tree is cultivated using a random subset of predictor variables. The trees are cultivated without pruning and all of the responses of individual trees are combined by the average to obtain the forecasts of the final output. This combination of results characterizes the algorithm as an ensemble and can be more precise than any single-constituent model. The RF classifies the predictor variables in decreasing order and evaluates them simultaneously, but it behaves like a “black box”, because the trees cannot be examined separately ([30,31]).
Start with a training set with n elements, ( x i , y i ) , i = 1 , , n , where x i is a vector of p predictor variables associated with the ith response variable. Consider the following steps:
1.
Obtain B bootstrap samples with a replacement of size n based on the training data. In general, the number of elements chosen in each sample is the order 2 / 3 of the n elements of the set provided for training. The remaining elements are called out-of-bag (OOB) samples;
2.
Select m < p predictor variables randomly;
3.
In the construction of each node of the tree, choose the best predictor among the selected m;
4.
For each unpruned tree, determine the predictor of Y, y ^ b ( x ) , where b = 1 , , B ;
5.
Aggregate all of the predictors as y ^ R F B ( x ) = 1 B b = 1 B y ^ b ( x ) .

3.3. Cross-Validation

Cross-validation has the objective of avoiding problems like overfitting and bias, and of verifying the capacity for the generalization of a model, i.e., its prediction performance with new data. The aforementioned division of the dataset into two disjoint subsets (training and testing) is the simplest cross-validation technique, called hold-out. Since the partition is performed randomly, high variability can be obtained depending on the training set utilized. The k-fold method [32] can be an interesting alternative.
In this method, n observations from the original sample D are divided into K disjoint subsets of observations, namely D 1 , , D K , each one having approximately equal size m k such that n = k = 1 K m k . The validation sample is composed of the partition D k , while the training sample consists of the other K 1 partitions not including the kth partition, i.e., the training set is given by D ( k ) = { D 1 , , D k 1 , D k + 1 , , D K } . This process is repeated iteratively K times until each of k = 1 , , K partitions is considered a validation sample. A general recommendation is to use K = 5 or K = 10 , where greater values of K imply larger training samples, and, hence, greater variances and higher computational costs, but can also generate smaller biases [33].

4. Simulation Study

A Monte Carlo simulation study (for sample sizes n = 50, 150, and 450) evaluates the consistency of the estimates in the proposed regression model, and the empirical distribution of the qrs is determined by the RS algorithm in the gamlss package. We generate r = 1000 samples from Equation (2) with two covariates x 1 Binomial ( 1 , 0.5 ) and t Uniform ( 0 , 1 ) under two scenarios:
Scenario 1 (homogeneous variance and cubic relationship): The first systematic component is μ = β 11 x 1 i + h ( t i ) , where h ( t i ) = 2 t i + sin ( 2 π t i ) to obtain a cubic relationship. The true values for the generation process are β 11 = 0.4 , log ( σ ) = 1.6 , log ( ν ) = 0.5 and log ( τ ) = 0.6 .
Scenario 2 (heterogeneous variance and quadratic relationship): The second systematic component is μ = β 11 x 1 i + h ( t i ) and σ = exp ( β 20 + β 21 x 1 i + β 22 t i ) , respectively, where h ( t i ) = sin ( π t i ) to obtain a quadratic relationship. The true values in the generation process are β 11 = 0.6 , β 20 = 1.3 , β 21 = 1.1 , β 22 = 0.9 , log ( ν ) = 0.4 and log ( τ ) = 0.5 .
The following steps are carried out:
(i)
Obtain x 1 i Binomial ( 1 , 0.5 ) and t i Uniform ( 0 , 1 ) ;
(ii)
Calculate μ and σ from their respective systematic components in each scenario;
(iii)
Set u i Uniform ( 0 , 1 ) ;
(iv)
Calculate the observations from Equation (2) using the previous steps;
(v)
Determine the qrs.
The average estimates (AEs), biases, and mean squared errors (MSEs) are found for each replication. Table 1 confirms that the AEs converge to the true parameters and the biases and the MSEs decay if n increases. So, the estimators are consistent. The plots in Figure 3 and Figure 4 reveal that the generated smoothed curves approximate the true curve when n increases, i.e., the estimator of the nonparametric part is also consistent. We emphasize that the quadratic and cubic forms are only used to illustrate possible relationships between variables. The model has the flexibility to accommodate different nonlinear relationships without the need to restrictively define its format. The normal probability plots in Figure 5 and Figure 6 reveal that the empirical distribution of the qrs tends to the standard normal distribution.

Predictive Performance

For Scenarios 1 and 2, we verify the predictive capacity of the EOLLN model, decision trees, and random forests. Each of the 1000 samples is divided into training and validation sub-samples, BY approximately 70% and 30%, respectively. The models are fitted to the training data, and the coefficient of determination (R 2 ), root mean square error (RMSE), and variance accounted for (VAF) are calculated using the test data. The coefficient R 2 evaluates the accuracy of the model fitting with values between 0 and 1, where the higher values mean the stronger explanatory power. The RMSE indicates the prediction errors, i.e., it shows discrepancies between predicted and observed values, where values closer to zero are considered adequate. The VAF is typically adopted to verify if the predictive model is rigorous [34]. The more accurate the prediction is, the VAF of the predictive model will be closer to 100 (i.e., lower variance). These metrics are:
R 2 = i = 1 n y i 2 i = 1 n ( y i y ^ i ) 2 i = 1 n y i 2 ,
RMSE = 1 n i = 1 n ( y i y i ^ ) 2 1 / 2 ,
VAF = 100 × 1 v a r ( y i y i ^ ) v a r ( y i ) ,
where y i ’s are the observed values for the validation set, n is its size, and y ^ i ’s are the predicted values.
These values can be noted in the boxplots for the 1000 replicates of each sample size in Figure 7, Figure 8 and Figure 9. The results are given below:
  • All models perform worse for Scenario 2, which can be justified by the presence of heterogeneous variance;
  • All metrics improve for larger sample sizes, for all models, and in both scenarios;
  • For the first scenario, better performance of the EOLLN regression model is notable, i.e., higher R 2 s, lower RMSEs, and higher VAFs;
  • On the other hand, for the second scenario, we can observe similar results for both models.

5. Application: Pseudostem Height Data

Consider a dataset from a homogeneous experimental design of 3000 m 2 conducted in the African Center for Banana and Plantain Researches (CARBAP) experimental station in Cameroon by plantain researchers, where nine varieties of plantain-like hybrids are found interesting and promising based on characteristics such as the structure of the bunch or tolerance to disease; they are: Batard (BA)—Natural and Giant; Big Ebanga (BE)—Natural and Medium; CRBP39 (CR)—Hybrid-a and Medium; D248 (DD)—Hybrid-a and Medium; D535 (DC)—Hybrid-a and Medium; Essong (ES)—Natural and Giant; French Clair (FC)—Natural and Medium; FHIA21 (FH)—Hybrid-b and Medium; and Mbouroukou no. 3 (MB)—Natural and Medium, where hybrid-a was obtained from the CIRAD-CARBAP breeding collaborative program and hybrid-b from the FHIA breeding program.
The experiment was conducted from the plantation (on 29 July 2009) on the harvest of the first crop cycle of each variety; the last mother plants were harvested in February 2011. So, this is mostly mother-plant data.
The completely randomized experimental design described in [35] has the pseudostem height as the response and variety as a single factor. Five replicates of each variety were considered, totaling 45 plants per variety. Cropping conditions such as mineral nutrition, fertilization, and irrigation are not taken into account. The pseudostem height was measured between the soil and the bottom of the “V” formed by the two leaves last emitted in the stage “complete flowering” or “flowering”, i.e., the day when the latest fertile hand has appeared. The dataset includes n = 355 observations of height, so not all varieties have records of the flowering period. So, the variables under study are:
  • y i : pseudostem height (in cm);
  • t i : planting–flowering period (in days);
  • x i 1 : varieties ( i = 1 , , 355 ).
The covariate x i 1 has nine levels, and then eight dummy variables ( d i j , j = 1 , 8 ) are defined.
Complementing the descriptive analysis shown in the introduction (Figure 1), Figure 10 displays the scatter plot of the pseudostem height as a function of days per variety. The varieties achieved complete flowering on different days, especially the ES variety.

5.1. The EOLLN Partially Linear Regression Model

The proposed regression model is compared with three sub-models (OLLN, Exp-N, and Normal) and the partially linear regression model based on the skew-normal distribution [2], whose pdf is ( y R )
f ( μ , σ , ν ) = 2 σ ϕ y μ σ Φ ν y μ σ ,
where μ R , σ > 0 and ν R .
The following systematic components are defined (for i = 1 , , 355 ):
μ i = β 10 + j = 1 8 β 1 j d i j + h ( t i ) and σ i = exp β 20 + j = 1 8 β 2 j d i j .
Note that the covariate x 1 (varieties) has nine levels (BA, BE, CR, DC, DD, ES, FC, FH, MB). Then, we create eight dummy variables ( d 1 , , d 8 ) taking the first level of the BA variety as a reference. For example, the dummy variable d 1 measures the effect of the BA variety level in relation to the BE level. The other dummy variables have similar interpretations.
Two well-known statistics (denoted by acronyms) and the global deviance (GD) are reported in Table 2. The new regression model gives the lowest values of these statistics, and then it can be chosen as the most appropriate to explain the pseudostem height. The likelihood ratio (LR) test for nested models provides p-values < 0.001 , thus confirming our findings.
We move up to the analysis from the fitted EOLLN partially linear regression model, where the maximum likelihood estimates (MLEs), standard errors (SEs), and p-values are given in Table 3. The nonlinear effects ( t i ) will be interpreted at the end through Figure 11a. Figure 11 shows the partial effects of the day variable on the parameters μ and σ . We conclude that the average of the pseudostem height is constant until (approximately) day 300. After this day, this average grows as the days go by; see Figure 11a. The variability of the pseudostem height grows when the days pass; see Figure 11b.
The qrs for the fitted regression models are compared by worm plots in Figure 12, thus supporting that our proposal yields the best fit. Figure 13 shows that the qrs are randomly distributed around zero and the majority of them are in the interval [−3,3]. In turn, Figure 13b indicates that nearly all points are within the simulated envelope. So, the new regression model is adequate. Further, Figure 14 provides the histograms with the marginally adjusted densities, and Figure 14b displays the estimated and empirical cumulative functions for each level of the variety. Hence, the proposed regression model suitably captures the distribution of this response variable.
Finally, some conclusions can be inferred from Table 3:
  • Interpretations for μ :
    -
    All varieties are significant in terms of mean pseudostem height compared to BA variety.
    -
    Only the ES variety ( β 15 ) has positive effects in relation to the BA variety, i.e., its mean is higher. Therefore, the other varieties have negative effects.
    -
    From Table 4, referring to the multiple comparisons, we note that there is no significant difference between the varieties DC-BE, FC-CR, and MB-DC for a 5% significance level. For other varieties, there is a significant difference. These results can be visualized graphically in Figure 14b.
  • Interpretations for σ :
    -
    Only the ES variety ( β 25 ) is significant at a 5% level of significance, i.e., its variability differs from the BA variety.
    -
    From the multiple comparisons in Table 4, it is noted that there is a significant difference between the DD-BE, ES-BE, DD-CR, ES-CR, DD-DC, ES-DC, MB-DC, FC-DD, FC-ES, FH-ES, and MB-ES varieties in relation to the pseudostem height variability for a 5% significance level. For other varieties, there is no significant difference.
    -
    Again, these results are consistent with the descriptive analysis.
Table Table 4 reveals comparisons between all varieties from which similar interpretations to these can be made.

5.2. Machine Learning Models

The R software is also employed for the machine learning models. The decision trees are found with the rpart function of the rpart package and the random forests with the randomForest function of the randomForest package by defining the hyperparameters ntree = 500 and mtry = 1. The argument ntree denotes the number of trees to be cultivated in the forest, and mtry is the number of predictor variables to be considered in the construction of each node in the tree.
We perform the cross-validation method, described in Section 3.3, for K = 5 and K = 10 . The K partitions are generated and the DT, RF, and EOLLN partially linear regression models are fitted to the training set.
For each model and K value, the average performance metrics (Equations (8)–(10)) are reported in Table 5, from which the results follow:
1.
Regression model × machine learning models: there are small improvements of the DT and RF models regarding R 2 and RMSE, but no improvement with regard to VAF;
2.
DT × RF: comparison of the two machine learning models reveals no significant improvements when changing from DT to RF;
3.
K = 5 × K = 10 : with regard to the values of K, no significant difference occurred in the values of these metrics.
For both comparisons, these only slightly varying results can be related to the small number of covariates. The values of p and n influence the performance of the machine learning models, as noted by [36], who studied the performance of the models based on trees according to the sizes of these quantities.
The OOB error rates stabilized quickly, i.e., with a few trees for both values of K. This fact is again related to the small number of covariates, since one of the forms of randomness of the algorithm is to consider arbitrary variables at each node of the tree. An interesting result that can be obtained with RF is the importance plot. An increase in the percentage of the mean squared error (MSE) indicates greater importance. Figure 15 shows the importance of the variables planting–flowering days and varieties in terms of plantain pseudostem height. It can be seen that the variable variety is most important in predicting pseudostem height.
Further, we fit the complete decision tree model shown in Figure 16. Each node of the tree indicates the corresponding value of its function and the percentage of data in that node. The initial node gives the general average of the response variable. The need for pruning through the CP is checked, but it turns out not necessary in practice. We obtain six decision rules:
1.
If varieties = DD or FH and days < 236 then y ^ i = 304.1667 ;
2.
If varieties = DD or FH and days ≥ 236 then y ^ i = 329.8936 ;
3.
If varieties = CR or FC then y ^ i = 358.9438 ;
4.
If varieties = BE or DC or MB then y ^ i = 388.6283 ;
5.
If varieties = BA or ES and days < 369.5 then y ^ i = 440.5636 ;
6.
If varieties = BA or ES and days ≥ 369.5 then y ^ i = 496.9333 .
We note that these rules agree with the descriptive analysis of the data. The scatterplot in Figure 10 shows that these rules hold for the variables and values of the days. Figure 14b also shows the agreement between the rules and the empirical cumulative functions of the varieties.
The objectives of the two proposed methods are very different, so the researcher’s purpose should be considered in choosing the best model. If the objective is to make predictions, the decision tree model can be the best, since it is much simpler. The decision rules for the prediction can be obtained easily and consulted rapidly. The RF is not recommended due to its greater complexity and absence of significant improvement in relation to the DT.
On the other hand, for researchers interested in making inferences along with predictions, we recommend the proposed regression model. The EOLLN partially linear regression model has the potential to supply information with regard to the variables studied, unlike the machine learning model. Besides this, it obtains equally satisfactory prediction results, so it can be used for both objectives.

6. Conclusions

This paper introduced a new partially linear regression model with the exponentiated odd log-logistic normal (EOLLN) distribution. It was motivated by an agronomic experiment involving banana varieties, where a nonlinear relationship occurred between the days from planting–flowering and the height of the pseudostem (response variable). Further, we noted heterogeneous variances for the covariable plant variety. The new EOLLN partially linear regression model was a suitable alternative to study the nonlinear relationship between the variables through nonparametric functions, and also to verify the effect of the variability of the varieties on the plant height through the parameter related to the variance of the distribution. We compared its predictive capacity with two machine learning algorithms: decision trees and random forests.
Additionally, the machine learning models were interesting alternatives to predict the response variable, because they did not need prior knowledge of the functional form between the response variable and the covariates. All of the methods obtained similar prediction performance, i.e., none stood out as the best predictor. The random forests achieved a stable OOB error rate with only a few trees, and the covariable variety stood out as the most important predictor of the plantain pseudostem height. However, although it is more robust than the decision tree model, it did not obtain better results regarding the predictive capacity of new values.
In this respect, for those wanting to make predictions, the decision tree model is recommended due to its simplicity. In turn, for researchers wanting to make inferences, we recommend the new regression model, which provides more information regarding the relationship of the variables under consideration, besides also having good prediction performance. The EOLLN partially linear regression model provides good inferences such as which varieties are significant regarding the mean and variance of the response, besides comparisons between the varieties, i.e., which of them differs from each other.
We carried out a simulation study under two scenarios that supply cubic and quadratic relations between the variables. The results showed that the new regression model was adequate to capture different nonlinear forms and provided consistent maximum likelihood estimators (MLEs).
In future works, we recommend analyzing other banana plant response variables due to the scarcity of studies related to this plant. We also suggest this model to analyze other datasets and problems in other areas of knowledge, where variables have a nonlinear relationship or the response variable is bimodal and/or skewed. Finally, we suggest comparing the new model with those of the machine learning methods to analyze datasets with a larger number of covariates (high-dimensional datasets).

Author Contributions

Conceptualization, G.M.R., E.M.M.O. and G.M.C.; methodology, G.M.R., E.M.M.O. and G.M.C.; software, G.M.R., E.M.M.O. and G.M.C.; validation, G.M.R., E.M.M.O. and G.M.C.; formal analysis, G.M.R., E.M.M.O. and G.M.C.; investigation, G.M.R., E.M.M.O. and G.M.C.; data curation, G.M.R., E.M.M.O. and G.M.C.; writing—original draft preparation, G.M.R., E.M.M.O. and G.M.C.; writing—review and editing, G.M.R., E.M.M.O. and G.M.C.; visualization, G.M.R., E.M.M.O. and G.M.C.; supervision, G.M.R., E.M.M.O. and G.M.C. All authors have read and agreed to the published version of the manuscript.

Funding

This study was financed in part by the Coordenação de Aperfeiçoamento de Pessoal de Nível Superior-Brasil (CAPES) and Conselho Nacional de Desenvolvimento Cientifico e Tecnológico. (CNPq).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data Availability at https://dataverse.cirad.fr/dataset.xhtml?persistentId=doi:10.18167/DVN1/CBUVWU (accessed on 1 September 2023).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vanegas, L.H.; Paula, G.A. A semiparametric approach for joint modeling of median and skewness. Test 2015, 24, 110–135. [Google Scholar] [CrossRef]
  2. Xu, D.; Zhang, Z.; Du, J. Skew-normal semiparametric varying coefficient model and score test. J. Stat. Comput. Simul. 2015, 85, 216–234. [Google Scholar] [CrossRef]
  3. Ramires, T.G.; Ortega, E.M.M.; Hens, N.; Cordeiro, G.M.; Paula, G.A. A flexible semiparametric regression model for bimodal, asymmetric and censored data. J. Appl. Stat. 2018, 45, 1303–1324. [Google Scholar] [CrossRef]
  4. Ramires, T.G.; Hens, N.; Cordeiro, G.M.; Ortega, E.M.M. Estimating nonlinear effects in the presence of cure fraction using a semi-parametric regression model. Comput. Stat. 2018, 33, 709–730. [Google Scholar] [CrossRef]
  5. Lee, J.; Sison-Mangus, M. A Bayesian semiparametric regression model for joint analysis of microbiome data. Front. Microbiol. 2018, 9, 522. [Google Scholar] [CrossRef] [PubMed]
  6. Dhekale, B.S.; Sahu, P.K.; Vishwajith, K.P.; Mishra, P.; Narsimhaiah, L. Application of parametric and nonparametric regression models for area, production and productivity trends of tea (Camellia sinensis) in India. Indian J. Ecol. 2017, 44, 192–200. [Google Scholar]
  7. Prataviera, F.; Ortega, E.M.M.; Cordeiro, G.M. An extended Maxwell semiparametric regression for censored and uncensored data. Commun. Stat.-Simul. Comput. 2023, 52, 3305–3326. [Google Scholar] [CrossRef]
  8. Prataviera, F.; Ortega, E.M.M.; Cordeiro, G.M.; Cancho, V.G. The exponentiated power exponential semiparametric regression model. Commun. Stat.-Simul. Comput. 2022, 51, 5933–5953. [Google Scholar] [CrossRef]
  9. Alonso, L.; Renard, F. A new approach for understanding urban microclimate by integrating complementary predictors at different scales in regression and machine learning models. Remote Sens. 2020, 12, 2434. [Google Scholar] [CrossRef]
  10. Oukawa, G.Y.; Krecl, P.; Targino, A.C. Fine-scale modeling of the urban heat island: A comparison of multiple linear regression and random forest approaches. Sci. Total Environ. 2022, 815, 152836. [Google Scholar] [CrossRef]
  11. Khan, M.A.; Shah, M.I.; Javed, M.F.; Khan, M.I.; Rasheed, S.; El-Shorbagy, M.A.; El-Zahar, E.R.; Malik, M.Y. Application of random forest for modelling of surface water salinity. Ain Shams Eng. J. 2022, 13, 101635. [Google Scholar] [CrossRef]
  12. Subeesh, A.; Bhole, S.; Singh, K.; Chandel, N.S.; Rajwade, Y.A.; Rao, K.V.R.; Kumar, S.P.; Jat, D. Deep convolutional neural network models for weed detection in polyhouse grown bell peppers. Artif. Intell. Agric. 2022, 6, 47–54. [Google Scholar] [CrossRef]
  13. Kuhn, M.; Johnson, K. Applied Predictive Modeling; Springer: New York, NY, USA, 2013; Volume 26, p. 13. [Google Scholar]
  14. Swennen, R.; Vuylsteke, D.; Ortiz, R. Phenotypic diversity and patterns of variation in West and Central African plantains (Musa spp., AAB group Musaceae). Econ. Bot. 1995, 45, 320–327. [Google Scholar] [CrossRef]
  15. Ortiz, R.; Madsen, S.; Vuylsteke, D. Classification of African plantain landraces and banana cultivars using a phenotypic distance index of quantitative descriptors. Theor. Appl. Genet. 1998, 96, 904–911. [Google Scholar] [CrossRef]
  16. Depigny, S.; Lescot, T.; Achard, R.; Diouf, O.; Côte, F.X.; Fonbah, C.; Sadom, L.; Tixier, P. Model-based benchmarking of the production potential of plantains (Musa spp., AAB): Application to five real plantain and four plantain-like hybrid varieties in Cameroon. J. Agric. Sci. 2017, 155, 888–901. [Google Scholar] [CrossRef]
  17. Alizadeh, M.; Tahmasebi, S.; Haghbin, H. The exponentiated odd log-logistic family of distributions: Properties and applications. J. Stat. Model. Theory Appl. 2020, 1, 29–52. [Google Scholar]
  18. Gleaton, J.U.; Lynch, J.D. Properties of generalized log-logistic families of lifetime distributions. J. Probab. Stat. Sci. 2006, 4, 51–64. [Google Scholar]
  19. Mudholkar, G.S.; Srivastava, D.K.; Kollia, G. A generalization of the Weibull distribution with application to the analysis of survival data. J. Am. Stat. Assoc. 1996, 91, 1575–1583. [Google Scholar] [CrossRef]
  20. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2022. [Google Scholar]
  21. Eilers, P.H.; Marx, B.D. Flexible smoothing with B-splines and penalties. Stat. Sci. 1996, 11, 89–121. [Google Scholar] [CrossRef]
  22. Rigby, R.A.; Stasinopoulos, D.M. Generalized additive models for location, scale and shape. J. R. Stat. Soc. Ser. C Appl. Stat. 2005, 54, 507–554. [Google Scholar] [CrossRef]
  23. Voudouris, V.; Gilchrist, R.; Rigby, R.; Sedgwick, J.; Stasinopoulos, D. Modelling skewness and kurtosis with the BCPE density in GAMLSS. J. Appl. Stat. 2012, 39, 1279–1293. [Google Scholar] [CrossRef]
  24. Lee, Y.; Nelder, J.A.; Pawitan, Y. Generalized Linear Models with Random Effects: Unified Analysis via H-Likelihood; Chapman & Hall/CRC: New York, NY, USA, 2006. [Google Scholar]
  25. Rigby, R.A.; Stasinopoulos, D.M. Automatic smoothing parameter selection in GAMLSS with an application to centile estimation. Stat. Methods Med. Res. 2014, 23, 318–332. [Google Scholar] [CrossRef] [PubMed]
  26. Atkinson, A.C. Plots, Transformations and Regression: An Introduction to Graphical Methods of Diagnostics Regression Analysis, 2nd ed.; Clarendon Press: Oxford, UK, 1987; 282p. [Google Scholar]
  27. Breiman, L.; Friedman, J.; Olshen, R.; Stone, C. Cart. Classification and Regression Trees; Routledge: New York, NY, USA, 1984. [Google Scholar]
  28. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  29. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning, 2nd ed.; Springer: New York, NY, USA, 2008. [Google Scholar]
  30. Prasad, A.M.; Iverson, L.R.; Liaw, A. Newer classification and regression tree techniques: Bagging and random forests for ecological prediction. Ecosystems 2006, 9, 181–199. [Google Scholar] [CrossRef]
  31. Rodriguez-Galiano, V.; Mendes, M.P.; Garcia-Soldado, M.J.; Chica-Olmo, M.; Ribeiro, L. Predictive modeling of groundwater nitrate pollution using random forest and multisource variables related to intrinsic and specific vulnerability: A case study in an agricultural setting (southern Spain). Sci. Total Environ. 2014, 476, 189–206. [Google Scholar] [CrossRef]
  32. Burman, P. A comparative study of ordinary cross-validation, v-fold cross-validation and the repeated learning-testing methods. Biometrika 1989, 76, 503–514. [Google Scholar] [CrossRef]
  33. Borra, S.; Di Ciaccio, A. Measuring the prediction error. A comparison of cross-validation, bootstrap and covariance penalty methods. Comput. Stat. Data Anal. 2010, 54, 2976–2989. [Google Scholar] [CrossRef]
  34. Facchini, L.; Betti, M.; Biagini, P. Neural network based modal identification of structural systems through output-only measurement. Comput. Struct. 2014, 138, 183–194. [Google Scholar] [CrossRef]
  35. Dépigny, S.; Tchotang, F.; Talla, M.; Fofack, D.; Essomé, D.; Ebongué, J.P.; Kengni, B.; Lescot, T. The Plantain-Optim dataset: Agronomic traits of 405 plantains every 15 days from planting to harvest. Data Brief 2018, 17, 671–680. [Google Scholar] [CrossRef]
  36. Genç, S.; Mendeş, M. Evaluating performance and determining optimum sample size for regression tree and automatic linear modeling. Arq. Bras. Med. Veterinária e Zootec. 2021, 73, 1391–1402. [Google Scholar] [CrossRef]
Figure 1. Plantain data: (a) Scatterplot between the pseudostem height and planting–flowering period and (b) Boxplot of the pseudostem height by variety.
Figure 1. Plantain data: (a) Scatterplot between the pseudostem height and planting–flowering period and (b) Boxplot of the pseudostem height by variety.
Axioms 12 01027 g001
Figure 2. Plots of the EOLLN ( ν , τ , μ , σ ) density.
Figure 2. Plots of the EOLLN ( ν , τ , μ , σ ) density.
Axioms 12 01027 g002
Figure 3. Scenario 1: (a) n = 50 , (b) n = 150 , (c) n = 450 .
Figure 3. Scenario 1: (a) n = 50 , (b) n = 150 , (c) n = 450 .
Axioms 12 01027 g003
Figure 4. Scenario 2: (a) n = 50 , (b) n = 150 , (c) n = 450 .
Figure 4. Scenario 2: (a) n = 50 , (b) n = 150 , (c) n = 450 .
Axioms 12 01027 g004
Figure 5. Normal probability plots for Scenario 1: (a) n = 50 , (b) n = 150 , (c) n = 450 .
Figure 5. Normal probability plots for Scenario 1: (a) n = 50 , (b) n = 150 , (c) n = 450 .
Axioms 12 01027 g005
Figure 6. Normal probability plots for Scenario 2: (a) n = 50 , (b) n = 150 , (c) n = 450 .
Figure 6. Normal probability plots for Scenario 2: (a) n = 50 , (b) n = 150 , (c) n = 450 .
Axioms 12 01027 g006
Figure 7. Boxplots of coefficient of determination for Scenarios: (a) 1 and (b) 2.
Figure 7. Boxplots of coefficient of determination for Scenarios: (a) 1 and (b) 2.
Axioms 12 01027 g007
Figure 8. Boxplots of RMSE for Scenarios: (a) 1 and (b) 2.
Figure 8. Boxplots of RMSE for Scenarios: (a) 1 and (b) 2.
Axioms 12 01027 g008
Figure 9. Boxplots of VAF for Scenarios: (a) 1 and (b) 2.
Figure 9. Boxplots of VAF for Scenarios: (a) 1 and (b) 2.
Axioms 12 01027 g009
Figure 10. Scatterplot of the pseudostem height and planting–flowering period by variety.
Figure 10. Scatterplot of the pseudostem height and planting–flowering period by variety.
Axioms 12 01027 g010
Figure 11. The terms of the new fitted regression model for (a) μ and (b) σ .
Figure 11. The terms of the new fitted regression model for (a) μ and (b) σ .
Axioms 12 01027 g011
Figure 12. Worm plots: (a) EOLLN, (b) OLLN, (c) Exp-N, (d) Normal.
Figure 12. Worm plots: (a) EOLLN, (b) OLLN, (c) Exp-N, (d) Normal.
Axioms 12 01027 g012
Figure 13. Residual plots from the new fitted linear regression model: (a) Index and (b) Normal probability plots with simulated envelope.
Figure 13. Residual plots from the new fitted linear regression model: (a) Index and (b) Normal probability plots with simulated envelope.
Axioms 12 01027 g013
Figure 14. (a) Histogram of the pseudostem height with estimated densities and (b) Empirical and estimated cumulative functions by variety.
Figure 14. (a) Histogram of the pseudostem height with estimated densities and (b) Empirical and estimated cumulative functions by variety.
Axioms 12 01027 g014
Figure 15. Plots of the variable importance for (a) K = 5 and (b) K = 10 .
Figure 15. Plots of the variable importance for (a) K = 5 and (b) K = 10 .
Axioms 12 01027 g015
Figure 16. Decision tree for Pseudostem height data.
Figure 16. Decision tree for Pseudostem height data.
Axioms 12 01027 g016
Table 1. Findings from the simulated new regression model.
Table 1. Findings from the simulated new regression model.
Scenario 1n = 50n = 150n = 450
ω TrueAEsBiasesMSEsAEsBiasesMSEsAEsBiasesMSEs
β 11 0.40.40−0.000.000.40−0.000.000.40−0.000.00
log ( σ ) −1.6−1.71−0.110.04−1.64−0.040.01−1.62−0.020.00
log ( ν ) 0.50.540.040.040.510.010.000.500.000.00
log ( τ ) 0.60.630.030.150.610.010.020.610.010.00
Scenario 2n = 50n = 150n = 450
ω TrueAEsBiasesMSEsAEsBiasesMSEsAEsBiasesMSEs
β 11 0.60.59−0.010.040.59−0.010.010.60−0.000.00
β 20 −1.3−1.50−0.200.15−1.37−0.070.03−1.32−0.020.01
β 21 1.11.210.110.081.130.030.021.110.010.01
β 22 0.90.960.060.200.920.020.050.900.000.02
log ( ν ) 0.40.40−0.000.030.40−0.000.010.39−0.010.00
log ( τ ) 0.50.530.030.110.520.020.020.500.000.01
Table 2. Adequacy measures.
Table 2. Adequacy measures.
ModelAICBICGD
EOLLN3244.213339.183195.17
OLLN3262.073349.773216.77
Exp-N3258.243339.553216.24
Normal3285.223362.663245.22
Skew-Normal3287.283368.603245.28
Table 3. Findings from the new fitted regression model.
Table 3. Findings from the new fitted regression model.
ω MLEsSEsp-Values ω MLEsSEsp-Values
β 10 399.134813.0847< 0.01 β 20 3.23290.4815< 0.01
β 11 −46.69315.6440< 0.01 β 21 0.26570.19170.1667
β 12 −80.60665.4382< 0.01 β 22 0.30850.23790.1955
β 13 −50.69336.0629< 0.01 β 23 0.45440.25680.0776
β 14 −133.69104.7972< 0.01 β 24 −0.27780.21340.1939
β 15 11.35754.57050.0134 β 25 −0.79940.2888< 0.01
β 16 −73.09265.0582< 0.01 β 26 0.19060.22220.3916
β 17 −104.51674.8448< 0.01 β 27 −0.00840.20680.9674
β 18 −42.50774.9276< 0.01 β 28 0.00870.18260.9621
log ( ν ) 1.92810.0447< 0.00001
log ( τ ) −0.75280.0493< 0.00001
Table 4. Comparing varieties according to the new regression model.
Table 4. Comparing varieties according to the new regression model.
Hypotheses H 0 Tests for the Location μ Tests for the Scale σ
MLEsSEsp-ValuesMLEsSEsp-Values
BE-BA−46.6935.644< 0.01 0.2660.1920.167
CR-BA−80.6075.438< 0.01 0.3090.2380.195
DC-BA−50.6936.063< 0.01 0.4540.2570.078
DD-BA−133.6914.797< 0.01 −0.2780.2130.194
ES-BA11.3574.5700.013−0.7990.289< 0.01
FC-BA−73.0935.058< 0.01 0.1910.2220.392
FH-BA−104.5174.845< 0.01 −0.0080.2070.967
MB-BA−42.5084.928< 0.01 0.0090.1830.962
CR-BE−33.9135.194< 0.01 0.0430.2030.833
DC-BE−4.0006.0110.5060.1890.2320.417
DD-BE−86.9984.476< 0.01 −0.5430.174< 0.01
ES-BE58.0515.943< 0.01 −1.0650.316< 0.01
FC-BE−26.3995.046< 0.01 −0.0750.1920.696
FH-BE−57.8245.017< 0.01 −0.2740.1800.128
MB-BE4.1855.0090.404−0.2570.1410.070
DC-CR29.9135.615< 0.01 0.1460.2660.584
DD-CR−53.0843.854< 0.01 −0.5860.213< 0.01
ES-CR91.9645.862< 0.01 −1.1080.364< 0.01
FC-CR7.5144.5800.102−0.1180.2310.610
FH-CR−23.9104.596< 0.01 −0.3170.2240.158
MB-CR38.0994.563< 0.01 −0.3000.1910.116
DD-DC−82.9984.956< 0.01 −0.7320.245< 0.01
ES-DC62.0516.356< 0.01 −1.2540.357< 0.01
FC-DC−22.3995.485< 0.01 −0.2640.2580.307
FH-DC−53.8235.463< 0.01 −0.4630.2480.063
MB-DC8.1865.4530.134−0.4460.2230.046
ES-DD145.0485.295< 0.01 −0.5220.3480.135
FC-DD60.5983.746< 0.01 0.4680.2060.024
FH-DD29.1743.776< 0.01 0.2690.1980.174
MB-DD91.1833.731< 0.01 0.2860.1590.073
FC-ES−84.4505.379< 0.01 0.9900.336< 0.01
FH-ES−115.8745.074< 0.01 0.7910.3120.012
MB-ES−53.8655.204< 0.01 0.8080.3160.011
FH-FC−31.4244.366< 0.01 −0.1990.2120.348
MB-FC30.5854.359< 0.01 −0.1820.1800.313
MB-FH62.0094.300< 0.01 0.0170.1680.919
Table 5. Predictive performance of the EOLLN models, decision tree, and random forest according to the metrics: R 2 , RMSE, and VAF.
Table 5. Predictive performance of the EOLLN models, decision tree, and random forest according to the metrics: R 2 , RMSE, and VAF.
ModelRMSER 2 VAFModelRMSER 2 VAF
DT ( K = 5 )25.3950.76376.75DT ( K = 10 )25.6720.75275.918
RF ( K = 5 )26.0550.74975.535RF ( K = 10 )25.9770.74274.904
EOLLN ( K = 5 )29.0890.68376.359EOLLN ( K = 10 )28.6670.68576.714
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rodrigues, G.M.; Ortega, E.M.M.; Cordeiro, G.M. New Partially Linear Regression and Machine Learning Models Applied to Agronomic Data. Axioms 2023, 12, 1027. https://doi.org/10.3390/axioms12111027

AMA Style

Rodrigues GM, Ortega EMM, Cordeiro GM. New Partially Linear Regression and Machine Learning Models Applied to Agronomic Data. Axioms. 2023; 12(11):1027. https://doi.org/10.3390/axioms12111027

Chicago/Turabian Style

Rodrigues, Gabriela M., Edwin M. M. Ortega, and Gauss M. Cordeiro. 2023. "New Partially Linear Regression and Machine Learning Models Applied to Agronomic Data" Axioms 12, no. 11: 1027. https://doi.org/10.3390/axioms12111027

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop