Cyclones are well-known natural catastrophes and intense weather events that inflict significant damage and loss of life. Predicting intensity properly and efficiently is critical in wind engineering to prevent and mitigate cyclone-induced disasters. However, inconsequential cyclone power or wind speed phases become difficult to assess and, as a result, tough to anticipate with greater accuracy and in less time using current approaches. Then, a unique deep-learning algorithm, known as LEGEMP, correctly and swiftly forecasts cyclone intensity.
To decrease the time complexity of data categorization, the feature selection procedure must be carried out initially to discover which features are significant. The suggested approach selects significant features using a Herfindahl correlative linear support vector regression and excludes irrelevant features to improve the performance of cyclone intensity prediction in less time. Herfindahl correlative linear support vector regression is a machine-learning approach that helps identify the essential characteristics to improve cyclone intensity prediction while requiring less processing time.
Secondly, hybridization of the Nesterov gradient descent Jaccardized deep multilayer perceptive classifier is employed to enhance cyclone intensity prediction performance with minimum time by estimating the testing and training parameters using the Jaccard similarity coefficient. Then, the soft step activation is applied to analyze the relevant value and predict the various cyclone intensities: minimum pressure, depression, deep depression, cyclonic storm, severe cyclonic storm, severe cyclonic storm, and super cyclonic storm. Finally, the Nesterov gradient descent method is applied to minimize the error by updating the weights. These two functions of the proposed LEGEMP technique are expressed in the following subsections.
3.1. Herfindahl Correlative Linear Support Vector Regression-Based Feature Selection
Feature selection is the most important machine-learning model, which greatly impacts the production of the classification model. Further features are supplied to the learning algorithm for a whole classification task. However, it is frequently the case that common features are irrelevant or redundant to the learning task, which minimizes the algorithm’s performance and leads to the problem of achieving better accuracy in the prediction process. As a result, it is essential to choose the relevant, necessary features and eradicate the unnecessary features from the dataset.
The proposed LEGEMP technique uses Herfindahl correlative linear support vector regression to find the relevant features. Support vector regression is a machine-learning technique for estimating the relationships with variables, i.e., features. It is linear regression used to find and closely fit the features according to an exact mathematical criterion.
Figure 2 illustrates the feature selection process for identifying the significant features with minimum time using Herfindahl correlative linear support vector regression. Initially, the number of features and data is taken from the dataset. Then, input is given to the linear support vector regression model to identify the data with selected features with the help of the Herfindahl correlation index. It is a statistical method used to find the correlation between the relevant and irrelevant features, as depicted in
Figure 3.
Initially, the features and the corresponding values were arranged in the form of a matrix with rows and columns, as given below:
From (1), a set of features, like relevant and irrelevant data, is represented as a matrix ‘F’ in which each column represents features and a row represents the feature data. In linear support vector regression, applying hyperplane is constructed for separating the original input vector space into two sets. After that, compute the relationship between features using the applied Herfindahl correlation function. The two marginal hyperplanes are constructed to split the input features into two sets according to the relationship between relevant and irrelevant features. All the subscripts of ‘f’ represent the rows and columns of features.
The regression model considers the set of training samples
, where ‘
’ represents the number of features matrix and ‘
‘represents the output. The classification result provides two classes as
, where
indicates the relevant features set and
indicates the irrelevant features set. The support vector model uses the hyperplane to find the relevant or irrelevant feature set. The hyperplane is a decision boundary between the two sets, where the features are separated on either side of the decision boundary with the help of the Herfindahl correlation index. The Herfindahl correlation index that is used to calculate the relationship between features is measured as
where
indicates a Herfindahl correlation index, ‘
’ indicates many features, and
represents a statistical variance.
where
denotes features. The correlation index provides the output value from 0 to 1. The hyperplane separates the relevant and irrelevant features above or below based on the correlation output. If the correlation results are higher, the hyperplane separates the relevant features above the decision boundary. If the correlation result is lesser, the hyperplane separates the irrelevant features below the decision boundary.
The output of support vector regression is given below:
In the above Equation (4), the positive results provide the maximum relationship between features, and the negative results present a lesser relationship between the features.
Figure 3 illustrates the yield of the linear support vector regression model. As represented in the figure,
denotes a hyperplane, i.e., decision boundary, and
and
denote a marginal between the two classes. These two margins are called the support vectors. The relevant features are separated into positive (i.e., +1), and irrelevant features are in negative classes (−1). The algorithm of the Herfindahl correlative linear support vector regression model is as follows. Algorithm 1 describes the different processes of data classification using the Herfindahl correlative linear support vector regression model for achieving better prediction outcomes with minimum time. As the input, the number of features and data is obtained. The original vector is then divided into two sets using the separating hyperplane in the regression model. The features are separated on either side of the hyperplane at every instant.
The Herfindahl correlation index is used to estimate the connection between the characteristics. When the link between the data is the strongest, the input characteristics are classified above the hyperplane. Otherwise, the input characteristics are grouped beneath the hyperplane. As a result, the necessary characteristics are identified, reducing the time complexity.
3.2. Nesterov Gradient Descent Jaccardized Deep Multilayer Perceptive Classifier for Cyclone Intensity Prediction
The second process of the proposed LEGEMP technique is to carry out the cyclone intensity prediction with selected features. The proposed technique uses the Nesterov gradient descent Jaccardized deep multilayer sensitive classifier with selected features. A deep multilayer perceptive classifier is a fully connected, feed-forward ANN that transforms any input dimension to the desired output. The multilayer deep perceptive classifier’s main advantage over the other deep-learning techniques is to resolve complex nonlinear issues. It also handles huge amounts of input data and provides accurate predictions with less time after training.
In Algorithm 1, Herfindahl correlative linear support vector regression employs a linear approach within the framework of the Herfindahl index, enhancing regression accuracy by capturing market concentration. It optimally balances support vectors to model correlations, aiding effective predictions in economic analysis.
// Algorithm 1: Herfindahl correlative linear support vector regression |
Input: cyclone dataset, number of features Output: select significant features |
Begin- 1
For each feature of fi from the dataset - 2
Construct feature matrix ‘F’ - 3
Construct hyperplane H - 4
Find two marginal hyperplanes m1, m2 - 5
Measure the Herfindahl correlation index ‘HCl’ - 6
If the relationship between (fi, fj) is high - 7
Y = +1 - 8
the feature is said to be a relevant feature - 9
else - 10
Y = −1 - 11
the feature is said to be an irrelevant feature - 12
End if - 13
End for - 14
Return (relevant features)
|
End |
Explanation of the Algorithm 1
A loop iterates over each feature in the dataset, denoted as fi.
A matrix F is created that contains the dataset’s features. Each row of F represents an instance, and each column represents a feature.
A hyperplane is established in the feature space, a subspace with one dimension less than the ambient space, used for classification or regression.
Two marginal hyperplanes (m1 and m2) are identified, potentially significant for further analysis or classification.
The Herfindahl correlation index (HCI), a measure quantifying the correlation between features, is calculated to assess feature distribution concentration or dispersion.
A conditional statement checks the HCI to evaluate the relationship between features fi and fj.
If HCI indicates a high relationship between features fi and fj, the variable Y is set to +1, marking fi as relevant.
If the relationship between features fi and fj is low (i.e., low correlation), the variable Y is set to −1, indicating fi as irrelevant.
The algorithm concludes with the end of the if–else statement, returning a list of relevant features based on the assessments made during the loop.
The perceptrons can process the given weighted sum of inputs, apply the activation function, and provide the final classification output. The relationship between perceptrons or neurons is known as the synapse. The perceptrons are a simplified computational model inspired by neurons, aiming to mimic certain aspects of neural behavior for specific computational tasks. In contrast, neurons are highly complex and specialized cells in biological nervous systems with many functions and capabilities.
Figure 4 illustrates the structural design of a deep multilayer perceptive learning classifier that includes an input layer, more than one output layer, and one hidden layer. The input layer receives the selected features input features of the cyclone with their data. Each datum is connected with a set of weights ‘
’ and added with bias ‘
’. The activity of the neuron is formulated as given below:
where ‘
’ specifies a neuron’s action at the input layer that weighted ‘
’ sum of the input data ‘
’ and adds to bias function ‘
’ stored, the value is ‘1’. Then, the input is transferred into the first hidden layer, where the deep-learning process is performed. More than one hidden layer is presented in the input and output layers. The hidden layer involves the small individual units called neurons or nodes. The motion of artificial neurons at the hidden layer is illustrated in
Figure 3.
Figure 5 illustrates a process of artificial neuron activity that receives the weighted sum of features added with the bias. Then, the input data were analyzed using the cyclone’s testing parameters and the Jacquard similarity method. The Jacquard coefficient is a statistical method used to compute the similarity between two sets of data, namely training data with testing parameters of the cyclone. The similarity is estimated as given below:
where
denotes a similarity coefficient,
represents the training parameters of the cyclone,
represents the testing parameters value of the cyclone, and
denotes a mutual dependence between the data.
The coefficient () returns the output value from 0 to 1. The similarity outcomes are given to the input of the activation function for analyzing the value and providing the different intensity levels of the cyclone. The proposed deep-learning classifier uses the soft step activation role for providing final intensity prediction results.
The soft step activation metrics assist the network in learning complex testing and training patterns in data. The soft step activation function provides the best-normalized output between 1 and 0. It makes an accurate prediction. It also provides a bounded absolute value.
where
represents a soft step activation, and ‘
’ indicates the similarity outcomes. The soft step activation function provides the final classification results.
The soft step activation provides ‘1’, indicating that the cyclone intensity level is correctly predicted based on the similarity between testing and training parameter values. In the learning process in perceptron, the error rate is measured between the predicted results and the target output. The error rate is calculated as follows:
where
represents the error rate,
indicates the target prediction outcomes, and ‘
’ denotes a predicted cyclone intensity output produced with the deep multilayer perceptive classifier. The weight is updated using the Nesterov accelerated gradient descent method to minimize the error.
where
where
is an updated weight,
is a current weight, and
denotes a learning rate
. A huge learning rate permits the deep classifier to learn faster than a smaller value; ‘
’ is a partial derivative of the error ‘
’ concerning current weight. ‘
’,
default value is
, and
is initialized at 0. The technique is repeated until the smallest error is found. Finally, the output layer of the multilayer perceptive classifier produces correct prediction results.
The different processes of cyclone intensity prediction with better accuracy and less time. Nesterov gradient descent using Jaccardization, a deep multilayer perceptive classifier, uses a variety of layers to assess the input data. The deep-learning classifier’s input layer is supplied with the chosen pertinent features and their corresponding data. Then, the weight is assigned to each input and added with the bias.
In Algorithm 2, Nesterov gradient descent refines the deep multilayer perceptive classifier for cyclone intensity prediction, integrating Jaccard similarity. This augments accuracy by adapting learning rates, ensuring efficient convergence, and refining the forecasting of cyclone strength, using advanced neural network architectures.
// Algorithm 2: Nesterov gradient descent Jaccardized deep multilayer perceptive classifier-based cyclone intensity prediction |
Input: Output: Improve the cyclone intensity prediction |
Begin- 1
with the data are collected at the input layer - 2
’ - 3
’ and add bias ‘’ - 4
’ - 5
End for - 6
For each training data with testing parameter data –[hidden layer] - 7
’ - 8
’ - 9
then - 10
Correctly predict the type of cyclone intensity - 11
End if - 12
End for - 13
For each prediction outcome - 14
- 15
- 16
’ - 17
- 18
Obtain the final prediction results with minimum error at the output layer - 19
Return (accurate cyclone intensity prediction) - 20
End for
|
End |
The subsequently hidden layer neuron receives the weighted input after that. The Jacquard index is applied to compute the similarity between the input data and the testing cyclone parameter value in that layer. Then, the estimated similarity value is given to the soft step activation function at the hidden layer. The activation function evaluates the similarity value and outputs either a ‘1’ or a ‘0’. The cyclone intensity level is correctly predicted if the output is ‘1’. Otherwise, the activation function returns ‘0’. The error rate is based on the squared difference between the target and the predicted output for each result. The initial weight is updated by applying a Nesterov accelerated gradient descent method to minimize the error. This process is continuously iterated until the algorithm reaches minimum error. Finally, the accurate cyclone intensity prediction results are displayed at the output layer.