Next Article in Journal
Parametric Study of the Hydrodynamic Characteristics of the Pumpjet Propulsor for the SUBOFF Submarine
Previous Article in Journal
Role of Shape and Kinematics in the Hydrodynamics of a Fish-like Oscillating Hydrofoil
Previous Article in Special Issue
Influence of Van Gent Parameters on the Overtopping Discharge of a Rubble Mound Breakwater
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Predictions of Wave Overtopping Using Deep Learning Neural Networks

1
Bachelor’s Program of Precision System Design, Feng Chia University, Taichung 407, Taiwan
2
Department of Civil Engineering, National Chung Hsing University, Taichung 402, Taiwan
*
Authors to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2023, 11(10), 1925; https://doi.org/10.3390/jmse11101925
Submission received: 7 September 2023 / Revised: 28 September 2023 / Accepted: 1 October 2023 / Published: 5 October 2023
(This article belongs to the Special Issue Advanced Studies in Breakwaters and Coastal Protection)

Abstract

:
Deep learning techniques have revolutionized the field of artificial intelligence by enabling accurate predictions of complex natural scenarios. This paper proposes a novel convolutional neural network (CNN) model that involves deep learning technologies, such as the bottleneck residual block, layer normalization, and dropout layer, to predict wave overtopping at coastal structures under a wide range of conditions. To optimize the performance of the CNN model, the hyperparameter tuning process via Bayesian optimization is used. The results of validation demonstrate that the proposed CNN model is highly accurate in estimating wave overtopping discharge from hydraulic and structural parameters. The testing accuracy of the overtopping predictions using a prototype dataset shows that the proposed CNN model outperforms those existing machine learning models. An example application of the CNN model is presented for predicting prototype overtopping considering various crest freeboards of coastal structures.

1. Introduction

The accurate assessment of wave overtopping discharges is crucial for the appropriate design of coastal structures that can prevent severe coastal flooding. Wave overtopping at coastal structures has been studied extensively through various methods, including field investigations, scale model measurements, and numerical simulations [1,2,3,4,5,6,7]. The CLASH (Crest Level Assessment of Coastal Structures) project [8] conducted numerous physical modeling and prototype investigations of wave overtopping discharge at various types of coastal structures. Various empirical formulas are included in the EurOtop manual [9] for the estimation of wave overtopping discharge. On the other hand, robust artificial neural network (ANN) models, utilizing hydraulic and structural parameters sourced from either the new EurOtop database [9] or the original CLASH database [10] as input variables, have demonstrated efficient performance in predicting wave overtopping discharge at coastal structures [4,7,11,12,13,14]. Each test in the EurOtop database is characterized by a reliability factor (RF) and complexity factor (CF) [10], which are used to select reliable data for training and improving ANN models.
Previous studies have proposed a variety of multilayer architectures for ANN models in the field of machine learning. Verhaeghe et al. [7] proposed a two-phase neural network model, including a classifier–quantifier scheme, to distinguish negligible from significant overtopping and avoid large overpredictions in areas of low overtopping. However, Zanuttigh et al. [11] argued that classifier–quantifier schemes do not truly improve ANN performance and instead increase the complexity of the ANN architecture while creating undesirable discontinuities in predictions. den Bieman et al. [12] presented a new model using an advanced machine learning technique called XGBoost (XGB) associated with feature engineering, which generally outperforms conventional ANN models and empirical formulas in terms of prediction accuracy. Bootstrap resampling techniques were employed in these machine learning models to evaluate the uncertainties associated with their predictions.
Accurately predicting wave overtopping discharge in complex scenarios remains a challenging task. As an alternative to machine learning artificial neural networks (ANNs), deep convolutional neural networks (CNNs) present a promising avenue to tackle this issue. Originally developed for identifying the characteristics of individual images from large datasets, CNNs have found applications in big data analysis and decision making [13,14]. Utilizing deep convolutional neural networks for the prediction of wave overtopping discharge represents a relatively recent research area, yet it has already exhibited significant potential. Through convolutional processing for feature extraction, CNNs can effectively capture the distinctive attributes of data representations and employ filters to accentuate the respective importance of each output task. As a result, CNNs may offer significant improvements in the prediction accuracy of wave overtopping discharge.
This article presents a CNN model using a serial bottleneck residual block architecture that incorporates convolution layers, layer normalization, and a dropout layer. The model is designed to determine the relevance of hydraulic and coastal structure inputs for accurately predicting wave overtopping discharge. This paper is organized as follows. Section 2 describes the new EurOtop database which is applied to train the CNN model. Then, we explain the model expression, network layer architecture, and hyperparameter optimization process used for training the CNN model. As applied in previous ANNs, the bootstrap resampling technique is also introduced in training the CNN model. Next, we discuss the accuracy of the CNN model predictions and compare them with those of existing ANN models. An example application of the model for predicting prototype overtopping is also presented. Finally, the conclusions of the study are addressed.

2. EurOtop Database

The proposed CNN model was trained using data from the new EurOtop database [9], which is an extensive dataset containing over 17,000 tests. This dataset includes nearly 13,500 records specifically related to wave overtopping discharge. This new EurOtop database is an extension of the original CLASH database, which was compiled from approximately 10,000 schematized tests conducted worldwide. By utilizing this comprehensive dataset, the CNN model can learn from a wide range of test scenarios and improve its predictive capabilities for wave overtopping discharge. Figure 1 depicts the general geometric and relevant hydraulic parameters of coastal structures. Full descriptions of the parameters, wave conditions, and structural cross sections in the database are provided in the literature [9,11,12,15,16]. The overtopping experiments in the database are categorized into seven groups based on the structure type and oblique wave attack conditions [9]; they are rock permeable straight slopes (group A), rock impermeable straight slopes (group B), armor units with straight slopes (group C), smooth and straight slopes (group D), structures with combined slopes and berms (group E), vertical walls (group F), and oblique wave attacks (group G).
The parameters selected for training the CNN model are listed in Table 1 and comprise three hydraulic parameters of waves and thirteen structural parameters as inputs and one parameter (wave overtopping discharge) as the output. The training dataset of the neural network model consists of data derived from laboratory tests conducted at various scales. To ensure the applicability of the model to prototype scenarios, the use of dimensionless parameters is advantageous. Following [9,11,16], the significant wave height and water depth at the structure toe are represented as the wave steepness Hm0,t/Lm−1,0,t and the relative water depth h/Lm−1,0,t, respectively, by dividing the wavelength Lm−1,0,t, in which Lm−1,0,t is calculated by 1.56T2m−1,0,t, and Tm−1,0,t is the spectral wave period at the structure toe. The parameters Bt, Gc, and B describing horizontal measures of structure widths are normalized by the wavelength Lm−1,0,t and the parameters ht, Du, Dd, Ac, Rc, and hb describing vertical measures of structure heights are normalized by the wave height Hm0,t. By incorporating dimensionless parameters, the neural network model can effectively capture the underlying physics and behavior of wave overtopping, regardless of the specific scale of the experiments.
Following [11,16], the wave overtopping discharge q in the database is scaled by a normalized formula for training, which is defined as:
q s = l o 𝑔 10 q a d m i n l o 𝑔 10 q a d m i n l o 𝑔 10 q a d m a x l o 𝑔 10 q a d
in which q a d is defined as
q a d = q 𝑔 H m 0 , t 3
where q is the original mean wave overtopping discharge value from the EurOtop database and 𝑔 is the gravitational acceleration (9.8 m/s2). Equation (2) represents a dimensionless formula known as the relative wave overtopping discharge. As elucidated in [13], the application of Equation (1) for normalization confines the fitted data within a range of 0 to 1, thereby mitigating the scale-dependent nature of the error. Subsequently, the predicted output value qs generated by the trained CNN is reverted to obtain the corresponding value of q.
A weight factor WF = (4 − RF)⋅(4 − CF) defined in [4] was adopted in previous works for the selection of reliable data to minimize the influence of environmental uncertainties and irregular structures. Higher WFs for the overtopping database indicate greater reliability and simpler structural schemes. As the same in earlier prediction methods [11,16], excluding the data of very unreliable (RF = 4) or complex (CF = 4), q = 0 , and missing values, a total of 8653 records labeled as core data in the overtopping database are used in the proposed CNN model. It is noted that the core data excluded the dataset of prototype overtopping.

3. Method Description

3.1. Model Description

In this section, we present how the proposed CNN model learns the inputs of the hydraulic and structural parameters corresponding to the output of wave overtopping discharge. The overtopping database contains many different structure types and related input parameters, but they may have the same distributions in the output parameters, as known as a covariate shift problem. The general neural network model may require specific feature engineering to select appropriate input parameters, preventing the model from possibly misclassifying the data. Alternatively, the CNN gains experience by observing a training set of input data and predicting the output without any feature assumptions. Accounting for this problem, the convolutional layer [13], layer normalization [17], and a dropout layer [18] are used for the proposed CNN, described as follows.
Convolutional Layer: This layer operates by traversing a compact filter across the input data, enabling the detection of patterns and distinctive features. The equation of the convolutional layers is expressed as follows:
x l = w · x l 1 + b = τ = 0 T w τ l x t τ + ϕ l 1 + b t
where the superscript l represents the index of each neutral network layer and t denotes the index of the input parameters. In Equation (3), x l 1 is a given input dataset, w is a set of one-dimensional convolutional filter coefficients, and ϕ denotes the shift number. Equation (3) can be described as the area under the function w[τ], which is weighted by the function x[−τ], which is shifted by an amount t. As t changes, the weighting function x[tτ] emphasizes different components of the input weighting function w[τ] to extract the significant features determined by the filter matrices. b[t] is the bias coefficient as the constant of a linear function.
Layer normalization layer: The main objective of layer normalization is to reduce the internal covariate shift that occurs during training, which helps stabilize the distribution of the inputs to a layer. Layer normalization works by normalizing the values within a layer, making their mean close to zero and their variance close to one. This normalization can speed up training and allow for the use of larger learning rates. Additionally, it helps with better gradient flow during backpropagation, which can lead to faster convergence and more stable training.
Dropout layer: It is used to prevent overfitting in neural networks. During training, a dropout layer randomly disables a fraction of the neurons or units in the layer by setting them to zero for that particular pass. By introducing noise and preventing neurons from relying too heavily on any particular input, dropout promotes learning more robust and generalized features. This regularization technique helps prevent overfitting and improves the generalization of neural networks, making them more capable of performing well on unseen data.
The proposed CNN consists of an input layer for inputting sequence data such as the parameters shown in Table 1, hidden layers, and an output layer as the fully connected layer to calculate a weighted sum of its inputs and is used to make predictions qs. In any feed-forward neural network, the middle layers are referred to as hidden layers because their inputs and outputs are masked by the activation function and final convolution. To make a deeper CNN model, a bottleneck residual block [19] involving convolution layers, layer normalization, and a dropout layer are applied to mitigate degradation (accuracy saturation) and the overfitting problem [20] caused by combining many hidden layers. It may also accelerate learning for our proposed CNN. For the architecture of the overall connection model with the aforementioned network representations, the bottleneck residual block adds a route for simple addition into each hidden layer connection as shown in Figure 2.

3.2. CNN Training with Hyperparameter Optimization

To obtain an optimal model, it is common practice to limit the number of learning iterations or set a stopping threshold for the objective loss function in order to prevent overfitting. In this study, we employ the hyperparameter optimal tuning process (HOTP) [21,22] to minimize the objective loss function and mitigate overfitting [20]. The objective loss function combines the mean square loss function [23] and the Wasserstein distances W [24] between two measures P q ^ and P q , which are the cumulative distribution functions of the model predictions q ^ and the observed values q, respectively. The objective loss function L can be mathematically defined as follows:
L = λ 1 1 I i = 1 I q ^ i q i 2 + λ 2 W s P q ^ , P q
In Equation (4), I is the subsample number to be randomly extracted from the training set. Each summand L is associated with the n-th observation in the subsample during training. Here, λ 1 and λ 2 are constraints on the sensitivity of the squared error loss and Wasserstein distances W s P q ^ , P q . To train the CNN, the adaptive movement estimation algorithm, a backward propagation-based iterative method developed by Adam [25], is employed for optimizing the objective function with desirable smoothness properties. The learning rate is one of the key hyperparameters when training a deep learning neural network model. It is determined through HOTP to find the value that results in the best model performance.
Figure 3 provides an overview of the hyperparameter optimization (HOTP) process utilized to develop an optimal CNN model. The primary objective of HOTP is to identify the hyperparameters that yield the best performance of the CNN model, while avoiding overfitting. It is important to emphasize that the validation dataset used during HOTP is distinct from the testing dataset and is not part of the training process.
In each iteration of HOTP, a new set of hyperparameters is evaluated by computing the associated loss using the validation dataset. The objective function aims to minimize this loss, enabling the identification of optimal hyperparameters that lead to a model with the lowest possible loss. Bayesian optimization [26] is employed in this study to conduct an exhaustive search of the hyperparameter space, focusing on a predetermined subset of hyperparameters. The optimization process involves exploring different combinations of hyperparameters based on previous observations, with the goal of refining the model's performance.
Once the CNN model is trained using HOTP and the optimal hyperparameters are determined, it can be utilized to make predictions for overtopping values. This comprehensive approach ensures that the CNN model is fine-tuned to achieve optimal performance and can provide accurate predictions for the given task.
To configure hyperparameters for a CNN, first we define the network's architectural aspects according to Figure 2 and initialize the corresponding hyperparameters randomly, such as the number of layers, filters, and size of neurons. Secondly, set general hyperparameters such as the neurons of convolution layers and fully connected layers, number of residual blocks, learning rate, batch size, dropout rate, and weighting coefficients of the loss function to fine-tune the training process. After establishing these foundational settings, fine-tune the hyperparameters with HOPT on a validation set. Table 2 displays a hypothetical interval of the hyperparameter space used during the HOTP process. The aim of HOTP is to identify the optimal values for these hyperparameters, which are crucial for the performance of the CNN model. Within Table 2, the optimal values of the hyperparameters are denoted by a star (*) symbol, indicating the configurations that yield the best performance for the CNN model.

4. Results and Discussion

4.1. Training Setup

For the CNN model in this study, the training process was carried out with specific configurations. In the context of neural network training, an epoch refers to a complete iteration over the entire training dataset. During an epoch, the model goes through multiple iterations, each involving a batch of training samples. The number of iterations per epoch is determined by the batch size. In this study, each batch size was limited to a maximum size of 256, while the remaining samples constituted the minimum quantity for the last batch size. The training was performed with a maximum of 5000 epochs, and the early stopping algorithm was employed to prevent overfitting. This algorithm continuously monitored the validation performance and stored the model with the best results when a predefined global training accuracy was achieved.
The training procedure took place on a computer equipped with a 3.4 GHz Intel Core i7-6700 CPU and two NVIDIA GeForce GTX 1080 Ti GPUs. The implementation of the CNN model was conducted using the Python programming language in conjunction with the TensorFlow library [27]. This setup ensured efficient training and allowed for the attainment of optimal outcomes.
In this study, the entire core dataset of the EurOtop overtopping database was utilized, following the approach outlined in [12]. The database was randomly divided into two portions: 80% for training the CNN model and 20% for validation. Similar to previous machine learning models [4,11,12,16], we employed the bootstrap resampling technique for the training of the CNN model. The bootstrap resampling technique involves creating N resamples from the original database. Each resample forms a distinct training and validation dataset. Subsequently, the CNN model is trained using each resample dataset, and the predicted overtopping discharge is obtained by calculating the ensemble mean of the predictions. This technique allows for the evaluation of the model’s performance under various training conditions and provides a robust estimation of its predictive capabilities.
When dealing with a substantial volume of data, increasing the number of iterations in bootstrap resampling typically results in more precise estimates derived from these resamples [28]. However, in the context of the overtopping database, which contains relatively modest-sized samples compared to fields with datasets comprising hundreds of thousands of data points or more, excessively increasing the number of bootstrap resampling iterations may not yield substantial improvements in predicting future outcomes. Moreover, this practice can potentially create a deceptive sense of certainty if each model within the ensemble exhibits notable levels of uncertainty.
In cases where a model already displays minimal variance in its prediction outcomes, reducing uncertainty can be accomplished by examining only a limited number of bootstrap samples. This approach not only aids in reducing training time but has also proven to be effective. Therefore, it becomes crucial to determine the appropriate number of bootstrap resamples required to construct a CNN model ensemble. In this study, we evaluate the model’s performance across various numbers of bootstrap resamples, ranging from N = 1 to N = 500. It is noteworthy that the scenario with N = 500 aligns with the same number of bootstrap resamples used in [4].
For the assessment of predicted quality, we employ the weighted root-mean-square error (RMSE) as the error criterion, as defined in [12]. The weighted RMSE is calculated using the following formula:
R M S E = 1 k K W F k k = 1 K W F k · l o 𝑔 10 q ^ k l o 𝑔 10 q k 2
where K is the observation number, and q k and q ^ k are the observed and predicted values, respectively. This error criterion takes into account the weights assigned to different variables or data points, allowing for a more comprehensive evaluation of the prediction accuracy. By considering the weighted RMSE, we can assess the performance of the model in a more nuanced and representative manner.
In Figure 4, the variation of the weighted root-mean-square error (RMSE) is depicted from the ensemble mean with bootstrap resample numbers ranging from N = 1 to N = 500, along with the corresponding running time. It can be observed that as the bootstrap resample number increases, the RMSE decreases. The RMSE value converges at N = 10, as shown in Figure 4. Table 3 compares the RMSE values for N = 1, 10, and 500, indicating that satisfactory accuracy can be achieved using N = 10. It is worth noting that the computational time is significantly different between N = 10 (44 seconds) and N = 500 (2155 seconds). This finding suggests that employing N = 10 bootstrap resamples balances accuracy and efficiency for the predictions. Figure 5 displays the predictions made by the trained CNN model using N = 10 for both the training and validation datasets. The proposed CNN model incorporated a strategy to mitigate overfitting, which was validated by the strong performance observed in Figure 5, where the training and validation results exhibited good agreement.

4.2. Verification on the Overtopping Database

Following the precedent of earlier ANN models [4,11,14], the trained CNN model is assessed for accuracy using the entire core dataset from the overtopping database. The trained CNN model can be accessed online at https://github.com/fcuyttsai/overtoppingEve.git/ (accessed on 1 September 2023). The predictions of existing ANNs can be directly estimated from their respective websites. These existing ANN models are the NN model [12] (online at https://www.deltares.nl/en/software/overtopping-neural-network/ (accessed on 1 May 2023)), the NNb model [11] (online at http://overtopping.ing.unibo.it/overtopping/ (accessed on 1 May 2023)), and the XGB model [12] (online at https://www.deltares.nl/en/software/overtopping-xgb/ (accessed on 1 May 2023)). In comparison to the deep learning CNN model, the NN, NNb, and XGB models fall within the realm of machine learning techniques. The NN model [4] employs the conventional back-propagation neural network as its learning method. Meanwhile, the NNb model [11], which utilizes an extended CLASH database as its training dataset, builds upon the fundamental NN model concept but integrates a quantitative classifier and three quantifiers. On the other hand, the XGB model [14], as an alternative to both the NN and NNb models, leverages an advanced machine learning technique known as XGBoost to predict overtopping discharge.
In Table 4, we present a comparison of the resulting weighted root-mean-square errors (RMSEs) for wave overtopping discharge predictions, along with the Pearson correlation coefficient (CC) and R2 as measures of agreement. The Pearson correlation coefficient (CC) is given by:
C C = i = 1 n x i x ̄ y i ȳ n 1 s x s y
where x i is the i-th value of observation, y i is the i-th value of prediction, x ̄ is the mean value of observation, ȳ is the mean value of prediction, s is the standard deviation of the observation, and s is the standard deviation of the prediction. R2 is the square value of CC that is the coefficient of determination.
As a result of the comparison, the RMSE value for the entire core dataset is 0.112 for the CNN model and 0.199 for the trained XGB. These values are significantly smaller than the RMSE values obtained from the trained NN (0.654) and NNb (0.639) models. Additionally, the CC and R2 values further demonstrate that both the present deep learning model (CNN) and advanced machine learning model (XGB) outperform the conventional machine learning models (NN and NNb). The corresponding quantile–quantile plot of wave overtopping discharge predictions for each model is displayed in Figure 6.
To demonstrate the adaptability of the current CNN model to different scenarios, we directly extract the results of Figure 6 for the groups A–G of the EurOtop database, based on structure type and oblique wave attack conditions. The predictions of each model are plotted in Figure 7a–g. The scatterplots in Figure 7a–g show a high correlation between the predictions and observations for each group using the CNN model. The results indicate that all models perform relatively well for group D, which represents the situation of coastal structures with smooth and straight slopes. The CNN model also performs better for group B data, despite it containing only a few observations compared to other categories. Regarding the oblique wave attack prediction (group G), it is noted by [12] that the low accuracy of the NN models may be due to the limited amount of training data, leading to an improvement in the XGB model by incorporating new oblique wave cases into the training data. However, as shown in Table 4, the group G dataset comprises 1107 cases, which is larger than one-eighth of the total dataset. It can be observed that the current CNN model achieves highly accurate oblique wave attack (group G) predictions without the need for additional datasets.

4.3. Testing/Prediction Using Prototype Overtopping Dataset

Designers expect prediction models to be applied to prototype scenarios. The testing performance of the trained CNN model is evaluated using the prototype overtopping dataset, which was not seen during the model’s training and validation processes. The EurOtop database provides 150 wave overtopping measurement records from prototype situations. These records include 11 measurements labeled as C-59.1 to C-59.11 [28] from 9 storms between 1999 and 2004 at Zeebrugge breakwater, 77 measurements labeled as G-18.1 to G-18.77 [29] from 7 storms between 2003 and 2004 at Ostia breakwater, and 62 measurements labeled as F-18.1 to F-18.39 and G-12.1 to G-12.23 [30] from other sources.
The testing results of the trained CNN model are depicted in Figure 8, where the predictions of the NN, NNb, and XGB models are estimated directly from their respective websites using the same data inputs. It is important to note that all these models utilized the entire core dataset (i.e., laboratory measured data) during the training and validation processes, with the prototype dataset remaining unseen. As illustrated in Figure 8 and Table 5, the CNN model exhibits better performance in testing when applied to the prototype dataset. It is worth noting that all models achieve high correlation coefficient (CC) values. However, due to the larger scale of the prototype, the RMSE values reported in Table 5 are understandably larger than those observed in the laboratory-scale data presented in Table 4. Furthermore, real-world prototype scenarios can introduce unanticipated variables such as noise, surface irregularities, wind, spray, currents, and the scale effect itself. These sources of uncertainty can contribute to more pronounced prediction errors.
The prototype overtopping database can be divided into three categories: armor units with straight slopes (C-59.1 to C-59.11), vertical walls (F-18.1 to F-18.39), and oblique wave attacks (G-12.1 to G-12.23 and G-18.1 to G-18.77). Figure 9 presents the testing results for these three types of datasets, extracted directly from the results of Figure 9. As mentioned in [11], their model (i.e., NNb model) underestimated the predicted values of prototype overtopping for the Ostia breakwater (i.e., G-18.1 to G-18.77), while the predictions for Zeebrugge (i.e., C-59.1 to C-59.11) fell within the confidence bands. Figure 9a–c show similar tendencies observed in [11] by the NNb model.

4.4. Application of CNN for Prototype Various Crest Freeboards

In an example scenario, each model was applied to predict the prototype wave overtopping discharge, considering various crest freeboards (Rc) of coastal structures. The predictions were based on the observation labeled as G-12.22 in the prototype situations of the EurOtop database. The input conditions are presented in Table 6. The predictions started with Rc = 3.34 m and gradually increased to 7.50 m to observe the trend of the wave overtopping discharge, with a measured observation available at Rc = 7.36 m.
In Figure 10, the depicted results illustrate the predicted outcomes of relative wave overtopping discharge concerning the relative crest freeboard for four different models: the current CNN, XGB, NN, and NNb models. These results reveal a consistent trend across all predictions, where a lower relative crest freeboard corresponds to a larger relative wave overtopping discharge. Along the projected curve, the predictions generated by the current CNN model closely align with the observed measurements at Rc = 7.36 m, specifically at Rc/Hm0,t = 4.41. In contrast, the predictions produced by the NN, NNb, and XGB models notably underestimate the relative wave overtopping discharge for various relative Rc/Hm0,t values.

5. Conclusions

In this paper, a novel convolutional neural network (CNN) model with deep learning technologies is introduced for the accurate prediction of wave overtopping discharge at various types of coastal structures, including those with complex geometries and under a wide range of wave conditions. The proposed CNN model incorporates various hidden layers, such as bottleneck residual blocks, which effectively extract features and scale input sequences. For the enhancement of prediction accuracy, hyperparameters of the model are optimized through Bayesian optimization to mitigate overfitting, and the model is trained utilizing the bootstrap resampling technique. The results demonstrate that the proposed CNN model outperforms traditional machine learning models, as evident from comparisons of quantile–quantile plots and root-mean-square errors for estimations on the overtopping database. Notably, the CNN model exhibits remarkable agreement with observations across various structure types and oblique wave attack conditions. The model also demonstrates satisfactory performance for prototype overtopping prediction, highlighting its value as a valuable tool for coastal engineers involved in the design and maintenance of coastal structures.

Supplementary Materials

Our CNN model can be accessed online at https://github.com/fcuyttsai/overtoppingEve.git/ (accessed on 1 September 2023).

Author Contributions

Conceptualization, Y.-T.T. and C.-P.T.; Methodology, Y.-T.T. and C.-P.T.; Data curation, Y.-T.T. and C.-P.T.; Validation, Y.-T.T. and C.-P.T.; Formal analysis, Y.-T.T. and C.-P.T.; Visualization, Y.-T.T. and C.-P.T.; Software, Y.-T.T. and C.-P.T.; Writing—original draft, Y.-T.T. and C.-P.T.; Writing—review and editing, Y.-T.T. and C.-P.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science and Technology Council of Taiwan for financial support under Grant Nos. MOST 109-2625-M-005-006-MY2, NSTC 111-2221-E-035-044, and MOST 109-2221-E-035-001-MY2.

Data Availability Statement

The authors do not have permission to share data.

Acknowledgments

The authors would like to thank the EurOtop team for providing the overtopping database.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Troch, P.; Geeraerts, J.; Van de Walle, B.; De Rouck, J.; Van Damme, L.; Allsop, W.; Franco, L. Full-scale wave-overtopping measurements on the Zeebrugge rubble mound breakwater. Coast. Eng. 2004, 51, 609–628. [Google Scholar] [CrossRef]
  2. Briganti, R.; Bellotti, G.; Franco, L.; De Rouck, J.; Geeraerts, J. Field measurements of wave overtopping at the rubble mound breakwater of Rome–Ostia yacht harbour. Coast. Eng. 2005, 52, 1155–1174. [Google Scholar] [CrossRef]
  3. Cáceres, I.; Sánchez-Arcilla, A.; Zanuttigh, B.; Lamberti, A.; Franco, L. Wave overtopping and induced currents at emergent low crested structures. Coast. Eng. 2005, 52, 931–947. [Google Scholar] [CrossRef]
  4. van Gent, M.R.A.; van den Boogaard, H.F.P.; Pozueta, B.; Medina, J.R. Neural network modelling of wave overtopping at coastal structures. Coast. Eng. 2007, 54, 586–593. [Google Scholar] [CrossRef]
  5. Losada, I.J.; Lara, J.L.; Guanche, R.; Gonzalez-Ondina, J.M. Numerical analysis of wave overtopping of rubble mound breakwaters. Coast. Eng. 2008, 55, 47–62. [Google Scholar] [CrossRef]
  6. Reeve, D.E.; Soliman, A.; Lin, P.Z. Numerical study of combined overflow and wave overtopping over a smooth impermeable seawall. Coast. Eng. 2008, 55, 155–166. [Google Scholar] [CrossRef]
  7. Verhaeghe, H.; De Rouck, J.; van der Meer, J. Combined classifier–quantifier model: A 2-phases neural model for prediction of wave overtopping at coastal structures. Coast. Eng. 2008, 55, 357–374. [Google Scholar] [CrossRef]
  8. De Rouck, J.; Verhaeghe, H.; Geeraerts, J. Crest level assessment of coastal structures—General overview. Coast. Eng. 2009, 56, 99–107. [Google Scholar] [CrossRef]
  9. EurOtop. Manual on Wave Overtopping of Sea Defences and Related Structures. An Overtopping Manual Largely Based on European Research, but for Worldwide Application; Van der Meer, J.W.; Allsop, N.W.H.; Bruce, T.; De Rouck, J.; Kortenhaus, A.; Pullen, T.; Schüttrumpf, H.; Troch, P.; Zanuttigh, B.: 2018. Available online: www.overtopping-manual.com (accessed on 1 November 2019).
  10. van der Meer, J.W.; Verhaeghe, H.; Steendam, G.J. The new wave overtopping database for coastal structures. Coast. Eng. 2009, 56, 108–120. [Google Scholar] [CrossRef]
  11. Zanuttigh, B.; Formentin, S.M.; van der Meer, J.W. Prediction of extreme and tolerable wave overtopping discharges through an advanced neural network. Ocean Eng. 2016, 127, 7–22. [Google Scholar] [CrossRef]
  12. den Bieman, J.P.; van Gent, M.R.A.; van den Boogaard, H.F.P. Wave overtopping predictions using an advanced machine learning technique. Coast. Eng. 2021, 166, 103830. [Google Scholar] [CrossRef]
  13. Heaton, J. Ian Goodfellow, Yoshua Bengio, Aaron Courville. Deep learning. Genet. Program. Evolvable Mach. 2018, 19, 305–307. [Google Scholar] [CrossRef]
  14. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural. Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  15. Formentin, S.M.; Zanuttigh, B.; van der Meer, J.W. The New EurOtop Neural Network Tool for an Improved Prediction of Wave Overtopping. In ICE Coasts, Marine Structures and Breakwaters; ICE Publishing: Leeds, UK, 2017. [Google Scholar]
  16. Formentin, S.M.; Zanuttigh, B.; van der Meer, J.W. A Neural Network Tool for Predicting Wave Reflection, Overtopping and Transmission. Coast. Eng. J. 2018, 59, 1750006-1–1750006-31. [Google Scholar] [CrossRef]
  17. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer Normalization. arXiv 2016, arXiv:1909.11855. [Google Scholar]
  18. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  19. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  20. Ying, X. An Overview of Overfitting and its Solutions. J. Phys. Conf. Ser. 2019, 1168, 022022. [Google Scholar] [CrossRef]
  21. Isabona, J.; Imoize, A.L.; Kim, Y. Machine Learning-Based Boosted Regression Ensemble Combined with Hyperparameter Tuning for Optimal Adaptive Learning. Sensors 2022, 22, 3776. [Google Scholar] [CrossRef]
  22. Sóbester, A.; Leary, S.J.; Keane, A.J. On the Design of Optimization Strategies Based on Global Response Surface Approximation Models. J. Glob. Optim. 2005, 33, 31–59. [Google Scholar] [CrossRef]
  23. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  24. Mueller, J.; Jaakkola, T. Principal Differences Analysis: Interpretable Characterization of Differences between Distributions. arXiv 2015, arXiv:1510.08956. [Google Scholar]
  25. Diederik, P.; Kingma, J.B. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference for Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  26. Scikit-Optimize. Sequential Model-Based Optimization in Python. Available online: https://scikit-optimize.github.io/ (accessed on 1 May 2023).
  27. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation, Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  28. Geeraerts, J.; Boone, C. Report on Full Scale Measurements Zeebrugge, 2nd Full Winter Season; Ghent University: Gent, Belgium, 2004. [Google Scholar]
  29. Franco, L.; Briganti, R.; Bellotti, G. Report on Full Scale Measurements, Ostia, 2nd Full Winter Season; Modimar: Rome, Italy, 2004. [Google Scholar]
  30. Pullen, T. Final Report on Laboratory Measurements, Samphire Hoe; HR Wallingford: Wallingford, UK, 2004. [Google Scholar]
Figure 1. Schematic of structure based on CLASH, including geometric and hydraulic parameters [9]. Hm0,t: spectra significant wave height at the structure toe; Tm−1,0,t: spectral wave period at the structure toe; h: water depth at the structure toe; ht: toe submergence; Bt: toe width; B: berm width; hb: berm submergence; Ac: crest freeboard of armor structure; Rc: crest freeboard of wall; Gc: crest width of armor structure; α u   a n d   α d   : angles of armor structure slope part above/below the berm; Du and Dd: sizes of the armor elements along up/down slopes; γ f u   a n d   γ f d : roughness factors for up/down slopes; β: incident angle of wave direction.
Figure 1. Schematic of structure based on CLASH, including geometric and hydraulic parameters [9]. Hm0,t: spectra significant wave height at the structure toe; Tm−1,0,t: spectral wave period at the structure toe; h: water depth at the structure toe; ht: toe submergence; Bt: toe width; B: berm width; hb: berm submergence; Ac: crest freeboard of armor structure; Rc: crest freeboard of wall; Gc: crest width of armor structure; α u   a n d   α d   : angles of armor structure slope part above/below the berm; Du and Dd: sizes of the armor elements along up/down slopes; γ f u   a n d   γ f d : roughness factors for up/down slopes; β: incident angle of wave direction.
Jmse 11 01925 g001
Figure 2. Architecture of proposed CNN with the combination of bottleneck residual blocks including convolutional, layer normalization, and dropout layers in serial for learning relevant feature values from inputs for output of wave overtopping discharge qs.
Figure 2. Architecture of proposed CNN with the combination of bottleneck residual blocks including convolutional, layer normalization, and dropout layers in serial for learning relevant feature values from inputs for output of wave overtopping discharge qs.
Jmse 11 01925 g002
Figure 3. Schematic of hyperparameter optimal tuning process (HOTP) to create the CNN model.
Figure 3. Schematic of hyperparameter optimal tuning process (HOTP) to create the CNN model.
Jmse 11 01925 g003
Figure 4. Comparisons of the RMSE and running time using different bootstrap resamples.
Figure 4. Comparisons of the RMSE and running time using different bootstrap resamples.
Jmse 11 01925 g004
Figure 5. Predictions by the trained CNN for the training and the validation dataset with 10 bootstrap resamples.
Figure 5. Predictions by the trained CNN for the training and the validation dataset with 10 bootstrap resamples.
Jmse 11 01925 g005
Figure 6. Comparison of predictions for the wave overtopping database predicted by the different models with the entire core dataset.
Figure 6. Comparison of predictions for the wave overtopping database predicted by the different models with the entire core dataset.
Jmse 11 01925 g006
Figure 7. Comparisons of predicted wave overtopping discharges of different models for various coastal structures and oblique wave attack conditions with the entire core data.
Figure 7. Comparisons of predicted wave overtopping discharges of different models for various coastal structures and oblique wave attack conditions with the entire core data.
Jmse 11 01925 g007aJmse 11 01925 g007b
Figure 8. Comparison of predictions for the prototype wave overtopping discharge for different models.
Figure 8. Comparison of predictions for the prototype wave overtopping discharge for different models.
Jmse 11 01925 g008
Figure 9. Comparisons of predicted wave overtopping discharges of different models for various coastal structures.
Figure 9. Comparisons of predicted wave overtopping discharges of different models for various coastal structures.
Jmse 11 01925 g009
Figure 10. Comparisons of the predicted relative wave overtopping discharge versus relative crest freeboard using different models for prototype situations.
Figure 10. Comparisons of the predicted relative wave overtopping discharge versus relative crest freeboard using different models for prototype situations.
Jmse 11 01925 g010
Table 1. CNN input and output parameters.
Table 1. CNN input and output parameters.
#ParametersDefinition of the Parameters
1Hydraulic
parameters of waves
Hm0,t/Lm−1,0,tWave steepness at the structure toe
2h/Lm1,0,tRelative water depth at the structure toe
3 β Wave obliquity
4Structure
parameters
ht/Hm0,tRelative submergence of the toe structure
5Bt/Lm1,0,tRelative width of the toe structure
6 c o t α d Cotangent of the structure slope below the berm
7 c o t α u Cotangent of the structure slope above the berm
8 γ f d Roughness factor for c o t α d
9 γ f u Roughness factor for c o t α u
10Dd/Hm0,tRelative size of the armor elements along c o t α d
11Du/Hm0,tRelative size of the armor elements along c o t α u
12Ac/Hm0,tRelative crest freeboard of armor structure
13Rc/Hm0,tRelative crest freeboard of wall
14Gc/Lm1,0,tRelative crest width of armor structure
15B/Lm1,0,tRelative berm width
16hb/Lm1,0,tRelative berm submergence
1Predicted
parameter
qsNormalized wave overtopping discharge per unit width
Table 2. Setup of hypothetical interval of hyperparameters for HOTP.
Table 2. Setup of hypothetical interval of hyperparameters for HOTP.
Hyperparameter NameValuesDefinition of the Hyperparameter
Convolution_filters[16, 24, 32, 64, 128, 256 *]Number of filters w in each convolution layer
Residual_block[1, 2 *]Number of residual block
Fully_connection_layer_neurons[32, 64, 128 *, 256, 512]Neurons in a fully connected layer
Learning_rate[0.0001 *, 0.0002, …, 0.1]Given range of learning rate for CNN training
Dropout_rate[0.1, 0.2, …0.7 *, 0.8, 0.9],Determines the probability of a neuron being dropped out
λ1[0.5, 0.6, …, 1.0 *]Given weights of RMSE term of L
λ2[0.1, 0.2,…, 0.5 *, …, 1.0]Given weights for constraint term of L
Table 3. The comparison of predicted RMSEs for the training and validation datasets using different numbers of bootstrap resamples.
Table 3. The comparison of predicted RMSEs for the training and validation datasets using different numbers of bootstrap resamples.
Dataset (Size)Bootstrap Resamples
N = 1N = 10N = 500
Training data (6922)0.2380.2300.225
Validation data (1731)0.2460.2140.215
Table 4. The comparison of predicted RMSEs of different models.
Table 4. The comparison of predicted RMSEs of different models.
Dataset (Size)ModelRMSECCR2
Entire core data (8653)NN0.6540.8690.756
NNb0.6390.8580.736
XGB0.1990.9900.981
CNN0.1120.9960.991
A: rock permeable straight slopes (1131)NN0.7020.7680.590
NNb0.6870.7910.625
XGB0.2470.9770.955
CNN0.1530.9890.978
B: rock impermeable straight slopes (104)NN0.6240.8640.747
NNb0.8230.7100.504
XGB0.3490.9840.968
CNN0.1360.9940.989
C: armor units with straight slopes (934)NN0.5060.8720.760
NNb0.5830.8580.737
XGB0.2190.9830.966
CNN0.1180.9910.983
D: smooth and straight slopes (2069)NN0.4800.9720.946
NNb0.3800.9430.889
XGB0.1200.9960.991
CNN0.0860.9970.994
E: structures with combined slopes and berms (1631)NN0.7840.8380.703
NNb0.6700.8770.770
XGB0.2040.9870.975
CNN0.0930.9970.993
F: vertical walls (1677)NN0.6850.8630.745
NNb0.7040.8450.714
XGB0.2060.9880.977
CNN0.1100.9940.988
G: oblique wave attacks (1107)NN0.7600.7770.603
NNb0.8720.7720.595
XGB0.2280.9830.965
CNN0.1280.9890.978
Table 5. The comparison of predicted RMSEs of different models for the prototype dataset.
Table 5. The comparison of predicted RMSEs of different models for the prototype dataset.
Dataset (Size)ModelRMSECCR2
All testing data (150)NN*0.9360.8740.764
NNb1.2180.8950.802
XGB0.6570.8680.753
CNN0.5550.9320.868
* For the NN model, it is out of range for the estimation of G-18.1 to G-18.77. This result only calculates 23 cases of G-12.1 to G-12.23 for the NN model.
Table 6. Conditions for estimating overtopping discharge in example scenario.
Table 6. Conditions for estimating overtopping discharge in example scenario.
ParameterValueParameterValueParameterValue
Hmo,t [m]1.67Rc [m]3.34 to 7.50Dd [m]0.60
hb [m]1.03Ac [m]3.34Du [m]0.00
ht [m]3.28Gc [m]1.00 γ f d 0.55
h [m]3.28B [m]8.00 γ f u 1.00
Lm−1,0,t [m]59.39cotαd1.00β [°]14
Bt [m]0.00cotαu0.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tsai, Y.-T.; Tsai, C.-P. Predictions of Wave Overtopping Using Deep Learning Neural Networks. J. Mar. Sci. Eng. 2023, 11, 1925. https://doi.org/10.3390/jmse11101925

AMA Style

Tsai Y-T, Tsai C-P. Predictions of Wave Overtopping Using Deep Learning Neural Networks. Journal of Marine Science and Engineering. 2023; 11(10):1925. https://doi.org/10.3390/jmse11101925

Chicago/Turabian Style

Tsai, Yu-Ting, and Ching-Piao Tsai. 2023. "Predictions of Wave Overtopping Using Deep Learning Neural Networks" Journal of Marine Science and Engineering 11, no. 10: 1925. https://doi.org/10.3390/jmse11101925

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop