Next Article in Journal
Floating Riverine Litter Flux to the White Sea: Seasonal Changes in Abundance and Composition
Previous Article in Journal
Recognition of Unsafe Onboard Mooring and Unmooring Operation Behavior Based on Improved YOLO-v4 Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Electric Consumption Forecast for Ships Using Multivariate Bayesian Optimization-SE-CNN-LSTM

1
Romantique, Contents AI Research Center, 27 Daeyeong-ro, Busan 49227, Republic of Korea
2
Division of Marine System Engineering, Korea Maritime and Ocean University, 727 Taejong-ro, Yeongdo-gu, Busan 49112, Republic of Korea
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2023, 11(2), 292; https://doi.org/10.3390/jmse11020292
Submission received: 26 December 2022 / Revised: 18 January 2023 / Accepted: 24 January 2023 / Published: 30 January 2023
(This article belongs to the Section Ocean Engineering)

Abstract

:
Many studies on reducing greenhouse gas emissions from ships have been conducted to reduce environmental pollution. Reducing the fuel oil consumption of traditional and green ships is a key focus of these studies. The fuel oil consumption of the ship depends on electric loads. Thus, ship power load estimation is necessary to develop methods for reducing the fuel oil consumption of ships. However, data accessibility for ship power load estimation is low, limiting the number of relevant studies. This study proposes a model for estimating the actual power load of ships using squeeze and excitation (SE), a convolutional neural network (CNN), and long short-term memory (LSTM). The electric load, power generated by the generator, power consumption of the reefer container, rudder angle, water speed, wind speed, and wind angle of a ship were measured in 10-minute increments for approximately 145 d. The existing parallel and direct CNN-LSTM power load estimation models were used to evaluate the performance of the proposed model. The proposed model had the lowest root mean square error (RMSE), mean absolute error (MAE), and mean absolute percentage error (MAPE), demonstrating the best ship power load estimation performance compared to existing power load estimation models.

1. Introduction

Various methods for reducing greenhouse gas emissions from ships have been studied in recent years. In particular, research on reducing fuel oil consumption of traditional ships and green ships, which do not use internal combustion engines but consume fuel oil, is increasingly being conducted [1,2,3,4,5]. Forecasts representing the actual load data of the ships are essential for these studies because they can evaluate the amount of electrical energy that the algorithm can reduce by estimating the power load of ships.
Green ships use fuel cells, batteries, eco-friendly power sources, and supercapacitors as power sources [6,7,8]. Power loads in traditional ships are classified into general power loads (consistently used) and intermittent heavy loads. Owing to the high electricity consumption of heavy loads, ensuring a sufficient fuel cell and battery capacity is necessary for the smooth operation of green ships [8,9,10,11]. Therefore, estimating the power load of ships is necessary to promote the utilization of green ships.
Unlike those related to the power load in ships [12,13], the trend, seasonality, and periodicity of shore power demand can be determined conveniently. Moreover, many studies have been conducted on regression, fuzzy estimation, random forest, support vector machines, and power demand estimation using artificial neural networks owing to the high accessibility of shore power demand data [14,15]. The power load in ships is intermittently high owing to heavy loads, such as the bow thrusters and cargo pumps [16]. Furthermore, research on ship power load estimation is limited because of low data accessibility.
Therefore, this study proposes a model for estimating the actual power load of ships using squeeze and excitation (SE), a convolutional neural network (CNN), and long short-term memory (LSTM). Moreover, this study used the parallel and direct CNN-LSTM power load estimation models presented in previous studies to evaluate the performance of the proposed model.

2. Theoretical Background

2.1. Convolutional Neural Network (CNN)

A CNN can analyze data features or patterns and can be classified as one-, two-, or three-dimensional (1D, 2D, or 3D), making them an excellent tool for analyzing time series data with a 1D structure, images, and video data, respectively. The ship data used in this study are time series data with a 1D structure; therefore, a 1D CNN is used to examine the spatial features within the ship data. A CNN comprises convolution and pooling layers [17,18]. Filters are used in the convolution layer to extract features of the input data. The following formula expresses the operation occurring in the convolution layer:
x k l = f ( b k l + i = 1 N l 1 w i k l 1 y i l 1 ) ,
where w i k l 1 is the weight value of the vectors, y i l 1 is the vector value input into the convolution layer operation, b k l represents the bias value, x k l is the value calculated using the activation function, and f is an activation function.

2.2. Long Short-Term Memory (LSTM)

A recurrent neural network (RNN) is a framework for processing time series data. It analyzes the relationship between the preceding and present data and estimates the succeeding data. However, a long-term dependency problem occurs, and past information is lost. LSTM, comprising a cell state, forget gate, input gate, and output gate, uses cell states to improve the limitations of RNNs. The cell state ensures that the input information remains unchanged. Figure 1 shows the inner structure of LSTM.
The forget gate calculates the output of the previous cell and current input data and removes them when judged to be unnecessary values. The input gate determines the information to be stored in the cell. The output gate updates the state values at a specific point in time. LSTM utilizes the following equations to perform operations:
f t = σ W f h t 1 ,   x t + b f
i t = σ W t h t 1 ,   x t + b i
C t ^ = t a n h W c h t 1 ,   x t + b c
C t = f t C t ^ + i t C t ^
O t = σ W o h t 1 ,   x t + b o
h t = O t t a n h C t ,
where f t is the forget gate, i t is the input gate, O t is the output gate, C t is the cell state or memory cell, h t 1 is the previous state, C t ^ is the candidate for the cell state at timestamp (t), and h t is the output state or next hidden state. σ and t a n h are the activation functions sigmoid and hyperbolic tangent, respectively. W t , W f , W c , W o are the respective weights, and b f , b i , b c , b o are the biases. h t 1 is the output of the previous LSTM block, and x t is the input at the current timestep.

2.3. Squeeze and Excitation (SE) Block

The SE block, which is behind the CNN [19], comprises a squeeze that extracts only relevant information and an excitation that recalibrates the squeezed data. Using the SE block enables features output by the CNN to be analyzed to emphasize significant features and suppress the insignificant ones. Global average pooling can be performed for squeeze and expressed with the following equation:
z c = F s q u c = 1 W k = j W u c i , j ,
where z c is the data resulting from the squeeze operations with the data input into the c channel, u c is the c channel data of the input data, F s q is the squeeze operation, and W is the size of the feature map. The squeezed data proceed to excitation, expressed as follows:
s = F e x z = f s i g W 2 f r e l u W 1 z c ,
where W 1   i s C r × C , W 2   i s C × C r , W 1 is a layer that reduces dimensions according to the r ratio, and W 2 is a layer that re-increases the data dimensions according to the r ratio. Further, f r e l u refers to the rectified linear unit (ReLU) activation function, and   f s i g is the sigmoid function.

2.4. Bayesian Optimization

The deep learning model consists of various hyperparameters, such as the number of neurons, learning rate, and batch size. Selecting the value of the hyperparameters is necessary to enhance the performance of the deep learning model. Grid or random search is typically used to find the appropriate hyperparameter values for the deep learning model. Grid search algorithms are characterized by increasing the search space with increasing the number of hyperparameters [20], indicating a disadvantage of requiring a substantial amount of central processing unit (CPU) or graphics processing unit (GPU) resources. Furthermore, random search algorithm functions are better than grid search algorithms to enhance the probability to find optimized hyperparameters but have lower performance than Bayesian optimization algorithms [21]. The field of machine learning uses intelligence optimization algorithms, such as the genetic algorithm (GA), simulated annealing algorithm (SA), particle swarm optimization (PSO), and Bayesian optimization algorithm. The Bayesian optimization algorithm has the advantage of finding optimized points while utilizing minimal iteration and computer resources [22,23,24,25,26,27,28]. In addition, it can determine the optimal solution to an unknown object function by creating an alternative model for the object function and hyperparameter pair [29,30,31]. It updates the hyperparameters sequentially and searches for the optimal hyperparameter combination.
p * = argmin   p P l o s s p ,
where P represents the set of all hyperparameters, p refers to the hyperparameter combinations that can be made using P , p * is the hyperparameter combinations derived through optimization, and l o s s ( · ) is the object function for the Bayesian optimization. If the commonly used root mean square error (RMSE) is used as the loss function, the following equation can be used:
l o s s p j = i = 1 n y i ^ p j y i 2 n ,
where p j refers to the jth combination among all hyperparameter combinations, y i refers to the ith true value, and y i ^ is the ith value estimated by the deep learning model using p j . An observed dataset, D , is produced for the Bayesian optimization operations, expressed as follows:
D = x i ,     y i ,   x i + 1 ,     y i + 1 ,
where x i   is the i th hyperparameter combination and y i is the error value occurring with the deep learning model using the i th hyperparameter combination. The dataset, D , can be used to obtain the posterior probability with the following equations:
p l o s s = G l o s s ; μ ,   K ,
P l o s s D = G l o s s ; μ l o s s D ,   K l o s s D ,  
where G indicates a Gaussian distribution, μ refers to the mean, and K represents the variance. Bayesian optimization uses previous sampling points to predict the form of the object function and finds the point where the hyperparameter values can derive the maximum result desired by the user. Figure 2 shows the flowchart of the Bayesian optimization algorithm.

2.5. Data Description

The power load of a container ship comprises the power load of the equipment used to propel the ship and hotel load. Determining the factors affecting the power load of the ship is essential for load estimation. Ships must maintain stable sailing routes; thus, all ships continuously control their rudder angle. Some ships use an electric hydraulic system (EHS) for rudder angle control. The EHS operates the hydraulic pump using an electric motor, which controls the rudder angle by moving the hydraulic cylinder. It utilizes the multivariate auto-regressive eXogeneous (MARX) model for rudder control, expressed as follows:
X n = m = 1 M A m X n m + m = 1 M B m Y n m + U n ,
where X n is a 2D vector that includes yaw and roll. U n is the Gaussian white noise, and X n m and Y n m are the differences between the measured input and output vector values. M is the M th value of command. Furthermore, the optimal control rule of the MARX model can be expressed as follows:
Y n = G Z n ,
where Y n is the optimal control law, G is the stationary gain, and Z n is the state vector.
The rudder angle set using the MARX model is maintained continuously. However, ships are subject to constant wave and wind effects, and the rudder angle is subject to change during severe wave and wind conditions. We collected data from actual merchant ships. The data measured in this study for ship load estimation are electric load, power generated by the generator, power consumption of the reefer container, rudder angle, water speed, wind speed, and wind angle. Every ship exploits the rudder angle for maneuvering, and the rudder control system requires electricity to maintain its system. According to the value of ship resistance, the rudder control system’s electricity consumption varies [32]. Additionally, water speed, wind speed, wind angle, and rudder angle are classified as resistances on the vessel [33]; therefore, we selected these data for this study because they can directly affect a ship’s electricity load. The data were measured and acquired in 10-min increments for approximately 145 d. A total of 20,935 data entries were used in this study. Table 1 lists the data used in this study.
Figure 3 shows the power load data of the ship measured in 10-min increments, with heavy loading occurring intermittently.

3. Proposed Model

3.1. Electric Consumption Forecast for Ships by a Multivariate CNN-LSTM Network with SE Using Bayesian Optimization

The datasets were scaled to accelerate model training. We selected the MinMaxScaler, expressed as follows:
M i n M a x S c a l e r : x = x min x max x min x ,
where x is the new value from the formula, x is the original value, and min x and max x are the minimum and maximum values of the column, respectively. The scaled data were input into the CNN for data feature extraction. The extracted features were input into the SE and LSTM. The SE designates weights to the input data in the CNN and classifies it as significant or non-significant. The LSTM estimates the power load using the data input into the CNN. The weighted values derived from the SE and power load estimation values outputted from the LSTM were input into the dense layer, resulting in the estimated power load of the ship. The estimated value was configured to allow the user to confirm the final value using inverse data scaling. Figure 4 shows the proposed model.

3.2. Model Training Process

The proposed CNN-LSTM with SE model was trained using the collected data, and classified into training, validation, and test datasets to prevent overfitting of the model. Figure 5 shows the ratio of the training, test, and validation datasets.
The data in the dataset were measured every 10 min. In this study, the power load in the next 10 min was forecast from the previous 50 min; thus, the time step was five. The CNN-LSTM with SE model uses the previously mentioned datasets to search for hyperparameters and estimate power values. The model forecasting steps are as follows:
Step 1.
Parameter initialization
The LSTM used in the proposed model used ReLU as the activation function. Table 2 shows the initial parameter configurations of the 1D convolution, SE, and LSTM layers.
The parameters in Table 2 were optimized in the hyperparameter searching stage. Furthermore, the RMSE, mean absolute error (MAE), and mean absolute percentage error (MAPE) were used to measure the model performance, expressed as follows [34]:
R M S E = 1 N i = 1 N y i y l ^ 2 ,
M A E = 1 N i = 1 N y i y l ^ ,
M A P E = 1 N i = 1 N y i y l ^ y i × 100 ,
where N refers to the total number of data utilizing the performance evaluation function, y i is the ith true value, and y l ^ is the ith estimated value. The RMSE penalizes large residual values using the square function. The MAE checks the mean of the absolute error by comparing the estimated value with the actual value to examine the two values against the absolute scaler. Finally, the MAPE evaluates the uniform estimation error of the forecast model as a percentage.
Step 2.
Searching hyperparameters
Bayesian optimization was used in searching for hyperparameters. The forecasting value and loss were calculated using an initial parameter value. The kernel function of Bayesian optimization was used to determine the optimized hyperparameter value. Table 3 shows the search space of the hyperparameters.
Step 3.
Forecasting
The test datasets were used to evaluate the performance of the proposed model with the optimal hyperparameters.

4. Results and Discussion

This section compares the performance of the proposed model with that of the parallel and direct CNN-LSTM models, primarily used in estimating conventional time series datasets [35,36]. The selected models were chosen based on previous research regarding electricity forecasting [37]. The parameters for the proposed model were obtained using the Bayesian optimization algorithm. Figure 6 shows the convergence plot of the proposed model with the Bayesian optimization algorithm. It can be observed that the Bayesian optimization algorithm rapidly reached the convergence point and minimum point of the loss function in the ninth iteration. Using a convergence plot confirmed that the Bayesian optimization algorithm applied to the proposed model was successfully implemented.
Table 4 lists the values of the optimal hyperparameters of the proposed model derived through Bayesian optimization.
Figure 7 shows the structures of the parallel and direct CNN-LSTM models used for the comparative evaluation.
The comparative evaluation experiments were performed using Python, Scikit-Learn, and Tensorflow libraries. Each model used the same train, test, and validation dataset ratio, with the power load in the next 10 min forecasted from the previous 50 min as the model training process. Figure 8 shows the actual data (true) and forecasting results for all models. The results of all models used for comparative evaluation were consistent with the actual data. However, the detailed comparative evaluations of the models in Figure 9 shows that, on average, severe amplitudes of the forecasted value occurring through forecasting appeared in the CNN-LSTM (direct) model. In particular, amplitudes were generated more intensely after the heavy loading section; values with higher differences than the actual data values were output. Furthermore, the forecasting capability was relatively poor in areas below 800 kW. The forecasting performance of the parallel CNN-LSTM model was better than that of the direct CNN-LSTM model but inferior to that of the proposed model. The latter was demonstrated by the unstable loading changes detected between 1000 kW and 2000 kW. Finally, the proposed model performed forecasting by reflecting the characteristics of the actual data values more smoothly than the parallel and direct CNN-LSTM models.
Table 5 summarizes the performances of all models evaluated using RMSE, MAE, and MAPE. The proposed model demonstrated the best output.

5. Conclusions

Research on artificial intelligence using ship data has been increasingly conducted in recent years. The LSTM model processes time series data smoothly, and the CNN model extracts data features. Conventional time series data estimation models use the LSTM and CNN models simultaneously. This study used the SE model to enhance the performance of existing time series data forecasting models. Furthermore, the hyperparameters ensuring the highest performance reduced the time spent on hyperparameter searching using Bayesian optimization algorithms. The primary conclusions are summarized as follows:
  • Bayesian optimization is suitable for determining hyperparameters for the proposed CNN-LSTM with the SE model by combining loss functions and hyperparameters. Compared with traditional methods, such as grid and random search, this process can determine the parameters more rapidly.
  • This study used the parallel and direct CNN-LSTM load estimation models to evaluate the performance of the proposed model. According to the comparative test results, the proposed model demonstrated the best ship power load forecasting performance.
  • More data are required to enhance the estimation performance of the proposed model for intermittent heavy loading.
We recognize that the analytical ability to forecast a ship’s electricity feature is essential to enhance the ability of the forecasting model. As well, this model can help ship’s simulation to make algorithms for reducing fuel oil consumption. However, the ability for heavy load forecasting is limited. Thus, we will research a co-occurrence forecasting model for a ship’s electricity in a follow-up study.

Author Contributions

Conceptualization, Methodology and Software, J.-Y.K.; Project administration, Funding acquisition, J.-S.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Korea Institute of Marine Science and Technology Promotion (KIMST) funded by the Korea Coast Guard (20190460).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Lee, J.-H.; Oh, J.-H.; Oh, J.-S. Battery Control Algorithm for Merchant Ships Using Load Usage Area. In Proceedings of the International Conference on Electrical, Computer, Communications and Mechatronics Engineering (ICECCME), Reduit, Mauritius, 7–8 October 2021. [Google Scholar]
  2. Yang, X.; Bai, G.; Schmidhalter, R. Shore to Ship Converter System for Energy Saving and Emission Reduction. In Proceedings of the 8th International Conference on Power Electronics—ECCE Asia, Jeju, Republic of Korea, 30 May 2011; pp. 2081–2086. [Google Scholar]
  3. Chin, C.S.; Tan, Y.-J.; Kumar, M.V. Study of Hybrid Propulsion Systems for Lower Emissions and Fuel Saving on Merchant Ship during Voyage. J. Mar. Sci. Eng. 2022, 10, 393. [Google Scholar] [CrossRef]
  4. Jang, J.-H.; Oh, J.-S. The Study on a Ship Energy Management System Applied Rechargeable Battery. J. Korean Soc. Mar. Eng. 2014, 38, 202–207. [Google Scholar] [CrossRef] [Green Version]
  5. Abdelmoula, A.-A.; Khalifa, M.; Mohamed, Y.; Mohammed, Q. Toward Energy Saving and Environmental Protection by Implementation of Autonomous Ship. In Proceedings of the 2018 19th IEEE Mediterranean Electrotechnical Conference (MELECON), Marrakech, Morocco, 2–7 May 2018. [Google Scholar]
  6. Nyanya, M.N.; Vu, H.B.; Schönborn, A.; Olçer, A.I. Wind and Solar Assisted Ship Propulsion Optimisation and Its Application to a Bulk Carrier. Sustain. Energy Technol. Assess. 2021, 47, 101397. [Google Scholar] [CrossRef]
  7. Jayasinghe, S.G.; Suryawanshi, U.; Sheikh, A.; Alahakoon, S. Fuel Saving with a Hybrid Power System for an Electric Ferry. In Proceedings of the 14th Conference on Industrial and Information Systems (ICIIS), Kandy, Sri Lanka, 18–20 December 2019; pp. 255–259. [Google Scholar]
  8. Kistner, L.; Bensmann, A.; Hanke-Rauschenbach, R. Optimal Design of Power Gradient Limited Solid Oxide Fuel Cell Systems with Hybrid Storage Support for Ship Applications. Energy Convers. Manag. 2021, 243, 114396. [Google Scholar] [CrossRef]
  9. Chen, H.; Zhang, Z.; Guan, C.; Gao, H. Optimization of Sizing and Frequency Control in Battery/Supercapacitor Hybrid Energy Storage System for Fuel Cell Ship. Energy 2020, 197, 117285. [Google Scholar] [CrossRef]
  10. Bassam, A.M.; Phillips, A.B.; Turnock, S.R.; Wilson, P.A. Development of a Multi-scheme Energy Management Strategy for a Hybrid Fuel Cell Driven Passenger Ship. Int. J. Hydrogen Energy 2017, 42, 623–635. [Google Scholar] [CrossRef] [Green Version]
  11. Wu, S.; Miao, B.; Chan, S.H. Feasibility Assessment of a Container Ship Applying Ammonia Cracker-Integrated Solid Oxide Fuel Cell Technology. Int. J. Hydrogen Energy 2022, 47, 27166–27176. [Google Scholar] [CrossRef]
  12. Dai, X.; Sheng, K.; Shu, F. Ship Power Load Forecasting Based on PSO-SVM. Math. Biosci. Eng. 2022, 19, 4547–4567. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Zhou, Q.; Sun, C.; Lei, S.; Liu, Y.; Song, Y. RBF Neural Network and ANFIS-Based Short-Term Load Forecasting Approach in Real-Time Price Environment. IEEE Trans. Power Syst. 2008, 23, 853–858. [Google Scholar] [CrossRef]
  14. Al-Hamadi, H.M. Long-Term Electric Power Load Forecasting Using Fuzzy Linear Regression Technique. In Proceedings of the IEEE Power Engineering and Automation Conference, Wuhan, China, 8–9 September 2011. [Google Scholar]
  15. Tian, M.; Alattas, K.; El-Sousy, F.; Alanazi, A.; Mohammadzadeh, A.; Tavoosi, J.; Mobayen, S.; Skruch, P. A New Short Term Electrical Load Forecasting by Type-2 Fuzzy Neural Networks. Energies 2022, 15, 3034. [Google Scholar] [CrossRef]
  16. Bakar, N.N.A.; Bazmohammadi, N.; Çimen, H.; Uyanik, T.; Vasquez, J.C.; Guerrero, J.M. Data-Driven Ship Berthing Forecasting for Cold Ironing in Maritime Transportation. Appl. Energy 2022, 326, 119947. [Google Scholar] [CrossRef]
  17. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D Convolutional Neural Networks and Applications: A Survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  18. Avci, O.; Abdeljaber, O.; Kiranyaz, S.; Hussein, M.; Inman, D.J. Wireless and Real-Time Structural Damage Detection: A Novel Decentralized Method for Wireless Sensor Networks. J. Sound Vib. 2018, 424, 158–172. [Google Scholar] [CrossRef]
  19. Karim, F.; Majumdar, S.; Darabi, H.; Harford, S. Multivariate LSTM-FCNs for Time Series Classification. Neural Netw. 2019, 116, 237–245. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Candelieri, A.; Giordani, I.; Archetti, F.; Barkalov, K.; Meyerov, I.; Polovinkin, A.; Sysoyev, A.; Zolotykh, N. Tuning Hyperparameters of a SVM-Based Water Demand Forecasting System through Parallel Global Optimization. Comput. Oper. Res. 2019, 106, 202–209. [Google Scholar] [CrossRef]
  21. Thiede, L.A.; Parlitz, U. Gradient Based Hyperparameter Optimization in Echo State Networks. Neural Netw. 2019, 115, 23–29. [Google Scholar] [CrossRef]
  22. Yeniay, Ö. Penalty Function Methods for Constrained Optimization with Genetic Algorithms. Math. Comput. Appl. 2005, 10, 45–56. [Google Scholar] [CrossRef] [Green Version]
  23. Stochino, F.; Lopez Gayarre, F.L. Reinforced Concrete Slab Optimization with Simulated Annealing. Appl. Sci. 2019, 9, 3161. [Google Scholar] [CrossRef] [Green Version]
  24. Wu, T.; Shi, X.; Liao, L.; Zhou, C.; Zhou, H.; Su, Y. A Capacity Configuration Control Strategy to Alleviate Power Fluctuation of Hybrid Energy Storage System Based on Improved Particle Swarm Optimization. Energies 2019, 12, 642. [Google Scholar] [CrossRef] [Green Version]
  25. Snoek, J.; Larochelle, H.; Adams, R.P. Practical Bayesian Optimization of Machine Learning Algorithms. Adv. Neural Inf. Process. Syst. 2012, 25, 2960–2968. [Google Scholar]
  26. Zhou, J.; Qiu, Y.; Zhu, S.; Armaghani, D.J.; Khandelwal, M.; Mohamad, E.T. Estimating of the TBM Advance Rate under Hard Rock Conditions Using XGBoost and Bayesian Optimization. Undergr. Space 2021, 6, 506–515. [Google Scholar] [CrossRef]
  27. Zhou, J.; Li, X.; Shi, X. Long-Term Prediction Model of Rockburst in Underground Openings Using Heuristic Algorithms and Support Vector Machines. Saf. Sci. 2012, 50, 629–644. [Google Scholar] [CrossRef]
  28. Zhou, J.; Qiu, Y.G.; Zhu, S.; Armaghani, D.J.; Li, C.; Nguyen, H.; Yagiz, S. Optimization of Support Vector Machine through the Use of Metaheuristic Algorithms in Forecasting TBM Advance Rate. Eng. Appl. Artif. Intell. 2021, 97, 104015. [Google Scholar] [CrossRef]
  29. Aoki, M. On Some Convergence Questions in Bayesian Optimization Problems. IEEE Trans. Autom. Control 1965, 10, 180–182. [Google Scholar] [CrossRef]
  30. Bergstra, J.; Bardenet, R.; Kégl, B.; Bengio, Y. Algorithms for Hyper-parameter Optimization. In Proceedings of the 25th Annual Conference on Neural Information Processing Systems, Granada, Spain, 12–15 December 2011; p. hal-00642998f. [Google Scholar]
  31. Shahriari, B.; Swersky, K.; Wang, Z.; Adams, R.P.; de Freitas, N. Taking the Human out of the Loop: A Review of Bayesian Optimization. Proc. IEEE 2016, 104, 148–175. [Google Scholar] [CrossRef] [Green Version]
  32. Kim, J.H.; Choi, J.E.; Choi, B.J.; Chung, S.H. Twisted Rudder for Reducing Fuel-Oil Consumption. Int. J. Nav. Archit. Ocean Eng. 2014, 6, 715–722. [Google Scholar] [CrossRef] [Green Version]
  33. Elkafas, A.G.; Elgohary, M.M.; Zeid, A.E. Numerical Study on the Hydrodynamic Drag Force of a Container Ship Model. Alex. Eng. J. 2019, 58, 849–859. [Google Scholar] [CrossRef]
  34. Ahmed, R.; Sreeram, V.; Mishra, Y.; Arif, M.D. A Review and Evaluation of the State-of-the-Art in PV Solar Power Forecasting: Techniques and Optimization. Renew. Sustain. Energy Rev. 2020, 124, 109792. [Google Scholar] [CrossRef]
  35. Sheng, D.D.; Tian, R.L.; Xun, G.; Yan, Y.; Shi, J.H. Traffic Flow Forecasting Based on Hybrid Deep Learning Framework. In Proceedings of the 12th International Conference on Intelligent Systems and Knowledge Engineering, Nanjing, China, 24–26 November 2017. [Google Scholar]
  36. Agga, A.; Abbou, A.; Labbadi, M.; Houm, Y.E.; Ou Ali, I.H.O. CNN-LSTM: An Efficient Hybrid Deep Learning Architecture for Predicting Short-Term Photovoltaic Power Production. Electr. Power Syst. Res. 2022, 208, 107908. [Google Scholar] [CrossRef]
  37. Kim, J.Y.; Lee, J.H.; Oh, J.H.; Oh, J.S. A Comparative Study on Energy Consumption Forecast Methods for Electric Propulsion Ship. J. Mar. Sci. Eng. 2021, 10, 32. [Google Scholar] [CrossRef]
Figure 1. Inner structure of long short-term memory (LSTM).
Figure 1. Inner structure of long short-term memory (LSTM).
Jmse 11 00292 g001
Figure 2. Flowchart of the Bayesian optimization algorithm.
Figure 2. Flowchart of the Bayesian optimization algorithm.
Jmse 11 00292 g002
Figure 3. Electric power load data of the ship.
Figure 3. Electric power load data of the ship.
Jmse 11 00292 g003
Figure 4. Convolutional neural network (CNN)-LSTM with squeeze and excitation (SE).
Figure 4. Convolutional neural network (CNN)-LSTM with squeeze and excitation (SE).
Jmse 11 00292 g004
Figure 5. Dataset classification.
Figure 5. Dataset classification.
Jmse 11 00292 g005
Figure 6. Convergence plot of the CNN-LSTM with the SE model by the Bayesian optimization algorithm.
Figure 6. Convergence plot of the CNN-LSTM with the SE model by the Bayesian optimization algorithm.
Jmse 11 00292 g006
Figure 7. Structures of CNN-LSTM models for comparison: (a) Parallel; (b) Direct.
Figure 7. Structures of CNN-LSTM models for comparison: (a) Parallel; (b) Direct.
Jmse 11 00292 g007
Figure 8. Comparison of models through forecasting with real data.
Figure 8. Comparison of models through forecasting with real data.
Jmse 11 00292 g008
Figure 9. Detailed comparison of models using forecasting with real data: (a) Forecasting after heavy loading; (b) Power load forecasting occurring from propulsion.
Figure 9. Detailed comparison of models using forecasting with real data: (a) Forecasting after heavy loading; (b) Power load forecasting occurring from propulsion.
Jmse 11 00292 g009
Table 1. List and features of the selected data.
Table 1. List and features of the selected data.
AttributeUnitRangeAverage
Electric LoadkW585–15,2001567.38
No. 1 Diesel Generator PowerkW0–3800240.98
No. 2 Diesel Generator PowerkW0–3800604.42
No. 3 Diesel Generator PowerkW0–3800323.9
No. 4 Diesel Generator PowerkW0–3800401.07
No. 1 Reefer Container LoadkW0–13014.19
No. 2 Reefer Container LoadkW0–13012.37
No. 3 Reefer Container LoadkW0–13024.17
No. 4 Reefer Container LoadkW0–13022.21
No. 5 Reefer Container LoadkW0–13021.51
No. 6 Reefer Container LoadkW0–13019.76
Rudder AngleDegree−38–38°−1.31
Water Speedm·s−1−4.36–16.669.85
Wind Speedm·s−10–24.86.23
Wind AngleDegree0–360°155.08
Table 2. Initial parameters of the CNN-LSTM with the SE model.
Table 2. Initial parameters of the CNN-LSTM with the SE model.
Parameter NameInitial Parameter Value
Number of first 1D convolution layer filter10
Size of first 1D convolution layer kernel3
Number of second 1D convolution layer filter10
Size of second 1D convolution layer kernel3
Number of neurons in excitation layer input10
Number of neurons in excitation layer output10
Number of neurons in first LSTM16
Number of neurons in second LSTM16
Number of neurons in third LSTM16
Value of learning rate10−6–10−3
Value of batch size16
Table 3. Search space of the CNN-LSTM with the SE model.
Table 3. Search space of the CNN-LSTM with the SE model.
Parameter NameSearch Space
Number of first 1D convolution layer filter10–128
Size of first 1D convolution layer kernel3–128
Number of second 1D convolution layer filter10–128
Size of second 1D convolution layer kernel3–128
Number of neurons in excitation layer input10–1024
Number of neurons in excitation layer output10–1024
Number of neurons in first LSTM16–1024
Number of neurons in second LSTM16–1024
Number of neurons in third LSTM16–1024
Value of learning rate10−6–10−3
Value of batch size16–256
Table 4. Selected optimized hyperparameters of the proposed CNN-LSTM with the SE model by Bayesian optimization.
Table 4. Selected optimized hyperparameters of the proposed CNN-LSTM with the SE model by Bayesian optimization.
Parameter NameSelected Hyperparameter Value
Number of first 1D convolution layer filter90
Size of first 1D convolution layer kernel3
Number of second 1D convolution layer filter128
Size of second 1D convolution layer kernel3
Number of neurons in excitation layer input1024
Number of neurons in excitation layer output194
Number of neurons in first LSTM512
Number of neurons in second LSTM528
Number of neurons in third LSTM194
Value of learning rate0.001
Value of batch size256
Table 5. Performances of the proposed CNN-LSTM with SE, CNN-LSTM (parallel), and CNN-LSTM (direct) models.
Table 5. Performances of the proposed CNN-LSTM with SE, CNN-LSTM (parallel), and CNN-LSTM (direct) models.
ModelRMSEMAEMAPE
CNN-LSTM with SE148.8190.415.73
CNN-LSTM (parallel)159.85100.296.52
CNN-LSTM (direct)202.78134.438.83
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, J.-Y.; Oh, J.-S. Electric Consumption Forecast for Ships Using Multivariate Bayesian Optimization-SE-CNN-LSTM. J. Mar. Sci. Eng. 2023, 11, 292. https://doi.org/10.3390/jmse11020292

AMA Style

Kim J-Y, Oh J-S. Electric Consumption Forecast for Ships Using Multivariate Bayesian Optimization-SE-CNN-LSTM. Journal of Marine Science and Engineering. 2023; 11(2):292. https://doi.org/10.3390/jmse11020292

Chicago/Turabian Style

Kim, Ji-Yoon, and Jin-Seok Oh. 2023. "Electric Consumption Forecast for Ships Using Multivariate Bayesian Optimization-SE-CNN-LSTM" Journal of Marine Science and Engineering 11, no. 2: 292. https://doi.org/10.3390/jmse11020292

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop