Next Article in Journal
Different Types of Near-Inertial Internal Waves Observed by Lander in the Intermediate-Deep Layers of the South China Sea and Their Generation Mechanisms
Previous Article in Journal
Ocean Current Prediction Using the Weighted Pure Attention Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Forecasting Shanghai Container Freight Index: A Deep-Learning-Based Model Experiment

1
Graduate School of Maritime Sciences, Kobe University, 5-1-1 Fukaeminami-machi, Higashinada, Kobe 658-0022, Japan
2
Faculty of Commerce, Takushoku University, 3-4-14 Kohinata, Bunkyo-ku, Tokyo 112-8585, Japan
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2022, 10(5), 593; https://doi.org/10.3390/jmse10050593
Submission received: 28 February 2022 / Revised: 22 April 2022 / Accepted: 25 April 2022 / Published: 27 April 2022
(This article belongs to the Section Ocean Engineering)

Abstract

:
With the increasing availability of large datasets and improvements in prediction algorithms, machine-learning-based techniques, particularly deep learning algorithms, are becoming increasingly popular. However, deep-learning algorithms have not been widely applied to predict container freight rates. In this paper, we compare a long short-term memory (LSTM) method and a seasonal autoregressive integrated moving average (SARIMA) method for forecasting the comprehensive and route-based Shanghai Containerized Freight Index (SCFI). The research findings indicate that the LSTM deep learning models outperformed SARIMA models in most of the datasets. For South America and the east coast of the U.S. routes, LSTM could reduce forecasting errors by as much as 85% compared to SARIMA. The SARIMA models performed better than LSTM in predicting freight movements on the west and east Japan routes. The study contributes to the literature in four ways. First, it presents insights for improving forecasting accuracy. Second, it helps relevant parties understand the trends of container freight markets for wiser decision-making. Third, it helps relevant stakeholders understand overall container shipping market trends. Lastly, it can help hedge against the volatility of freight rates.

1. Introduction

Robust market forecasting is a critical and practical requirement in the management of shipping companies [1]. Prominent stakeholders in maritime business, such as carriers, freight forwarders, and shippers, rely on container freight-rate forecasts for operational decision making. Accordingly, various organizations continue to conduct periodic studies [2,3] on forecasting, and numerous researchers have actively studied improved prediction models.
Building reliable and robust forecasting models is essential in predicting market behavior and movements. Several techniques have been developed to build models that can estimate and forecast future time points, which can aid calculated decision-making to reduce risk and increase returns. The time-series forecasting method is becoming increasingly popular in various industries, including shipping.
In this study, we analyzed methods for predicting the Shanghai Containerized Freight Index (SCFI), which was chosen as the research object for three reasons. First, it has been studied in detail because it is one of the few container shipping freight indices that covers most of the global container trade volumes. It targets spot freight markets, which reflect up-to-date market trends and market equilibrium, and it provides high-frequency time-series data forecasts that reflect fluctuations in spot freight rates in Shanghai’s export container transport market. To calculate the SCFI, the Shanghai Shipping Exchange collects spot freight cost, together with insurance and freight (CIF) rates from the export container market, based on 13 shipping routes departing from the port of Shanghai on a container-yard-to-container-yard basis. The focus of this research, therefore, was the comprehensive SCFI, which is the weighted average of the 13 shipping routes.
Second, the SCFI is an underlying asset in freight derivatives [4]. In addition, time-series data forecasting, such as that provided by the SCFI, is highly important for business managers because these data reflect the overall trends in the corresponding container shipping market and provide implications for its future state.
Third, to date, the application of models for forecasting container freight rates has been relatively limited. Due to the high complexity, irregularity, randomness, and nonlinearity of time-series data, conventional econometric models cannot achieve a satisfactory forecasting accuracy.
Overall, few studies have been conducted on forecasting container freight rates by applying deep-learning algorithms. The goal of this study was to propose a novel LSTM-based forecasting model. The results of our evaluations complement the existing literature and indicate the need to develop more efficient forecasting techniques.
Our study resulted in three primary contributions. First, its findings can provide insights on improving forecasting accuracy. Second, the results of the study can aid in the understanding of freight trend forecasting in the corresponding container shipping market and enhance decision-making rationality in the shipping sector. Finally, the study provides implications for future investment in forward freight agreements or other risk-hedging freight derivatives.
The remainder of this paper is organized as follows: In Section 2, we review the literature on forecasting approaches in the shipping sector. Section 3, Section 4 and Section 5 present the properties of the associated data and describe the selected forecasting methods. In Section 6, we present our experimental results, including an assessment of forecasting accuracy in a given context. Finally, Section 7 presents discussions on our key findings and highlights the relevance of our work for practice and future research.

2. Literature Review

Forecasting shipping markets, particularly container freight-rate movements, has always been challenging. Over the past two decades, the global container shipping market has experienced turbulence due to several historical events. The first such event was the destruction of the shipping conference system in 2008 for port calls across all trades in the European Union (EU), marking the end of the cartel of shipping companies comprising the liner conference system [5]. The second was the global financial crisis of 2007–2008, which caused the container shipping freight market to significantly fluctuate along a generally downward-trending trajectory [2]. The third was the development of an unstable market environment characterized by the mergers and acquisitions of shipping lines, reformations of shipping alliances, and the commissioning of large container ships during the past decade, all of which contributed to heavy fluctuations in freight rates [6,7]. Recently, in the wake of the COVID-19 pandemic and the Suez Canal blockage, container shipping markets have become more unpredictable.
The methodologies applied in modeling container shipping include conventional econometric models, artificial intelligence (AI) models, or combinations of both, often referred to as hybrid models. Examples of research on econometric models include that of Chou et al. [8], in which a vector autoregression (VAR) model was used to forecast container trade volumes to Taiwan; Xie et al. [9] used autoregressive integrated moving average (ARIMA), SARIMA, and least-squares support vector regression (LSSVR) models to forecast container port throughputs; Schulze and Prinz [10], who reported that SARIMA models produced better results; and Kawasaki and Matsuda [11,12], who assessed the applicability of SARIMA and VAR models in forecasting container trade volumes.
Examples of AI model research include that of Chen and Chen [13], in which a genetic programming (GP) model was shown to outperform an ARIMA model by approximately 30%. An example hybrid model study is that of Xiao et al. [14], in which a transfer forecasting model guided by a discrete particle swarm optimization (TF-DPSO) algorithm was shown to outperform several existing models in terms of forecasting performance.
In recent years, other examples of hybrid models have emerged. Huang et al. [15] proposed a combination of projection pursuit regression (PPR) and GP algorithms. Xie et al. [16] proposed several hybrid approaches based on the LSSVR model and reported that their hybrid models achieved better forecasting performance than preprocessing methods such as SARIMA. Mo et al. [17] developed a hybrid model that applied SARIMA to the linear part of the data and support vector regression (SVR), back-propagation (BP), and GP to the nonlinear part; their research results indicated that the performance of the hybrid model was better than the other evaluated models.
Many approaches focus on forecasting container trade volumes and/or port throughput, but there are relatively few methods for forecasting container freight rates, especially the SCFI. Stopford [1] and Luo et al. [18] were pioneers in forecasting container freight rates by using supply and demand factors. Koyuncu and Tavacıoğlu [19] compared SARIMA and Holt–Winters Methods in forecasting the SCFI, and they concluded that the SARIMA model provided comparatively better results than the existing freight-rate forecasting models while performing short-term monthly forecasts. Chen et al. [20] applied a decomposition–ensemble method that combines empirical mode decomposition (EMD) and the grey wave forecasting model to forecast the China Container Freight Index (CCFI), and they found that the proposed method performed better than random walk and autoregressive moving average model (ARMA) in multi-step-ahead prediction. Munim and Schramm [2] recently deployed an autoregressive conditional heteroscedasticity (ARIMARCH) model to forecast container freight rates in Asia–North Europe routes. They observed that the ARIMARCH model provided better results than existing freight-rate forecasting models while enabling short-term forecasts on weekly and monthly bases. In his most recent research, Munim [21] reported that a state-space Trigonometric seasonality, Box–Cox transformation, ARMA errors, Trend and Seasonal components (TBATS) model outperformed seasonal neural network autoregression (SNNAR) and SARIMA models.
Like real-world time-series data, a container freight index features high complexity, irregularity, randomness, and nonlinearity. It is often difficult for conventional methods such as ARIMA to achieve a high prediction accuracy. As a result, models based on artificial neural networks (ANNs) are gaining increasing attention because of their ability to effectively manage the nonlinearity of time-series data [22]. In addition, machine learning methods can be used to build nonlinear prediction models using large quantities of historical time-series data, making it possible to obtain prediction results that are more accurate than those of conventional statistical models through repeated and iterative training and learning to approximate real models.
The machine learning (ML) methods that have been applied include SVR and ANN. These methods have strong nonlinear function approximation abilities and can be applied to tree-based ensemble learning [23]. Hassan et al. [24] proposed a new approach based on typical time-series and ML to forecast freight demand in the US market. Their model self-enhances through a reinforcement learning framework applied over a rolling horizon. Barua et al. [25] thoroughly reviewed ML models applied in international freight transportation management (IFTM), stating that ML is a powerful tool that enables better prediction and more robust support in IFTM.
A review of the existing literature that focuses on applied methodologies is summarized in Table 1. To the best of our knowledge, the LSTM model has not been previously applied in container shipping-related research.
This study was intended to contribute to the existing literature by filling these research gaps. We propose a deep-learning-based LSTM model for forecasting SCFI time-series data.

3. Deep Learning

Deep learning is a subfield of machine learning techniques concerned with ANNs. The most popular deep learning algorithms include convolutional neural networks (CNNs), recurrent neural networks (RNN), and stacked auto-encoders (SAE), among others. LSTM is an RNN architecture that can process a sequence of inputs. Since its introduction by Hochreiter and Schmidhuber [27], it has been refined by many researchers.

3.1. Model Description

As a special type of RNN, LSTM processes the input in a sequential manner by computing the output from the input of the previous step. However, typical RNNs suffer from vanishing (and exploding) gradients arising from the repeated use of recurrent weight matrices. As they are calculated using the chain rule, RNN gradients must undergo continuous matrix multiplications during the backpropagation process, which causes the gradients to either exponentially shrink (vanish) or exponentially blow up (explode). LSTM does not suffer from this problem because its models introduce gating functions to prevent gradient vanishing.
The LSTM is illustrated in Figure 1. It comprises a cell, input gate, output gate, and forget gate. The three gates regulate the flow of information into and out of the cell, and the cell remembers the values over any time interval. The gating functions enable the network to determine the extent to which the gradient vanishes and to obtain values at each time step. In other words, LSTM can process time-series data as a unit and can store, discard (forget), or add important information for making predictions, thus making it suitable for the analysis of time-series data.

3.1.1. Forget Gate (F)

The first step in LSTM is to identify the information to be discarded. The forget gate determines which information from the long-term memory is not required and should be discarded. This is performed by multiplying the incoming long-term memory with a forget vector generated by the current input and the incoming short-term memory:
F t = σ ( X t W x F + C t 1 W h F + B F )
where X t   is the current input vector, C t 1 is the output from the previous time step, W h F is the weight, and B F is the bias in the forget gate layer.

3.1.2. Input Gate (I)

The second step in LSTM is the input gate that determines whether the new information is to be stored in the long-term memory. It works with information based on current input and short-term memory from the previous time step. The input gate is defined as follows:
I t = σ ( X t W x I + C t 1 W h I + B I )
where X t is the input vector, C t 1 is the output from the previous time step, W h I is the weight, and B I is the bias at the input gate layer.

3.1.3. Output Gate (O)

The output gate computes the current input, previous short-term memory, and newly-computed long-term memory to produce a new short-term memory or hidden state to be passed on to the cell in the next time step:
O t = σ ( X t W x O + C t 1 W h O + B O )
where X t is the input vector, C t 1 is the output from the previous time step, W h O   is the weight, and B O is the bias in the output gate layer.

3.1.4. Activating Function

A hyperbolic tangent function (tanh) was used as the activating function in the proposed model. The tensor G applied to the input gate is expressed as follows:
G t = t a n h ( X t W x G + C t 1 W h G + B G )
The new cell that stores long-term memory C t is then given by:
C t = G t I t + F t C t 1
where ⊗ is the point-wise product.
The short- and long-term memories produced by the three gates are carried over to the next cell, and the neural network functions by repeating this process. The output of each time step obtained from the short-term memory is also called the hidden state, whose layer H is defined as:
H t = O t t a n h C t
where ⊗ is the point-wise product.
Unlike standard neural networks, LSTM has feedback connections, which enable it to remember values from earlier stages for future use. This ability to store information over a period of time is useful for managing time-series data.
LSTM has been applied to not only the processing of single data points for uses such as image captioning and generation but also the processing of entire data sequences for machine translation.

4. Data

Data Description

The dataset used in this study contained time-series data extracted from the composite SCFI from its inception in 2009 to April 2020. The total sample size was 548 for each data series, with 7672 data points (Table 2). The time granularity of the data was one week.
First, a stationarity check of dataset was conducted using the Dickey–Fuller (DF) test. Time-series data are said to be stationary if their statistical properties, such as the mean and variance, remain constant over time. In DF testing, the null hypothesis is that the time series is non-stationary, and the test results comprise a test statistic and some critical values relating to the different confidence levels. If the test statistic is less than the critical value, the null hypothesis can be rejected and the series can be considered stationary.
In this study, we chose to apply an augmented version of the DF test that is often applied to larger and more complicated sets of time-series models. The augmented DF (ADF) statistic we used is a negative number that firmly rejects the hypothesis that there is a unit root at some level of confidence as it becomes smaller. For a stationary series, the p-value (0 ≤ p ≤ 1) should be as low as possible and the critical values at different confidence intervals should be close to the test statistic value.
Prior to the treatment, the ADF test statistic was less than 5% of the critical value (Table 3); therefore, the stationarity of the time series was rejected with 95% confidence.
We conducted differencing, seasonal, and trend decomposition treatments using locally estimated scatterplot smoothing (LOESS STL) [28] to decompose trends and seasonality (Figure 2).
After treatment, the ADF test statistic was significantly lower than the 1% critical value (Table 3), indicating that the adjusted time series was close to stationary.
For the implementation, the dataset was split into training and testing sets. Cross-validation was performed with a time step of three for the training dataset (equaling 76% of the total dataset size). The testing dataset (24% of the total dataset size) was used to evaluate the performance of the models outside the training set while avoiding overfitting.

5. Methods

For comparison, we built a SARIMA model and a LSTM model.

5.1. SARIMA Model

The SARIMA model is one of the most important and widely used time-series models. The ARIMA model is the most general class of models for time-series forecasting because it can represent several different types of time series, that is, pure autoregressive (AR, often denoted as p ), pure moving average (MA, often denoted as q ), and integrated versions of stationary series (I, often denoted as d ). ARIMA accepts data that are either non-seasonal or have the seasonal component removed, e.g., data that are seasonally adjusted via methods such as seasonal differencing. A non-seasonal ARIMA model can be classified as A R I M A p , d , q , where:
p : trend autoregressive order.
d : trend difference order (the number of non-seasonal differences needed for stationarity).
q : trend moving average order (the number of lagged forecast errors in the prediction equation).
The general form of the A R M A p , q equation for forecasting time series y is given as:
y t = c + θ 1 y t 1 + θ 2 y t 2 + + θ p y t p + ε t φ 1 ε t 1 φ 2 ε t 2 φ q ε t q
where c is a constant;   y t and ε t are the actual value and random error at time period t , respectively; and θ i i = 1 , , p and φ j j = 0 , , q   are the AR and MA parameters, respectively. The error terms ε t are assumed to be independently and evenly distributed, with a mean of zero and a constant variance of σ 2 . If q = 0 , then Equation (7) becomes an AR model of order p ; if p = 0 , the model reduces to an MA model of order q . To consider the integration (I) part, d is defined in terms of y , the nth difference of Y , as follows:
If d = 0 : y t = Y t ;
If d = 1: y t = Y t 1 ;
If d = 2 : y t = ( Y t Y t 1 ) Y t 1 Y t 2 = Y t 2 Y t 1 + Y t 2 .
Frequent seasonal effects come into play in many time-series datasets, especially in the case of SCFI data because of the seasonal demand of container freight. In response, an SARIMA model can be formulated by including additional seasonal terms in the ARIMA model [29]. As an ARIMA model does not support seasonal data that reflect repeating cycles within a time series, we implemented an SARIMA model to reflect the clear seasonality of the SCFI (e.g., [30]), The seasonal part of the model comprised terms that are very similar to the non-seasonal components but include the backshifts of the seasonal period. In addition to the three hyperparameters p , d , and q in the ARIMA model, four seasonal elements P , D , Q , S were configured as follows:
P : seasonal autoregressive order
D : seasonal difference order.
Q : seasonal moving average order
S : number of time steps for a single seasonal period.
The general form of the SARIMA model is denoted as A R I M A p , d , q   P , D , Q s , where p is the non-seasonal AR order, d is the non-seasonal differencing, q is the non-seasonal MA order, P is the seasonal AR order, D is the seasonal differencing, Q is the seasonal MA order, and s is the time span of the repeating seasonal pattern. The SARIMA model can be written as follows:
y t = c + i = 1 p θ 1 y t 1 + i = 1 P Θ i y t i s + ε t j = 1 q φ j ε t j j = 1 Q Φ j ε t j s
where p , q , P , Q , θ i , and φ j are as defined previously and Θ i   i = 1 ,   ,   i s   and Φ j j = 1 ,   ,   j s are their seasonal counterparts.
Box and Jenkins [31] proposed the use of the ACF and PACF of the sample data as the basic tools for identifying the order of an ARIMA model. The ACF is used to measure the amount of linear dependence between observations in a time series that are separated by a lag of p , whereas the PACF is used to determine the number of necessary autoregressive terms ( q ). Using this approach, it is possible to identify the preliminary values of p , d , q , P , D , and Q . Parameter d is the order of the difference that depends on the stationarity of the time series. To assess stationarity, the order of differencing ( d ) needs to stationarize the series and remove the gross features of seasonality is determined; if d = 0 , the data do not tend to fluctuate over the long term, that is, the model is already stationary.

5.1.1. Autocorrelation Function (ACF)

The ACF is a measure of the correlation between a time series and a lagged version of itself. For instance, the ACF at a lag of five would compare the series at time instant ‘ t 1 t n with the same series at ‘ t 1 5 t n 5 ( where   t 1 5 and t n 5 are the end points). The value of q was obtained from the ACF at y = 0 .

5.1.2. Partial Autocorrelation Function (PACF)

The PACF measures the correlation between a time series and a lagged version of itself after eliminating all variations that have already been explained by intervening comparisons. For example, the PACF at a lag of five checks the correlation described in Section Data Description. but removes the effects already explained by lags one to four. The value of p was obtained from the PACF at y = 0 . Plots of the ACF and PACF are shown in Figure 3.

5.1.3. Parameters for SARIMA Model

The values used for the SARIMA model are summarized in Table 4.

5.2. The Proposed LSTM Architecture

The LSTM parameters include a sequence length, which determines how long the LSTM method should remember information and the dropout, that floats between zero and one. The method also counteracts overfitting and includes other parameters that control training. The values are listed in Table 5. The LSTM model assessed in this study applied the mean-squared loss function over 200 epochs. Here, “epoch” is a hyperparameter, wherein one epoch corresponding to an iteration comprises one forward and one backward pass through the neural network model. Because an entire epoch is too large to be fed into a computer in one step, it is often divided into several smaller batches. We used Adam as our optimizer to control the number of iterations using an early stopping criterion [32].
We tested different parameters of the LSTM model, and the parameters that generated the best results were applied and reported for each freight index. The LSTM model used a neural network with one input layer, 120 hidden layers (dimensions), and one output layer.

5.3. Assessment Metric

The RMSE, which is often used to evaluate trained models for accuracy, is the standard deviation of the residuals or differences between the predicted and observed values. The formula for computing the RMSE is as follows:
RMSE = 1 n i = 1 n y i y ^ i 2
where n is the total number of observations, y i is the observed value, and y ^ i is the predicted value. The main benefit of using the RMSE is that it penalizes large errors and scales the scores into the same units as the forecast values (i.e., per week for this study).
A smaller RMSE indicates a lesser noise and, therefore, a trained model with higher accuracy.

6. Results

The prediction algorithms were implemented using Python version 3.7.3 in MacOS Catalina (10.15.3, MacBook Pro; processor: 2.4 GHz Quad-Core Intel Core i5; memory: 16 GB, 2 × 133 MHz LPDDR3).
The selected SARIMA and LSTM models with the lowest predication errors were used to predict freight indices. Figure 4 shows the prediction of the comprehensive SCFI using SARIMA and LSTM.
We compared the results of different time steps and reported those that generated better performance. The results of the average RMSEs obtained using the rolling SARIMA and LSTM models, presented in Table 6, indicate that for all major routes, the LSTM models significantly outperformed the SARIMA models. The highest forecasting error reduction was 85% for the SAM and SAF routes, reflecting the practices observed in current market. Deep-sea shipping routes (NCMP, MED, USWC, USEC, Persian, ANZ, WAF, SAF, and SAM) are of long distance, require large vessels and containers, and are limited in the number of companies that can participate. In addition, because of the high freight rates of these shipping routes, the majority of shipments are moved under fixed long-term freight contracts while the rest are moved under spot rates. Freight rates under the long-term contracts for these routes are fixed and generally lower than spot rates. Although the SCFI is an indicator of spot rates, shippers decide whether to carry cargo at spot or contract rates depending on market conditions. Therefore, both short-term changes and long-term trends in contracted freight rates affect spot rates.
LSTM is capable of capturing the patterns of both long-term trends such as yearly pattern and short-term trends such as weekly patterns, which explains why LSTM outperformed SARIMA in forecasting deep-sea shipping routes. In addition, deep-sea shipping routes, such as the Europe, America, South America, and South Africa routes, are subject to the influence of multiple expected and unforeseen factors. Deep-learning-based algorithms accommodating both long-term and short-term memories are therefore more suitable for this type of shipping route.
Notably, there were also two trade lanes, namely JPNW and JPNE, where the SARIMA models performed better than LSTM. For short-sea shipping routes, linear models, such as SARIMA, often generate better performance. In addition to the small proportion of long-term contracts, the effect of unanticipated factors is small, which can be attributed to the high explanatory power of the ordinary seasonal fluctuations.
For KOR, there was no significant improvement in the RMSE between the LTSM and SARIMA. These routes have shorter distances than other routes, with relatively large numbers of shipping companies and a few small companies participating. As a result, the average size of the ships in operation is small. For example, as of 2018, there were at least 40 companies on the route between Japan and China and at least 24 companies between China and Korea. The average vessel size is only approximately 1000 TEUs on the Japan–China and China–Korea routes. In addition, Chinese state-owned shipping companies occupy significant positions on all the routes. These carriers sometimes transport cargo without profit, and the base rates sometimes even fall below USD 0. Differences in market environments might influence the forecast accuracy of SARIMA and LTSM.

7. Discussions and Future Studies

Time-series forecasting has been a popular research topic in many fields over the past few decades. The accuracy of time-series forecasting is fundamental to many decision processes. Therefore, concerted efforts have been made to improve the accuracy of forecasting models. In addition to conventional econometric methods, the collection of machine-learning-based techniques for time-series forecasting has grown in recent years. However, there have been no comparative studies between conventional and machine-learning-based models in terms of their ability to forecast the comprehensive SCFI.
The authors of this study presented a deep-learning-based LSTM model for forecasting the SCFI. The results indicated that the LSTM model is effective in predicting freight-rate movement for deep-sea routes, which are easily affected by various factors during each voyage. For example, in predicting freight indices for South America and South Africa routes, LSTM reduced the forecasting error by 76% compared to an SARIMA model. However, for short-sea routes, the SARIMA models outperformed the LSTM models.
Our results are significant due to four major reasons. First, while the study complements the existing literature, LSTM has, to the best of our knowledge, not been previously applied to predict the SCFI.
Second, our results provide insights for improving forecasting accuracy. With the growing availability of large datasets and improvements in algorithms, deep-learning algorithms are becoming increasingly popular in many fields, including shipping. Determining the accuracy and power of these newly introduced approaches relative to conventional methods is of significant interest. The study showed that LSTM is superior to SARIMA in predicting the SCFI for some deep-sea routes and hence highlights the relative advantage of newly introduced approaches. One of the significant contributions of this study is that it highlights that the preferred method may depend on the market conditions of the shipping routes.
Third, our research findings can help relevant parties to understand the overall trends in the container shipping market. The SCFI is one of the most cited metrics for assessing the health and conditions of global trade. Global top carriers refer to the SCFI in their annual reports [33]. The composite SCFI reached its highest value of 5109 on 7 January 2022. This booming market is partially a result of decreased supply and increased demand owing to COVID-19 and the proliferation of e-commerce. During the COVID-19 crisis, carriers blanked large numbers of sailings, which led to a decrease in capacity. Further surges in the demand to prepare for the subsequent spread of infectious diseases have led to a worldwide lack of available container boxes. These conditions contribute to a state of market instability, and better prediction accuracy is essential for planning, managing, and optimizing the use of resources [4]. By offering a better forecasting model, our study provides governments, shippers, carriers, and analysts with insights into the mechanisms of container freight and the health of the container shipping industry.
Fourth, our results have implications for future investments. As the SCFI is often used as the underlying asset in freight derivatives, such as forward freight agreements (FFAs) [4], better predictions of the index would lead to better decision-making by the shipping industry, trading companies, and shippers regarding the use of freight derivatives to hedge against the volatility of freight rates.
Relevant data samples are still limited due to the short history of the SCFI. Future studies may focus on experimenting with other forecast techniques to further improve forecasting accuracy.

Author Contributions

Conceptualization, E.H.; methodology, E.H.; software, E.H.; validation, E.H. and T.M.; data curation, T.M.; writing—original draft preparation, E.H. and T.M.; writing—review and editing, E.H. and T.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Japan Society for the Promotion of Science (JSPS KAKENHI), grant numbers 17K03686, 20K22129, 21H01564 and the Takushoku University Research Fund.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Stopford, M. Maritime Economics 3e; Routledge: London, UK, 2008. [Google Scholar]
  2. Munim, Z.H.; Schramm, H.-J. Forecasting container shipping freight rates for the Far East—Northern Europe trade lane. Marit. Econ. Logist. 2016, 19, 106–125. [Google Scholar] [CrossRef]
  3. Drewry. Container Forecaster Q2/2020. 2020. Available online: https://www.drewry.co.uk/ (accessed on 22 April 2022).
  4. Kavussanos, M.G.; Visvikis, I.D.; Dimitrakopoulos, D.N. Freight Markets and Products. In Handbook of Multi-Commodity Markets and Products: Structuring, Trading and Risk Management; Wiley: Hoboken, NJ, USA, 2014; pp. 355–398. [Google Scholar]
  5. Hirata, E. Contestability of Container Liner Shipping Market in Alliance Era. Asian J. Shipp. Logist. 2017, 33, 27–32. [Google Scholar] [CrossRef]
  6. Hirata, E. A non-structural approach to assess competitive conditions in container liner shipping market: 2009–2014. Int. J. Shipp. Transp. Logist. 2018, 10, 500–513. [Google Scholar] [CrossRef]
  7. Matsuda, T.; Hirata, E.; Kawasaki, T. Monopoly in Container Shipping Market: An Econometric Approach. Marit. Bus. Rev. 2021, ahead-of-print. [Google Scholar] [CrossRef]
  8. Chou, C.-C.; Chu, C.-W.; Liang, G.-S. A modified regression model for forecasting the volumes of Taiwan’s import containers. Math. Comput. Model. 2008, 47, 797–807. [Google Scholar] [CrossRef]
  9. Xie, G.; Zhang, N.; Wang, S. Data characteristic analysis and model selection for container throughput forecasting within a decomposition-ensemble methodology. Transp. Res. Part E Logist. Transp. Rev. 2017, 108, 160–178. [Google Scholar] [CrossRef]
  10. Schulze, P.M.; Prinz, A. Forecasting container transshipment in Germany. Appl. Econ. 2009, 41, 2809–2815. [Google Scholar] [CrossRef] [Green Version]
  11. Kawasaki, T.; Matsuda, T.; Hanaoka, S. An Applicability of SARIMA Model for Forecasting Container Movement from East Asia to U.S. J. Jpn. Logist. Soc. 2013, 21, 167–174. [Google Scholar]
  12. Kawasaki, T.; Matsuda, T. An impact duration of economic indicator on container movement. J. Jpn. Logist. Soc. 2014, 22, 141–148. [Google Scholar]
  13. Chen, S.-H.; Chen, J.-N. Forecasting container throughputs at ports using genetic programming. Expert Syst. Appl. 2010, 37, 2054–2058. [Google Scholar] [CrossRef]
  14. Xiao, J.; Xiao, Y.; Fu, J.; Lai, K.K. A transfer forecasting model for container throughput guided by discrete PSO. J. Syst. Sci. Complex. 2014, 27, 181–192. [Google Scholar] [CrossRef]
  15. Huang, A.; Lai, K.K.; Li, Y.; Wang, S. Forecasting container throughput of Qingdao port with a hybrid model. J. Syst. Sci. Complex. 2014, 28, 105–121. [Google Scholar] [CrossRef]
  16. Xie, G.; Wang, S.; Zhao, Y.; Lai, K.K. Hybrid approaches based on LSSVR model for container throughput forecasting: A com-parative study. Appl. Soft Comput. 2013, 13, 2232–2241. [Google Scholar] [CrossRef]
  17. Mo, L.; Xie, L.; Jiang, X.; Teng, G.; Xu, L.; Xiao, J. GMDH-based hybrid model for container throughput forecasting: Selective combination forecasting in nonlinear subseries. Appl. Soft Comput. 2018, 62, 478–490. [Google Scholar] [CrossRef]
  18. Luo, M.; Fan, L.; Liu, L. An econometric analysis for container shipping market. Marit. Policy Manag. 2009, 36, 507–523. [Google Scholar] [CrossRef]
  19. Koyuncu, K.; Tavacioğlu, L. Forecasting Shanghai Containerized Freight Index by Using Time Series Models. Mar. Sci. Technol. Bull. 2021, 10, 426–434. [Google Scholar] [CrossRef]
  20. Chen, Y.; Liu, B.; Wang, T. Analysing and forecasting China containerized freight index with a hybrid decomposition–ensemble method based on EMD, grey wave and ARMA. Grey Syst. Theory Appl. 2021, 11, 358–371. [Google Scholar] [CrossRef]
  21. Munim, Z.H. State-space TBATS model for container freight rate forecasting with improved accuracy. Marit. Transp. Res. 2022, 3, 100057. [Google Scholar] [CrossRef]
  22. Tealab, A. Time series forecasting using artificial neural networks methodologies: A systematic review. Future Comput. Inform. J. 2018, 3, 334–340. [Google Scholar] [CrossRef]
  23. Munim, Z.H.; Schramm, H.-J. Forecasting container freight rates for major trade routes: A comparison of artificial neural networks and conventional models. Marit. Econ. Logist. 2021, 23, 310–327. [Google Scholar] [CrossRef]
  24. Hassan, L.A.H.; Mahmassani, H.S.; Chen, Y. Reinforcement learning framework for freight demand forecasting to support operational planning decisions. Transp. Res. Part E Logist. Transp. Rev. 2020, 137, 101926. [Google Scholar] [CrossRef]
  25. Barua, L.; Zou, B.; Zhou, Y. Machine learning for international freight transportation management: A comprehensive review. Res. Transp. Bus. Manag. 2020, 34, 100453. [Google Scholar] [CrossRef]
  26. Ubaid, A.; Hussain, F.K.; Charles, J. Machine Learning-Based Regression Models for Price Prediction in the Australian Container Shipping Industry: Case Study of Asia-Oceania Trade Lane. In International Conference on Advanced Information Networking and Applications; Springer: Berlin/Heidelberg, Germany, 2020; pp. 52–59. [Google Scholar]
  27. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  28. Cleveland, R.B.; Cleveland, W.S.; McRae, J.E.; Terpenning, I. STL: A seasonal-trend decomposition. J. Off. Stat. 1990, 6, 3–73. [Google Scholar]
  29. Hyndman, R.J.; Athanasopoulos, G. Forecasting: Principles and Practice; Otexts: 2018. Available online: https://books.google.com/books?hl=en&lr=&id=_bBhDwAAQBAJ&oi=fnd&pg=PA7&dq=+Forecasting:+Principles+and+Practice&ots=TijWtg0QGO&sig=Ebcpx6Ib1uyLHVOUL7iEV8slLTc#v=onepage&q=Forecasting%3A%20Principles%20and%20Practice&f=false (accessed on 22 April 2022).
  30. Jeon, J.-W.; Duru, O.; Yeo, G.-T. Modelling cyclic container freight index using system dynamics. Marit. Policy Manag. 2020, 47, 287–303. [Google Scholar] [CrossRef]
  31. Box, G.E.; Jenkins, G.M.; Reinsel, G.C.; Ljung, G.M. Time Series Analysis: Forecasting and Control; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  32. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980v9. [Google Scholar]
  33. Flexport. The History of the Shanghai Containerized Freight Index. 2016. Available online: https://www.flexport.com/blog/shanghai-containerized-freight-index-scfi-history/ (accessed on 22 April 2022).
Figure 1. Architecture of LSTM (author-created).
Figure 1. Architecture of LSTM (author-created).
Jmse 10 00593 g001
Figure 2. STL decomposition of data.
Figure 2. STL decomposition of data.
Jmse 10 00593 g002
Figure 3. ACF and PACF.
Figure 3. ACF and PACF.
Jmse 10 00593 g003
Figure 4. Plots of (a) SARIMA and (b) LSTM prediction results for comprehensive SCFI.
Figure 4. Plots of (a) SARIMA and (b) LSTM prediction results for comprehensive SCFI.
Jmse 10 00593 g004
Table 1. Literature review summary of forecasting techniques in the shipping sector.
Table 1. Literature review summary of forecasting techniques in the shipping sector.
Reference ApplicationMethodFindings
Chou et al. [8]Container trade volumesVARA modified regression model for forecasting Taiwan’s inbound container volumes was proposed.
Kawasaki and Matsuda [11]Container trade volumesSARIMAThe residual had no autocorrelation, and the defined model could reproduce volumes of container movement with high accuracy.
Kawasaki and Matsuda [12]Container trade volumesVARThe VAR model clarified the length of time that U.S. economic indicators affect the container trade movement from East Asia to the U.S.
Chen and Chen [13]Container port throughputDecomposition approach, SARIMA, GPThe GP model predictions were approximately 32–36% better than the decomposition approach and SARIMA.
Huang et al. [15]Container port throughputPPR with GP, ANN, SARIMA, and PPRThe proposed PPR-with-GP method significantly outperformed ANN, SARIMA, and PPR models.
Mo et al.
[17]
Container port throughputSARIMA, SVR, BP, and GP A hybrid model was developed by applying SARIMA for the linear part and SVR, BP, and GP for the nonlinear part of the data. Results showed that the hybrid model performed better than others.
Xiao et al. [14]Container port throughputTF-DPSO, ARIMA, and Elman Neural network (ENN)The forecasting performance of TF-DPSO was better than that of several existing models.
Xie et al. [16]Container port throughputLSSVR and OLSThe proposed hybrid LSSVR approach achieved better forecasting performance than OLS.
Xie et al. [9]Container port throughputsARIMA, SARIMA, and LSSVRData characteristic analysis results suggest that hybrid models can perform better than other methods.
Schulze and Prinz [10] Container port throughput and trade volumesSARIMA and exponential smoothing approachThe SARIMA approach yielded slightly better-modeled values of container throughput than the exponential smoothing approach.
Luo et al. [18]Container freightThree-stage Least Squared (3SLS) The overall model could explain more than 90% of the fleet capacity and freight-rate variations.
Munim and Schramm [2]Container freight—FE–FE route of SCFIARIMARCHThe ARIMARCH model provided better results than the existing freight-rate forecasting models.
Munim and Schramm [23]Container freight—SCFI, 4 routesARIMA, VAR, ANN, and SVROverall, VAR/VEC models outperformed ARIMA and ANN in training-sample forecasts.
Al Haji Hassan et al. [24]Freight demand in the U.S.time-series models with reinforcement learning frameworkA new approach building on typical time-series and machine learning models to forecast the freight demand in the U.S. market was proposed.
Ubaid et al. [26]Container freight, Asia/Oceania routeSVR, RFR, and GBRThe GBR model outperformed other models, with a test accuracy of 84%.
Chen et al. [20]Container freight—CCFIDecomposition–ensemble method based on EMD and grey wave forecasting modelThe proposed method performed better than random walk and ARMA in multi-step-ahead prediction.
Koyuncu and Tavacıoğlu [19]Container freight—SCFISARIMA and Holt–Winters MethodsThe SARIMA model provided better results than the existing freight-rate forecasting models while performing short-term forecasts on a monthly rate.
Munim [21]Container freight—CCFISARIMA, SNNAR, and the state-space TBATS modelThe TBATS model or a combination of TBATS and SARIMA forecasts outperformed SARIMA and SNNAR, as well as their combinations.
Source: Author-compiled (models in bold letters indicate machine-learning-based models).
Table 2. Description of data.
Table 2. Description of data.
IndexCountMeanStandard DeviationMinimum25%50%75%Maximum
SCFI (comprehensive index54899432440079595111172885
NCMP (North Europe)548102751320573890712184452
MED (Mediterranean)548107050819576194412624298
USEC (U.S. East Coast)54818736307251457177421154080
USWC (U.S. West Coast)548306874414482563309934555049
Persian (Persian Gulf and Red Sea)5487392982115307119021995
ANZ (Australia/New Zealand)5487983782495427609582490
WAF (East/West Africa)54820236978411624194323236630
SAF (South Africa)54894938130774886910813307
SAM (South America)54817131081991097160921198907
JPNW (West Japan)5482537577215231332382
JPNE (East Japan)5482597066214238333389
SEA (Southeast Asia)54820411953145185235974
KOR (Korea)5481623786125162191234
Table 3. Results of DF test for SCFI.
Table 3. Results of DF test for SCFI.
Before TreatmentAfter Treatment
Test Statistic−0.1205−4.6303
p-value0.9474−0.0001
Number of Lags Used1217
Number of Observations Used550544
Table 4. Parameter ranges used for the SARIMA model.
Table 4. Parameter ranges used for the SARIMA model.
Non-SeasonalSeasonal
Parameters p Q d P Q D
312011
Table 5. Values used for respective LSTM model parameters.
Table 5. Values used for respective LSTM model parameters.
ParameterValue
Sequence Length120
Dropout0.0 (do not drop for the linear transformation of the inputs)
Epochs200
Input dimension120
Activation FunctionTanh
Recurrent Activation FunctionSigmoid
Optimization FunctionAdam
LossMean squared error
Table 6. RMSE results.
Table 6. RMSE results.
RMSESARIMALSTMTime StepsReduction in RMSE
SCFI49.1317.62564%
NCMP124.2628.43977%
MED129.8439.01970%
USWC144.5226.47982%
USEC140.0020.97485%
Persian55.329.98382%
ANZ54.889.46483%
WAF110.3822.45780%
SAF70.7615.15379%
SAM223.8032.58385%
JPNW3.197.759−143%
JPNE2.717.738−185%
SEA25.7810.86358%
KOR7.666.9939%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hirata, E.; Matsuda, T. Forecasting Shanghai Container Freight Index: A Deep-Learning-Based Model Experiment. J. Mar. Sci. Eng. 2022, 10, 593. https://doi.org/10.3390/jmse10050593

AMA Style

Hirata E, Matsuda T. Forecasting Shanghai Container Freight Index: A Deep-Learning-Based Model Experiment. Journal of Marine Science and Engineering. 2022; 10(5):593. https://doi.org/10.3390/jmse10050593

Chicago/Turabian Style

Hirata, Enna, and Takuma Matsuda. 2022. "Forecasting Shanghai Container Freight Index: A Deep-Learning-Based Model Experiment" Journal of Marine Science and Engineering 10, no. 5: 593. https://doi.org/10.3390/jmse10050593

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop