Next Article in Journal
Crypto Asset Portfolio Selection
Next Article in Special Issue
The Non-Fungible Token (NFT) Market and Its Relationship with Bitcoin and Ethereum
Previous Article in Journal / Special Issue
Recent Development in Fintech: Non-Fungible Token
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Modeling and Forecasting Cryptocurrency Closing Prices with Rao Algorithm-Based Artificial Neural Networks: A Machine Learning Approach

by
Sanjib Kumar Nayak
1,
Sarat Chandra Nayak
2,* and
Subhranginee Das
3
1
Department of Computer Application, VSS University of Technology, Burla, Sambalpur 768018, India
2
Department of Artificial Intelligence and Machine Learning, CMR College of Engineering & Technology, Hyderabad 501401, India
3
Department of Computer Science and Engineering, Koneru Lakshmaiah Education Foundation (KL University), Hyderabad 500075, India
*
Author to whom correspondence should be addressed.
FinTech 2022, 1(1), 47-62; https://doi.org/10.3390/fintech1010004
Submission received: 20 September 2021 / Revised: 9 November 2021 / Accepted: 10 November 2021 / Published: 30 December 2021
(This article belongs to the Special Issue Recent Development in Fintech)

Abstract

:
Artificial neural networks (ANNs) are suitable procedures for predicting financial time series (FTS). Cryptocurrencies are good investment assets; therefore, the effective prediction of cryptocurrencies has become a trending area of research. Capturing inherent uncertainties associated with cryptocurrency FTS with conventional methods is difficult. Though ANNs are the better alternative, fixing the optimal parameters of ANNs is a tedious job. This article develops a hybrid ANN through Rao algorithm (RA + ANN) for the effective prediction of six popular cryptocurrencies such as Bitcoin, Litecoin, Ethereum, CMC 200, Tether, and Ripple. Six comparative models such as GA + ANN, PSO + ANN, MLP, SVM, LSE, and ARIMA are developed and trained in a similar way. All these models are evaluated through the mean absolute percentage of error (MAPE) and average relative variance (ARV) metrics. It is found that the proposed RA + ANN generated the lowest MAPE and ARV values, statistically different as compared with existing methods mentioned above, and hence can be recommended as a potential financial instrument for predicting cryptocurrencies.

1. Introduction

Cryptocurrency is a virtual or digital currency that is meant to be a means of exchange for online transactions to purchase goods and services. An online ledger is used with strong cryptography for securing online transactions. Cryptos evince a lot of interest today because they are traded for-profit and because of the rich and growing valuations in relatively shorter periods of time. While these currencies are unregulated, many of the international governments and countries, after an early period of denial, are now showing signs of the adoption of some of the other forms of cryptocurrencies for transactions. Without any bank or any central authority, these cryptos are built on a decentralized or distributed peer-to-peer network. Cryptocurrencies have certain special characteristics [1]. They use decentralized control, so central authority is not required.
To avoid dispute, it ensures pseudo-anonymity. Since the reproduction of digital currency is easy, double-spending attack protection is ensured.
While Bitcoin was the earliest and most popular one [2], there are also many others, which are continuing to grow in user base, popularity, and market share, such as Ethereum, Tether, CMC 200, Ripple, Polkadot, XRP, Cardano, Chainlink, Litecoin, Bitcoin Cash, Binance Coin, etc. As the cryptocurrency market prices are growing fast and behave similarly to stock market price movement, investors, as well as researchers, are concerned about the proper prediction of cryptocurrencies market values. The principle of cryptography is the basis of the decentralized digital currency having an open source. The relevance of cryptocurrencies as an emerging market in the financial world is escalating significantly. Nowadays, more and more financial organizations are getting concerned about cryptocurrency trading. The trading opportunities and profitability in the cryptocurrency market are studied in [3]. A survey on systems of cryptocurrency is elaborated in [4]. The characteristics of this market, such as the high availability of market data, high volatility, smaller capitalization, and decentralized control, are studied in the literature [5]. Despite the fact that this market acts similarly to other stock markets in terms of intrinsic volatility, investor confidence has been reflected in it [6,7,8]. The feasibility of this market in relation to other financial markets is documented in the literature [9,10]. Similarly, to another stock market, cryptocurrency market prices behaved arbitrarily and with high nonlinearity and dynamics. Though few computational intelligence models are available for the prediction of cryptocurrencies, sophisticated methods with accurate prediction abilities are still missing and need to be searched. From the literature, it is observed that different stock market predictions have been made successively by different ANN techniques [11,12,13,14,15,16]. Motivated by such works in the last few years, different soft computing methods are suggested in the literature to predict cryptocurrencies indices. Soft computing techniques, such as Long short-term memory neural networks and MLP [17], Random walk theory [18], Support vector machines [19], etc., are already used to forecast cryptocurrencies. The recurrent neural network was also used for entrapping threats in cryptocurrency [20]. Few machine learning and statistical methodologies are found in the literature for the prediction of Bitcoin [21,22,23,24,25]. Several data-driven and machine learning approaches are developed in the literature for industrial and agricultural applications. A data-driven prognosis approach for the prediction of run-to-failure low-speed slew bearing vibration data is proposed in [26]. A machine learning-based method is proposed in [27] for clay prediction technology. A hybrid architecture is proposed in [28] for the minimization of energy consumption and elapsed time for IoT workloads. The forecasting accuracy of a neural model is greatly subjective to the network magnitude and learning method. Gradient descent learning is a common method for neural network training. The typical downsides of this technique are that it has a slow convergence rate, imperfect learning, and is prone to local minima. It adds computational overhead to the model. Therefore, evolutionary optimization techniques are used to train the ANN and claim superior to [29,30]. The improper selection of algorithm-specific parameters of evolutionary optimization methods may land them at local optima or inaccurate solutions. To eradicate this, a good number of swarm and evolutionary optimization techniques are groomed up and heavily used for ANN training. These techniques are broadly termed nature-inspired optimization algorithms because they mimic natural processes. Certain algorithms, such as genetic algorithms, differential evolution, and evolution strategy, are based on the theory of natural evolution. Few algorithms simulate the behavior of swarms, insects, etc., such as particle swarm optimization, ant colony optimization, artificial bee colony optimization, bacterial foraging optimization, monarch butterfly optimization, and so on. Some other techniques are bio-geography optimization, gravitational search algorithm, multi-verse optimization [31], fireworks algorithm [32,33], chemical reaction optimization [34,35], etc. Though such nature-inspired optimization techniques have been used for solving several complex problems, their efficiency is determined by well-adjustable learning parameters. In the quest to search for the optimal global solution, the selection of appropriate learning parameters makes the technique difficult to use. Choosing optimal learning parameters for any particular problem requires a number of trial-and-error methods. The improper selection of algorithm-specific parameters may land the search operation at local optima or inaccurate solutions. As a result, an optimization technique with higher approximation capability but fewer learning parameters will be of relevance for better forecasting accuracy. Recently, few methods have been developed for constrained and unconstrained optimization problems, which are claimed to be simple as well as parameter-free algorithms, such as teaching learning-based optimization [36] and Jaya algorithm [37]. With the objective of developing simple and effective optimization techniques, Rao proposed metaphor-less algorithms called Rao algorithms [25], where algorithm-specific learning parameters are not required, i.e., these algorithms are free from learning parameters. Only the initialization of population size is enough. The algorithms keep on finding the best and the worst solution from the selected population and initiate random interaction among the candidate solutions. Very soon, these algorithms are applied to mechanical system components to obtain the design optimization [14,38,39], estimation of photovoltaic cell [40,41], optimal coordination of overcurrent relays [42], etc. However, Rao algorithms are not explored for the optimization of ANN parameters. This article attempts to test out the appropriateness of Rao algorithms on finding the optimal parameters of ANNs and applying the resultant method for stock closing price prediction.
The objectives of this article are highlighted as follows:
  • We are designing an efficient ANN forecast for the nearly precise prediction of cryptocurrencies such as Bitcoin, Litecoin, Ethereum, CMC 200, Tether, and Ripple.
  • Suitable tuning of ANN parameters (i.e., weight and bias) by RA, thus forming a hybrid model (i.e., RA + ANN) to overcome the limitations of derivative-based optimization techniques.
  • We are evaluating the performance of the RA + ANN forecast through two performance metrics, MAPE and ARV.
The flow of the article is as follows. A concise description of ANNs is presented in Section 2. The proposed RA + ANN is described in Section 3. The experimental data are presented in Section 4. Section 5 contains the summarized results from the experimentation and analysis of outcomes, followed by concluding remarks.

2. Artificial Neural Network

An ANN simulates the way the human brain works, making it different from conventional digital computers [43]. ANNs are also capable of complex data processing having neurons as computational units. ANNs have the capabilities of getting good approximation solutions to intractable problems, and they can also solve a very complex problem by decomposing it into smaller tasks, which makes them different from conventional computing.
The neural networks gain training by detecting the patterns in the data. As the human brain consists of billions of neurons, fully interconnected by synapses, similarly ANNs consist of hundreds of processing units as artificial neurons, which are fully connected through neural links. The Schematic block diagram (Figure 1) represents the basic ANN architecture with some hidden layers of neurons and one output layer neuron.
Supervised learning is used for error correction in this case, i.e., the expected response for the given inputs is submitted at the output neuron to calculate the error. In the time-series data to speculate one-day-ahead index, a single output unit model is used. A linear transfer function is used in the input layer. The hidden and output layer neurons contain nonlinear activation functions. Here, we have taken nonlinear activation as sigmoidal function in (1), i.e.,:
  y out = 1 1 + e λ y in
where y out and y in represent the output and input of the neuron y , respectively, and λ acts as the sigmoidal gain. In the input layer, let there be n neurons. The input layer stands for the input variable of the problem. Each input variable input layer contains one-one node. The hidden layers are helpful in capturing the nonlinearity among variables. In the hidden layer output y j   is computed using (2) for each neuron j,
y j = f ( b j + i = 1 n w ij x i )  
where f stands for a nonlinear activation function, b j is the bias value, x i represents the i th component of the input vector,   w ij is the synaptic weight connecting i th input neuron and j th hidden neuron.
Suppose there are m numbers of neurons present in this hidden layer, then for the next hidden layer these m outputs become the input. Then (3) for each neuron j of the next hidden layer can be represented as,
  y j = f ( b j + i = 1 m w ij y i )
This signal flows in the forward direction through each hidden layer until it reaches the node of the output layer.
For a single output neuron (4) is used to calculate the output y esst
y esst = f ( b o + j = 1 m v j y j )
where b o is the bias for output node, the weighted sum calculated as in (2) is y j , v j   represents the synaptic weight between j th hidden neuron and output neuron.
Given a set of training samples S = { x i , y i } i = 1 N to train the ANN, let y i be the output of i th input sample, and y esst is the computed output of the same i th input, then using (5) the error is calculated as,
  Error i = y i   y esst
The error value that is produced by n th training sample at the output of neuron i is defined by (6) as,
Error i ( n ) = y i ( n )   y esst ( n )
Then the instantaneous error at neuron i is defined by (7) as,
ε i ( n ) = 1 2 Error i 2 ( n )
Hence the total instantaneous error of the whole network, ε ( n ) can be defined by the following (8) as,
ε ( n ) = i ϵ C ε i ( n )
Here C represents the set containing all the output layer neurons. In this paper, we have considered one neuron in the output layer. Therefore, here we can represent the network error in the neural network model as in (9),
  ε ( n ) = ε i ( n )  
The average error over the whole training sample in the neural network model can be defined by (10) as,
ε av ( N ) = n = 1 N ε ( n )
The optimal weight vector is computed by the weight update algorithm using error signal Error i . In the training phase, the training set input vectors are repeatedly presented to the neural network model to update the weights and the biases by training algorithm. If the error is small enough per epoch, then no further training is required. Unfortunately, this criterion may lead to local minimum. Therefore, the performance of the network is tested for generalization performance after each iteration of training. Once the generalization performance is sufficient or adequate, the learning process is stopped. Therefore, the intent is to reduce the total error of Equation (5) with an optimized set of weight and bias vector of the ANN.

3. Proposed RA + ANN Based Forecasting

Rao algorithms are newly developed population-based optimization algorithms. Unlike other nature-inspired optimization techniques, these are not mimicking the behavior of swarms, animals, birds, or any physical or chemical phenomenon. Therefore, they are claimed as metaphor-less optimization techniques by the inventor [25]. Any algorithm-specific parameters are not required by these algorithms. Only specifying the number of candidate solutions, design variables, and termination criteria is sufficient to automate the search process. The techniques are simple, do not necessitate human intervention, and provide effective solutions to complicated problems. The algorithms are presented as follows.
Let f ( w ) be an objective function that needs to be optimized (i.e., the error function to be minimized here). Let the population (search space of ANN model) have n number of candidate solutions (weight and bias vector), each of which has m number of design variables (weight values). Each candidate solution has a fitness value (i.e., error value), and the lower the error signal, the better is the solution. At any iteration i , let the best and worst solution of the population be f ( w ) b e s t   and f ( w ) w o r s t respectively. The current value of a candidate solution (weight vector) at i t h   iteration is updated as per the following Equations (11)–(13) as,
  W j , k , i = W j , k , i + r a n d 1 , j , i ( W j , b e s t , i W j , w o r s t , i )
W j , k , i = W j , k , i + r a n d 1 , j , i ( W j , b e s t , i W j , w o r s t , i ) + r a n d 2 , j , i ( | W j , k , i   o r   W j , l , i | | W j , l , i   o r   W j , k , i | )  
W j , k , i = W j , k , i + r a n d 1 , j , i ( W j , b e s t , i | W j , w o r s t , i | ) + r a n d 2 , j , i ( | W j , k , i   o r   W j , l , i | W j , l , i   o r   W j , k , i )  
where:
  • W j , k , i   = the value of j t h variable of k t h solution ( k = 1 , 2 , 3 , , n ) at i t h iteration.
  • W j , k , i   = the modified value of j t h variable of k t h solution ( k = 1 , 2 , 3 , , n ) at i t h iteration.
  • W j , b e s t , i = j t h variable value of the best solution in i t h   iteration.
  • W j , w o r s t , i = j t h variable value of the worst solution i t h   iteration.
  • r a n d 1 , j , i   and r a n d 2 , j , i are two random values in [0,1].
The terms “ W j , k , i   o r   W j , l , i ” in (11)–(13) represent the fitness comparison of k t h candidate solution with that of l t h solution, which is randomly drawn from the population. An information exchange occurs based on their fitness values. This ensures the communication among the candidate solutions. Based on the concept discussed, the RA + ANN-based forecasting can be explained by Algorithm 1. The pictorial description is given below (Figure 2).
Algorithm 1: RA + ANN-based forecasting.
1. Set population size (n), No. of design variables (m), and Termination criteria.
2. Initialization of population.
3. Set training and test data using sliding window.
4. Normalization of training and test data.
  /* Model Training */
5. While (! = termination criteria)
       For each candidate solution (W) in the population.
          Supply train data and W to ANN.
          Compute ANN output.
          Error = expected output–estimated output
          Fitness = accumulated error.
       End
          Identify best and worst solution.
          Update population using any Equations (8)–(10).
6. End
  /* Model Testing */
7. Feed test data and best solution to ANN.
          Calculate the model output error value.
8. Reiterate Steps 2–6 for all training and test patterns and preserve total error.
The RA + ANN forecasting model works as follows: the cryptocurrency’s historical closing prices are collected. Approximately 60% of the total input data are pre-owned for training the neural net, and after the network is trained, the remaining 40% of the data is used for testing the correctness of the proposed model. The closing price data are collected and stored as time-series data. We have used the sliding window approach to generate the training and testing data sets. The input for the network is normalized in the range [0,1]. The normalized input is fed into the network to obtain an output. The difference between the computed output (estimated) and expected output (actual) gives an error. The Rao algorithm is used to update the weights to obtain the minimum error. The performance of the network is tested for generalization performance after each iteration of training. Here, Rao algorithms work on identifying the best and worst solution and interaction between contestants in the search space, i.e., the population. The lack of algorithm-specific parameters eliminates any human interventions. In Algorithm 1, the size of the population, design variables, and stopping criteria are needed to be fixed at the beginning. The size of the sliding window used to form training and test patterns is a matter of experimentation.

4. Cryptocurrency Data

The models are evaluated on experimenting with the cryptocurrency’s historical closing prices collected from https://www.CryptoDataDownload.com (accessed on 26 March 2021). The currencies are Bitcoin, Litecoin, Ethereum, CMC 200, Tither, and Ripple. The data are collected during the period from 1 January 2019 to 25 March 2021. A condensed statistic of the six datasets is gathered in Table 1. The price series are plotted in Figure 3. The raw data are normalized using the sigmoid data normalization method [44].

5. Experimental Results and Analysis

To evaluate the models, we used them to forecast the prices of four cryptocurrency datasets separately. To access the capacity of the RA + ANN method, five other methods such as ANN trained by genetic algorithm (GA + ANN), particle swarm optimization trained ANN (PSO + ANN), MLP, SVM, ARIMA, and least squared estimator (LSE) are developed in this study and a comparison is performed.

5.1. Model Input Selection and Normalization

A time series approach is used in this study. A rolling window method is used to generate the train and test patterns from the dataset. The method is depicted in Figure 4.
Each sample in the dataset is considered as a data point on the time series. A window of fixed size is rolled over the time series. On each movement, an old datapoint is dropped and a new datapoint is included, as shown in Figure 4. The number of datapoints included by the window at any instant of time is considered as one training sample for the model. The width of the window is determined experimentally. For example, one train/test set is formed by the rolling window of width three, as follows.
x ( k ) x ( k + 1 ) x ( k + 2 ) x ( k + 1 ) x ( k + 2 ) x ( k + 3 ) x ( k + 2 ) x ( k + 3 ) x ( k + 4 ) T r a i n i n g   d a t a x ( k + 3 ) x ( k + 4 ) x ( k + 5 ) T a r g e t
x ( k + 3 ) x ( k + 4 ) x ( k + 5 ) T e s t   d a t a x ( k + 6 ) T a r g e t
Each input pattern is then normalized to scale the data into same range for each input feature to diminish the bias [44,45]. The t a n h normalization method as in Equation (14) is used to standardize the input data. The mean and standard deviation of a training pattern are represented as µ and σ respectively.
x ^ = 0.5 ( t a n h ( 0.01 ( x μ ) σ ) + 1 )

5.2. Performance Evaluation Metrics

The accuracy of all models is accessed through two mesures, i.e., Mean Absolute Percentage of Errors (MAPE) and Average Relative Variance (ARV), as in (15) and (16).
MAPE = 1 N i = 1 N | x i x ^ i | x i × 100 %
ARV = i = 1 N ( x ^ i x i ) 2 i = 1 N ( x ^ i X ¯ ) 2  

5.3. Experimental Setting

The train and test patterns are generated as described in Section 5.1, and the same patterns are fed to all seven models separately. To overcome the biases of the models, each model is simulated twenty times, and the mean error value is recorded for comparisons. An ANN with one hidden layer is used as the base network. The input layer size is equal to the rolling window size, and the output layer has only one neuron, as there is one target output. However, the numbers of neurons in the hidden layer are decided experimentally. An inadequate number of hidden neurons may produce poor accuracy, whereas the excess number of such neurons adds computational overhead. Therefore, the size of the hidden layer impacts the model performance a lot and must be decided carefully. During experimentation, different possible values for the model parameters were tested, and the best values were recorded. The suitable parameter values obtained during the simulation process are called simulated parameters. For the MLP model, the learning rate (α) and momentum factor (µ) are set to 0.2 and 0.5, respectively. The number of iterations was fixed to 400, and gradient descent-based backpropagation learning was adopted for training. For SVM, we used the radial basis function as the kernel. For PSO + ANN, the total number of the particle was set to 100, the number of iterations was 250, learning factors were fixed to 3 each. Similarly, in the case of GA + ANN, the three critical parameters, i.e., population size, crossover, and mutation probabilities, are considered as 80, 0.7, and 0.02, respectively. The stopping condition is the maximum number of generations, i.e., 200. Binary encoding is used for individual encoding, and elitism is used as the selection method. However, the RA + ANN used only 50 candidate solutions and 100 number iterations to reach the global optima. All the experimentations are performed with a system of Intel core i3 CPU, 2.27GHz and 8.0 GB memory, and a Matlab-2015 program writing environment.

5.4. Simulation Results and Discussion

The mean values from twenty simulations for each forecast are recorded and considered as the model performance. The MAPE and ARV values produced by seven models from six FTS are encapsulated in Table 2. It may be seen that the RA + ANN has MAPE values lower than others. For example, MAPE of RA + ANN is 0.0300 from Bitcoin, 0.0322 from Litecoin, 0.0397 from Ripple, and 0.0385 from Ethereum. It also achieved the lowest ARV values for all six FTS datasets. The overall performance of RA + ANN is better than others. Furthermore, to show the goodness of the proposed RA + ANN, the estimated prices are plotted against the expected and depicted in (Figure 5, Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10), respectively. It is apparent that the predicted prices are very close to the actual prices and follows the tendency accurately. For the sake of clear visibility, we plotted the estimated prices against actual prices for one financial year data (approximately 252 financial days).
We also conducted two well-known statistical significance tests, such as the Wilcoxon signed test and the Diebold–Mariano test, to establish the difference between the RA + ANN forecast and comparative methods statistically considering the MAPEs from all the models. The test outcomes are summarized in Table 3. The h = 1 indicates the rejection of the null hypothesis that the proposed and comparative forecasts are not statistically different. In the case of the Diebold–Mariano test, if the statistic falls beyond the range of −1.976 to +1.976, then the null hypothesis will be rejected and h = 1. These results are in support of the rejection of the null hypothesis and prove that there is a significant difference between RA + ANN and other methods.
From the simulation studies, comparative analysis, and significance test results, the following major points are drawn.
  • The RA + ANN-based forecast was found quite capable in capturing the inherent dynamism and uncertainties associated with cryptocurrency data.
  • The hybridization of RA and ANN achieved improved forecasting accuracy compared to others.
  • The outcomes from the statistical test justified the significant difference between RA + ANN and others.

6. Conclusions

A hybrid ANN through Rao algorithm-based optimization (RA + ANN) is developed in this article using a single hidden layer ANN as the base neural architecture and RA as the search technique for exploiting the optimal ANN parameters. The Rao algorithms are recently proposed metaphor-less optimization techniques that are simple and do not need an algorithm-specific parameter to be adjusted. This article explored the modality of RA on training the ANN for forecasting cryptocurrency data such as Bitcoin, Litecoin, Ethereum, CMC 200, Tether, and Ripple. The RA + ANN predictability is compared with that of GA + ANN, PSO + ANN, MLP, SVM, ARIMA, and LSE in terms of MAPE and ARV. The proposed RA + ANN model has a faster convergence rate and better generalization ability compared to gradient descent learning. Furthermore, to ensure the effectiveness of RA + ANN, the Wilcoxon signed test and Diebold–Mariano test are conducted. From the comparative analysis of experimental results and outcomes of statistical significance tests, it is concluded that the RA + ANN-based forecasting is good enough to follow the dynamic trend of cryptocurrencies in comparison with the other forecasts under consideration. In the future, more ANN structures and sophisticated learning approaches may be investigated. Other factors such as the economic and technical determinants along with the closing prices may be used for achieving better forecasting accuracy. The proposed RA + ANN framework can also be applied for the efficient prediction of other existing FTS.

Author Contributions

Conceptualization: S.C.N.; data curation: S.C.N.; formal analysis: S.C.N.; investigation: S.C.N.; methodology: S.C.N.; implementation: S.C.N.; resources: S.D.; software: S.D.; supervision: S.C.N.; validation: S.D.; visualization: S.C.N.; writing—original draft: S.C.N.; and S.K.N.; writing—review and editing: S.K.N. and S.C.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

The datasets analyzed and experimented during the current study are available at https://www.CryptoDataDownload.com (accessed on 26 March 2021), which is openly available for access by researchers.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lansky, J. Possible state approaches to cryptocurrencies. J. Syst. Integr. 2018, 9, 19–31. [Google Scholar] [CrossRef]
  2. Nakamoto, S. Bitcoin: A peer-to-peer electronic cash system. Decentralized Bus. Rev. 2019, 21260. Available online: https://www.debr.io/article/21260.pdf (accessed on 26 March 2021).
  3. Kyriazis, N.A. A survey on efficiency and profitable trading opportunities in cryptocurrency markets. J. Risk Financ. Manag. 2019, 12, 67. [Google Scholar] [CrossRef] [Green Version]
  4. Mukhopadhyay, U.; Skjellum, A.; Hambolu, O.; Oakley, J.; Yu, L.; Brooks, R. A brief survey of cryptocurrency systems. In Proceedings of the 2016 14th annual conference on privacy, security and trust (PST), Auckland, New Zealand, 12–14 December 2016; pp. 745–752. [Google Scholar]
  5. Ferreira, M.; Rodrigues, S.; Reis, C.I.; Maximiano, M. Blockchain: A tale of two applications. Appl. Sci. 2018, 8, 1506. [Google Scholar] [CrossRef] [Green Version]
  6. Grinberg, R. Bitcoin: An innovative alternative digital currency. Hastings Sci. Technol. Law J. 2012, 4, 159. [Google Scholar]
  7. Mai, F.; Shan, Z.; Bai, Q.; Wang, X.; Chiang, R.H. How does social media impact Bitcoin value? A test of the silent majority hypothesis. J. Manag. Inf. Syst. 2018, 35, 19–52. [Google Scholar] [CrossRef]
  8. Sapuric, S.; Kokkinaki, A. Bitcoin Is Volatile! Isn’t That Right? Proceedings of the International Conference on Business Information Systems, Larnaca, Cyprus, 22–23 May 2014; Springer: Cham, Switzerland, 2014; pp. 255–265. [Google Scholar]
  9. Corelli, A. Cryptocurrencies and exchange rates: A relationship and causality analysis. Risk 2018, 6, 111. [Google Scholar] [CrossRef] [Green Version]
  10. Trabelsi, N. Are there any volatility spill-over effects among cryptocurrencies and widely traded asset classes? J. Risk Financ. Manag. 2018, 11, 66. [Google Scholar] [CrossRef] [Green Version]
  11. Billah, M.; Waheed, S.; Hanifa, A. Stock market prediction using an improved training algorithm of neural network. In Proceedings of the 2016 2nd International Conference on Electrical, Computer & Telecommunication Engineering (ICECTE), Rajshahi, Bangladesh, 8–10 December 2016; pp. 1–4. [Google Scholar]
  12. Nayak, S.C.; Misra, B.B.; Behera, H.S. An adaptive second order neural network with genetic-algorithm-based training (ASONN-GA) to forecast the closing prices of the stock market. Int. J. Appl. Metaheuristic Comput. (IJAMC) 2016, 7, 39–57. [Google Scholar] [CrossRef]
  13. Nayak, S.C.; Misra, B.B.; Behera, H.S. Efficient forecasting of financial time-series data with virtual adaptive neuro-fuzzy inference system. Int. J. Bus. Forecast. Mark. Intelligence 2016, 2, 379–402. [Google Scholar] [CrossRef]
  14. Rebane, J.; Karlsson, I.; Papapetrou, P.; Denic, S. Seq2Seq RNNs and ARIMA models for cryptocurrency prediction: A comparative study. In Proceedings of the SIGKDD Fintech’18, London, UK, 19–23 August 2018. [Google Scholar]
  15. White, H. Economic prediction using neural networks: The case of IBM daily stock returns. In Proceedings of the IEEE 1988 International Conference on Neural Networks, San Diego, CA, USA, 24–27 July 1988; Volume 2, pp. 451–458. [Google Scholar]
  16. Sureshkumar, K.K.; Elango, N.M. Performance analysis of stock price prediction using artificial neural network. Glob. J. Comput. Sci. Technol. 2012. [Google Scholar]
  17. Misnik, A.; Krutalevich, S.; Prakapenka, S.; Borovykh, P.; Vasiliev, M. Impact Analysis of Additional Input Parameters on Neural Network Cryptocurrency Price Prediction. In Proceedings of the 2019 XXI International Conference Complex Systems: Control and Modeling Problems (CSCMP), Samara, Russia, 3–6 September 2019; pp. 163–167. [Google Scholar]
  18. Jay, P.; Kalariya, V.; Parmar, P.; Tanwar, S.; Kumar, N.; Alazab, M. Stochastic neural networks for cryptocurrency price prediction. IEEE Access 2020, 8, 82804–82818. [Google Scholar] [CrossRef]
  19. Akyildirim, E.; Goncu, A.; Sensoy, A. Prediction of cryptocurrency returns using machine learning. Ann. Oper. Res. 2020, 297, 3–36. [Google Scholar] [CrossRef]
  20. Yazdinejad, A.; HaddadPajouh, H.; Dehghantanha, A.; Parizi, R.M.; Srivastava, G.; Chen, M.Y. Cryptocurrency malware hunting: A deep recurrent neural network approach. Appl. Soft Comput. 2020, 96, 106630. [Google Scholar] [CrossRef]
  21. Alessandretti, L.; ElBahrawy, A.; Aiello, L.M.; Baronchelli, A. Anticipating cryptocurrency prices using machine learning. Complexity 2018, 2018, 8983590. [Google Scholar] [CrossRef]
  22. Corbet, S.; Lucey, B.; Urquhart, A.; Yarovaya, L. Cryptocurrencies as a financial asset: A systematic analysis. Int. Rev. Financ. Anal. 2019, 62, 182–199. [Google Scholar] [CrossRef] [Green Version]
  23. McNally, S.; Roche, J.; Caton, S. Predicting the price of bitcoin using machine learning. In Proceedings of the 2018 26th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP), Cambridge, UK, 21–23 March 2018; pp. 339–343. [Google Scholar]
  24. Nakano, M.; Takahashi, A.; Takahashi, S. Bitcoin technical trading with artificial neural network. Phys. A Stat. Mech. Its Appl. 2018, 510, 587–609. [Google Scholar] [CrossRef]
  25. Rao, R. Rao algorithms: Three metaphor-less simple algorithms for solving optimization problems. Int. J. Ind. Eng. Comput. 2020, 11, 107–130. [Google Scholar] [CrossRef]
  26. Caesarendra, W.; Pratama, M.; Kosasih, B.; Tjahjowidodo, T.; Glowacz, A. Parsimonious network based on a fuzzy inference system (PANFIS) for time series feature prediction of low speed slew bearing prognosis. Appl. Sci. 2018, 8, 2656. [Google Scholar] [CrossRef] [Green Version]
  27. Helfer, G.A.; Barbosa, J.L.; Alves, D.; da Costa, A.B.; Beko, M.; Leithardt, V.R. Multispectral cameras and machine learning integrated into portable devices as clay prediction technology. J. Sens. Actuator Netw. 2021, 10, 40. [Google Scholar] [CrossRef]
  28. Alazeb, A.; Panda, B.; Almakdi, S.; Alshehri, M. Data Integrity Preservation Schemes in Smart Healthcare Systems That Use Fog Computing Distribution. Electronics 2021, 10, 1314. [Google Scholar] [CrossRef]
  29. Nayak, S.C.; Misra, B.B.; Behera, H.S. On developing and performance evaluation of adaptive second order neural network with ga-based training (asonn-ga) for financial time series prediction. In Advancements in Applied Metaheuristic Computing; IGI Global: Hershey, PA, USA, 2018; pp. 231–263. [Google Scholar]
  30. Nayak, S.C.; Misra, B.B.; Behera, H.S. Efficient financial time series prediction with evolutionary virtual data position exploration. Neural Comput. Appl. 2019, 31, 1053–1074. [Google Scholar] [CrossRef]
  31. Mirjalili, S.; Mirjalili, S.M.; Hatamlou, A. Multi-verse optimizer: A nature-inspired algorithm for global optimization. Neural Comput. Appl. 2016, 27, 495–513. [Google Scholar] [CrossRef]
  32. Nayak, S.C.; Das, S.; Misra, B.B. Development and Performance Analysis of Fireworks Algorithm-Trained Artificial Neural Network (FWANN): A Case Study on Financial Time Series Forecasting. In Handbook of Research on Fireworks Algorithms and Swarm Intelligence; IGI Global: Hershey, PA, USA, 2020; pp. 176–194. [Google Scholar]
  33. Tan, Y.; Zhu, Y. Fireworks Algorithm for Optimization. In Proceedings of the International Conference in Swarm Intelligence, Beijing, China, 12–15 June 2010; Springer: Berlin, Germany, 2010; pp. 355–364. [Google Scholar]
  34. Alatas, B. ACROA: Artificial chemical reaction optimization algorithm for global optimization. Expert Syst. Appl. 2011, 38, 13170–13180. [Google Scholar] [CrossRef]
  35. Nayak, S.C.; Misra, B.B. A chemical-reaction-optimization-based neuro-fuzzy hybrid network for stock closing price prediction. Financ. Innov. 2019, 5, 38. [Google Scholar] [CrossRef]
  36. Rao, R.V.; Savsani, V.J.; Vakharia, D.P. Teaching–learning-based optimization: An optimization method for continuous non-linear large scale problems. Inf. Sci. 2012, 183, 1–15. [Google Scholar] [CrossRef]
  37. Rao, R. Jaya: A simple and new optimization algorithm for solving constrained and unconstrained optimization problems. Int. J. Ind. Eng. Comput. 2016, 7, 19–34. [Google Scholar]
  38. Rao, R.V.; Pawar, R.B. Constrained design optimization of selected mechanical system components using Rao algorithms. Appl. Soft Comput. 2020, 89, 106141. [Google Scholar] [CrossRef]
  39. Rao, R.V.; Pawar, R.B. Self-adaptive multi-population Rao algorithms for engineering design optimization. Appl. Artif. Intell. 2020, 34, 187–250. [Google Scholar] [CrossRef]
  40. Premkumar, M.; Babu, T.S.; Umashankar, S.; Sowmya, R. A new metaphor-less algorithms for the photovoltaic cell parameter estimation. Optik 2020, 208, 164559. [Google Scholar] [CrossRef]
  41. Wang, L.; Wang, Z.; Liang, H.; Huang, C. Parameter estimation of photovoltaic cell model with Rao-1 algorithm. Optik 2020, 210, 163846. [Google Scholar] [CrossRef]
  42. Jabir, H.A.; Kamel, S.; Selim, A.; Jurado, F. Optimal Coordination of Overcurrent Relays Using Metaphor-less Simple Method. In Proceedings of the 2019 21st International Middle East Power Systems Conference (MEPCON), Cairo, Egypt, 17–19 December 2019; pp. 1063–1067. [Google Scholar]
  43. Haykin, S. Neural Networks and Learning Machines 3/E; Pearson Educ: Delhi, India, 2010. [Google Scholar]
  44. Nayak, S.C.; Misra, B.B.; Behera, H.S. Impact of data normalization on stock index forecasting. Int. J. Comput. Inf. Syst. Ind. Manag. Appl. 2014, 6, 257–269. [Google Scholar]
  45. Nayak, S.C.; Misra, B.B. Extreme learning with chemical reaction optimization for stock volatility prediction. Financ. Innov. 2020, 6, 16. [Google Scholar] [CrossRef]
Figure 1. Basic Architecture of ANN. (Source: Authors).
Figure 1. Basic Architecture of ANN. (Source: Authors).
Fintech 01 00004 g001
Figure 2. RA + ANN-based forecasting.
Figure 2. RA + ANN-based forecasting.
Fintech 01 00004 g002
Figure 3. Bitcoin, Litecoin, Ethereum, Ripple, CMC 200, and Tether FTSs.
Figure 3. Bitcoin, Litecoin, Ethereum, Ripple, CMC 200, and Tether FTSs.
Fintech 01 00004 g003
Figure 4. Rolling window method for input pattern generation.
Figure 4. Rolling window method for input pattern generation.
Fintech 01 00004 g004
Figure 5. Forecasted prices v/s actual from Bitcoin.
Figure 5. Forecasted prices v/s actual from Bitcoin.
Fintech 01 00004 g005
Figure 6. Forecasted prices v/s actual from Litecoin.
Figure 6. Forecasted prices v/s actual from Litecoin.
Fintech 01 00004 g006
Figure 7. Forecasted prices v/s actual from Ethereum.
Figure 7. Forecasted prices v/s actual from Ethereum.
Fintech 01 00004 g007
Figure 8. Forecasted prices v/s actual from Ripple.
Figure 8. Forecasted prices v/s actual from Ripple.
Fintech 01 00004 g008
Figure 9. Forecasted prices v/s actual from CMC 200.
Figure 9. Forecasted prices v/s actual from CMC 200.
Fintech 01 00004 g009
Figure 10. Forecasted prices v/s actual from Tether.
Figure 10. Forecasted prices v/s actual from Tether.
Fintech 01 00004 g010
Table 1. Statistical summary of four cryptocurrencies.
Table 1. Statistical summary of four cryptocurrencies.
StatisticBitcoinLitecoinEthereumRippleCMC 200Tether
Minimum68.43001.16000.43480.00280.97420.9742
Mean1.4826 × 10320.4966147.78430.09841.00291.0029
Median482.81003.910012.02000.00791.00171.0017
Variance8.7535 × 1062.2240 × 1036.9765 × 1040.10282.5621 × 10−52.5621 × 10−5
Maximum1.9497 × 104358.34001.3964 × 1033.38001.05361.0536
Standard deviation2.9593 × 10347.1594264.13080.32060.00510.0051
Skewness3.53944.16272.38445.80391.94171.9417
Kurtosis15.967121.59258.512242.946919.379719.3797
Correlation coefficient0.001300.00211−0.00520.00270.00390.0035
Table 2. MAPE and ARV from all models.
Table 2. MAPE and ARV from all models.
MODELBITCOINRIPPLEEHTEREUMLITECOINCMC200TETHER
MAPEARVMAPEARVMAPEARVMAPEARVMAPEARVMAPEARV
RA + ANN0.03000.00570.03970.00510.03850.00390.03220.00540.02750.00270.00650.0055
GA + ANN0.03220.00750.04570.01430.04950.00570.04540.00580.02990.00530.00870.0173
PSO + ANN0.04540.00610.04730.00530.04390.00630.03940.00790.02870.01640.00820.0094
SVM0.03940.00650.04760.00680.04630.00590.08370.00600.04280.03420.00930.0107
MLP0.04720.00770.04970.01740.05930.04670.17260.02460.04790.02770.03290.0637
ARIMA0.06060.00830.08280.01630.07580.01930.09400.00950.13270.03360.08520.2953
LSE0.09460.10050.08760.13730.25570.07360.21630.02660.52830.27340.10860.4066
Table 3. Results from statistical significance test.
Table 3. Results from statistical significance test.
Proposed ForecastComparative Forecastp and h Values from Wilcoxon Signed Test
NASDAQBSEDJIAHSINIKKEI
RA + ANNGA + ANN2.5034 × 10−5
(h = 1)
3.1432 × 10−4
(h = 1)
2.2441 × 10−3
(h = 1)
3.0134 × 10−5
(h = 1)
2.2763 × 10−4
(h = 1)
PSO + ANN3.1362 × 10−3
(h = 1)
4.2634 × 10−5
(h = 1)
4.3004 × 10−3
(h = 1)
6.2515 × 10−3
(h = 1)
1.7253 × 10−2
(h = 1)
SVM4.2418 × 10−4
(h = 1)
2.1505 × 10−3
(h = 1)
3.3410 × 10−2
(h = 1)
2.990152
(h = 0)
1.6728 × 10−1
(h = 1)
MLP1.5728 × 10−3
(h = 1)
0.008524
(h = 0)
6.2803 × 10−2
(h = 1)
7.2095 × 10−2
(h = 1)
4.2021 × 10−3
(h = 1)
ARIMA4.3007 × 10−4
(h = 1)
3.7726 × 10−2
(h = 1)
1.7362 × 10−5
(h = 1)
3.3282 × 10−3
(h = 1)
1.7062 × 10−4
(h = 1)
LSE1.5338 × 10−3
(h = 1)
0.01821
(h = 0)
6.2033 × 10−2
(h = 1)
7.0025 × 10−2
(h = 1)
3.2521 × 10−3
(h = 1)
p and h values from Diebold–Mariano test
GA + ANN2.09755
(h = 1)
2.34220
(h = 1)
2.0007
(h = 1)
1.9820
(h = 1)
2.2037
(h = 1)
PSO + ANN2.22043
(h = 1)
2.36020
(h = 1)
1.9873
(h = 1)
1.9815
(h = 1)
−1.9900
(h = 1)
SVM2.36105
(h = 1)
1.99247
(h = 1)
1.97695
(h = 1)
1.98577
(h = 1)
1.9792
(h = 1)
MLP2.7329
(h = 1)
1.9800
(h = 1)
−3.05263
(h = 1)
−1.99265
(h = 1)
2.52655
(h = 1)
ARIMA1.9909
(h = 1)
2.61227
(h = 1)
−3.40582
(h = 1)
1.96035
(h = 0)
2.45884
(h = 1)
LSE2.0728 × 10−3
(h = 1)
2.88524
(h = 1)
3.2800 × 10−2
(h = 1)
5.1095 × 10−2
(h = 1)
4.32171 × 10−3
(h = 1)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nayak, S.K.; Nayak, S.C.; Das, S. Modeling and Forecasting Cryptocurrency Closing Prices with Rao Algorithm-Based Artificial Neural Networks: A Machine Learning Approach. FinTech 2022, 1, 47-62. https://doi.org/10.3390/fintech1010004

AMA Style

Nayak SK, Nayak SC, Das S. Modeling and Forecasting Cryptocurrency Closing Prices with Rao Algorithm-Based Artificial Neural Networks: A Machine Learning Approach. FinTech. 2022; 1(1):47-62. https://doi.org/10.3390/fintech1010004

Chicago/Turabian Style

Nayak, Sanjib Kumar, Sarat Chandra Nayak, and Subhranginee Das. 2022. "Modeling and Forecasting Cryptocurrency Closing Prices with Rao Algorithm-Based Artificial Neural Networks: A Machine Learning Approach" FinTech 1, no. 1: 47-62. https://doi.org/10.3390/fintech1010004

Article Metrics

Back to TopTop