Next Article in Journal
A Novel Dynamic S-Box Generation Scheme Based on Quantum Random Walks Controlled by a Hyper-Chaotic Map
Next Article in Special Issue
Identification of Parameters of Conservative Multispecies Lotka-Volterra System Based on Sampled Data
Previous Article in Journal
New One-Parameter Over-Dispersed Discrete Distribution and Its Application to the Nonnegative Integer-Valued Autoregressive Model of Order One
Previous Article in Special Issue
Parameter Identification of the Discrete-Time Stochastic Systems with Multiplicative and Additive Noises Using the UD-Based State Sensitivity Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Approach to Modeling Incommensurate Fractional Order Systems Using Fractional Neural Networks

1
FraCAL Lab., The University of the South Pacific, Laucala Campus, Suva 1168, Fiji
2
Lab. LTI, University of Picardie Jules Verne, 80000 Amiens, France
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(1), 83; https://doi.org/10.3390/math12010083
Submission received: 28 November 2023 / Revised: 21 December 2023 / Accepted: 25 December 2023 / Published: 26 December 2023
(This article belongs to the Special Issue New Trends on Identification of Dynamic Systems)

Abstract

:
This research explores the application of the Riemann–Liouville fractional sigmoid, briefly R L F σ , activation function in modeling the chaotic dynamics of Chua’s circuit through Multilayer Perceptron (MLP) architecture. Grounded in the context of chaotic systems, the study aims to address the limitations of conventional activation functions in capturing complex relationships within datasets. Employing a structured approach, the methods involve training MLP models with various activation functions, including R L F σ , sigmoid, swish, and proportional Caputo derivative P C σ , and subjecting them to rigorous comparative analyses. The main findings reveal that the proposed R L F σ consistently outperforms traditional counterparts, exhibiting superior accuracy, reduced Mean Squared Error, and faster convergence. Notably, the study extends its investigation to scenarios with reduced dataset sizes and network parameter reductions, demonstrating the robustness and adaptability of R L F σ . The results, supported by convergence curves and CPU training times, underscore the efficiency and practical applicability of the proposed activation function. This research contributes a new perspective on enhancing neural network architectures for system modeling, showcasing the potential of R L F σ in real-world applications.

1. Introduction

Fractional calculus (FC) serves as an extension of classical calculus, expanding the concepts of integral and derivative. The increased interest in fractional-order integrals and derivatives is due to their unique non-locality and memory characteristics [1,2]. Recent research in practical systems modeling highlights a growing fascination among scholars with fractional-order calculus [3,4,5]. The use of fractional-order operators has significantly improved the precision of mathematical models in representing the behavior of various real dynamical systems. Examples include applications in Lithium-Ion battery State of Charge (SoC) [6], a flexible model for supercapacitors [7,8], the transient behavior of photovoltaic solar panels [9], understanding the mechanism of atrial fibrillation [10], thermal dynamics of buildings [11], analysis of viscoelasticity in asphalt mixtures [12], behavior of magneto-active elastomers in magnetic fields [13], and modeling macroeconomic systems [14].
Moving on, the domain of system identification has been a focal point of research for numerous decades. Historically, scholars relied on conventional methodologies such as linear systems theory and signal processing to model and scrutinize intricate systems [15,16,17], as well as advanced methods including the predictor-corrector compact difference scheme [18] and others [19,20]. However, the surge in data accessibility and computational capabilities has led to a rising inclination toward nonlinear techniques. Including additive noise in modeling is also broadening the scope of the problem [21]. The adoption of Artificial Neural Networks (ANNs) for system identification has gained prominence due to their adeptness in modeling intricate nonlinear systems. ANNs can approximate any continuous function with remarkable precision, rendering them an ideal instrument for system identification [22,23,24,25]. Nevertheless, the training of ANNs can be computationally demanding and necessitates substantial data volumes. Recently, Fractional Neural Networks (FNNs) have emerged as a promising realm for modeling complex systems that manifest non-local behaviors [26,27,28]. FC offers a potent tool for modeling non-local systems, a prevalent occurrence in various engineering and scientific applications. The amalgamation of ANNs and FC for system identification has garnered increasing interest. FNNs have been suggested as a means to synergize the strengths of both ANNs and FC in modeling complex systems characterized by non-local behaviors [29,30,31,32,33].
Additionally, fractional-order models can be categorized into two groups: commensurate and incommensurate models [34,35]. Each state variable may possess a distinct order in incommensurate state-space models, while in commensurate models, the derivative orders for all state variables are identical [36]. An incommensurate fractional-order system (IFOS) is a dynamic system or process exhibiting fractional-order behavior, which is not a rational number [37]. Some real-world systems and processes cannot be accurately described using integer-order models; instead, they exhibit fractional-order dynamics. Fractional-order systems (FOSs) are described by fractional differential equations (FDEs), where the differentiation order is a number, typically between 0 and 1 [38]. IFOSs with order values that are irrational can lead to particularly intricate and challenging behaviors in the system’s response, as the non-integer order introduces memory effects and long-range dependencies that are not present in integer-order systems [39,40]. The study of these systems is still an active and evolving area of research, and their applications range from modeling physical phenomena to improving control strategies for complex processes.
Finally, this paper proposes extending the applications of the developed fractional sigmoid activation function for behavior identification and prediction. For numerical validation, the incommensurate fractional-order Chua’s system is used, and a comparative analysis is presented. Additionally, this paper strives to maximize performance in system modeling, achieve faster convergence, and reduce CPU training time when using neural networks. Furthermore, Chua’s circuit is known for its complex and chaotic behavior. Its dynamics exhibit nonlinearity and sensitivity to initial conditions, making it a challenging and interesting system to study. The chaotic nature of the system provides a rigorous test for the predictive capabilities of the proposed fractional neural network. Chaotic systems are known for their sensitivity to perturbations, and successfully predicting their behavior can showcase the robustness of the proposed model. From the above comprehensive discussion, this study’s vital contributions are,
  • To incorporate the newly developed fractional activation function into system modeling, more specifically for estimating the output of Chua’s model using simulation data.
  • To compare the state-of-the-art method with available classical activation functions and existing fractional activation functions.
  • To evaluate and verify the new FNN-based black-box modeling of Chua’s circuit based on data reduction and trainable parameter reduction.

2. Fractional Theory and Proposed Activation Function

The study of FC involves a variety of techniques and methods, such as the Grunwald–Letnikov fractional difference operator, the Riemann–Liouville fractional integral and derivative, the Caputo fractional derivative [41], and the newly developed definition known as the conformable fractional derivative (CFD) [42]. In this paper, the CFD defined below is used.
Given a function f : R R , the improved Riemann–Liouville-type conformable fractional derivative ( R L CFD) [43] of f of order α , is defined by:
a R L T α ( t ) = lim ε 0 ( 1 α ) f ( t ) + α f ( t + ε ( t a ) 1 α f ( t ) ε
Consider first the classical sigmoid function as
σ ( t ) = 1 1 + e t ,
Using (1) and taking f ( t ) = e t , one can obtain the function
0 R L T α ( t ) = ( 1 α ) ( e t ) α t 1 α e t ,
Substituting (3) with the exponential function in (2), we obtain
R L F σ ( t ) = 1 1 + 0 R L T α ( t ) .
Table 1 portrays the equation of the proposed fractional sigmoid activation function and its derivative. Figure 1a shows the characteristics of the R L F σ ( t ) with different values of α . For smaller values of α , the slope gradually becomes steeper, and for α = 1 , the traditional sigmoid activation function is achieved. Likewise, the derivative of R L F σ ( t ) in Figure 1b shows that the value of α determines the amplitude of the curve.
Table 2 provides a concise overview of different sigmoid-based activation functions commonly used in NNs, giving insight into their key characteristics. These characteristics include the presence of diminishing gradients, the ability to promote better generalization, computational efficiency, and the speed of convergence during training. The table clearly compares the properties of both standard and fractional activation functions, highlighting the advantages and limitations of each.

3. Problem Statement

Fractional-Order Chua’s Circuit

Chua’s circuit, also known as Chua’s oscillator, is a simple electronic circuit that exhibits nonlinear dynamical phenomena, such as bifurcation and chaos [47], as illustrated in Figure 2. Chua’s circuit comprises various components, including resistors, capacitors, and operational amplifiers, as shown in Figure 3. Its crucial element is Chua’s diode, a two-terminal nonlinear device with a piecewise-linear current-voltage characteristic. This nonlinearity is what allows the circuit to exhibit chaotic oscillations. Chaotic systems, like this one, display unpredictable and highly sensitive behaviors to initial conditions, making them useful in various applications, such as secure communications and random number generation [48,49].
FDEs provide a mathematical framework to describe systems with memory, complex dynamics, and power-law behaviors that traditional integer-order differential equations cannot adequately capture [50]. The incommensurate fractional-order Chua’s system with disturbance [51] is defined as:
D 0.98 x 1 = a 1 ( x 2 x 1 m 1 x 1 f ( x 1 ) ) + D 1 ( x , t ) , D 0.98 x 2 = x 1 x 2 + x 3 + D 2 ( x , t ) , D 0.94 x 3 = q 2 x 2 a 3 x 3 + u + D 3 ( x , t ) , y = x 1
where D 1 ( x , t ) = 0.005 sin ( 1.5 t ) cos ( 0.6 x 1 x 2 ) , D 2 ( x , t ) = 0.04 cos ( 1.5 t ) cos ( 0.6 x 2 x 3 ) , D 3 ( x , t ) = 0.04 sin ( 1.5 t ) sin ( 1.5 x 1 2 x 3 ) are the external disturbances, f ( x 1 ) = 1 2 ( m 0 m 1 ) x 1 + 1 x 1 1 is the nonlinear function, a 1 = 10.725 , a 2 = 10.593 , a 3 = 0.268 are the system parameters, m 0 = 0.7872 , m 1 = 1.1726 and the initial value is x ( 0 ) = ( 1.2 , 0.1 , 0.9 ) T .

4. Dataset Description and Neural Model Design

The Chua’s circuit was modeled and simulated for 60 s using MATLAB. The simulated data consists of 12,000 samples taken at a sampling rate of 0.005 s. Figure 4 presents a visual representation of the input and output data of the Chua’s circuit.
The MLP architecture is utilized to test and validate the results using proposed functions compared to utilizing all activation functions presented in Table 2 to estimate the chaotic output of Chua’s circuit model. The architecture consists of an input layer with one node, two hidden layers, and an output layer with one output node. Figure 5 displays the general architecture used to map the input and output data of Chua’s circuit.
First, for each comparison, only the hidden layer activation functions are changed, while the output layer activation functions (linear) remain constant. The proposed function-based MLP models are trained for 25 epochs, whereas other MLP models are trained for 50 epochs. Also, based on initial results, in relation to the convergence curves, even though the conventional model reaches maximum accuracy within 25 epochs, it is not flexible enough to interpret the complex relationship between the dataset and suffers heavily when predicting the outputs. Therefore, other models are trained for 50 epochs for maximum training and prediction accuracy. Table 3 presents the fixed parameters of the MLP model used for training and validation.

5. Numerical Investigations

This section presents a comparative analysis of the MLP model utilizing the proposed R L F σ , sigmoid [44], swish [45], and P C σ [46] functions. The evaluation is carried out by estimating the chaotic dynamics of the Chua’s circuit using 80% of the data for training and 20% for testing. The analysis focuses on two key metrics: Mean Squared Error (MSE) and the Coefficient of Determination ( R 2 ). Three sets of analyses are performed to compare the performance of all models comprehensively. Firstly, the model is trained with all dataset samples and the complete MLP model. Secondly, the impact of dataset size is examined. Lastly, model parameter reduction is considered. The effectiveness of R L F σ is also evaluated in terms of its influence on the convergence rate, adaptability, and the CPU training time required to achieve maximum accuracy. The experiments were conducted on an Anaconda platform, utilizing the Jupyter Notebook environment, on a computer system featuring an 8th generation Intel i7 processor with a clock speed of 4.0 GHz and 16 GB of RAM.

5.1. Case 1: Full Model Analysis

All MLP models are trained and tested with full 12,000 sample data and complete network architecture with 609 trainable parameters.
Table 4 presents the numerical comparison of the proposed sigmoid MLP model with other MLP models based on R 2 and MSE values for case 1. It can be observed that the proposed model outperforms all the other models in both the training and testing phases, achieving higher accuracy values. When comparing the error values, the proposed function reduces the error by 65% compared to the classical model, 67% compared to the enhanced classical model, and 85% compared to the fractional model. This significant reduction in error percentage contributes significantly to the prediction accuracy.
Figure 6 illustrates the training convergence curve of all the models. It can be observed that the proposed model has faster convergence as it achieves higher accuracy and a low loss value right from epoch 1. In comparison, the proposed model achieves 40% more accuracy than all other models at epoch 1 and converges quickly to maximum accuracy within 25 epochs. Moving on, Figure 7 below shows the estimated output of Chua’s circuit by all models and compares it with the real output. It can be observed that, due to the superior performance of the proposed model, it is able to estimate the output more accurately. The key difference that can be observed is that all other models suffer from accurately detecting the chaotic and abrupt peak changes in the circuit dynamics.
Table 5 presents the CPU training time to achieve maximum accuracy for all the models in case 1. Due to the fast convergence properties of the proposed function, the CPU training time is reduced drastically. It can be observed that the proposed model trains approximately 50% faster than the classical model, 54% faster than the enhanced classical model, and 60% faster than the fractional model.

5.2. Case 2: Dataset Size Impact

For this analysis, the model is trained and tested with a 50% reduction in data samples, and the trainable network parameters remained the same as in case 1.
Table 6 presents the numerical analysis of case 2. In regards to training and testing performance, it can be observed that the proposed model again outperforms all the other models even with a 50% reduced dataset size. In comparison to case 1, the proposed model accuracy only drops by 1% on average and other models accuracy drops by 3% on average. When comparing the error value, the proposed function reduces error by 60% when compared to the classical model, 63% when compared to the enhanced classical model, and 84% when compared to the fractional model.
Figure 8 illustrates the training convergence curve of case 2. Again, the proposed model displays faster convergence as it achieves 83% more accuracy on average compared to all other models at epoch 1. Figure 9 below shows the estimated output based on a reduced training dataset size. It can be observed that even with limited information about the system, the proposed model is able to estimate the output more accurately, whereas the other models suffer greatly when compared to the predictions of case 1.
Table 7 presents the CPU training time to achieve maximum accuracy for case 2. It can be observed that the proposed model trains approximately 45% faster than the classical model, 53% faster than the enhanced classical model, and 60% faster than the fractional model.

5.3. Case 3: Network Parameter Reduction

For this analysis, instead, the network trainable parameters are reduced by 80%, which brings down the model with only 111 trainable parameters, and the model is trained and tested using full 12,000 data samples.
Furthermore, Table 8 presents the numerical analysis of case 3. Similar results are obtained, as seen in the previous two cases, whereby the proposed model outperforms all the other models in both the training and testing phases. When looking at the accuracies obtained in case 1, the proposed model still achieves similar accuracies, whereas the other models suffer greatly with an average of a 5% reduction in accuracies. When comparing the error values, the proposed function reduces error by 84% compared to the classical model, 83% compared to the enhanced classical model, and 89% compared to the fractional model.
Figure 10 illustrates the training convergence curve of case 3. Again, it can be observed that the proposed model has faster convergence. In comparison, the proposed model achieves 95% more accuracy compared to all other models at epoch 1, even with reduced trainable parameters. Figure 11 similarly shows the estimated output. Once again, even with a reduced network architecture, the proposed model is able to efficiently and accurately predict the dynamics of the circuit.
Table 9 presents the CPU training time to achieve maximum accuracy for case 3. It can be observed that the proposed model trains approximately 46% faster than the classical models, 55% faster than the enhanced classical models, and 66% faster than the fractional models.

6. Discussion

The study encompasses three distinct cases, each shedding light on different aspects of the proposed activation function’s performance. In Case 1, where the full model analysis is conducted with the complete dataset, R L F σ emerges as the standout performer. It consistently outperforms traditional activation functions such as sigmoid, swish, and P C σ , achieving higher R 2 values and significantly reducing MSE. The faster convergence of the proposed model from the early epochs, coupled with notably lower CPU training time, underscores its efficiency in capturing the intricate dataset relationships.
Moving on to Case 2, which explores the impact of dataset size reduction, the proposed activation function continues to showcase resilience. Even with a 50% reduction in data samples, R L F σ maintains superiority in terms of R 2 values and MSE compared to other models. The convergence curve further illustrates its robust learning capability with limited data, suggesting adaptability to scenarios where obtaining comprehensive datasets may be challenging.
In Case 3, where network parameter reduction is implemented by 80%, R L F σ once again stands out, maintaining high accuracy and outperforming other models in terms of R 2 values and MSE. The model’s ability to generalize and accurately capture system dynamics, even with a simplified architecture, highlights the efficacy of the proposed activation function.
In conclusion, the comprehensive numerical analyses consistently support the superiority of the proposed function across various scenarios. Its efficiency in modeling Chua’s circuit dynamics, adaptability to reduced datasets, and resilience to network parameter reductions make it a promising candidate for real-world applications such as weather pattern prediction, fluid dynamics, and biological systems modeling. The study provides valuable insights into the specific application of the proposed activation function and prompts considerations for broader applications and avenues for future research. An essential direction for continued research involves the H1-norm error analysis of a robust ADI method applied to graded meshes for three-dimensional sub-diffusion problems. This exploration aims to enhance our understanding of the method’s accuracy and reliability in addressing complex scenarios. Additionally, investigating an efficient ADI difference scheme for the nonlocal evolution problem in three-dimensional space represents a crucial avenue for advancement. This scheme can contribute to developing effective numerical approaches for handling nonlocal phenomena, expanding the applicability of such methods in various scientific domains.

Author Contributions

Conceptualization, U.M.; methodology, M.K., U.M. and G.C.; software, M.K.; validation, M.K.; formal analysis, M.K.; investigation, M.K.; resources, U.M. and M.K.; data curation, M.K.; writing—original draft preparation, M.K.; writing—review and editing, U.M. and M.K.; visualization, M.K. and U.M.; supervision, U.M. and G.C.; project administration, U.M.; funding acquisition, U.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

No data except that in the paper.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANNArtificial Neural Network
CFDConformable Fractional Derivative
FCFractional Calculus
FDEFractional Differential Equation
FNNFractional Neural Network
FOSFractional Order System
IFOSIncommensurate Fractional-order System
MSEMean Squared Error
SoCState of Charge

References

  1. Li, X.; He, J.; Wen, C.; Liu, X.K. Backstepping-based adaptive control of a class of uncertain incommensurate fractional-order nonlinear systems with external disturbance. IEEE Trans. Ind. Electron. 2021, 69, 4087–4095. [Google Scholar] [CrossRef]
  2. Pishro, A.; Shahrokhi, M.; Sadeghi, H. Fault-tolerant adaptive fractional controller design for incommensurate fractional-order nonlinear dynamic systems subject to input and output restrictions. Chaos Solitons Fractals 2022, 157, 111930. [Google Scholar] [CrossRef]
  3. Cattani, C.; Srivastava, H.M.; Yang, X.J. Fractional Dynamics; Walter de Gruyter GmbH & Co KG: Berlin, Germany, 2015. [Google Scholar]
  4. Tavazoei, M.; Asemani, M.H. Robust stability analysis of incommensurate fractional-order systems with time-varying interval uncertainties. J. Frankl. Inst. 2020, 357, 13800–13815. [Google Scholar] [CrossRef]
  5. Kothari, K.; Mehta, U.; Prasad, R. Fractional-Order System Modeling and its Applications. J. Eng. Sci. Technol. Rev. 2019, 12, 1–10. [Google Scholar] [CrossRef]
  6. Sabatier, J.; Guillemard, F.; Lavigne, L.; Noury, A.; Merveillaut, M.; Francico, J.M. Fractional models of lithium-ion batteries with application to state of charge and ageing estimation. In Proceedings of the Informatics in Control, Automation and Robotics: 13th International Conference, ICINCO 2016, Lisbon, Portugal, 29–31 July 2016; Springer: Berlin/Heidelberg, Germany, 2018; pp. 55–72. [Google Scholar]
  7. Prasad, R.; Kothari, K.; Mehta, U. Flexible Fractional Supercapacitor Model Analyzed in Time Domain. IEEE Access 2019, 7, 122626–122633. [Google Scholar] [CrossRef]
  8. Prasad, R.; Mehta, U.; Kothari, K. Various analytical models for supercapacitors: A mathematical study. Resour.-Effic. Technol. 2020, 1, 1–15. [Google Scholar]
  9. AbdelAty, A.M.; Radwan, A.G.; Elwakil, A.S.; Psychalinos, C. Transient and steady-state response of a fractional-order dynamic PV model under different loads. J. Circuits Syst. Comput. 2018, 27, 1850023. [Google Scholar] [CrossRef]
  10. Ugarte, J.P.; Tobon, C.; Lopes, A.M.; Machado, J.T. Atrial rotor dynamics under complex fractional order diffusion. Front. Physiol. 2018, 9, 975. [Google Scholar] [CrossRef]
  11. Chen, L.; Basu, B.; McCabe, D. Fractional order models for system identification of thermal dynamics of buildings. Energy Build. 2016, 133, 381–388. [Google Scholar] [CrossRef]
  12. Lagos-Varas, M.; Movilla-Quesada, D.; Arenas, J.; Raposeiras, A.; Castro-Fresno, D.; Calzada-Pérez, M.; Vega-Zamanillo, A.; Maturana, J. Study of the mechanical behavior of asphalt mixtures using fractional rheology to model their viscoelasticity. Constr. Build. Mater. 2019, 200, 124–134. [Google Scholar] [CrossRef]
  13. Nadzharyan, T.; Kostrov, S.; Stepanov, G.; Kramarenko, E.Y. Fractional rheological models of dynamic mechanical behavior of magnetoactive elastomers in magnetic fields. Polymer 2018, 142, 316–329. [Google Scholar] [CrossRef]
  14. Tarasov, V.E.; Tarasova, V.V. Macroeconomic models with long dynamic memory: Fractional calculus approach. Appl. Math. Comput. 2018, 338, 466–486. [Google Scholar] [CrossRef]
  15. Mehta, U.; Bingi, K.; Saxena, S. Applied Fractional Calculus in Identification and Control; Springer: Singapore, 2022. [Google Scholar] [CrossRef]
  16. Pappalardo, C.M.; Guida, D. System identification algorithm for computing the modal parameters of linear mechanical systems. Machines 2018, 6, 12–32. [Google Scholar] [CrossRef]
  17. Prasad, V.; Mehta, U. Modeling and parametric identification of Hammerstein systems with time delay and asymmetric dead-zones using fractional differential equations. Mech. Syst. Signal Process. 2022, 167, 108568. [Google Scholar] [CrossRef]
  18. Jiang, X.; Wang, J.; Wang, W.; Zhang, H. A predictor–corrector compact difference scheme for a nonlinear fractional differential equation. Fractal Fract. 2023, 7, 521. [Google Scholar] [CrossRef]
  19. Wang, W.; Zhang, H.; Jiang, X.; Yang, X. A high-order and efficient numerical technique for the nonlocal neutron diffusion equation representing neutron transport in a nuclear reactor. Ann. Nucl. Energy 2024, 195, 110163. [Google Scholar] [CrossRef]
  20. Zhang, H.; Yang, X.; Tang, Q.; Xu, D. A robust error analysis of the OSC method for a multi-term fourth-order sub-diffusion equation. Comput. Math. Appl. 2022, 109, 180–190. [Google Scholar] [CrossRef]
  21. Ivanov, D.; Yakoub, Z. Overview of Identification Methods of Autoregressive Model in Presence of Additive Noise. Mathematics 2023, 11, 607. [Google Scholar] [CrossRef]
  22. Maroli, J.M. Generating discrete dynamical system equations from input–output data using neural network identification models. Reliab. Eng. Syst. Saf. 2023, 235, 109198. [Google Scholar] [CrossRef]
  23. Forgione, M.; Muni, A.; Piga, D.; Gallieri, M. On the adaptation of recurrent neural networks for system identification. Automatica 2023, 155, 111092. [Google Scholar] [CrossRef]
  24. Yamada, K.; Maruta, I.; Fujimoto, K. Subspace State-Space Identification of Nonlinear Dynamical System Using Deep Neural Network with a Bottleneck. IFAC-PapersOnLine 2023, 56, 102–107. [Google Scholar] [CrossRef]
  25. Kumar, P.; Rawlings, J.B.; Wenzel, M.J.; Risbeck, M.J. Grey-box model and neural network disturbance predictor identification for economic MPC in building energy systems. Energy Build. 2023, 286, 112936. [Google Scholar] [CrossRef]
  26. Maiti, M.; Sunder, M.; Abishek, R.; Bingi, K.; Shaik, N.B.; Benjapolakul, W. Recent advances and applications of fractional-order neural networks. Eng. J. 2022, 26, 49–67. [Google Scholar] [CrossRef]
  27. Cao, J.; Udhayakumar, K.; Rakkiyappan, R.; Li, X.; Lu, J. A comprehensive review of continuous-/discontinuous-time fractional-order multidimensional neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 5476–5496. [Google Scholar] [CrossRef] [PubMed]
  28. Viera-Martin, E.; Gómez-Aguilar, J.; Solís-Pérez, J.; Hernández-Pérez, J.; Escobar-Jiménez, R. Artificial neural networks: A practical review of applications involving fractional calculus. Eur. Phys. J. Spec. Top. 2022, 231, 2059–2095. [Google Scholar] [CrossRef] [PubMed]
  29. Aguilar, C.Z.; Gómez-Aguilar, J.; Alvarado-Martínez, V.; Romero-Ugalde, H. Fractional order neural networks for system identification. Chaos Solitons Fractals 2020, 130, 109444. [Google Scholar] [CrossRef]
  30. Kaslik, E.; Sivasundaram, S. Dynamics of fractional-order neural networks. In Proceedings of the 2011 International Joint Conference on Neural Networks, San Jose, CA, USA, 31 July–5 August 2011; pp. 611–618. [Google Scholar]
  31. Chen, M.R.; Chen, B.P.; Zeng, G.Q.; Lu, K.D.; Chu, P. An adaptive fractional-order BP neural network based on extremal optimization for handwritten digits recognition. Neurocomputing 2020, 391, 260–272. [Google Scholar] [CrossRef]
  32. Sheng, D.; Wei, Y.; Chen, Y.; Wang, Y. Convolutional neural networks with fractional order gradient method. Neurocomputing 2020, 408, 42–50. [Google Scholar] [CrossRef]
  33. Rahimkhani, P. Numerical solution of nonlinear stochastic differential equations with fractional Brownian motion using fractional-order Genocchi deep neural networks. Commun. Nonlinear Sci. Numer. Simul. 2023, 126, 107466. [Google Scholar] [CrossRef]
  34. Tavazoei, M.S.; Haeri, M. Chaotic attractors in incommensurate fractional order systems. Phys. D Nonlinear Phenom. 2008, 237, 2628–2637. [Google Scholar] [CrossRef]
  35. Tavazoei, M.; Asemani, M.H. Stability analysis of time-delay incommensurate fractional-order systems. Commun. Nonlinear Sci. Numer. Simul. 2022, 109, 106270. [Google Scholar] [CrossRef]
  36. Monje, C.A.; Chen, Y.; Vinagre, B.M.; Xue, D.; Feliu-Batlle, V. Fractional-Order Systems and Controls: Fundamentals and Applications; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  37. Gong, P.; Han, Q.L.; Lan, W. Finite-time consensus tracking for incommensurate fractional-order nonlinear multiagent systems with directed switching topologies. IEEE Trans. Cybern. 2020, 52, 65–76. [Google Scholar] [CrossRef] [PubMed]
  38. Wang, B.; Ding, J.; Wu, F.; Zhu, D. Robust finite-time control of fractional-order nonlinear systems via frequency distributed model. Nonlinear Dyn. 2016, 85, 2133–2142. [Google Scholar] [CrossRef]
  39. Wang, C.; Liang, M.; Chai, Y. Adaptive control of a class of incommensurate fractional order nonlinear systems with input dead-zone. IEEE Access 2019, 7, 153710–153723. [Google Scholar] [CrossRef]
  40. Shahvali, M.; Naghibi-Sistani, M.B.; Modares, H. Distributed consensus control for a network of incommensurate fractional-order systems. IEEE Control Syst. Lett. 2019, 3, 481–486. [Google Scholar] [CrossRef]
  41. De Oliveira, E.C.; Tenreiro Machado, J.A. A review of definitions for fractional derivatives and integral. Math. Probl. Eng. 2014, 2014, 238459. [Google Scholar] [CrossRef]
  42. Khalil, R.; Al Horani, M.; Yousef, A.; Sababheh, M. A new definition of fractional derivative. J. Comput. Appl. Math. 2014, 264, 65–70. [Google Scholar] [CrossRef]
  43. Gao, F.; Chi, C. Improvement on conformable fractional derivative and its applications in fractional differential equations. J. Funct. Spaces 2020, 2020, 5852414. [Google Scholar] [CrossRef]
  44. Dubey, S.R.; Singh, S.K.; Chaudhuri, B.B. Activation functions in deep learning: A comprehensive survey and benchmark. Neurocomputing 2022, 503, 92–108. [Google Scholar] [CrossRef]
  45. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for activation functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  46. Altan, G.; Alkan, S.; Baleanu, D. A novel fractional operator application for neural networks using proportional Caputo derivative. Neural Comput. Appl. 2023, 35, 3101–3114. [Google Scholar] [CrossRef]
  47. Hartley, T.T.; Lorenzo, C.F.; Qammer, H.K. Chaos in a fractional order Chua’s system. IEEE Trans. Circuits Syst. I Fundam. Theory Appl. 1995, 42, 485–490. [Google Scholar] [CrossRef]
  48. Petráš, I. A note on the fractional-order Chua’s system. Chaos Solitons Fractals 2008, 38, 140–147. [Google Scholar] [CrossRef]
  49. Zhu, H.; Zhou, S.; Zhang, J. Chaos and synchronization of the fractional-order Chua’s system. Chaos Solitons Fractals 2009, 39, 1595–1603. [Google Scholar] [CrossRef]
  50. Shishkina, E.; Sitnik, S. Basics of fractional calculus and fractional order differential equations. In Transmutations, Singular and Fractional Differential Equations with Applications to Mathematical Physics; Mathematics in Science and Engineering; Shishkina, E., Sitnik, S., Eds.; Academic Press: Cambridge, MA, USA, 2020; Chapter 2; pp. 53–84. [Google Scholar]
  51. Cao, B.; Nie, X.; Cao, J. Event-triggered adaptive neural networks tracking control for incommensurate fractional-order nonlinear systems with external disturbance. Neurocomputing 2023, 554, 126586. [Google Scholar] [CrossRef]
Figure 1. Characteristics of new fractional activation functions. (a) R L F σ activation function; (b) Derivative of R L F σ of order α .
Figure 1. Characteristics of new fractional activation functions. (a) R L F σ activation function; (b) Derivative of R L F σ of order α .
Mathematics 12 00083 g001
Figure 2. Chua’s circuit chaotic dynamics.
Figure 2. Chua’s circuit chaotic dynamics.
Mathematics 12 00083 g002
Figure 3. Chua’s circuit.
Figure 3. Chua’s circuit.
Mathematics 12 00083 g003
Figure 4. Chua’s circuit input/output simulated data.
Figure 4. Chua’s circuit input/output simulated data.
Mathematics 12 00083 g004
Figure 5. MLP architecture.
Figure 5. MLP architecture.
Mathematics 12 00083 g005
Figure 6. MLP convergence curve of all models for case 1.
Figure 6. MLP convergence curve of all models for case 1.
Mathematics 12 00083 g006
Figure 7. Chua’s circuit output prediction for case 1.
Figure 7. Chua’s circuit output prediction for case 1.
Mathematics 12 00083 g007
Figure 8. MLP convergence curve of all models for case 2.
Figure 8. MLP convergence curve of all models for case 2.
Mathematics 12 00083 g008
Figure 9. Chua’s circuit output prediction for case 2.
Figure 9. Chua’s circuit output prediction for case 2.
Mathematics 12 00083 g009
Figure 10. MLP convergence curve of all models for case 3.
Figure 10. MLP convergence curve of all models for case 3.
Mathematics 12 00083 g010
Figure 11. Chua’s circuit output prediction for case 3.
Figure 11. Chua’s circuit output prediction for case 3.
Mathematics 12 00083 g011
Table 1. Proposed fractional sigmoid function and its derivative.
Table 1. Proposed fractional sigmoid function and its derivative.
TypeFunctionDerivative
R L F σ ( t ) 1 1 + 0 R L T α ( t ) R L F σ ( t ) ( 1 R L F σ ( t ) )
Table 2. Comparison of activation functions.
Table 2. Comparison of activation functions.
Activation FunctionDiminishing GradientsBetter GeneralizationComputational EfficiencyFaster Convergence
R L F σ NoYesNoYes
σ [44]YesNoNoNo
Swish [45]NoNoNoNo
P C σ [46]NoYesNoNo
Table 3. Network hyper-parameters for training and validation.
Table 3. Network hyper-parameters for training and validation.
Loss M S E = 1 N i = 1 N ( y i y i ) 2
Hidden layers L = 2
Neurons n = 16 per layer
Learning rate η = 0.001
Batch size m = 32
Fractional order α = 0.3
Table 4. Performance comparison of R L F σ for case 1.
Table 4. Performance comparison of R L F σ for case 1.
R 2 (Training) R 2 (Testing)MSE ( × 10 3 )
sigmoid0.96040.9584196.8
swish0.95780.9558207.8
P C σ 0.90050.9008475.3
Proposed0.98600.985369.4
Table 5. Training time to achieve maximum accuracy for case 1.
Table 5. Training time to achieve maximum accuracy for case 1.
Activation FunctionCPU Time (Hrs:Mins:Secs)
sigmoid00:02:05
swish00:02:23
P C σ 00:02:54
Proposed00:01:01
Table 6. Performance comparison of R L F σ for case 2.
Table 6. Performance comparison of R L F σ for case 2.
R 2 (Training) R 2 (Testing)MSE ( × 10 3 )
sigmoid0.93090.9292346.7
swish0.92560.9247374.2
P C σ 0.87200.8729849.1
Proposed0.97030.9709137.3
Table 7. Training time to achieve maximum accuracy for case 2.
Table 7. Training time to achieve maximum accuracy for case 2.
Activation FunctionCPU Time (Hrs:Mins:Secs)
sigmoid00:01:02
swish00:01:12
P C σ 00:01:26
Proposed00:00:34
Table 8. Performance comparison of R L F σ for case 3.
Table 8. Performance comparison of R L F σ for case 3.
R 2 (Training) R 2 (Testing)MSE ( × 10 3 )
sigmoid0.89120.8845538.9
swish0.90460.9022492.4
P C σ 0.85840.8554748.9
Proposed0.98280.982283.8
Table 9. Training time to achieve maximum accuracy case 3.
Table 9. Training time to achieve maximum accuracy case 3.
Activation FunctionCPU Time (Hrs:Mins:Secs)
sigmoid00:01:16
swish00:01:32
P C σ 00:02:02
Proposed00:00:41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kumar, M.; Mehta, U.; Cirrincione, G. A Novel Approach to Modeling Incommensurate Fractional Order Systems Using Fractional Neural Networks. Mathematics 2024, 12, 83. https://doi.org/10.3390/math12010083

AMA Style

Kumar M, Mehta U, Cirrincione G. A Novel Approach to Modeling Incommensurate Fractional Order Systems Using Fractional Neural Networks. Mathematics. 2024; 12(1):83. https://doi.org/10.3390/math12010083

Chicago/Turabian Style

Kumar, Meshach, Utkal Mehta, and Giansalvo Cirrincione. 2024. "A Novel Approach to Modeling Incommensurate Fractional Order Systems Using Fractional Neural Networks" Mathematics 12, no. 1: 83. https://doi.org/10.3390/math12010083

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop