Next Article in Journal
Using Laplace Residual Power Series Method in Solving Coupled Fractional Neutron Diffusion Equations with Delayed Neutrons System
Previous Article in Journal
Fractal Dimension and Nuclear Magnetic Resonance Characteristics of Surfactants for Coal Gas Desorption
Previous Article in Special Issue
Effective Fractal Dimension at 2d-3d Crossover
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Intelligent Model for Supercapacitors with a Graphene-Based Fractal Electrode to Investigate the Cyclic Voltammetry

1
Metallurgical and Materials Engineering Department, Faculty of Technology, Fırat University, Elazig 23200, Turkey
2
Software Engineering Department, Faculty of Technology, Fırat University, Elazig 23200, Turkey
*
Author to whom correspondence should be addressed.
Fractal Fract. 2023, 7(3), 218; https://doi.org/10.3390/fractalfract7030218
Submission received: 19 January 2023 / Revised: 23 February 2023 / Accepted: 24 February 2023 / Published: 26 February 2023
(This article belongs to the Special Issue Advances in Multiparticle Fractal Aggregation)

Abstract

:
The aim of this study was to investigate the performance measurement of supercapacitors using the electrochemical properties of cyclic voltammetry (CV). The use of CV is crucial in evaluating the electrochemical performance of supercapacitors and determining the surface area of the catalyst with regard to the fractal properties of the electrode. The study specifically focused on the CV behavior of a supercapacitor formed by a cobalt-doped ceria/reduced graphene oxide (Co-CeO2/rGO) fractal nanocomposite, and its assessment was conducted using a machine learning (ML) model with the enhanced XGBoost. The model was trained using an experimental open-source dataset. The results showed that the proposed XGBoost model had a superior ability to predict the CV behavior of the supercapacitor, with nearly perfect results for the MAE, RMSE, and R-squared metrics, which are effective at evaluating the performance of regression models. With the successful design of the proposed intelligent prediction model, the study is expected to provide valuable insights into forming novel nanocomposite forms with high accuracy and minimal need for experiments.

1. Introduction

Supercapacitors have features such as a superior power density, a long cycle life, fast charge–discharge rates, maintenance-free features, a simpler packaging method, and compatibility with integrated circuits [1,2,3,4]. Therefore, supercapacitors have proven their feasibility as excellent storage components. The integration of supercapacitors into a self-power system has been widely employed in a range of portable microelectronic devices and sensing areas. The performance of supercapacitors is heavily influenced by the structural characteristics and chemical and physical properties of the electrode materials used. Thus, the design and screening of high-performance supercapacitor materials is a fundamental necessity for supercapacitor development [4,5,6].
Recent advances in nanomaterials have proven that the electrochemical performance of materials can be significantly improved by tuning their nanostructure and synthetic conditions [7,8]. In the past, many types of nanomaterials have been used for electrochemical energy storage devices, including nanoparticles, nanorods, nanotubes, thin films, and nanofibers. Nanofibers have been used as electrodes due to their high porosity, flexibility, lightness, large surface area, and resistance to aggregation [7,9].
Many classes of materials have been tried to fabricate supercapacitor electrodes. These materials include carbon [10,11], metal oxides [12,13], conductive polymers [14], metal sulfides, metal nitrides, and metal oxynitrides [15]. Metal oxides and metal nitrides are suitable for supercapacitor electrode production. However, oxynitride and its equivalents are being investigated more as supercapacitor electrodes [15]. Among the transition metal oxides, ruthenium oxide (RuO 2 ), manganese dioxide (MnO 2 ), nickel oxides (NiO), and cobalt oxide (Co 3 O 4 ) have been studied as electrode materials to date [16,17]. Simultaneously, because of their adjustable surface area, chemical stability, high electrical conductivity, and outstanding mechanical performance, graphene oxide (GO) and functionalized GO have been studied for supercapacitor applications [17,18]. Graphene consists of hexagonal carbon atoms that have a large surface area of 2630 m 2 g 1 . It has an electrical conductivity of 10 3 Sm 1 and a theoretical gravimetric capacitance of 550 Fg 1 . Due to these properties, the supercapacitor has shown great potential as an electrode material. Graphene-based supercapacitors have excellent cycling stability and rate capability [19]. Although the surface area of graphene oxide is less than that of graphene, graphene oxide exhibits a higher capacitance. This is due to the oxygen-containing functional groups on the surface [20].
It is desirable to obtain high specific capacitance by using low-cost electrode materials. CeO 2 nanostructures with high porosity and surface area may be produced, and it is simple to modify their form. Additionally, their nanocomposites may work in concert to enhance electrochemical performance [21]. On the other hand, the cobalt element exhibits good electrocatalytic activity between cobalt nanocrystals and various rare earth elements on other oxides due to its superior properties [22]. Pure CeO 2 has low capacitance (<100 Fg 1 ) in different electrolytic environments. The specific capacitance of CeO 2 can be improved with conductive carbon material supports. The excellent conductivity, porosity, and large surface area of carbon materials can effectively compensate for a number of cerium dioxide’s defects as an electrode material. For this purpose, it was aimed to create a hybrid material class with different properties by adding rGO to the composite and creating excellent physicochemical properties.
In their study, Wang et al. prepared a new electrode with high specific capacity and excellent cycling performance with Flammulina-velutipes-like CeO 2 /Co 3 O 4 /rGO nanoparticles on nickel foam substrate (CCGN) by hydrothermal synthesis and annealing [23]. Afza et al. proposed reduced graphene oxide (rGO) nanocomposites prepared by the hydrothermal method. They stated that rGO-based CeO 2 nanomaterials have an excellent charge-carrying capacity for dual applications in the fields of photocatalysis and electrocatalysis [24]. Veeresha et al. created highly scalable porous hierarchical microspheres composed of polyaniline nanofibers (PANIs), reduced graphene oxide (rGO), and cerium oxide nanorods (CNRs) synthesized by the spray drying method to obtain a functional and structural synergistic effect. They proposed a composite supercapacitor material with superior electrochemical properties by combining the three components together as a 3D porous hierarchical structure that maximizes ion and charge transport while reducing agglomeration [25].
The graphene-based materials have specific features such as being functional in many different aspects, and they have widespread usage in related fields. Reduced graphene oxide has a wide range of applications. In the study [26], the authors synthesized triple MoS 2 -rGO-Cu 2 O (MG-Cu) composites with NO 2 sensing properties. An attempt was made to design an efficient supercapacitor electrode material by using the synergistic effects of molybdenum disulfide (MoS 2 ) and rGO in [27]. Farshadnia et al. proposed an important analysis in their experiments to increase the porosity and surface area by first synthesizing a spongy graphene oxide nanostructure. Then, they fixed CoNi 2 S 4 and MoS 2 nanocomposites on porous graphene oxide to increase the capacity and improve its performance as a substrate. Finally, they integrated all the components to produce the final nanocomposite. The presence of metal sulfides as electroactive materials has promised a synergistic effect for use in supercapacitors by accelerating ion/electron diffusion rates and expanding active sites [28].
The fractal theory is a mathematical concept that describes the repetition of patterns on different scales in a self-similar manner. It is often used to describe complex systems with a hierarchical structure, where smaller patterns are similar to larger ones. In the field of materials science, the fractal theory has been used to understand the microstructures of materials, including fractal nanocomposites [29].
Fractal nanocomposites are a new class of materials that are characterized by their hierarchical structures, which are composed of multiple levels of fractal patterns. They are often made by incorporating nanoparticles into a polymer matrix, and their properties are influenced by both the size and distribution of the nanoparticles. In the context of supercapacitor design, fractal nanocomposites have been explored as potential electrode materials. This is because they have several desirable properties that make them attractive for supercapacitor applications, such as high surface area, good electrical conductivity, and excellent mechanical stability [30].
One of the main advantages of fractal nanocomposites in supercapacitor design is their high surface area, which is due to the fractal structure of the nanoparticles. This high surface area provides a large number of active sites for the storage of charge, which is important for improving the capacitance of the electrode. Another advantage of fractal nanocomposites is their good electrical conductivity, which is achieved through the use of nanoparticles with high electrical conductivity. This allows for efficient charge transport within the electrode, which is essential for achieving high power and energy density in supercapacitors. As a consequence, fractal nanocomposites have shown promise in the design of supercapacitors due to their high surface area, good electrical conductivity, and excellent mechanical stability. Further research is needed to fully understand the potential of fractal nanocomposites for supercapacitor applications and to optimize their performance for practical use [31]. In our study, we handled a fractal nanocomposite in terms of electrode design and performance measurement model. However, the fabrication of it and the details of the synthesis parameters were out of the scope of the proposed study. Due to the purposes of providing the machine learning model, the following sections of the article will not focus on the fractal design itself.
The supercapacitor system’s behavior is so complicated that it limits the possibility of obtaining the desired behavior with basic intuition using an Edisonian method. With its transient flow, supercapacitors’ power supply analysis requires tools such as numerical models, simulations, machine learning, and so on, apart from the trial and error methods [32,33]. Accurate application of those technologies in the layouts of storage behavior could result in fabricating more effective devices with high energy capability. Thanks to the superior modeling characteristic of artificial-intelligence (AI)-based techniques, the performance measurement of supercapacitors scoping the efficiency, reliability, and safety can better be figured out. Indeed, in nanotechnology, the physical, chemical, and electrical characteristics of nanomaterials could better be presented with the help of AI [34,35]. Besides AI-supported methods recently being preferred in supercapacitors, their important role in the formulation of material structures in materials engineering has taken place in the related literature [36,37]. At the center of the AI-based supercapacitor performance measurement tools lies the artificial neural network (ANN) methods. The conventional ANN prediction and classification models have widespread usage with their many different forms of architectures in complex analysis in different fields such as power electronics, energy management, and data-driven solutions to black box estimation problems. Indeed, ANNs are black boxes that find solutions on their own. That is, once a solution is discovered, it is impossible to determine how that solution was obtained. The ANN itself does not guarantee prediction or solution convergence. A learning model based on an ANN architecture can approximate a specific function with the correct parameters (also known as hyper-parameters) until it achieves a sufficient output. Such a remedy, however, is not always possible [38]. Modern machine learning techniques overcome these disadvantages of conventional algorithms. Amid growing interest in the existing machine learning techniques used in practice, gradient tree boosting excels in a variety of applications. It is utilized not just as a standalone predictor but is also integrated into the real-world production workflow. Chen and Guestrin described the extreme gradient boosting algorithm, a scalable machine learning system for tree boosting, which they named XGboost [39]. On a single machine, their proposed algorithmic system operates greater than 10-times faster than the up-to-date favored solutions with the capability of scaling billions of samples in distributed or memory-restricted scenarios. The most-significant component of XGBoost’s success is its scalability in all settings [39,40,41]. The widespread usage of the XGboost model inspired us to investigate its capability of prediction on supercapacitor performance conditions. To the best of the authors’ knowledge, there is not any reported study accessible aiming at the prediction of the performance of energy storage devices including batteries, supercapacitors, and other similar devices with the help of the XGboost model [39,42,43].
Supercapacitors are gaining attention as a new form of energy storage, but the main challenge is developing high-performance electrode materials. The proposed paper focused on analyzing carbon-material-based nanocomposites for use in supercapacitors, which can be improved by integrating additional materials. The analysis suggests that a specially designed electrode material is a suitable alternative for excellent performance [44]. The article also discusses new challenges and trends in supercapacitor design using artificial intelligence and machine learning methods.
Cyclic voltammetry (CV) is an important technique in the performance analysis of supercapacitors because it can provide information on the electrochemical behavior of the supercapacitor, including its capacitance, charge–discharge behavior, and kinetics. In a supercapacitor, the charge is stored on the surface of the electrodes rather than in the bulk of the material, as in a traditional battery. This means that the capacitance of the supercapacitor depends on the surface area of the electrodes, the nature of the electrode–electrolyte interface, and the kinetics of the charge–discharge process [45]. CV can be used to study the capacitance of a supercapacitor by measuring the current response to a potential sweep. The shape of the CV curve can provide information on the capacitance and the rate of charge–discharge. The capacitance can be calculated by the area under the CV curve, which is known as the charge stored. In addition, CV can be used to study the kinetics of the charge–discharge process by measuring the rate of the current response to a potential sweep. This can be used to understand the rate-limiting steps in the charge–discharge process and optimize the performance of the supercapacitor [46].
In cyclic voltammetry, an electrode is scanned through a potential range while the current is measured, and then, the potential is scanned in the opposite direction. The resulting plot of the current vs. potential is called a voltammogram, and it can provide information about the kinetics and thermodynamics of the redox reactions taking place [45]. The relationship between fractal structures and cyclic voltammetry is that fractals can be used to model the electrochemical behavior of a system. For example, a fractal model can be used to describe the mass transport of redox species to an electrode surface during a cyclic voltammetry experiment. This can be used to understand the kinetics of the electrochemical reaction and to optimize the conditions for the experiment. The proposed study deals with the CV performance analysis of such a fractal carbon electrode, whose synthesis process is described in the study [34].
Machine learning (ML) is used in the modeling of experimental data because it can provide a flexible and efficient way to analyze large and complex datasets. ML algorithms can be used to identify patterns and relationships in the data that might not be immediately apparent. This can help to extract meaningful information from the data and to understand the underlying mechanisms of the system being studied. One of the main advantages of using ML in the modeling of experimental data is that it can handle a large number of variables and can automatically identify important features and interactions. This can be particularly useful when the data are high-dimensional and when the underlying relationships are not well understood [47]. Another advantage is that ML models can be easily updated and refined as new data become available, which is important for applications such as predictive modeling and control systems. ML can also be used to optimize the experimental conditions to improve the performance of a system, by using techniques such as Bayesian optimization. ML offers a powerful and flexible approach to analyzing experimental data, allowing extracting insights and knowledge that would be difficult or impossible to obtain with traditional modeling techniques [46].
In this study, we took advantage of the effectiveness and robustness of a modern machine learning method, XGboost, to make an intelligent model picturing the cyclic voltammetry (CV) behavior of a nanostructured supercapacitor. The supercapacitor CV dataset was a public experimental dataset obtained from a published study [34], which also proposed a model design for CV behavior prediction. The investigated dataset consists of three subsets, which include experimental data of supercapacitors with different structures. Inspired by the study of Parwaiz et al., our aim was to enhance the data analysis and modeling strategy on the experimental supercapacitor dataset to reveal distinctive outcomes for research experts in their production stages. The benchmark study focused on the design of a Co-CeO 2 /rGO nanocomposite-structured supercapacitor and the prediction of several responses using the conventional models of the ANN and random forest algorithm. Our study also included an exploratory data analysis (EDA) besides a comprehensive prediction model of XGboost with superior performances compared to the benchmark study [34]. The importance and originality of this study focus on exploring the experimental data deeply for supercapacitor CV behavior prediction as a performance measurement tool along with an AI-based model. Figure 1 states the general framework of our study. In this holistic view, we can track the dataset handling, the core process of the study, which is called eXplainable AI, and the decision-making process. There are several ways in which this study makes an original contribution:
  • Instead of designing complex methods for building models for supercapacitor performance measurement by a practical approach, we considered a robust, yet effective solution as a profitable enhancement in this field.
  • Aiming to increase the predictive power of CV behavior, the proposed model is faster and relies on decision trees, reinforced by gradient boosting (an enhanced version of the conventional gradient boosting trees) with a low computational cost and fewer parameters.
The reader should bear in mind that the study was based on an enhanced AI prediction model for supercapacitors’ performance measurement instead of focusing on fabrication conditions. The remaining parts of the text are arranged as follows: In Section 2, a brief explanation is presented regarding the relationship between fractals and graphene-based nanocomposites. Section 3 is the data definition part and also includes the EDA for the used experimental dataset. Extreme gradient boosting trees are explained in the methodology, which is Section 4. Section 5 investigates the experimental findings of three XGboost models for each dataset. Conclusions and future work are presented in Section 6 along with suggestions and the limitations of the study.

2. Fractals and Graphene-Based Nanocomposites

To achieve the optimal technological development and energy gains, materials must be tailored in a way that maximizes their property–structure relationship. Increasing the defect density of graphene through a transition from 1D linear edges to fractal edges is one way to accomplish this. This approach highlights the fractal nature inherent in graphene structures, which is important to consider when designing graphene-based materials [48,49].
In addition to fractalization, incorporating graphene into nanocomposites is another promising avenue for improving the properties and functionalities of graphene-based materials. For instance, the high surface area of graphene nanocomposites allows for increased interaction with other materials, resulting in improved mechanical, electrical, and thermal properties. Furthermore, the addition of graphene to polymers and other materials can enhance their strength, stiffness, and thermal conductivity [49].
As a comprehensive framework, combining the fractalization of graphene edges with the incorporation of graphene into nanocomposites holds great potential for advancing a wide range of technological applications, including energy storage, catalysis, and electronics. By leveraging the unique properties of graphene and tailoring its structure in a strategic manner, we can unlock new possibilities for material design and innovation.
In the literature, there are studies discussing the use of fractal-like structures for supercapacitor applications [50,51]. In the study [50], a metal–oxide electrode was synthesized in three different morphologies with similar specific surface areas: fern, flake, and microsphere. The fractal dimensions of these morphologies were estimated from electrochemical impedance spectroscopy. The capacitive surface charge storage contribution from cyclic voltammetry increased with the fractal dimensions from the microspheres to the ferns. The study suggested that fractal-like structures are beneficial for supercapacitor applications by promoting capacitive surface charge storage. The experiments of the same study aimed to investigate the effect of the fractal dimension on the charge storage performance of the related metal–oxide electrodes.
In the study [51], the preparation and application of a nanocomposite in a two-electrode supercapacitor were described. Upon the formation of the composite, nanomaterials with a flower-petal-like shape were produced. The symmetric two-electrode supercapacitor exhibited excellent properties suitable for use in supercapacitor applications. Thus, the fractal structured nanocomposites were of a great importance to the supercapacitor design.
In addition to all those fractal studies in the supercapacitor literature, cyclic voltammetry itself has also a significant importance for the calculation of the fractal dimensions using the I-V signals. There are also studies in which the computation of the fractal dimension of a surface is performed by analyzing cyclic voltammograms obtained when the surface is utilized as the working electrode in an electrochemical cell [52,53]. The separation between the peaks observed in such voltammograms is highly influenced by the fractal dimension of the electrode. This relationship between fractals and cyclic voltammetry is significant, as it enables researchers to obtain information about the surface properties of electrodes from the voltammograms produced during electrochemical experiments. Cyclic voltammetry can be used to determine the fractal dimension of a surface when it is used as the working electrode in an electrochemical cell. The separation between peaks in the resulting voltammograms is heavily influenced by the fractal dimension of the electrode, providing valuable information about its surface properties [53].
Our study focused on developing models for cyclic voltammetry (CV) data, which plays a crucial role in calculating the fractal structures of electrodes. Specifically, we gathered a dataset from a graphene-based fractal electrode design, which will be discussed in the following section. By proposing CV prediction models, our study makes a significant contribution to researchers investigating fractal electrodes. This work will be particularly relevant for those seeking to better understand the behavior and properties of such electrodes in electrochemical systems.

3. Dataset Description

The dataset we analyzed in this study as a backbone consists of three different applications of supercapacitors. In the benchmark study [34], the authors presented the experimental findings of the electrochemical characterization of Co-CeO 2 /rGO nanocomposite. The electrolyte type of the produced supercapacitor was 0.5 M Na 2 SO 4 , and the experimental data for performance analysis were gained from various doping rates of cobalt (Co) or graphene oxide (GO) concentrates [34].
AI-based methods are data-driven; thus, the cyclic voltammetry (CV) behavior prediction model was designed utilizing the experimental dataset gathered for the Co-CeO2/rGO nanocomposite. The CV behavior prediction models were shaped with the three sub-datasets. Before going into the details of the exploratory data analysis (EDA), the following lines will describe the three sub-sets of the dataset of this study.

3.1. Dataset A

The first sub-dataset, which is called Dataset A, consists of 8000 samples. In Dataset A, the CV nanocomposite doping rate values are 0, 1%, 3%, 5%, and 7% at 10 mVs 1 scan rates. As in all sub-datasets, we have three inputs as the voltage value, Volt (V), the type of phase oxidation and reduction, and different doping concentration rates of 0, 1%, 3%, 5%, and 7%. The single output of all the dataset packages is the output current (A) value.
Table 1 shows some statistical values of Dataset A with the sample count, mean, standard deviation (Std), minimum (Min), maximum (Max), and the distribution rates of the samples as 25%, 50%, and 75% metrics. In Figure 2, we can see the data distribution of Dataset A with a stacked-line plot. Here, we can clearly see the output current variation with the input parameters.

3.2. Dataset B

The second sub-dataset, which is called Dataset B, consists of 4800 samples. In Dataset B, the CV nanocomposite doping rate values are 30%, 45%, and 75%. In Table 2, we can see some important statistical values of Dataset B. As in all sub-datasets, the type values are binary as 0 and 1. In this dataset, one can see that the output current bandwidth is too narrow with the small current values.
In Figure 3, the data distribution of Dataset B is given with a stacked-line plot. Here, we can clearly see that the output current variation range is very narrow.

3.3. Dataset C

The last sub-dataset, which is called Dataset C consists of 10,000 samples. In Dataset C, the CV nanocomposite doping rate values are again 0, 1%, 3%, 5%, and 7%. Table 3 gives some important statistical values of Dataset C. Here, the variation of the output current keeps its rate in a narrow band opposite Dataset A.
Figure 4 shows the data distribution of Dataset C in a stacked-line form. Dataset C is the larger of the sub-sets in the study with its ten-thousand samples. In response to the output current, here in Dataset C, we can see rapid changes compared to the first two sub-datasets.

3.4. Exploratory Data Analysis

In this sub-section, we conducted a deeper analysis shaping our data in detail. Starting from Dataset A, the important relations of the features (inputs) of our datasets and how they reflect the outputs were handled through the following lines. EDA is the method of determining what the data can represent to us, and we utilize it to discover patterns, connections, or anomalies that can help us with our ongoing analysis. While there are several ways to employ an EDA, in our study, we used tools that investigate the correlation matrix with the correlation values, scatter plots, pairs plots, and probability distributions of the datasets.
Firstly, we describe the correlation matrix of the datasets in our study. Figure 5, Figure 6 and Figure 7 indicate the details of the correlation values of the features of the three datasets, respectively. The color scaling from dark to light means an increasing correlation between the attributes. In light of the correlation matrix presentations, we can see that the biggest correlation values are generated with the Volt column/feature of all three datasets. However, it also cannot be ignored that the other two features have also an impact on the output current. In the experiments, the effect of the Volt feature was handled to see whether it was as strong as it seemed in the correlations or not.
To outline a better understanding, we sorted the correlation values of the features with a table demonstration. In Table 4, one can see the sorted correlations values of the inputs to the Current response. As we can see here, in all sub-datasets, the dominant feature is the Volt feature linking to the current response. However, in Dataset C, the Type feature has a very close correlation value with Volt. The Concentration feature is less correlated with the Current response. The EDA revealed important characteristics of the data distribution before the modeling stage.
Going deeper into the EDA, we can investigate the scatter plots of the features with a common y-axis of the output data, i.e., Current. A scatter plot visualization represents the data samples for two separate numeric variables by using dots [54]. Each dot on the horizontal and vertical axes represents a value for a single data point. Scatter plots were utilized to investigate the relations between dataset features. Figure 8, Figure 9 and Figure 10 visualizes the inner relations of the features with the scatter plots. In the graphics of the Current feature, we can see the partial forms because of the variable flow of the output current measures. In the scatter plot visualization, the y-axis is preferred to be our output value, which is the Current. Because we should see the data distribution of other features with reference to the output response, in support of the previous data, we can say that the relationship between Current and Volt is more evident than the others. The data distribution in all Current vs. Volt scatters plots shows an uphill pattern from left to right; this situation indicates a positive relationship between Current and Volt. As the Volt values increase (moving to the right), the Current values tend to increase (moving up). The other two features have their self-gradual distribution.
The pair plots are one of the most-useful tools (also known as a scatter plot matrix) in EDA. A pair plot shows the distribution of a single variable, as well as the relationships between two variables. Pair plots are excellent tools for identifying trends for further investigation. The pair plots are based on two fundamental forms of figures: the scatter plot and the histogram. The diagonal line histogram shows the distribution of a single variable, but the upper and lower triangular scatter plots illustrate the connection between the two samples [54].
In Figure 11, we can see the partial distributions of the dataset features. Here, in Dataset A, the concentration value of 3% includes the greatest sample size of the output variable Current. The Type feature has two leveled distributions of 0–1. It is clearly seen that the oxidation and reduction types are equally distributed. In Figure 12, with higher concentration values of Dataset B, it is clear that the output values are distributed more equally between those concentrations. Figure 13 shows an even distribution of the Concentration and Volt features.
Lastly, we ended the EDA process with density estimations. With binning and counting observations, a histogram attempts to mimic the underlying probability density function that created the data. Kernel density estimation (KDE) offers an alternative approach to the same problem. A KDE plot, rather than utilizing discrete bins, provides a smooth observation distribution with a Gaussian kernel, yielding a continuous density estimate. Figure 14, Figure 15 and Figure 16 illustrate the density and probability plots of Dataset A, Dataset B, and Dataset C concerning the output (target) column of Current. In Figure 14, we can see a normal distribution, but the output is a bit far from a linear distribution. It does not follow a linear trend. In Figure 15, we can see a bi-modal distribution, which means that the output has two peaks. This points to Dataset B having two groups of the Current distribution. However, in Dataset C, we sense a more linear form in the output distribution. Figure 16 also illustrates a bi-modal distribution form for the output current of Dataset C. The output current of Dataset C has the most linear-like distribution among all of them. Making an overall statement about the sub-datasets, we can conclude that the supercapacitor form in Dataset C addresses the possibility of more stable CV behavior.

4. Machine Learning Model—The XGBoost Algorithm

With its very popular place in machine learning applications, the XGBoost algorithm mainly relies on decision tree logic. However, as its enhanced version, the main advantage is based on the tree boosting methodology. As an upper step, Chen and Guestrin [39] enriched the gradient tree boosting with a greedily added function for optimization. The XGBoost algorithm involves a combined learning method. The combined learning method merges multiple learning models in order for the integrated model to have a more powerful generalization capability to achieve more reliable modeling outcomes. XGBoost is an amendment to the boosting algorithm founded on the gradient descent tree, which is composed of multiple decision tree iterations. XGBoost initially forms multiple classification and regression tree (CART) models to predict the data and then combines those trees as a unique tree model. The model will proceed to iteratively update, and the brand-new tree model formed in each repetition will fit the residual of the earlier tree. Since the number of trees increases, the complexity of the integrated model will progressively increase until it addresses the complexity of the data themselves, whereupon the training succeeds in the most-reliable results. To provide a brief explanation of the XGBoost methodology, we followed the definition of the studies [39,55] in the following lines. Equation (1) is the basis of the XGBoost algorithm model, where f t ( x i ) = w q ( x ) states the CART space, w q ( x ) is the sample score of x, the model prediction is achieved by accumulation, and q describes the composition of each tree, T is the tree number, and each f t matches an independent tree composition q and leaf weight.
y ^ i = φ ( x i ) = t = 1 T f t ( x i )
In XGBoost, the inner decision tree uses a conventional regression tree. Concerning the squared loss function, the split node of the regression tree resembles the residual. For the global loss function (gradient descent), the split node of the regression tree matches the appraised value of the residual. Hence, the accuracy of XGBoost will be higher. Equation (2) states the iterative method of residual fitting. In (2), y i ^ ( t 1 ) is the prediction value of the ith sample after t 1 repetitions. y i ^ ( 0 ) is the initial set of the i-th data points.
y i ^ ( 0 ) = 0 y i ^ ( 1 ) = f 1 ( x i ) = y i ^ ( 0 ) + f 1 ( x i ) y i ^ ( 2 ) = f 1 ( x i ) + f 2 ( x i ) = y i ^ ( 1 ) + f 2 ( x i ) y i ^ ( t ) = k = 1 T f k ( x i ) = y i ^ ( t 1 ) + f t ( x i )
The objective optimization function in the XGBoost method, namely the loss function (3), can be achieved by the iterative manner of the residuals. Concerning the general loss function, XGBoost will implement a second-order Taylor expansion to explore more information about the gradient and simultaneously eliminate the constant term; therefore, the gradient descent method can be thoroughly trained. Equations (4) and (5) state the loss function of the t-th step, where g i and h i stand for the first and second derivatives.
f t o b j = i = 1 n l y i , y ^ i t + i = 1 t Ω ( f i ) = y ^ i ( t 1 ) + f t ( x i ) = i = 1 n l y · y ^ i t + Ω ( f i ) + C
g i = y ^ i ( t 1 ) l y i , y i ^ ( t 1 )
h i = y ^ i ( t 1 ) 2 l y i , y i ^ ( t 1 )
Ω ( f ) = γ T + 1 2 λ i = 1 n ω j 2
Unlike other methods, XGBoost uses a regularization term Ω ( f ) (6) to block over-fitting and considerably increase the accuracy of the model. The Ω ( f ) function describes the model complexity of the tree. The smaller the function output, the more powerful the generalization capability of the tree is. ω j stands for the weight on the jth leaf node of the tree model f; T states the cumulative number of leaf nodes of the tree model; γ shows the penalty term of the L1 regularity; λ is similarly the penalty term of the L2 regularity, also a design parameter for the algorithm. Consequently, the objective function definitions (7)–(9) are achieved, where I j = { i | q ( x i ) = j } pictures the sample set upon the jth leaf node [39,55]:
f o b j = 1 2 j = 1 T G j 2 H j + λ + γ T
G j = i ϵ I j g i
H j = i ϵ I j h i
Figure 17 addresses a general and briefly explained flow chart about the XGBoost algorithm in a single view. When K trees are designed after training, the features of the samples in the prediction will have a related leaf node in every tree, and each leaf node refers to a score. Finally, the related scores of each tree are summed up to build the recognition prediction value of each sample [56].

5. Experimental Results

In this section, we present and discuss the obtained experimental results for the CV prediction of supercapacitor performance measurement in three datasets of the study. In our code laboratory setup, we used the Canonical Ubuntu 20.04 operating system (OS) installed on a workstation. Ubuntu is an open-source OS using the Linux kernel and based on Debian. The workstation had an Intel Xeon E5-2620 v4 @ 2.10 GHz dual-core processor with 32 GB of RAM and an Nvidia Quadro M4000 8 GB GPU.
In our study, the model design was coded on a state-of-the-art AI platform named H 2 O [57] using the Jupyter Lab environment [58]. The programming language was based on the H 2 O Python module [59] and the necessary Python 3.8 packages.
The H 2 O platform makes AI and machine learning research rapid and effective. It makes the researcher’s study time and cost effective, and with its built-in grid search algorithm, the designed models’ hyper-parameters are optimized automatically. Even if a train–test split method is preferred in experiments, H 2 O models use k-fold cross-validation in their training procedure, pick the best model, and run the test operation to generate the final performance metrics. In our experiments, we performed these complex processes in only 60 s. The programming duration was selected as 60 s intentionally because we wanted to emphasize our algorithm’s fast and cost-effective structure over the benchmark study. The proposed model, XGBoost, ran in 60 s of experiment time including the training and testing phases to achieve the performance metrics that we demonstrate in the next lines. Before discussing the performance of our proposed prediction model for the supercapacitor CV behavior, we provide the selected model’s unique parameters to set it up in the code design of the H 2 O estimator models.
To describe the XGBoost model parameters, it is very logical to start with the booster parameter. It establishes the learner type. Ordinarily, this might be a tree or a linear function. In the event of trees, the model will be composed of an ensemble of trees. When it is set as the linear booster, it will be designed with a weighted sum of linear functions. In our case, the model was optimized with a dropouts meet multiple additive regression tree (DART) booster [60]. The normalization type of our model was set to “tree”, which provides the trees having the same weight as each of the dropped trees in the DART process. The seed option was set for the randomization of the datasets as “1234”. The models solve dour problems with 50 trees.
To provide a balanced evaluation and even level for separate datasets, we determined our metrics set as the mean-squared error (MSE), root-mean-squared error (RMSE), mean absolute error (MAE), and coefficient of determination, named R-squared [15,34]. We can explore the definitions of the metrics below as stated in (10):
MSE = 1 n t = 1 n e t 2 , where e t = A t p t RMSE = 1 n t = 1 n e t 2 MAE = 1 n t = 1 n e t R-squared = t = 1 n e t 2 t = 1 n ( A t p t ¯ )
where n is the sample number and e t is the error between the actual value ( A t ) and predictions of the model ( p t ). In the R-squared metric, p t ¯ states the mean value of the predicted data. Outliers might have an undesirable impact on the value of these evaluation criteria. To avoid this issue, the MAE was used as a balanced measure, which handles all errors uniformly [34]. The closer to zero they are, the MSE, RMSE, and MAE metrics point out a great performance. On the other hand, for R-squared, a great performance is obtained by being near 1.
In Table 5, we can examine the training results of the three datasets with the achieved performance values. Here, we can also see the sample number of the training datasets, for which we used the 80% and 20% split ratio for all the datasets. In the training stage, the R-squared metric was not investigated. Meanwhile, the readers should keep in mind that our training procedure also included five-fold cross-validation, which means our model was tested while it was being trained.
Table 5 shows us that in the training stage, Dataset B had the best performance on every metric. Following it, Dataset C achieved the second position in the training. This proved the former explanations about the data distributions we interpreted from the EDA.
In Table 6, the results of the tests are presented along with the R-squared metric and the number of samples in the testing dataset. If we examine the performance of the model based on the metrics of the MSE, RMSE, and MAE, it can be observed that Dataset B and Dataset C exhibited the best scores among the datasets. However, when considering the R-squared metric, Dataset A outperformed the other datasets. This difference in performance can be attributed to the larger sample size of Dataset A. Nevertheless, it can be concluded that our proposed intelligent model for predicting the CV behavior of supercapacitors demonstrated superior performance and achieved the objectives outlined in our motivation.
To provide a more comprehensive analysis of our results, additional visual explanations were prepared and are presented in Figure 18. This figure shows the distribution of the actual vs. predicted values and the residual error plots for all sub-datasets included in our study. The visual representation provides a clear picture of the performance of the proposed model and highlights its strengths and limitations.
The results obtained from the proposed machine learning-based prediction model for the output current of supercapacitors clearly demonstrated the effectiveness of the model. This can be easily seen from the visual presentations provided in Figure 18. The figures show the distribution of the actual vs. predicted values and the residual error plots for all sub-datasets included in our study.
Upon close examination of the figures, it becomes apparent that the model was able to accurately predict the output current of the supercapacitors. This can be seen from the strong correlation between the actual and predicted values, as well as the low residual errors. In other words, the model produced results with a high degree of accuracy and low prediction errors.
This level of performance is a testament to the robustness of the machine learning algorithms and techniques used in the model, as well as the quality of the data used for training and testing. The high correlation and low errors indicate that the model was able to capture the underlying patterns and relationships in the data and to generalize these patterns to make accurate predictions on new, unseen data.
In conclusion, the results obtained from the proposed model demonstrated that it is an effective solution for predicting the output current of supercapacitors. The model’s accuracy and low prediction errors make it a valuable tool for researchers, engineers, and other professionals in the field of energy storage and power electronics.
Figure 19 provides a visual representation of the actual output currents and the predicted output currents on the same plot. This presentation is useful for evaluating the performance of the prediction model and for clearly showing how the predicted outputs track the actual data.
From the Figure 19, it can be observed that, for all datasets, the predicted output currents provided a superior tracking performance compared to the actual data. This indicates that the model was able to effectively capture the underlying patterns and relationships between the input features and the output currents and to use these patterns to make accurate predictions on new data.
In order to gain a deeper understanding of the model’s performance, an additional test was conducted using a single-input prediction model that only included the Volt feature as its input. The results of this test were presented in the EDA, where it was shown that the Volt feature had a significant impact on the output compared to the other two features. However, in order to investigate the effect of the less dominant features on the outputs, an additional test was conducted using Dataset A and the single-input model.
This additional test provided valuable insights into the performance of the prediction model and highlighted the importance of considering all relevant features in the training and testing of machine learning models. The results of this test can be used to further optimize the model and improve its performance, making it an even more effective tool for predicting the output currents of supercapacitors. Please see Table 7 for metrics and detailed results.
Contrary to the commonly held belief that the Volt feature alone would provide sufficient performance, the results of the single-input model test showed that this is not the case. The R-squared value of the distribution revealed a negative correlation between the actual and predicted values, indicating that the model’s performance was poor when only considering the Volt feature.
To further understand the performance of the proposed model, a comparison was made with the benchmark study [34]. In this comparison, the mean values of the RMSE and MAE metrics were selected and calculated from the results obtained in the benchmark study. The results from the testing stage of our proposed model were then compared to these mean values.
The comparison with the benchmark study provided valuable insights into the performance of the proposed model and highlighted its strengths and weaknesses. The comparison also helped to validate the results obtained from the model and demonstrate its effectiveness compared to other similar studies in the field.
In conclusion, the results of the comparison with the benchmark study and the single input model test provided a clear picture of the performance of the proposed model. While the model may have some limitations, it still represents a valuable tool for predicting the output currents of supercapacitors and provides a solid foundation for future research and development in the field.
The comparison between the proposed XGBoost-based machine learning prediction model and the benchmark study is presented in Table 8. The results showed that the proposed model outperformed the benchmark study in terms of the RMSE and MAE metrics, commonly used for evaluating the performance of supercapacitor CV behavior prediction models.
The benchmark study proposed a conventional artificial neural network (ANN) method and a random forest algorithm for the same datasets. While the ANN is well-known for its iterative learning algorithm, the random forest is a basic version of the decision tree models. However, the proposed model uses the state-of-the-art XGBoost model, which offers a low-computational-cost implementation. With very few parameters and a 60 s runtime, the proposed model outperformed the benchmark study on almost all datasets with its RMSE and MAE values. The only exception was the RMSE value of Dataset B, where the proposed model fell just behind the ANN, but with a very small difference.
The results of the comparison demonstrated the strong performance of the proposed method and its superiority over conventional methods in supercapacitor performance measurement tool design. The proposed model is more efficient and offers a rapid process with a few parameter designs, making it an attractive solution for supercapacitor performance prediction.
The current study successfully demonstrated the implementation of a data-driven method for predicting the electrode effects on supercapacitor performance and revealed the most-prominent parameters that affect the behavior of supercapacitors. In the future, the proposed model may facilitate further analyses and potential works on the optimization of carbon-based electrodes in various supercapacitor applications. Overall, the results of this study make a valuable contribution to the field of supercapacitor performance prediction and open up new avenues for further research and development.

6. Conclusions

This article explored the application and validation of the XGBoost machine learning model as a performance measurement tool for supercapacitors using Co-CeO 2 /rGO fractal nanocomposites as the electrode material. The methodology of the study focused on improving the current state-of-the-art AI-based models for supercapacitor performance measurement. The XGBoost model stands out with its fast and efficient processing capabilities, and its reliability was confirmed by the R-squared, MAE, MSE, and RMSE metrics. As far as the authors are aware, this study is the first to introduce the use of the XGBoost model as a performance measurement tool for supercapacitors, and its simple structure makes it easy to implement. The results from the proposed model showed a very high level of accuracy and performance with regression values close to perfect and low RMSE, MAE, and R-square scores. This makes the XGBoost model a major game-changer in the field of supercapacitor performance prediction. The experimental results from the proposed intelligent model have the potential to be integrated with batteries and fuel cells in the near future for the development of controlled energy-related practices. The CV behavior played a critical role in determining the fractal properties of the material, and the proposed model outperformed previous benchmark studies in this aspect. This article represents a significant advancement in the field of supercapacitor performance measurement by combining AI-based measurement tools with nanotechnology. The results of this study will be invaluable in the development of novel nanocomposites for use as electrodes in supercapacitors and other energy storage devices for electrical and electronic applications. The authors’ future research plans include the use of novel prediction models to further improve the accuracy of the data analysis and to create more effective nanocomposites for use in supercapacitors. This work will be of great interest to experts in the field of supercapacitor performance measurement and energy storage technologies.

Author Contributions

Conceptualization, F.U. and N.K.; methodology, F.U. and N.K.; software, F.U.; validation, F.U. and N.K.; formal analysis, F.U.; investigation, F.U. and N.K.; resources, F.U. and N.K.; data curation, F.U.; writing—original draft preparation, F.U. and N.K.; writing—review and editing, F.U. and N.K.; visualization, F.U. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available in [34].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alagar, S.; Madhuvilakku, R.; Mariappan, R.; Piraman, S. Nano-architectured porous Mn2O3 spheres/cubes vs rGO for asymmetric supercapacitors applications in novel solid-state electrolyte. J. Power Sources 2019, 441, 227181. [Google Scholar] [CrossRef]
  2. Xiao, X.; Li, T.; Yang, P.; Gao, Y.; Jin, H.; Ni, W.; Zhan, W.; Zhang, X.; Cao, Y.; Zhong, J.; et al. Fiber-based all-solid-state flexible supercapacitors for self-powered systems. ACS Nano 2012, 6, 9200–9206. [Google Scholar] [CrossRef] [PubMed]
  3. Wu, Z.S.; Feng, X.; Cheng, H.M. Recent advances in graphene-based planar micro-supercapacitors for on-chip energy storage. Natl. Sci. Rev. 2014, 1, 277–292. [Google Scholar] [CrossRef] [Green Version]
  4. Liu, P.; Wen, Y.; Huang, L.; Zhu, X.; Wu, R.; Ai, S.; Xue, T.; Ge, Y. An emerging machine learning strategy for the assisted-design of high-performance supercapacitor materials by mining the relationship between capacitance and structural features of porous carbon. J. Electroanal. Chem. 2021, 899, 115684. [Google Scholar] [CrossRef]
  5. Qian, L.; Guo, F.; Jia, X.; Zhan, Y.; Zhou, H.; Jiang, X.; Tao, C. Recent development in the synthesis of agricultural and forestry biomass-derived porous carbons for supercapacitor applications: A review. Ionics (Kiel) 2020, 26, 3705–3723. [Google Scholar] [CrossRef]
  6. Zhu, T.; Wills, R.G.; Lot, R.; Ruan, H.; Jiang, Z. Adaptive energy management of a battery-supercapacitor energy storage system for electric vehicles based on flexible perception and neural network fitting. Appl. Energy 2021, 292, 116932. [Google Scholar] [CrossRef]
  7. Aihemaitituoheti, R.; Alhebshi, N.A.; Abudula, T. Effects of Precursors and Carbon Nanotubes on Electrochemical Properties of Electrospun Nickel Oxide. Molecules 2021, 26, 5656. [Google Scholar] [CrossRef]
  8. Becherif, M.; Ramadan, H.; Ayad, M.; Hissel, D.; Desideri, U.; Antonelli, M. Efficient start–up energy management via nonlinear control for eco–traction systems. Appl. Energy 2017, 187, 899–909. [Google Scholar] [CrossRef]
  9. Lu, X.; Wang, C.; Favier, F.; Pinna, N. Electrospun Nanomaterials for Supercapacitor Electrodes: Designed Architectures and Electrochemical Performance. Adv. Energy Mater. 2017, 7, 1601301. [Google Scholar] [CrossRef]
  10. Ghosh, S.; Santhosh, R.; Jeniffer, S.; Raghavan, V.; Jacob, G.; Nanaji, K.; Kollu, P.; Jeong, S.K.; Grace, A.N. Natural biomass derived hard carbon and activated carbons as electrochemical supercapacitor electrodes. Sci. Rep. 2019, 9, 16315. [Google Scholar] [CrossRef] [Green Version]
  11. Down, M.P.; Rowley-Neale, S.J.; Smith, G.C.; Banks, C.E. Fabrication of Graphene Oxide Supercapacitor Devices. ACS Appl. Energy Mater. 2018, 1, 707–714. [Google Scholar] [CrossRef] [Green Version]
  12. Rani, B.; Sahu, N.K. Electrochemical properties of CoFe2O4 nanoparticles and its rGO composite for supercapacitor. Diam. Relat. Mater. 2020, 108, 107978. [Google Scholar] [CrossRef]
  13. Talluri, B.; Ghosh, S.; Rao, G.R.; Thomas, T. Nanocomposites of digestively ripened copper oxide quantum dots and graphene oxide as a binder free battery-like supercapacitor electrode material. Electrochim. Acta 2019, 321, 134709. [Google Scholar] [CrossRef]
  14. Li, H.; Wang, J.; Chu, Q.; Wang, Z.; Zhang, F.; Wang, S. Theoretical and experimental specific capacitance of polyaniline in sulfuric acid. J. Power Sources 2009, 190, 578–586. [Google Scholar] [CrossRef]
  15. Ghosh, S.; Rao, G.R.; Thomas, T. Machine learning-based prediction of supercapacitor performance for a novel electrode material: Cerium oxynitride. Energy Storage Mater. 2021, 40, 426–438. [Google Scholar] [CrossRef]
  16. Winter, M.; Brodd, R.J. What are batteries, fuel cells, and supercapacitors? Chem. Rev. 2004, 104, 4245–4269. [Google Scholar] [CrossRef] [Green Version]
  17. Suneetha, R.B.; Selvi, P.; Vedhi, C. Synthesis, structural and electrochemical characterization of Zn doped iron oxide/grapheneoxide/chitosan nanocomposite for supercapacitor application. Vacuum 2019, 164, 396–404. [Google Scholar] [CrossRef]
  18. Xia, J.; Chen, F.; Li, J.; Tao, N. Measurement of the quantum capacitance of graphene. Nat. Nanotechnol. 2009, 4, 505–509. [Google Scholar] [CrossRef]
  19. Ge, W.; Ma, Q.; Wang, W.; Jia, F.; Song, S. Synthesis of three-dimensional reduced graphene oxide aerogels as electrode material for supercapacitor application. Chem. Phys. 2021, 543, 111096. [Google Scholar] [CrossRef]
  20. Korkmaz, S.; Kariper, İ.A. Graphene and graphene oxide based aerogels: Synthesis, characteristics and supercapacitor applications. J. Energy Storage 2020, 27, 101038. [Google Scholar] [CrossRef]
  21. Das, H.T.; Dutta, S.; Das, N.; Das, P.; Mondal, A.; Imran, M. Recent trend of CeO2-based nanocomposites electrode in supercapacitor: A review on energy storage applications. J. Energy Storage 2022, 50, 104643. [Google Scholar] [CrossRef]
  22. Veeresha, G.; Krishnamurthy, G.; Shivakumar, M. Cobalt nanocrystals doped on CeO2/RGO nanocomposite for supercapacitor applications. Inorg. Chem. Commun. 2022, 138, 109232. [Google Scholar] [CrossRef]
  23. Wang, Z.; Zhao, K.; Lu, S.; Xu, W. Application of flammulina-velutipes-like CeO2/Co3O4/rGO in high-performance asymmetric supercapacitors. Electrochim. Acta 2020, 353, 136599. [Google Scholar] [CrossRef]
  24. Afza, N.; Shivakumar, M.; Alam, M.W.; Kumar, A.N.; Bhatt, A.S.; Murthy, H.A.; Ravikumar, C.; Mylarappa, M.; Selvanandan, S. Facile hydrothermal synthesis of cerium oxide/rGO nanocomposite for photocatalytic and supercapacitor applications. Appl. Surf. Sci. Adv. 2022, 11, 100307. [Google Scholar] [CrossRef]
  25. Jeyaranjan, A.; Sakthivel, T.S.; Neal, C.J.; Seal, S. Scalable ternary hierarchical microspheres composed of PANI/rGO/CeO2 for high performance supercapacitor applications. Carbon 2019, 151, 192–202. [Google Scholar] [CrossRef]
  26. Ding, Y.; Guo, X.; Kuang, D.; Hu, X.; Zhou, Y.; He, Y.; Zang, Z. Hollow Cu2O nanospheres loaded with MoS2/reduced graphene oxide nanosheets for ppb-level NO2 detection at room temperature. J. Hazard. Mater. 2021, 416, 126218. [Google Scholar] [CrossRef]
  27. Kumar, P.; Sharma, S.; Jabeen, S.; Samra, K.S. Hybrid microwave annealing assisted synthesis of MoS2-RGO nanostructures: Optimization and characterization for application in supercapacitors. Electrochim. Acta 2022, 426, 140738. [Google Scholar] [CrossRef]
  28. Farshadnia, M.; Ensafi, A.A.; Mousaabadi, K.Z.; Rezaei, B. Design and synthesis of three-dimensional CoNi2S4@ MoS2@ rGO nanocomposites and its application in electrochemical supercapacitors. J. Alloys Compd. 2022, 906, 164278. [Google Scholar] [CrossRef]
  29. Mwema, F.M.; Jen, T.C.; Kaspar, P. Fractal Theory in Thin Films: Literature Review and Bibliometric Evidence on Applications and Trends. Fractal Fract. 2022, 6, 489. [Google Scholar] [CrossRef]
  30. Ehsani, A.; Bigdeloo, M.; Assefi, F.; Kiamehr, M.; Alizadeh, R. Ternary nanocomposite of conductive polymer/chitosan biopolymer/metal organic framework: Synthesis, characterization and electrochemical performance as effective electrode materials in pseudocapacitors. Inorg. Chem. Commun. 2020, 115, 107885. [Google Scholar] [CrossRef]
  31. Vatansever, D.; Haciismailoglu, M. Pencil graphite/polypyrrole supercapacitors: The structural and chemical changes of the electrodes after capacitive measurements. Mater. Chem. Phys. 2023, 295, 127073. [Google Scholar] [CrossRef]
  32. Guan, C.; Liu, J.; Wang, Y.; Mao, L.; Fan, Z.; Shen, Z.; Zhang, H.; Wang, J. Iron oxide-decorated carbon for supercapacitor anodes with ultrahigh energy density and outstanding cycling stability. ACS Nano 2015, 9, 5198–5207. [Google Scholar] [CrossRef]
  33. Lipus, J.; Krukiewicz, K. Challenges and limitations of using charge storage capacity to assess capacitance of biomedical electrodes. Measurement 2022, 191, 110822. [Google Scholar] [CrossRef]
  34. Parwaiz, S.; Malik, O.A.; Pradhan, D.; Khan, M.M. Machine-Learning-Based Cyclic Voltammetry Behavior Model for Supercapacitance of Co-Doped Ceria/rGO Nanocomposite. J. Chem. Inf. Model. 2018, 58, 2517–2527. [Google Scholar] [CrossRef] [PubMed]
  35. Haris, M.; Hasan, M.N.; Qin, S. Early and robust remaining useful life prediction of supercapacitors using BOHB optimized Deep Belief Network. Appl. Energy 2021, 286, 116541. [Google Scholar] [CrossRef]
  36. Singla, V.K.; Aggarwal, V.; Gupta, S. Artificial Intelligence and Nanotechnology: A Super Convergence. In Artificial Intelligence; CRC Press: Boca Raton, FL, USA, 2021; pp. 1–8. [Google Scholar]
  37. Zou, C.; Li, J.; Wang, W.Y.; Zhang, Y.; Lin, D.; Yuan, R.; Wang, X.; Tang, B.; Wang, J.; Gao, X.; et al. Integrating data mining and machine learning to discover high-strength ductile titanium alloys. Acta Mater. 2021, 202, 211–221. [Google Scholar] [CrossRef]
  38. Gilanifar, M.; Wang, H.; Cordova, J.; Ozguven, E.E.; Strasser, T.I.; Arghandeh, R. Fault Classification in Power Distribution Systems Based on Limited Labeled Data using Multi-Task Latent Structure Learning. Sustain. Cities Soc. 2021, 73, 103094. [Google Scholar] [CrossRef]
  39. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining; KDD 16; Association for Computing Machinery: New York, NY, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef] [Green Version]
  40. Doshi, M.; Ning, X. A Data-Driven Framework for Buckling Analysis of Near-Spherical Composite Shells Under External Pressure. J. Appl. Mech. 2021, 88, 081007. [Google Scholar] [CrossRef]
  41. Yang, P.; Wu, S.; Wu, H.; Lv, D.; Zou, W.; Chu, L.; Shao, Y.; Wu, S. Prediction of bending strength of Si3N4 using machine learning. Ceram. Int. 2021, 47, 23919–23926. [Google Scholar] [CrossRef]
  42. Shalauddin, M.; Akhter, S.; Basirun, W.J.; Anuar, N.S.; Akbarzadeh, O.; Mohammed, M.A.; Johan, M.R. Carboxylated nanocellulose dispersed nitrogen doped graphene nanosheets and sodium dodecyl sulfate modified electrochemical sensor for the simultaneous determination of paracetamol and naproxen sodium. Measurement 2022, 194, 110961. [Google Scholar] [CrossRef]
  43. Zhang, Z.; Huang, Y.; Qin, R.; Ren, W.; Wen, G. XGBoost-based on-line prediction of seam tensile strength for Al-Li alloy in laser welding: Experiment study and modelling. J. Manuf. Process. 2021, 64, 30–44. [Google Scholar] [CrossRef]
  44. L’vov, P.E.; Sibatov, R.T.; Yavtushenko, I.O.; Kitsyuk, E.P. Time-Fractional Phase Field Model of Electrochemical Impedance. Fractal Fract. 2021, 5, 191. [Google Scholar] [CrossRef]
  45. Singh, A.; Kumar, D.; Singh, B.; Shinde, V.; Kaur, R. Fractal Analysis of Pure and Fe-Doped Manganese Oxide Supercapacitor Electrodes. Prot. Met. Phys. Chem. Surf. 2022, 58, 991–998. [Google Scholar] [CrossRef]
  46. Suantai, S.; Sabir, Z.; Raja, M.A.Z.; Cholamjiak, W. A Stochastic Bayesian Neural Network for the Mosquito Dispersal Mathematical System. Fractal Fract. 2022, 6, 604. [Google Scholar] [CrossRef]
  47. Chen, R.; Tang, X.; Li, X. Adaptive Stochastic Gradient Descent Method for Convex and Non-Convex Optimization. Fractal Fract. 2022, 6, 709. [Google Scholar] [CrossRef]
  48. Vu, C.C.; Kim, J. Fractal structures in flexible electronic devices. Mater. Today Phys. 2022, 27, 100795. [Google Scholar] [CrossRef]
  49. Janowska, I.; Lafjah, M.; Papaefthymiou, V.; Pronkin, S.; Ulhaq-Bouillet, C. Edges fractal approach in graphene–Defects density gain. Carbon 2017, 123, 395–401. [Google Scholar] [CrossRef]
  50. Jaiswal, R.; Chaturvedi, G.; Sharma, V.; Ilangovan, S.; SarojiniAmma, S.; Ajeesh, K.; Tatiparti, S.S.V. Are Fractal-Like Structures Beneficial for Supercapacitor Applications? A Case Study on Fe2O3 Negative Electrodes. Energy Fuels 2022, 37, 1275–1287. [Google Scholar] [CrossRef]
  51. Ben Gouider Trabelsi, A.; M Elsayed, A.; Alkallas, H.; Al-Noaimi, M.; Kusmartsev, F.; Rabia, M. A Fractal, Flower Petal-like CuS-CuO/G-C3N4 Nanocomposite for High Efficiency Supercapacitors. Coatings 2022, 12, 1834. [Google Scholar] [CrossRef]
  52. Dhillon, S.; Kant, R. Quantitative roughness characterization and 3D reconstruction of electrode surface using cyclic voltammetry and SEM image. Appl. Surf. Sci. 2013, 282, 105–114. [Google Scholar] [CrossRef]
  53. Strømme, M.; Niklasson, G.A.; Granqvist, C.G. Voltammetry on fractals. Solid State Commun. 1995, 96, 151–154. [Google Scholar] [CrossRef]
  54. Cook, D.; Reid, N.; Tanaka, E. The Foundation is Available for Thinking about Data Visualization Inferentially. Harv. Data Sci. Rev. 2021, 3. [Google Scholar] [CrossRef]
  55. Liu, B.; Tan, X.; Jin, Y.; Yu, W.; Li, C. Application of RR-XGBoost combined model in data calibration of micro air quality detector. Sci. Rep. 2021, 11, 15662. [Google Scholar] [CrossRef] [PubMed]
  56. Guo, R.; Zhao, Z.; Wang, T.; Liu, G.; Zhao, J.; Gao, D. Degradation state recognition of piston pump based on ICEEMDAN and XGBoost. Appl. Sci. 2020, 10, 6593. [Google Scholar] [CrossRef]
  57. H2O.ai. H2O: Scalable Machine Learning Platform, Version 3.30.0.6; 2020. Available online: https://docs.h2o.ai/h2o/latest-stable/h2o-docs/faq/general.html#i-am-writing-an-academic-research-paper-and-i-would-like-to-cite-h2o-in-my-bibliography-how-should-i-do-that (accessed on 18 January 2012).
  58. Kluyver, T.; Ragan-Kelley, B.; Pérez, F.; Granger, B.E.; Bussonnier, M.; Frederic, J.; Kelley, K.; Hamrick, J.B.; Grout, J.; Corlay, S.; et al. Jupyter Notebooks-a Publishing Format for Reproducible Computational Workflows; JupyterLab: Amsterdam, The Netherlands, 2016; Volume 2016. [Google Scholar]
  59. H2O.ai. H2O: Python Interface for H2O, Python Package Version 3.30.0.6; 2020. Available online: https://docs.h2o.ai/h2o/latest-stable/h2o-docs/faq/general.html#i-am-writing-an-academic-research-paper-and-i-would-like-to-cite-h2o-in-my-bibliography-how-should-i-do-that (accessed on 18 January 2012).
  60. Rashmi, K.; Gilad-Bachrach, R. DART: Dropouts meet Multiple Additive Regression Trees. In Proceedings of the 18th International Conference on Artificial Intelligence and Statistics (AISTATS), San Diego, CA, USA, 9–12 May 2015; pp. 489–497. [Google Scholar]
Figure 1. General framework of the proposed study.
Figure 1. General framework of the proposed study.
Fractalfract 07 00218 g001
Figure 2. Distribution plot of Dataset A.
Figure 2. Distribution plot of Dataset A.
Fractalfract 07 00218 g002
Figure 3. Distribution plot of Dataset B.
Figure 3. Distribution plot of Dataset B.
Fractalfract 07 00218 g003
Figure 4. Distribution plot of Dataset C.
Figure 4. Distribution plot of Dataset C.
Fractalfract 07 00218 g004
Figure 5. Correlation matrix of Dataset A.
Figure 5. Correlation matrix of Dataset A.
Fractalfract 07 00218 g005
Figure 6. Correlation matrix of Dataset B.
Figure 6. Correlation matrix of Dataset B.
Fractalfract 07 00218 g006
Figure 7. Correlation matrix of Dataset C.
Figure 7. Correlation matrix of Dataset C.
Fractalfract 07 00218 g007
Figure 8. Scatter plots of Dataset A.
Figure 8. Scatter plots of Dataset A.
Fractalfract 07 00218 g008
Figure 9. Scatter plots of Dataset B.
Figure 9. Scatter plots of Dataset B.
Fractalfract 07 00218 g009
Figure 10. Scatter plots of Dataset C.
Figure 10. Scatter plots of Dataset C.
Fractalfract 07 00218 g010
Figure 11. Pair distribution plots of Dataset A.
Figure 11. Pair distribution plots of Dataset A.
Fractalfract 07 00218 g011
Figure 12. Pair distribution plots of Dataset B.
Figure 12. Pair distribution plots of Dataset B.
Fractalfract 07 00218 g012
Figure 13. Pair distribution plots of Dataset C.
Figure 13. Pair distribution plots of Dataset C.
Fractalfract 07 00218 g013
Figure 14. Density and probability plots of Dataset A.
Figure 14. Density and probability plots of Dataset A.
Fractalfract 07 00218 g014
Figure 15. Density and probability plots of Dataset B.
Figure 15. Density and probability plots of Dataset B.
Fractalfract 07 00218 g015
Figure 16. Density and probability plots of Dataset C.
Figure 16. Density and probability plots of Dataset C.
Fractalfract 07 00218 g016
Figure 17. XGBoost flow chart.
Figure 17. XGBoost flow chart.
Fractalfract 07 00218 g017
Figure 18. Performance plots of the datasets.
Figure 18. Performance plots of the datasets.
Fractalfract 07 00218 g018
Figure 19. Actual vs Predicted performance tracking plots of the datasets.
Figure 19. Actual vs Predicted performance tracking plots of the datasets.
Fractalfract 07 00218 g019
Table 1. Important statistical values of Dataset A.
Table 1. Important statistical values of Dataset A.
StatisticsVolt (V)Type (0–1)Concentration (%)Current (A)
Count8000800080008000
Mean0.40000.49933.20000.0230
Std0.23090.50002.56140.4425
Min0.00000.00000.0000−1.3920
25%0.20000.00001.0000−0.0001
50%0.40000.00003.00000.0001
75%0.60001.00005.00000.0004
Max0.80001.00007.00002.3280
Table 2. Important statistical values of Dataset B.
Table 2. Important statistical values of Dataset B.
StatisticsVolt (V)Type (0–1)Concentration (%)Current (A)
Count4800480048004800
Mean0.40000.499350.00000.0002
Std0.23090.500018.71020.0006
Min0.00000.000030.00000.0008
25%0.20000.000030.0000−0.0003
50%0.40000.000045.0000−0.0004
75%0.60001.000075.00000.3280
Max0.80001.000075.00000.0020
Table 3. Important statistical values of Dataset C.
Table 3. Important statistical values of Dataset C.
StatisticsVolt (V)Type (0–1)Concentration (%)Current (A)
Count10,00010,00010,00010,000
Mean−0.30000.499550.0000−0.0025
Std0.28860.50002.56130.0062
Min−0.80000.00000.0000−0.0163
25%−0.55000.00001.0000−0.0079
50%−0.30000.00003.0000−0.0007
75%−0.05001.00005.00000.0019
Max0.20001.00007.00000.0112
Table 4. Sorted correlation values of dataset features to the current response.
Table 4. Sorted correlation values of dataset features to the current response.
FeaturesDataset ADataset BDataset C
Type−0.404418−0.8911720.442978
Concentration−0.0041330.0414620.000176
Volt0.1464030.3313260.796888
Table 5. Performance results of supercapacitor datasets: training phase.
Table 5. Performance results of supercapacitor datasets: training phase.
MetricsDataset ADataset BDataset C
MSE 1.9783 × 10 5 4.1428 × 10 3 7.1261 × 10 8
RMSE0.0044 6.4364 × 10 5 0.0002
MAE0.0018 4.5503 × 10 5 0.0001
Training samples641838528037
Table 6. Performance results of supercapacitor datasets: testing phase.
Table 6. Performance results of supercapacitor datasets: testing phase.
MetricsDataset ADataset BDataset C
MSE 7.3254 × 10 5 4.6650 × 10 9 8.9026 × 10 8
RMSE0.0085 6.8300 × 10 5 0.0002
MAE0.0027 4.9252 × 10 5 0.0002
R-squared0.99960.98950.9977
Testing samples15829481963
Table 7. Performance results of the supercapacitor with just the Volt input for Dataset A.
Table 7. Performance results of the supercapacitor with just the Volt input for Dataset A.
MetricsTrainingTestingPrevious Testing
MSE0.18730.2033 7.3254 × 10 5
RMSE0.43280.45090.0085
MAE0.22630.23770.0027
R-squared-−0.030830.9996
Samples641815821582
Table 8. Overall performance comparison with the benchmark study.
Table 8. Overall performance comparison with the benchmark study.
MetricsRandom Forest Model [34]ANN Model [34]Our XGBoost Model
Dataset A
RMSE0.13080.00470.0085
MAE0.10430.04580.0027
Dataset B
RMSE0.14330.00870.0000
MAE0.09170.02270.0000
Dataset C
RMSE0.14810.00470.0000
MAE0.11590.03460.0002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kati, N.; Ucar, F. An Intelligent Model for Supercapacitors with a Graphene-Based Fractal Electrode to Investigate the Cyclic Voltammetry. Fractal Fract. 2023, 7, 218. https://doi.org/10.3390/fractalfract7030218

AMA Style

Kati N, Ucar F. An Intelligent Model for Supercapacitors with a Graphene-Based Fractal Electrode to Investigate the Cyclic Voltammetry. Fractal and Fractional. 2023; 7(3):218. https://doi.org/10.3390/fractalfract7030218

Chicago/Turabian Style

Kati, Nida, and Ferhat Ucar. 2023. "An Intelligent Model for Supercapacitors with a Graphene-Based Fractal Electrode to Investigate the Cyclic Voltammetry" Fractal and Fractional 7, no. 3: 218. https://doi.org/10.3390/fractalfract7030218

Article Metrics

Back to TopTop