Next Article in Journal
Revisiting Gradient Boosting-Based Approaches for Learning Imbalanced Data: A Case of Anomaly Detection on Power Grids
Previous Article in Journal
PCB Component Detection Using Computer Vision for Hardware Assurance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Breast and Lung Anticancer Peptides Classification Using N-Grams and Ensemble Learning Techniques

by
Ayad Rodhan Abbas
*,
Bashar Saadoon Mahdi
and
Osamah Younus Fadhil
Department of Computer Science, University of Technology-Iraq, Baghdad 10066, Iraq
*
Author to whom correspondence should be addressed.
Big Data Cogn. Comput. 2022, 6(2), 40; https://doi.org/10.3390/bdcc6020040
Submission received: 26 February 2022 / Revised: 29 March 2022 / Accepted: 11 April 2022 / Published: 12 April 2022
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
Anticancer peptides (ACPs) are short protein sequences; they perform functions like some hormones and enzymes inside the body. The role of any protein or peptide is related to its structure and the sequence of amino acids that make up it. There are 20 types of amino acids in humans, and each of them has a particular characteristic according to its chemical structure. Current machine and deep learning models have been used to classify ACPs problems. However, these models have neglected Amino Acid Repeats (AARs) that play an essential role in the function and structure of peptides. Therefore, in this paper, ACPs offer a promising route for novel anticancer peptides by extracting AARs based on N-Grams and k-mers using two peptides’ datasets. These datasets pointed to breast and lung cancer cells assembled and curated manually from the Cancer Peptide and Protein Database (CancerPPD). Every dataset consists of a sequence of peptides and their synthesis and anticancer activity on breast and lung cancer cell lines. Five different feature selection methods were used in this paper to improve classification performance and reduce the experimental costs. After that, ACPs were classified using four classifiers, namely AdaBoost, Random Forest Tree (RFT), Multi-class Support Vector Machine (SVM), and Multi-Layer Perceptron (MLP). These classifiers were evaluated by applying five well-known evaluation metrics. Experimental results showed that the breast and lung ACPs classification process provided an accurate performance that reached 89.25% and 92.56%, respectively. In terms of AUC, it reached 95.35% and 96.92% for both breast and lung ACPs, respectively. The proposed classifiers performed competently somewhat equally in AUC, accuracy, precision, F-measures, and recall, except for Multi-class SVM-based feature selection, which showed superior performance. As a result, this paper significantly improved the predictive performance that can effectively distinguish ACPs as virtual inactive, experimental inactive, moderately active, and very active.

1. Introduction

Cancer or malignant tumors are common terms for many diseases that can affect any part of the human body. Cancer has one discriminant feature: the rapid growth and production of abnormal cells that grow beyond their actual boundaries, infecting adjoining parts of the human body and spreading to other parts; this process is called metastasis. Metastases are the primary cause of millions of cancer deaths every year, especially breast and lung cancer [1]. Conventional cancer treatment procedures, such as radiotherapy and chemotherapy, are costly and frequently have harmful side effects on normal cells. Furthermore, cancer cells can develop resistance to currently available anticancer chemotherapeutic medicines [2]. As a result, developing and anticipating innovative therapies with specific mechanisms is critical. ACPs are special molecules compared to the real chemotherapy arsenal accessible to treat cancer. They show a spectrum of action patterns co-existing in some cancers [3]. Therefore, they represent an alternative to conventional chemotherapy. In addition, ACPs’ activity depends on their amino acids’ type, number, and structure. They effectively destroy the cancer cell structure, thereby inhibiting the proliferation and growth of cancer cells and inducing apoptosis throughout the electrostatic interaction with the cancer cell membrane according to its structure, ACPs can be classified into four types: α-helical, β-pleated sheets, random coil, and cyclic. ACPs’ design represents a challenge to researchers since they are expected to be selective to tumor cells without influencing normal body functions [4].
Traditionally, peptides can be divided into two types: oligopeptides and polypeptides, where oligopeptides have few amino acids, from 7 to 30 (e.g., A, R, N, D, C, Q, E, G, H, I, L, K, M, F, P, S, T, W, Y, and V), and includes dipeptides, tripeptides, and tetrapeptides. In contrast, polypeptides have many amino acids with more than 20 residues [5]. ACPs are divided into two categories: those toxic to both cancerous and normal cells with little indication of selectivity, and those toxic to cancer cells but not to normal mammalian cells or erythrocytes [6].
It is difficult to identify distinct ACPs quickly and effectively to understand their anticancer mechanisms better and develop new anticancer drugs. Many experimental methods for identifying and developing novel ACPs have been developed. However, they are usually time-consuming, expensive, and difficult to perform at high throughput. As a result, the enormous therapeutic importance of ACPs requires the development of highly effective prediction algorithms [7,8].
Recently, many different experimental methods have been proposed to classify APCs using either machine or deep learning.
However, there are certain limitations associated with the previous methods. The main limitation includes the type and quantity of dataset selection. In other words, previous methods focused on the binary classification of ACPs only (ACPs and non-ACPs), regardless of the ACPs type (e.g., cancer or lung). This creates a difficult situation for the experimental scientist in testing a particular type and classification of ACPs. Another important drawback of the previous methods is that AARs, which play a fundamental role in the function and structure of the peptides, are neglected, and this causes an inability to distinguish between peptides having similar structures but different activities. The issues above motivate us to propose a method that addresses the above-mentioned limitations. Therefore, this paper combines Natural Language Processing (NLP) and ensemble learning algorithms to classify breast and lung ACPs. Furthermore, we used two datasets (breast and lung cancer datasets) to perform a statistical evaluation and improve the robustness and reliability of the average performance. This paper has three contributions: first, different sizes of N-gram and k-length substrings (k-mers) were used to analyze the sequence of amino acids, specially AARs. Second, the proposed ensemble learning model improved the average prediction performance compared to that of any single machine learning model. Third, feature selection enabled ensemble learning to train faster by reducing the complexity of a model.
The rest of this paper is organized as follows. In Section 2, related works are discussed. In Section 3, the breast and lung databases are presented. In Section 3, feature extraction for peptides is illustrated. In Section 4, five feature selection methods are introduced. Section 5 provides the details of four classifiers. Section 6 demonstrates all experimental results. The last section presents the conclusions.

2. Related Work

This section highlights the works on ACPs classification using well-known techniques, either traditional machine or deep learning techniques. All works only classify ACPs into two labels: ACPs (positive) and non-ACPs (negative), using different datasets, as shown in Table 1.
Machine learning techniques require hand-crafted feature extraction to represent peptide sequences for classification tasks. These features are divided into five groups: binary profiles, composition-based features, structure-based features, patterns, and evolutionary information. In contrast, the composition-based features are amino acid index, amino acid composition, dipeptide composition, and physicochemical-based features [9].
Recently, some methods in the classification of ACPs have been developed. For example, iACP describes a sequence-based predictor built by optimizing the g-gap dipeptide components. Rigorous cross-validations have proved that the new predictor has significantly outperformed existing predictors for accuracy and overall stability. Furthermore, it uses the main concept of SVM to build a separating hyper-plane that maximizes the difference between the positive and negative datasets [10].
MLACP constructed SVM and RFT algorithms for predicting ACPs from amino acid sequence parameters using the Hajisharifi dataset [11]. SAP used only 400-dimension features with g-gap dipeptide features that were collected from the Hajisharifi dataset, and then the irrelevant features were removed using the features selection method. When SVM or LibD3C are used, the 400-dimension features perform better than the RFT model [12]. mACPpred applied the feature selection approach on seven feature encodings collected from the independent dataset. These important features are then fed into an SVM classifier to develop the prediction model [13]. ACPs are used to extract a 19-dimensional feature model with lower dimensions and better performance than those of specific existing approaches. The selected features are then fed into three classifiers, SVM, RFT, and LibD3C, using the Hajisharifi dataset [14]. In cACP-2LFS, three different nature encoding approaches are used to extract essential features from peptide sequences. However, K-space amino acid pair (KSAAP) is applied to extract highly linked and valuable descriptors. In addition, a unique two-level feature selection (2LFS) technique is used to select important features and reduce the dimensionality of the proposed descriptors [15]. AntiCP 2.0 is an improved version of AntiCP, which was created to predict anticancer peptides using various input features accurately and implement the Extra Tree classifier model on two datasets: main and alternate datasets [16].
Unfortunately, these exciting techniques rely on one-experience-based, hand-crafted features, which have two limitations; The ability to represent models to a certain extent and the lack of adaptability to a different dataset, which limits the improvement of predictive performance and affects the robustness of predictive models [17]. Therefore, many works used deep learning for extracting and combining spatial features from various datasets. For example, ACP-DL constructed a deep learning long short-term memory (LSTM) neural network model to effectively predict novel ACPs using a feature representation method incorporating binary profile features and a k-mer sparse matrix [18].
Another method, xDeep-AcPEP, identifies effective ACPs in rational peptide design for therapeutic purposes based on CNN to predict biological activity (EC50, LC50, IC50, and LD50) against six tumor cells, including those of the breast, colon, cervix, lung, skin, and prostate [19]. Finally, three methods, like ACP-MHCNN, DLFF-ACP, and ACPNet, extract and combine discriminative features from different information sources using CNN to distinguish ACPs. These methods employ three different types of peptide sequence information, peptide physicochemical properties, and auto-encoding features linking the training process [20,21,22].
In the context of ACPs classification, deep learning is not automatically suitable for small data situations to make accurate decisions. It might not be the best solution for not large-scale datasets.
Here, we used the biological activity data of two famous tissue types (breast and lung) from CancerPPD to train and test using the ensemble learning method to improve the classification task. In addition, we distinguished breast and lung ACPs as virtual inactive, experimental inactive, moderately active, and very active, rather than two binary classes (ACPs and non-ACPs). Consequently, this paper significantly improves the predictive performance compared with that of other state-of-the-art methods.

3. Datasets

This paper used two datasets of peptides pointed to breast and lung cancer. These peptides’ datasets were assembled and curated manually from the Cancer Protein and Peptides Database (CancerPPD) [23] by authors in [24] to predicate biological activities such as LD50 (Lethal Dose 50%), LC50 (Lethal Concentration 50%), EC50 (Median Effective Concentration), and IC50 (Inhibitory Concentration 50%) against lung and breast cancer cells.
Experimentally, peptides’ activity of anticancer was tested; only 53 peptides in the CancerPPD database possess low micromolar activity (EC50, IC50, or LC50 < 5 μM). Therefore, the authors in [24] split this anticancer activity of peptides into four classes according to μM: virtual inactive, experimental inactive, moderately active, and very active, as shown in Table 2; In other words, these classes represent the activities against breast and lung cancer cells. The breast cancer dataset contains 949 peptides, while the lung cancer dataset contains 901 peptides.
In the case of ties, the less active class was chosen (>50 μM). Because the CancerPPD is biased towards active peptide annotation, a set of virtual inactive peptides have been built by randomly collecting and extracting 750 alpha-helical peptides from crystalline structures recorded in the Protein Data Bank (7–30 amino acids) [24]. As a result, the total number of inactive peptides (virtual and experimental inactive) for breast and lung cancers was 833 and 803, respectively. However, this created a state of imbalance in the dataset. To overcome this problem, we used five well-known evaluation metrics: AUC, accuracy, precision, F-measures, and recall.

4. Features Extraction for Peptides

Amino acid repeats are abundant in every peptide sequence because these repeats reflect the inherent biological properties of the peptide [25]. In this paper, we applied an effective approach, namely k-mers, to extract these biological properties of the peptide. k-mers are primarily used with nucleotides (i.e., A, T, G, and C) in computational genomics [26], but we used k-mers with 20 amino acids. The primary approach is to break down the peptide into four different types of k-mers (k = 1, 2, 3, and 4) that represent the following amino acid profiles:
  • A monopeptide has a single amino acid (k = 1).
  • A dipeptide has two amino acids (k = 2).
  • A tripeptide has three amino acids (k = 3).
  • A tetrapeptide has four amino acids (k = 4).
The permutation index of k-mers can be represented as peptide function P k . P stands for a peptide sequence of 20 amino acids. For example, if we convert a peptide sequence (FAKALAKLAKKLL) into a single character k = 1 (one amino acid), F is one of the 20 possible elements of monopeptide. Similarly, FA has 400 possible elements of dipeptides; FAK has 8000 possible elements of tripeptides. Finally, FAKA has 160,000 possible elements of tetrapeptides.
In peptides classification, N-Grams is a consecutive subsequence of the primary structure of a peptide sequence of a length n, extensively used in NLP tasks. Where n is also the number of amino acids that are extracted from each candidate peptide sequence.
Each sequence from the ACPs dataset is divided into an overlapping subsequence of 2-mers, 3-mers, and 4-mers to extract encoded N-Grams. After that, from each of these k-mers, we extract N-Grams of sizes 2, 3, and 4, as shown in Figure 1. In this paper, unigrams (n = 1) are skipped because they do not reflect any biological representations, thus, we skipped n = 1 when (k = 1).
The weka word2vec tool was utilized to compute vector representations of a peptide, then, the number of features extraction depended on the n values using the following formula:
No .   of   features = k = 1 n p k
For example, if we consider n = 3 (3-Grams), then the total number of peptides features of k = 1, …, 3, will be:
No .   of   featuers = p 1 + p 2 + p 3

5. Proposed Feature Selection Methods

The major problem of the N-Grams model is that it generates features space with high dimensions and extreme spacing. In other words, most of these features have zero or empty values, and others give a number of occurrences of a given subsequence in the peptide. Furthermore, many relevant features have a high correlation degree with the class and low correlation with other features. Thus, these features provide essential information about the functions and structure of the peptide and positively affect the performance of the classifier and running time. In contrast, irrelevant or redundant features are removed from the original features vector. Therefore, in this paper, five features selection methods were successfully applied to make a comparative study of these methods and investigate the best performance of the proposed model.
These methods are Information gain [27], Gini index [28], Chi-square (X2) [29], Relief [30], and Correlation-Based Feature Selection (CBFS) [31].

6. Classifiers

The architecture of the proposed model consists of four stages: feature extraction, feature selection, classification, and voting process. The framework of the model is shown in Figure 2.
In the first stage, we extracted important features using k-mers and encoded N-Grams from each peptide sequence. Experimentally, to choose the best value for k in a k-mers model, it is necessary to find the suitable trade-off between the accuracy and the execution time. Therefore, one possible approach is to increase the value of k by one from k = 2 to k = 4. At the same time, k = 1 or n = 1 is skipped because they do not reflect any biological representations. In contrast, we chose k = 4 since it is a common choice with large training corpora and it gives higher accuracy than other k values do, whereas a k = 2 is often used with smaller ones. Subsequently, the optimal k or n values can be considered a set leading to the highest prediction accuracy.
In the second stage, five features selection methods were successfully applied to investigate the best performance of the proposed model. Thirty-five highest ranked features out of 8420 features of the 3-Grams profile were selected using the breast ACPs dataset. Thirty-one highest ranked features out of 168,420 features of the 4-Grams profile were selected using the lung ACPs dataset. In this paper, computing time was not an issue. To get high accuracy, we considered a loop in which features were gradually (one by one) added to the prediction model depending on their importance weight (highest ranked). As a result, the best feature set could be defined as one that led to the best prediction accuracy.
In the third stage, four well-known classifiers were applied, namely, AdaBoost [32], RFT [33], Multi-class SVM [34], and MLP [35], using orange data mining software [36]. In this paper, the prediction model, 5-fold cross-validation, was used with a training size of 66% and a testing size of 33%. All experiments were run on a computer with an Intel(R) Core (TM) i5-6300U CPU 2.50 GHz using 8 GB of RAM, running Windows 10 Pro. The key parameters of each classifier are shown in Table 3.
Finally, these classifiers were combined using the voting process model to enhance the proposed model’s performance by choosing the best classifier.

7. Results and Discussion

7.1. Performance Comparison with Different Amino Acid Profiles

For comparison with different N-Grams (2-Grams, 3-Grams, and 4-Grams), the average of the classification results of breast and lung ACPs using four classifiers are recorded in Figure 3 and Figure 4, respectively. In addition, these classifiers were adopted to execute five-fold cross-validation without using feature selection methods. As we can see in Figure 3, the 3-Grams breast ACPs classification achieved the best performance, attaining an average AUC, Accuracy, F1-measures, Precision, and Recall of 94.08%, 89.28%, 88.60%, 88.19%, and 89.28, respectively. However, 2-Grams and 4-Grams had good classification results. In contrast, the experimental metrics in Figure 4 show that the 4-Grams lung ACPs classification achieved the best performance in AUC; it was 94.94%. At the same time, 2-Grams lung ACPs classification achieved the best performance and average Accuracy, F1-measures, Precision, and Recall of 91.51%, 91.18%, 91.35%, and 91.51, respectively. However, 3-Grams had excellent classification results.

7.2. Performance Comparison with Different Features Selection Methods References

Selecting the AARs that are extracted from each candidate peptide sequence is extremely important. According to the validity of all previous experiments that directly impacted the classification process, we used the optimal values of 3-Grams and 4-Grams because they gave high AUC, Accuracy, Precision, F-measures, and Recall as they were used in large training corpora. After that, we applied five feature selection methods that used only 3-Grams and 4-Grams to facilitate comparison with these methods. Figure 5 shows the breast ACPs classification performance by selecting 35 highest ranked features out of 8420 features of the 3-Grams profile.
Similarly, Figure 6 shows the lung anticancer peptides classification performance by selecting 31 highest ranked features out of 168,420 features of the 4-Grams profile.
All experimental results in Figure 5 show that the AUC, Accuracy, F1-measures, Precision, and Recall of the feature selection-based breast ACPs classification were about 0.29%, 1.09%, 1.4%, 1.93%, and 1.09% higher than those of the breast ACPs classification without feature selection. Similarly, all experimental results in Figure 6 show that the AUC, Accuracy, F1-measures, Precision, and Recall of the feature selection-based lung ACPs classification were about 0.74%, 0.75%, 0.57%, 0.29%, and 0.75% higher than those of the lung ACPs classification without feature selection. As a result, the average results of four classifiers on the five-fold cross-validation showed that Information gain, Gini index, X2, Relief, and CBFS were relatively stable, and CBFS had the best overall effect. Thus, optimal feature selection reached the level of best performance with 35 and 31 features using the breast and lung cancer dataset.

7.3. Performance Comparison with Multiple Classifiers

Because of the validity of all previous experiments that directly impacted the classification process, we chose only CBFS to classify breast and lung breast ACPs using four classifiers (AdaBoost, RFT, Multi-class SVM, and MLP). First, monopeptide, dipeptide, and tripeptide acids were extracted from the breast ACPs dataset, and we selected the 35 highest ranked features based on CBFS, as shown in Figure 7. Second, monopeptide, dipeptide, tripeptide, and tetrapeptides acids were extracted from the lung ACPs dataset, and we selected the 31 highest ranked features based on CBFS shown in Figure 8. Conventionally, these extracted features were fed to four different classifiers. In the prediction models, five-fold cross-validation was used with a training size of 66% and testing size of 33%. Multi-class SVM showed superior performance on breast ACPs dataset than other classifiers did in terms of AUC, Accuracy, F1-measures, Precision, and Recall with 95.45%, 88.62%, 87.67%, 87.96%, and 88.62, respectively. In contrast, Multi-class SVM also showed superior performance on lung ACPs compared to that of other classifier datasets in terms AUC, Accuracy, F1-measures, Precision, and Recall with 96.92%, 92.56%, 92.12%, 92.60%, and 92.65, respectively. Overall, the above analysis indicates that all four classifiers’ performances were similar regardless of the datasets.

7.4. Performance Comparison with State of the Art

This section compares the proposed method with MLACP, cACP-2LFS, xDeep-AcPEP, and DLFF-ACP as state of the art [11,15,19,21], as shown in Table 4. It was not easy to directly compare the previous methods for two reasons. First, all mentioned methods in Section 2 did not consider any cancer types (e.g., breast or lung cancer). The other reason was that the purposes of these methods were almost identical, focusing only on the classification of ACPs into binary classes (non-ACP and ACP). However, we compared previous works, which used the same database (CancerPPD). In terms of accuracy, the experimental results showed as follows:
  • The proposed breast ACPs method outperformed all the mentioned methods, except for cACP-2LFS, because a minimal dataset was used (150 non-ACP 150 ACP) with 20 features only.
  • The proposed lung ACPs method outperformed all mentioned methods.
In this paper, we used the biological activity data on two famous tissue types (breast and lung) and distinguished breast and lung ACPs as virtual inactive, experimental inactive, moderately active, and very active, rather than two binary classes (ACPs and non-ACPs).
As a result, the classification outcomes reveal that our proposed method achieved improved performance accuracy compared to that of the state of the art so far.

8. Conclusions

This paper proposes a new model that combines NLP and ensemble learning algorithms to efficiently and accurately classify breast and lung ACPs. Specifically, we offer a novel feature extraction method by extracting AARs using N-Grams. Then, five feature selection methods were used to improve classification performance and reduce the experimental costs. Finally, we utilized four classifiers, AdaBoost, RFT, Multi-class SVM, and MLP, and applied them on two datasets. Experimental results reveal that (1) The 3-Grams profile (monopeptide, dipeptide, and tripeptide acids) of the breast ACPs classification achieved the best performance relative to that of other profiles. (2) The 4-Grams profile (monopeptide, dipeptide, tripeptide, and tetrapeptides acids) of the lung ACPs classification achieved the best performance compared to that of other profiles in terms of AUC. At the same time, 2-Grams also achieved the best performance but with other classification matrices. (3) All feature selection methods like Information gain, Gini index, X2, Relief, and CBFS were relatively stable, and CBFS had the best overall effect in breast and lung cancer. (4) The performances of the four classifiers were similar regardless of the datasets, but Multi-class SVM-based CBFS showed superior performance in the breast and lung classification. As a result, the impact on classification performance depended on four factors: the size of N-Grams, feature selection methods, the number of selected features, and machine learning algorithms. Thus, this paper significantly improves the predictive performance that can effectively distinguish ACPs as virtual inactive, experimental inactive, moderately active, and very active.

Author Contributions

Conceptualization, A.R.A.; methodology, A.R.A.; software, A.R.A.; validation, A.R.A., B.S.M. and O.Y.F.; formal analysis, A.R.A. and B.S.M.; investigation, A.R.A.; resources, A.R.A.; data curation, A.R.A.; writing original draft preparation, A.R.A.; writing review and editing, A.R.A., B.S.M. and O.Y.F.; visualization, A.R.A., B.S.M. and O.Y.F.; supervision, A.R.A.; project administration, A.R.A.; funding acquisition, A.R.A., B.S.M. and O.Y.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data supporting reported results can be found at: http://crdd.osdd.net/raghava/cancerppd/ (accessed on 1 January 2021).

Conflicts of Interest

The authors declare no conflict of interests.

References

  1. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef] [PubMed]
  2. Torre, L.A.; Bray, F.; Siegel, R.; Ferlay, J.; Lortet-Tieulent, J.; Jemal, A. Global cancer statistics, 2012. CA Cancer J. Clin. 2015, 65, 87–108. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Gaspar, D.; Veiga, A.S.; Castanho, M.A.R.B. From antimicrobial to anticancer peptides. A review. Front. Microbiol. 2013, 4, 294. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Chiangjong, W.; Chutipongtanate, S.; Hongeng, S. Anticancer peptide: Physicochemical property, functional aspect and trend in clinical application. Int. J. Oncol. 2020, 57, 678–696. [Google Scholar] [CrossRef] [PubMed]
  5. Jakubczyk, A.; Karaś, M.; Rybczyńska-Tkaczyk, K.; Zielińska, E.; Zieliński, D. Current trends of bioactive peptides—New sources and therapeutic effect. Foods 2020, 9, 864. [Google Scholar] [CrossRef]
  6. Harris, F.; Dennison, S.R.; Singh, J.; Phoenix, D.A. On the selectivity and efficacy of defense peptides with respect to cancer cells. Med. Res. Rev. 2013, 33, 190–234. [Google Scholar] [CrossRef]
  7. Basith, S.; Manavalan, B.; Shin, T.H.; Lee, D.Y.; Lee, G. Evolution of machine learning algorithms in the prediction and design of anticancer peptides. Curr. Protein Pept. Sci. 2020, 21, 1242–1250. [Google Scholar] [CrossRef]
  8. Wei, L.; Zhou, C.; Chen, H.; Song, J.; Su, R. ACPred-FL: A sequence-based predictor using effective feature representation to improve the prediction of anticancer peptides. Bioinformatics 2018, 34, 4007–4016. [Google Scholar] [CrossRef]
  9. Pande, A.; Patiyal, S.; Lathwal, A.; Arora, C.; Kaur, D.; Dhall, A.; Mishra, G.; Kaur, H.; Sharma, N.; Jain, S. Computing wide range of protein/peptide features from their sequence and structure. BioRxiv 2019, 599126. [Google Scholar] [CrossRef]
  10. Chen, W.; Ding, H.; Feng, P.; Lin, H.; Chou, K.-C. iACP: A sequence-based tool for identifying anticancer peptides. Oncotarget 2016, 7, 16895–16909. [Google Scholar] [CrossRef] [Green Version]
  11. Manavalan, B.; Basith, S.; Shin, T.H.; Choi, S.; Kim, M.O.; Lee, G. MLACP: Machine-learning-based prediction of anticancer peptides. Oncotarget 2017, 8, 77121. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Xu, L.; Liang, G.; Wang, L.; Liao, C. A novel hybrid sequence-based model for identifying anticancer peptides. Genes 2018, 9, 158. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Boopathi, V.; Subramaniyam, S.; Malik, A.; Lee, G.; Manavalan, B.; Yang, D.-C. mACPpred: A support vector machine-based meta-predictor for identification of anticancer peptides. Int. J. Mol. Sci. 2019, 20, 1964. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Li, Q.; Zhou, W.; Wang, D.; Wang, S.; Li, Q. Prediction of anticancer peptides using a low-dimensional feature model. Front. Bioeng. Biotechnol. 2020, 8, 892. [Google Scholar] [CrossRef]
  15. Akbar, S.; Hayat, M.; Tahir, M.; Chong, K.T. cACP-2LFS: Classification of anticancer peptides using sequential discriminative model of KSAAP and two-level feature selection approach. IEEE Access 2020, 8, 131939–131948. [Google Scholar] [CrossRef]
  16. Agrawal, P.; Bhagat, D.; Mahalwal, M.; Sharma, N.; Raghava, G.P.S. AntiCP 2.0: An updated model for predicting anticancer peptides. Brief. Bioinform. 2021, 22, bbaa153. [Google Scholar] [CrossRef] [PubMed]
  17. He, W.; Wang, Y.; Cui, L.; Su, R.; Wei, L. Learning embedding features based on multisense-scaled attention architecture to improve the predictive performance of anticancer peptides. Bioinformatics 2021, 37, 4684–4693. [Google Scholar] [CrossRef]
  18. Yi, H.-C.; You, Z.-H.; Zhou, X.; Cheng, L.; Li, X.; Jiang, T.-H.; Chen, Z.-H. ACP-DL: A deep learning long short-term memory model to predict anticancer peptides using high-efficiency feature representation. Mol. Ther.-Nucleic Acids 2019, 17, 1–9. [Google Scholar] [CrossRef] [Green Version]
  19. Chen, J.; Cheong, H.H.; Siu, S.W.I. xDeep-AcPEP: Deep Learning Method for Anticancer Peptide Activity Prediction Based on Convolutional Neural Network and Multitask Learning. J. Chem. Inf. Model. 2021, 61, 3789–3803. [Google Scholar] [CrossRef]
  20. Ahmed, S.; Muhammod, R.; Adilina, S.; Khan, Z.H.; Shatabda, S.; Dehzangi, A. ACP-MHCNN: An Accurate Multi-Headed Deep-Convolutional Neural Network to Predict Anticancer peptides. BioRxiv 2020, 11, 23676. [Google Scholar] [CrossRef]
  21. Cao, R.; Wang, M.; Bin, Y.; Zheng, C. DLFF-ACP: Prediction of ACPs based on deep learning and multi-view features fusion. PeerJ 2021, 9, 11906. [Google Scholar] [CrossRef] [PubMed]
  22. Sun, M.; Yang, S.; Hu, X.; Zhou, Y. ACPNet: A Deep Learning Network to Identify Anticancer Peptides by Hybrid Sequence Information. Molecules 2022, 27, 1544. [Google Scholar] [CrossRef] [PubMed]
  23. Tyagi, A.; Tuknait, A.; Anand, P.; Gupta, S.; Sharma, M.; Mathur, D.; Joshi, A.; Singh, S.; Gautam, A.; Raghava, G.P.S. CancerPPD: A database of anticancer peptides and proteins. Nucleic Acids Res. 2015, 43, D837–D843. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Grisoni, F.; Neuhaus, C.S.; Hishinuma, M.; Gabernet, G.; Hiss, J.A.; Kotera, M.; Schneider, G. De novo design of anticancer peptides by ensemble artificial neural networks. J. Mol. Model. 2019, 25, 112. [Google Scholar] [CrossRef]
  25. Luo, H.; Nijveen, H. Understanding and identifying amino acid repeats. Brief. Bioinform. 2014, 15, 582–591. [Google Scholar] [CrossRef] [Green Version]
  26. Zhang, J.; Guo, J.; Yu, X.; Yu, X.; Guo, W.; Zeng, T.; Chen, L. Mining k-mers of various lengths in biological sequences. In Bioinformatics Research and Applications, Proceedings of the 13th International Symposium, ISBRA 2017, Honolulu, HI, USA, 29 May–2 June 2017; Cai, Z., Daescu, O., Li, M., Eds.; Springer International Publishing: Berlin/Heidelberg, Germany, 2017. [Google Scholar]
  27. Azhagusundari, B.; Thanamani, A.S. Feature selection based on information gain. Int. J. Innov. Technol. Explor. Eng. 2013, 2, 18–21. [Google Scholar]
  28. Liu, H.; Zhou, M.; Lu, X.S.; Yao, C. Weighted Gini index feature selection method for imbalanced data. In Proceedings of the 2018 IEEE 15th International Conference on Networking, Sensing and Control (ICNSC), Zhuhai, China, 27–29 March 2018. [Google Scholar]
  29. Zhai, Y.; Song, W.; Liu, X.; Liu, L.; Zhao, X. A chi-square statistics based feature selection method in text classification. In Proceedings of the 2018 IEEE 9th International Conference on Software Engineering and Service Science (ICSESS), Beijing, China, 23–25 November 2018. [Google Scholar]
  30. Rosario, S.F.; Thangadurai, K. RELIEF: Feature selection approach. Int. J. Innov. Res. Dev. 2015, 4, 218–224. [Google Scholar]
  31. Hall, M.A. Correlation-based feature selection of discrete and numeric class machine learning. In Proceedings of the Seventeenth International Conference on Machine Learning, Stanford, CA, USA, 29 June–2 July 2000; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 2000. [Google Scholar]
  32. An, T.K.; Kim, M.H. A new diverse AdaBoost classifier. In Proceedings of the 2010 International Conference on Artificial Intelligence and Computational Intelligence, Sanya, China, 23–24 October 2010. [Google Scholar]
  33. Liaw, A.; Wiener, M. Classification and regression by random Forest. R News 2002, 2, 18–22. [Google Scholar]
  34. Chamasemani, F.F.; Singh, Y.P. Multi-class support vector machine (SVM) classifiers—An application in hypothyroid detection and classification. In Proceedings of the 2011 Sixth International Conference on Bio-Inspired Computing: Theories and Applications, Penang, Malaysia, 27–29 September 2011. [Google Scholar]
  35. Taud, H.; Mas, J.F. Multilayer perceptron (MLP). In Geomatic Approaches for Modeling Land Change Scenarios, 1st ed.; Camacho Olmedo, M.T., Paege, M., Mas, J.F., Escobar, F., Eds.; Springer: Cham, Switzerland, 2018; pp. 451–455. [Google Scholar]
  36. Demšar, J.; Curk, T.; Erjavec, A.; Gorup, Č.; Hočevar, T.; Milutinovič, M.; Možina, M.; Polajnar, M.; Toplak, M.; Starič, A.; et al. Orange: Data Mining Toolbox in Python. J. Mach. Learn. Res. 2013, 14, 2349–2353. [Google Scholar]
Figure 1. Encoded N-Grams extraction using k-mers.
Figure 1. Encoded N-Grams extraction using k-mers.
Bdcc 06 00040 g001
Figure 2. The architecture of the proposed model.
Figure 2. The architecture of the proposed model.
Bdcc 06 00040 g002
Figure 3. Performance evaluation of the breast ACPs classification with different N-Grams.
Figure 3. Performance evaluation of the breast ACPs classification with different N-Grams.
Bdcc 06 00040 g003
Figure 4. Performance evaluation of the lung ACPs classification with different N-Grams.
Figure 4. Performance evaluation of the lung ACPs classification with different N-Grams.
Bdcc 06 00040 g004
Figure 5. Performance of five feature selection methods on the breast ACPs using only 35 features.
Figure 5. Performance of five feature selection methods on the breast ACPs using only 35 features.
Bdcc 06 00040 g005
Figure 6. Performance of five feature selection methods on the lung ACPs using only 31 features.
Figure 6. Performance of five feature selection methods on the lung ACPs using only 31 features.
Bdcc 06 00040 g006
Figure 7. Performance of four classifiers using the breast ACPs.
Figure 7. Performance of four classifiers using the breast ACPs.
Bdcc 06 00040 g007
Figure 8. Performance of four classifiers using the lung anticancer peptides.
Figure 8. Performance of four classifiers using the lung anticancer peptides.
Bdcc 06 00040 g008
Table 1. Summary of the related work.
Table 1. Summary of the related work.
References No.MethodDatasetNo. of Non-ACP vs. ACPClassifier
[10]iACPHajisharifi206 non-ACP 138 ACPSVM
[11]MLACPCancerPPD206 non-ACP 138 ACPSVM and RFT
[12]SAPHajisharifi206 non-ACP 138 ACPSVM
[13]mACPpredIndependent dataset157 non-ACP 157 ACPSVM
[14]Li, Qingwen, et al.Hajisharifi206 non-ACP 138 ACPSVM, RFT, and LibD3C
[15]cACP-2LFSCancerPPD150 non-ACP 150 ACP FKNN, SVM and RFT
[16]AntiCP 2.0Main dataset861 non-ACP 861 ACPExtra Tree
[18]ACP-DLACP740 and ACP240364/111 non-ACP 376/129 ACP LSTM
[19]xDeep-AcPEPCancerPPD65 non-ACP 85 ACPCNN
[20]ACP-MHCNNIndependent dataset364/111 non-ACP 376/129 ACPCNN
[21]DLFF-ACPCancerPPD65 non-ACP 85 ACPCNN
[22]ACPNetIndependent dataset364/111 non-ACP 376/129 ACPRNN
Table 2. Description of anticancer peptides dataset.
Table 2. Description of anticancer peptides dataset.
Cancer TypesClassEC50, IC50, or LC50No. of Peptides
breast cancervirtual inactive>50 μM750
breast cancerexperimental inactive>50 μM83
breast cancermoderately activeup to 50 μM98
breast cancervery active≤5 μM18
lung cancervirtual inactive>50 μM750
lung cancerexperimental inactive>50 μM52
lung cancermoderately activeup to 50 μM75
lung cancervery active≤5 μM24
Table 3. Key parameters of four classifiers.
Table 3. Key parameters of four classifiers.
ClassifierKey Parameters
AdaBoostBase estimator: Tree
Number of estimators: 50
Learning rate: 1
Classification algorithm: SAMME.R
Regression loss function: Linear
RFTNumber of trees:10
Number of attributes at each split:5
Limit depth of individual tree: 3
Don’t split subset smaller than: 5
Multi-class SVMsCost©: 1
Regression loss epsilon (ɛ): 0.10
Kernel: RBF
Numerical tolerance: 0.0010
Iteration limit: 100
Multi-Layer PerceptronNeurons in hidden layers: 100
Activation function: ReLu
Solver: Adam
Regularization: 0.0001
Maximum number of iterations: 200
Table 4. Performance comparison with the state of the art.
Table 4. Performance comparison with the state of the art.
MethodClassifierNo. of FeaturesAccuracyPredictions
MLACPSVM and RFT 20, 400, 5, and 11 88.7% 206 non-ACP 138 ACP
cACP-2LFS FKNN, SVM and RFT 20 93.72% 150 non-ACP 150 ACP
xDeep-AcPEPCNN-82.42%65 non-ACP 85 ACP
DLFF-ACPCNN-82%65 non-ACP 85 ACP
Proposed breast ACPsEnsemble learning3589.25%750 virtual inactive, 83 experimental inactive, 98 moderately active, and 18 very active
Proposed lung ACPsEnsemble learning3195.35%750 virtual inactive, 52 experimental inactive, 75 moderately active, and 24 very active
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Abbas, A.R.; Mahdi, B.S.; Fadhil, O.Y. Breast and Lung Anticancer Peptides Classification Using N-Grams and Ensemble Learning Techniques. Big Data Cogn. Comput. 2022, 6, 40. https://doi.org/10.3390/bdcc6020040

AMA Style

Abbas AR, Mahdi BS, Fadhil OY. Breast and Lung Anticancer Peptides Classification Using N-Grams and Ensemble Learning Techniques. Big Data and Cognitive Computing. 2022; 6(2):40. https://doi.org/10.3390/bdcc6020040

Chicago/Turabian Style

Abbas, Ayad Rodhan, Bashar Saadoon Mahdi, and Osamah Younus Fadhil. 2022. "Breast and Lung Anticancer Peptides Classification Using N-Grams and Ensemble Learning Techniques" Big Data and Cognitive Computing 6, no. 2: 40. https://doi.org/10.3390/bdcc6020040

Article Metrics

Back to TopTop