Next Article in Journal
Pedestrian Navigation Method Based on Machine Learning and Gait Feature Assistance
Next Article in Special Issue
Investigation of Machine Learning Approaches for Traumatic Brain Injury Classification via EEG Assessment in Mice
Previous Article in Journal
A Review of Thin-Film Magnetoelastic Materials for Magnetoelectric Applications
Previous Article in Special Issue
Machine Learning Techniques Applied to Dose Prediction in Computed Tomography Tests
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Class-Specific Feature Selection for Cancer Detection with Gene Expression Profile Data of Platelets

1
College of Electrical & Electronic Engineering, Wenzhou University, Wenzhou 325035, China
2
Department of Planning & Finance, Wenzhou University, Wenzhou 325035, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(5), 1528; https://doi.org/10.3390/s20051528
Submission received: 27 January 2020 / Revised: 4 March 2020 / Accepted: 7 March 2020 / Published: 10 March 2020
(This article belongs to the Special Issue Multimodal Data Fusion and Machine-Learning for Healthcare)

Abstract

:
A novel multi-classification method, which integrates the elastic net and probabilistic support vector machine, was proposed to solve this problem in cancer detection with gene expression profile data of platelets, whose problems mainly are a kind of multi-class classification problem with high dimension, small samples, and collinear data. The strategy of one-against-all (OVA) was employed to decompose the multi-classification problem into a series of binary classification problems. The elastic net was used to select class-specific features for the binary classification problems, and the probabilistic support vector machine was used to make the outputs of the binary classifiers with class-specific features comparable. Simulation data and gene expression profile data were intended to verify the effectiveness of the proposed method. Results indicate that the proposed method can automatically select class-specific features and obtain better performance of classification than that of the conventional multi-class classification methods, which are mainly based on global feature selection methods. This study indicates the proposed method is suitable for general multi-classification problems featured with high-dimension, small samples, and collinear data.

Graphical Abstract

1. Introduction

Cancer detection plays a vital role in improving the survival rate and living quality of patients. Methods of detecting cancer mainly include clinical symptom analysis, pathological analysis of cancer tissue, radiology, biochemical examination and molecular diagnosis, etc. At present, using the approach to detect the gene expression profile data has emerged as a potential method for detecting cancer. For example, it is reported that the gene expression profile data of platelets can be used to detect cancer [1,2]. Platelets are the primary cell type of blood and known for their hemostatic property. Platelets play a crucial role in the process of tumor growth and metastasis, thereby altering their gene expression profile in cancer patients [1,2]. It shows that the platelets may be used as an all-in-one bio-source for the diagnosis of different types of cancer and precancerous tumors. Using the gene expression profile data of platelets to detect cancer has the unique advantage of micro-invasive molecular diagnostics, overcoming the difficulty of obtaining tissues [3].
Cancers are complex diseases, and are related to many etiological factors, especially in the early stage of cancer development [4,5]. It is hard to observe the expression levels of several marker genes for effective cancer diagnosis. Therefore, classification methods via machine learning algorithms are used to detect cancer with the gene expression profile data [6,7]. However, the gene expression profile data encompasses the expression levels of many genes assayed in relatively few samples. This is also called the “large p , small n ” problem [8,9,10,11], which has been a difficulty in the field of machine learning. For the expression profile data with the “large p , small n ” problem, feature selection has a direct effect on the generalization and the training efficiency of the classifier. The conventional methods of feature selection, such as Relieff and support vector machine recursive feature elimination (SVM-RFE), apply either single-variable ranking or recursive variable selection to reduce the number of features that are finally substituted into the classifier [12,13,14,15,16,17,18,19]. As a regularized regression method with automatic feature selection, the elastic net [20,21] naturally mitigates the “large p, small n” problem. Besides, the gene expression profile data is collinear as multiple genes tend to function together, e.g., in the shared biological pathways, and thus the correlations in expression level among those genes can be high. As for collinear data, the elastic net is able to select groups of features, rather than individual features, which are highly correlated.
In the field of machine learning, researches on the feature selection method for the multi-classification problem is mostly based on global features. It is actually the same case for cancer detection with gene expression profile data. In addition to the global feature selection method, there is another way to select features for the multi-classification problem, which is the class-specific feature selection method. There are few studies investigating on solving multi-class classification problems from the angle of the class-specific feature selection method. The class-specific feature selection method is related to the one-vs-all (OVA) strategy [22], which is commonly used in the multi-class classification problem. With the OVA strategy, the multi-class classification problem is decomposed into multiple binary classification problems, whose features can be optimized individually, and the selected features are called the class-specific features. Therefore, the class-specific features for these obtained binary classification problems are different, and increases the optimization space of the classification algorithm relative to the one based on the global features. Besides, the multi-class classification algorithm proposed on the basis of class-specific features is more complex than that based on the global features, and that how to fuse the outputs of the binary classifiers based on the class-specific features to get the final classification result of the initial multi-classification problem is an ongoing concern.
In this study, it was focused on the problem of how to use the class-specific feature selection method to enhance the performance of the classification algorithm in cancer detection with gene expression profile data. In this work, it was presented a new multi-class classification algorithm, which integrates the elastic net and probabilistic support vector machine. The OVA strategy was used to decompose the multi-class classification problem of cancer detection with gene expression profile data into multiple binary classification problems. The elastic net was conducted to select class-specific features for the obtained binary classification problems, on consideration of the characteristics of the elastic net and gene expression profile data. Besides, the probabilistic support vector machine was employed to construct the binary classifiers with the selected class-specific features. The probabilistic support vector machine inherits advantages of support vector machine (SVM), such as excellent generalization ability for the problem with nonlinear, small sample and high dimension. The probabilistic support vector machine converts the decision output of SVM into the form of probability, which makes the outputs of the binary classifiers based on the class-specific features comparable. The simulation data and gene expression profile data were obtained to verify the effectiveness of the proposed method and the results were compared with the conventional methods, such as SVM classifier combined with global feature selection methods, including Relieff and SVM-RFE [23,24,25].

Statement of Societal Impact

The blood-based liquid biopsies belong to a noninvasive cancer detection technology receiving increasing attention nowadays. The blood plate is an excellent biomarker among various biomarker sources for blood-based liquid biopsies. The mRNA profiles of tumor-educated blood plates have been used to train classifiers to detect cancers. This paper proposed a new idea that embedding elastic net in classifiers for solving the problem of detecting cancer using tumor-educated platelets dataset. We adopt classifiers of SVM and probabilistic SVM to implement this idea, named as ESVM and EPSVM respectively.
The proposed classifiers are applied to detection cancer with the transcriptomic data derived from tumor-educated platelets from individuals. The experiment results show that the proposed methods have the advantage of automatically selecting the feature subsets to improve the generalization ability of classifiers comparing to the conventional multi-classification methods.
The rest of the context is organized as follows: the related theories are described in Section 2. Original contributions are introduced in Section 3. The experimental procedures are presented in Section 4. Section 5 covers the results and discussion, and Section 6 gives the conclusion.

2. Related theories

2.1. Support Vector Machine and Probabilistic Support Vector Machine

SVM [26,27,28] is an excellent machine learning algorithm based on the statistical learning theory. A separating hyperplane was constructed from a low-dimensional space by SVM classifier as Equation (1) shown:
w x + b = 0 .
In doing that, it seeks the support vectors to support the hyperplane, using training data D = { x i , y i } i = 1 n , where x i R p is a feature vector and y i { 1 , 1 } is its class label. It is a constrained minimization problem, where w and b can be optimized by the Lagrange approach. The class label is predicted as Equation (2) shown
y = sgn ( f ( x ) ) = sgn ( i = 1 n y i a i K ( x i , x ) + b ) ,
where a i is Lagrange multiplier and K ( x i , x ) is a kernel function. Using kernel tricks, SVM essentially transforms the linearly non-separable problems in the low-dimensional space into linearly separable ones in the high-dimensional space.
However, the standard SVM classifier does not provide such probabilistic output. To address this need, the output of the SVM classifier has been calibrated into a posterior probability. Here, a typical probabilistic SVM algorithm (PSVM) was introduced in briefly [29,30].
With the data D = { x i , y i } i = 1 n , signals of + 1 and N + denote the label and number of positive samples, and signals of 1 and N denote the label and number of negative samples respectively. The decision output of SVM is converted into the posterior probability by a sigmoid function as Equation (3) shown.
P ( y = 1 | x ) P A , B ( y = 1 | f ( x ) ) = 1 1 + exp ( A f ( x ) + B ) ,
where f ( x ) is the decision output of SVM classifier given in Equation (2). The parameters A and B are determined in Equation (4) by minimizing the negative log-likelihood problem
min   F ( A ,   B ) = i = 1 n t i log ( p i ) + ( 1 t i ) log ( 1 p i )
where,
p i = P A , B ( f ( x i ) ) ,   and   t i = { N + + 1 N + + 2 , if y i = 1 1 N + 2 , if y i = 1 , i = 1 , , n .

2.2. Feature Selection Methods

The feature selection methods can be divided into two categories: filter algorithms and wrap algorithms [12,13,14,31]. The filter algorithms are independent of the specific classification algorithm and directly use the statistical characteristics of data to select features. The advantage of filter algorithms is a high efficiency, but the shortcoming is not effective in removing redundant features. The wrap algorithms are based on classifiers and take the performance of classifiers as the evaluation index of selected features. The wrap algorithms have the advantage of high accuracy, but a drawback of large computation costs for training classifiers repeatedly.

2.2.1. Relieff Algorithm

Relieff [32,33] is a classic filter algorithm. Relieff randomly selects a sample R from the training set and then searches K nearest neighbors from the same class, called near hits and K nearest neighbors form the different classes, called near misses. For a feature, if R is closer to the near-hits than the near misses, then this feature is beneficial to distinguish R from the different classes and the weight of this feature should be increased. Conversely, the weight of this feature should be reduced. This process is repeated several times to get the average weight of this feature. According to the weights of all features, Relieff ranks the importance of all features from most to least significant.

2.2.2. SVM-RFE

SVM-RFE [34] is a representative of wrap algorithms. SVM-RFE constructs SVM classifiers through the training sets. The performance of the SVM classifier is related to the weight vector W . Therefore, SVM-RFE ranks the importance of features according to their impacts on the weight vector W . SVM-RFE assumes that all features are important in the initial run and then removes one feature with the lowest ranking coefficient by the way of iteration. After removing all features one by one, SVM-REF gives a ranking of the importance of all features from most to least significant.

2.2.3. Elastic Net

Let Y = [ y 1 , , y n ] T denote the response vector, and X = [ x 1 ; ; x n ] denote the feature matrix. The objective function of the elastic net is given by:
β ^ = ARG min { Y X β 2 + λ [ α | β | 1 + ( 1 α ) β 2 ] } ,
where λ > 0 controls the amount of shrinkage and α ( 0 , 1 ) controls the balance between the L1 penalty term and the L2 penalty term [20].
With α = 1 , Equation (6) corresponds to the least absolute shrinkage and selection operator (LASSO) [35,36], which uses the L1 norm as the penalty term. With the natural attribute of the L1 penalty term, LASSO can shrink some of the estimated coefficients exactly to zero. Therefore, we can automatically select the features with non-zero regression coefficients as informative ones. However, for the “large p , small n ” problem, LASSO would only select at most n features. Besides, LASSO tends to select only one feature among these features that are highly correlated. Therefore, for the gene expression profile data with problems of high dimension, small sample, and collinearity, LASSO is not an effective feature selection method. With α = 0 , Equation (6) corresponds to ridge regression [37], which uses the L2 norm as the penalty term. Ridge regression has the advantage in estimating coefficients for collinear data, but it cannot really shrink estimated coefficients to zero. The elastic net uses a hybrid penalty term of the L1 and L2 norms. It is known that the elastic net outperforms LASSO method in the highly correlated data with the “large p , small n ” problem.
Relieff and SVM-RFE are based on ranking methods, which produce a series of candidate nested feature subsets according to the threshold values. The elastic net is based on the sparse attribute of the L1 penalty term, which can form a set of candidate feature subset directly according to the values of λ . Therefore, the elastic net is more flexible than Relieff and SVM-RFE in creating the candidate feature subsets.

3. The Proposed Method of Embedding Elastic Net in PSVM

The multi-class classification problems can be decomposed into a series of binary classification problems by OVA strategy, where one class is treated as the positive class and the rest classes are treated as the negative class. These binary classification problems are solved with global features in general. However, for a special positive class, the global features may not be the best class-specific features. If the class-specific features of positive classes are similar, then using the global features are suitable for these binary classification problems. However, if the class-specific features of positive classes are very different from one another, then using class-specific features to solve these binary classification problems will achieve better results than using global features. The idea of solving the multi-class classification problems was demonstrated by using class-specific features by the OVA strategy in Figure 1.
As shown in Figure 1, the global features may not be the optimal features for the multi-class classification problem, and using the class-specific features of each binary classification problem obtained by the decomposition strategy would be easier to distinguish the positive class from the negative class. Therefore, if the effects of the binary classifiers can be improved by the class-specific features, then the classification result of the multi-class classification problem can be improved as well.
Using gene expression profile data for detecting cancers is a kind of multi-class classification and these data encounter the following problems: a large number of highly correlated features and limited data samples. Considering the characteristics of expression profile data and class-specific feature selection method, an idea of embedding the elastic net was proposed in the probabilistic SVM to improve the effect of cancer detection, and the working principle is discussed as follows. For the multi-classification problem, the OVA strategy was used to decompose the initial problem into multiple binary classification problems. For each binary classification problem, the elastic net is built using the class label as the dependent variable and features as independent variables. A series of λ are taken into the elastic net to generate coefficients with different degree of sparsity indicating candidate class-specific feature sets. Then these candidate class-specific feature sets are fed into binary classifiers to test their effectiveness and the one achieves the highest accuracy with cross-validation (CV) is chosen as the optimal class-specific feature set. In this way, multiple class-specific feature sets can be obtained for the multi-class classification problem.
Using the OVA strategy to deal with the multi-class classification problem, the predicted class is taken from the binary classifier with maximum confidence:
class = arg max i = 1 , , m f i
where f i is the estimated confidence measure of the i-th classifier. Using global features, if the SVM classifier is chosen, then the f i can be measured as the decision output of the SVM classifier. However, using the class-specific features, the decision outputs of the SVM classifiers based on different features are not comparable. In order to solve this problem, it is to choose the PSVM classifier, which changes the decision output of the SVM classifier into probability, and this algorithm was named as EPSVM. The probabilistic outputs of the PSVM classifiers normalize the decision outputs of the SVM classifiers, which make the decision outputs of SVM classifiers comparable regardless of whether they are fed with the same features or not. The flowchart of using class-specific features to solve the multi-class classification problem with elastic net and EPSVM is given in Figure 2.

4. Experiments

The simulation data and gene expression profile data were provided here, to verify the effect of proposed EPSVM. The SVM classifier combined with global feature selection methods, such as SVM-RFE and Relieff is used to evaluate the performance of EPSVM. The SVM and PSVM are taken from LIBSVM [38]. All calculations of these experiments were performed in MATLAB 2016.

4.1. Datasets

4.1.1. Simulation Data

The performance of the proposed algorithm is related to the number of classes. Therefore, the simulation dataset is designed to gradually increase the number of classes to continuously show the advantage of the proposed algorithm. The simulation dataset includes nine data subsets produced in the following way. The data i   ( 1 i 9 ) contains i + 2 classes and each class contains 30 samples. The features of data i are produced by two normal distributions, which are F 1 ~ N ( 2 , 1 ) and F 2 ~ N ( 0 , 1 ) . For the class l   ( 1 l i + 2 ) , the features from [ 1 + ( l 1 ) × 20 ] to [ l × 20 ] are produced by F 1 ~ N ( 2 , 1 ) and the rest features from [ 1 ] to [ 20 × ( i + 2 ) ] are produced by F 2 ~ N ( 0 , 1 ) . The features from [ 20 × ( i + 2 ) + 1 ] to [ 40 × ( i + 2 ) ] are the repeat of features from [ 1 ] to [ 20 × ( i + 2 ) ] . The feature dimension of data i is 40 × ( i + 2 ) . All the features are added with random noise F 3 ~ N ( 0 , 2 ) . Taking data 1 for example, its structure is shown in Table 1.

4.1.2. Platelets Data for Cancer Detection

The gene expression profile data of platelets was measured by RNA-Sequencing technology [39,40], which is a second-generation sequencing technology and has become an important tool in gene expression research. The platelet data was downloaded from the NCBI Gene Expression Omnibus (GEO) database with the accession number GSE68086. The downloaded platelet data has already been pre-processed and can be directly used for testing the performance of the classification algorithm. If readers are interested in the acquisition, actual measurement, and pretreatment methods of platelet data can read the Reference [41] The platelets data contains healthy donors (HD) and patients with various types of cancer including pancreatic cancer (PAAD), non-small cell lung carcinoma (NSCLC), hepatobiliary cancer (HBC), colorectal cancer (CRC), glioblastoma (GBM), and breast cancer (BrCa) with expression levels of 57,737 genes. The platelet data was divided into six data with different classes, and then for the obtained data, the ANOVA test was used to filter the irrelevant features (p > 0.001) [42,43]. Table 2 provides the details of the platelet data after processing. Besides, the average gene expression levels of each class in dataset 15 are shown in Figure 3, and it can be seen that the gene expression profile data is seriously overlapped.

4.1.3. Microarray Data for Cancer Detection

The microarray technology [44] is widely used in molecular biology research. Several microarray data have been used to analyze differential gene expression on the transcription level between normal cells and cancer cells [45]. The microarray data were downloaded from http://www.gems-system.org/ and their information is shown in Table 3.

4.2. Parameter Settings

For each data, two-thirds of the samples in each class are randomly chosen to form the training set and the rest samples are reserved for independent testing. In the training process, the parameters are set as follows. We adopt the Gaussian kernel for the classifiers of SVM and PSVM and use the grid search method to find the best parameter setting of ( c , g ) in the space { 2 10 , 2 9 , , 2 9 , 2 10 } × { 2 15 , 2 9 , , 2 9 , 10 15 } . The candidate class-specific feature subsets of EPSVM are evaluated by the PSVM classifier using 5-fold cross-validation. The number of nearest neighbors in Relieff is set to 5. The kernel type of SVM-RFE is set to the linear kernel. For the ranking features produced by the global feature selection method (SVM-RFE or Relief), the candidate feature subsets are evaluated by the SVM classifier using 5-fold cross-validation. The rest parameters are the defaults. In the testing process, considering the class imbalance problem of the data, parameter of not error rate (NER) is taken as an evaluation index, which is defined as
N E R i = S E i + S P i 2
N E R = i = 1 m N E R i m ,
where N E R i , S E i , and S P i are the not error rate, sensitivity [46] and specificity [47] of the i-th binary classification problem obtained by the OVA strategy and N E R is the not error rate of the multi-class classification problem. This process is intended to repeat 30 times and the average N E R is taken as the evaluation index.

5. Results and Discussion

5.1. Simulation Experiment

According to the generating method of simulation data, the features of the data i   ( 1 i 9 ) were divided into i + 2 regions and each region includes 20 features. The features in the same region are independent and the features that i + 2 regions apart are corrected. So the simulation data are collinear, high dimensional and small sample. The features that can separate class l   ( 1 l i + 2 ) in data i from the rest classes are located in two regions. The first region includes features from [ 1 + ( l 1 ) × 20 ] to [ l × 20 ] and the second region includes features from [ 1 + ( l + i + 1 ) × 20 ] to [ 1 + ( l + i + 2 ) × 20 ] and the features of these two regions are correlated.
In this work, it uses one experiment of data 1 to show the feature selection results of the three methods. Figure 4 shows that EPSVM selects accurately the class-specific features in the two regions for each class. Figure 5 shows that the selected global features by the feature methods of SVM-RFE and Relieff, and one can see that these selected features are located in all regions.
The classification results of the simulation data are shown in Table 4, which shows that the performance of the EPSVM classifier performs better than the other two methods as increasing the number of classes. This is because the class-specific features for the binary classification problems are different. Therefore, the global features are suboptimal for binary classification problems. The differences between the global feature and the class-specific features of the binary classification problem increased as the number of the class is increased. For the data with fewer classes, such as data 1 to 5, EPSVM is a little better than the other methods while for the data with more classes, such as data 6 to 9, EPSVM is much better than the other methods. In thinking further of the OVA strategy, one can find that only the features reflecting the differences between positive class and negative class are effective for the obtained binary classification problem. Therefore, the useless features that only reflect the differences within the negative class, are automatically removed by the class-specific feature selection method, which makes the feature more compact than the global features.

5.2. Cancer Detection with the Gene Expression Profile

The performance of the elastic net for the binary classification problem with gene expression profile data is a key factor for the success of the proposed method. We choose the dataset 10 (HD+BrCa) to show the effect of the elastic net comparing with the other two feature selection methods in detail. According to the ranking of p-values, we gradually increase the dimensionality of the feature dimension to test the performance of the three feature selection methods. Table 5 shows that the results produced by the three feature selection methods are comparable regarding improving the generalization performance of classifiers, but the elastic net is more stable than Relieff and SVM-RFE when the size of feature search space changes considerably.
Table 5 shows that the elastic net is well suitable for feature selection in binary classification problem of platelet data. We will further investigate the effect of the elastic net for feature selection in the multi-class classification problem of platelet data. In the multiple classification problems, the feature selection and classifier design are more complicated. The experiment results of the three methods for the platelet data with the different number of classes are shown in Table 6. We can see that for the platelet data with fewer classes, such as data 11 to 13, the performance of EPSVM is slightly better than Relieff+SVM and SVM-RFE+SVM. However, for the platelets data with more classes, such as data 14 to 15, EPSVM shows a relatively great advantage comparing to other methods. Therefore, the proposed method works well for platelets data.
We plot the first two principal component scores of dataset 11 with raw features and selected features in one experiment in Figure 6. One can see that the cancer classes (BrCa class and CRC class) are strongly overlapping and partially overlap with the HD class. Therefore, the key difficulty of platelet data relies on how to improve the classification result among the cancer classes. According to Figure 6, we can see that the class-specific features determined by EPSVM make the class separate from other classes more clearly than the global features, especially for the cancer classes.
The OVA strategy comes with an inherent flaw that the positive class and negative class are imbalanced [45,46]. Table 7 and Table 8 show the SN and SP of the three methods for dataset 10 and data 14 respectively. We can see that EPSVM shows an obvious advantage in SN when the data including more classes. For the data with more classes, the positive class and negative class are more unbalanced using the OVA strategy. It seems that the EPSVM using class-specific features is more resistant to the impact of class-imbalance than the traditional method using global features under the OVA strategy for the multi-class classification problem.
Besides, the penalty parameter α ( 0 , 1 ) of the elastic net determines how much weight should be given to either | β | 1 or β 2 . It is known that an elastic net with α closer to 1 performs similarly with the LASSO. Especially, the feature selection results of the proposed method for Data 10 in one experiment is shown in Figure 7. We can see that the proposed method with smaller α selects fewer features, but the differences between them are small compared to the feature dimension of the data. Therefore, according to Table 5 and Figure 7, we can find that the proposed method has good robustness towards to α .
We also use microarray data to further demonstrate the validity of the proposed method. Table 9 shows that EPSVM achieves the best performance in most microarray data. Especially for the microarray data including more classes, such as the data of Brain_Tumor1, Tumors_9, and Tumors11, EPSVM is significantly better than the other two methods.
In general, the gene expression profile data has the problems of high dimension, small sample and collinearity. With more classes, the classification problem becomes harder to be dealt with by the global features. The proposed EPSVM can achieve better effects of cancer detection with gene expression profile data than the traditional method based on global feature selection methods.

6. Conclusions

In this work, a novel multi-class classification method was proposed to detect cancer diseases with the gene expression profile data, which has the problem of high dimension, small sample, and collinearity. In the proposed method, the class-specific feature selection, OVA strategy and probabilistic support vector machine were innovatively combined to enhance the effect of cancer detection with gene expression profile data. The experiment results show that the proposed method can achieve better classification results than conventional methods based on global feature selection methods. So that the proposed method is suitable for the problem of cancer detection with gene expression profile data.
In general, the proposed method is suitable for multi-class classification problems, where the class-feature set of each class is significantly different from others. Besides, the idea behind the proposed method in this paper can be easily extended to the combination of other feature selection methods and classifiers, which is flexibly in solving the multi-feature and multi-class classification problems encountered in practice. However, for the adopted classifiers, how to normalize the outputs of classifiers with different feature subsets is a key problem. The following work will further study the effect of different normalization methods on the class-feature based multi-class classification method.

Author Contributions

Conceptualization, Y.S.; data curation, G.H. and L.-m.Y.; formal analysis, G.H. and Y.S.; funding acquisition, G.H. and L.-m.Y.; investigation, G.H.; methodology, G.H. and L.-m.Y.; project administration, L.-m.Y.; resources, Y.S. and L.-m.Y.; software, L.-m.Y.; writing—original draft, G.H.; writing—review and editing, Y.S. and L.-m.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by Zhejiang natural science foundation of China (LQ20F030059) and National Natural Science Foundation of China (NO.61705168), Wenzhou bureau science and technology project of China (NO.S20170003, NO.G20190024), and the Graduate Scientific Research Foundation of Wenzhou University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wan, J.C.; Massie, C.; Garcia-Corbacho, J.; Mouliere, F.; Brenton, J.D.; Caldas, C.; Pacey, S.; Baird, R.; Rosenfeld, N. Liquid biopsies come of age: Towards implementation of circulating tumour DNA. Nat. Rev. Cancer 2017, 17, 223–238. [Google Scholar] [CrossRef]
  2. Newman, A.M.; Lovejoy, A.F.; Klass, D.M.; Kurtz, D.M.; Chabon, J.J.; Scherer, F.; Stehr, H.; Liu, C.L.; Bratman, S.V.; Say, C. Integrated digital error suppression for improved detection of circulating tumor DNA. Nat. Biotechnol. 2016, 34, 547–555. [Google Scholar] [CrossRef] [PubMed]
  3. Hanash, S.M.; Baik, C.S.; Kallioniemi, O. Emerging molecular biomarkers—Blood-based strategies to detect and monitor cancer. Nat. Rev. Clin. Oncol. 2011, 8, 142–150. [Google Scholar] [CrossRef] [PubMed]
  4. Borghei, Y.-S.; Hosseini, M.; Dadmehr, M.; Hosseinkhani, S.; Ganjali, M.R.; Sheikhnejad, R. Visual detection of cancer cells by colorimetric aptasensor based on aggregation of gold nanoparticles induced by DNA hybridization. Anal. Chim. Acta 2016, 904, 92–97. [Google Scholar] [CrossRef] [PubMed]
  5. McCormack, V.; Menya, D.; Munishi, M.; Dzamalala, C.; Gasmelseed, N.; Leon Roux, M.; Assefa, M.; Osano, O.; Watts, M.; Mwasamwaja, A. Informing etiologic research priorities for squamous cell esophageal cancer in Africa: A review of setting-specific exposures to known and putative risk factors. Int. J. Cancer 2017, 140, 259–271. [Google Scholar] [CrossRef] [Green Version]
  6. Huang, G.; Yuan, M.; Chen, M.; Li, L.; You, W.; Li, H.; Cai, J.J.; Ji, G. Integrating multiple fitting regression and Bayes decision for cancer diagnosis with transcriptomic data from tumor-educated blood platelets. Analyst 2017, 142, 3588–3597. [Google Scholar] [CrossRef]
  7. Cestarelli, V.; Fiscon, G.; Felici, G.; Bertolazzi, P.; Weitschek, E. CAMUR: Knowledge extraction from RNA-seq cancer data through equivalent classification rules. Bioinformatics 2016, 32, 697–704. [Google Scholar] [CrossRef]
  8. Randall, L.; Sundrum, R. Large mass hierarchy from a small extra dimension. Phys. Rev. Lett. 1999, 83, 3370. [Google Scholar] [CrossRef] [Green Version]
  9. Ji, G.; Huang, G.; Yang, Z.; Wu, X.; Chen, X.; Yuan, M. Using consensus interval partial least square in near infrared spectra analysis. Chemom. Intell. Lab. Syst. 2015, 144, 56–62. [Google Scholar] [CrossRef]
  10. Chen, X.; Liu, K.; Cai, J.; Zhu, D.; Chen, H. Identification of heavy metal-contaminated Tegillarca granosa using infrared spectroscopy. Anal. Methods 2015, 7, 2172–2181. [Google Scholar] [CrossRef]
  11. Yuan, L.-M.; Mao, F.; Chen, X.; Li, L.; Huang, G. Non-invasive measurements of ‘Yunhe’pears by vis-NIRS technology coupled with deviation fusion modeling approach. Postharvest Biol. Technol. 2020, 160, 111067. [Google Scholar] [CrossRef]
  12. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  13. Lefakis, L.; Fleuret, F. Jointly informative feature selection made tractable by gaussian modeling. J. Mach. Learn. Res. 2016, 17, 1–39. [Google Scholar]
  14. Hulse, J.V.; Khoshgoftaar, T.M.; Napolitano, A.; Wald, R. Threshold-based feature selection techniques for high-dimensional bioinformatics data. Netw. Model. Anal. Health Inform. Bioinform. 2012, 1, 47–61. [Google Scholar] [CrossRef] [Green Version]
  15. Abeel, T.; Helleputte, T.; Van de Peer, Y.; Dupont, P.; Saeys, Y. Robust biomarker identification for cancer diagnosis with ensemble feature selection methods. Bioinformatics 2010, 26, 392–398. [Google Scholar] [CrossRef]
  16. Rakotomamonjy, A. Variable selection using SVM-based criteria. J. Mach. Learn. Res. 2003, 3, 1357–1370. [Google Scholar]
  17. Robnik-Šikonja, M.; Kononenko, I. Theoretical and empirical analysis of ReliefF and RReliefF. Mach. Learn. 2003, 53, 23–69. [Google Scholar] [CrossRef] [Green Version]
  18. Tang, L.J.; Jiang, J.H.; Wu, H.L.; Shen, G.L.; Yu, R.Q. Variable selection using probability density function similarity for support vector machine classification of high-dimensional microarray data. Talanta 2009, 79, 260–267. [Google Scholar] [CrossRef]
  19. Wu, D.; Chen, X.; Zhu, X.; Guan, X.; Wu, G. Uninformative variable elimination for improvement of successive projections algorithm on spectral multivariable selection with different calibration algorithms for the rapid and non-destructive determination of protein content in dried laver. Anal. Methods 2011, 3, 1790–1796. [Google Scholar] [CrossRef]
  20. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. B (Stat. Methodol.) 2005, 67, 301–320. [Google Scholar] [CrossRef] [Green Version]
  21. Algamal, Z.Y.; Lee, M.H. Regularized logistic regression with adjusted adaptive elastic net for gene selection in high dimensional cancer classification. Comput. Boil. Med. 2015, 67, 136–145. [Google Scholar] [CrossRef] [PubMed]
  22. Guan, X.; Liang, J.; Qian, Y.; Pang, J. A multi-view OVA model based on decision tree for multi-classification tasks. Knowl. Based Syst. 2017, 138, 208–219. [Google Scholar] [CrossRef]
  23. Zhang, Y.; Ding, C.; Li, T. Gene selection algorithm by combining relief and mRMR. BMC Genom. 2008, 9, S27. [Google Scholar] [CrossRef] [Green Version]
  24. Duan, K.-B.; Rajapakse, J.C.; Wang, H.; Azuaje, F. Multiple SVM-RFE for gene selection in cancer classification with expression data. IEEE Trans. Nanobiosci. 2005, 4, 228–234. [Google Scholar] [CrossRef]
  25. Zhang, X.; Lu, X.; Shi, Q.; Xu, X.; Hon-Chiu, E.L.; Harris, L.N.; Iglehart, J.D.; Miron, A.; Liu, J.S.; Wong, W.H. Recursive SVM feature selection and sample classification for mass-spectrometry and microarray data. BMC Bioinform. 2006, 7, 197. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Vapnik, V.N.; Vapnik, V. Statistical Learning Theory; Wiley: New York, NY, USA, 1998; Volume 1. [Google Scholar]
  27. Gonen, M.; Tanugur, A.G.; Alpaydin, E. Multiclass posterior probability support vector machines. IEEE Trans. Neural Netw. 2008, 19, 130–139. [Google Scholar] [CrossRef] [PubMed]
  28. Dogan, U.; Glasmachers, T.; Igel, C. A unified view on multi-class support vector classification. J. Mach. Learn. Res. 2016, 17, 1–32. [Google Scholar]
  29. Madevska-Bogdanova, A.; Nikolik, D.; Curfs, L. Probabilistic SVM outputs for pattern recognition using analytical geometry. Neurocomputing 2004, 62, 293–303. [Google Scholar] [CrossRef]
  30. Duan, K.; Keerthi, S.S. Which Is the Best Multiclass SVM Method? An Empirical Study. Mult. Classif. Syst. 2005, 3541, 278–285. [Google Scholar]
  31. Yuan, L.M.; Chen, X.; Lai, Y.; Xi, C.; Shi, Y.; Zhu, D.; Li, L. A Novel Strategy of Clustering Informative Variables for Quantitative Analysis of Potential Toxics Element in Tegillarca Granosa Using Laser-Induced Breakdown Spectroscopy. Food Anal. Methods 2017, 1–12. [Google Scholar] [CrossRef]
  32. Kira, K.; Rendell, L.A. The feature selection problem: Traditional methods and a new algorithm. In Proceedings of the Tenth National Conference on Artificial Intelligence, San Jose, CA, USA, 12–16 July 1992; pp. 129–134. [Google Scholar]
  33. Kononenko, I. Estimating attributes: Analysis and extensions of RELIEF. In Proceedings of the European Conference on Machine Learning on Machine Learning, Catania, Italy, 6–8 April 1994; pp. 171–182. [Google Scholar]
  34. Guyon, I.; Weston, J.; Barnhill, S.; Vapnik, V. Gene Selection for Cancer Classification using Support Vector Machines. Mach. Learn. 2002, 46, 389–422. [Google Scholar] [CrossRef]
  35. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. B (Methodol.) 1996, 267–288. [Google Scholar] [CrossRef]
  36. Lockhart, R.; Taylor, J.; Tibshirani, R.J.; Tibshirani, R. A significance test for the lasso. Ann. Stat. 2014, 42, 413. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Hoerl, A.E.; Kennard, R.W. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics 2000, 42, 80–86. [Google Scholar] [CrossRef]
  38. Chang, C.-C.; Lin, C.-J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2011, 2, 27. [Google Scholar] [CrossRef]
  39. Wang, Z.; Gerstein, M.; Snyder, M. RNA-Seq: A revolutionary tool for transcriptomics. Nat. Rev. Genet. 2009, 10, 57–63. [Google Scholar] [CrossRef]
  40. Tarazona, S.; García-Alcalde, F.; Dopazo, J.; Ferrer, A.; Conesa, A. Differential expression in RNA-seq: A matter of depth. Genome Res. 2011, 2213–2223. [Google Scholar] [CrossRef] [Green Version]
  41. Best, M.G.; Sol, N.; Kooi, I.; Tannous, J.; Westerman, B.A.; Rustenburg, F.; Schellen, P.; Verschueren, H.; Post, E.; Koster, J. RNA-Seq of tumor-educated platelets enables blood-based pan-cancer, multiclass, and molecular pathway cancer diagnostics. Cancer Cell 2015, 28, 666–676. [Google Scholar] [CrossRef] [Green Version]
  42. Cuevas, A.; Febrero, M.; Fraiman, R. An anova test for functional data. Comput. Stat. Data Anal. 2004, 47, 111–122. [Google Scholar] [CrossRef]
  43. Özdemir, B.C.; Pentcheva-Hoang, T.; Carstens, J.L.; Zheng, X.; Wu, C.-C.; Simpson, T.R.; Laklai, H.; Sugimoto, H.; Kahlert, C.; Novitskiy, S.V. Depletion of carcinoma-associated fibroblasts and fibrosis induces immunosuppression and accelerates pancreas cancer with reduced survival. Cancer Cell 2014, 25, 719–734. [Google Scholar] [CrossRef] [Green Version]
  44. Heller, M.J. DNA microarray technology: Devices, systems, and applications. Annu. Rev Biomed. Eng. 2002, 4, 129–153. [Google Scholar] [CrossRef] [Green Version]
  45. Kallioniemi, O.-P.; Wagner, U.; Kononen, J.; Sauter, G. Tissue microarray technology for high-throughput molecular profiling of cancer. Hum. Mol. Genet. 2001, 10, 657–662. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  46. Rodionova, O.Y.; Oliveri, P.; Pomerantsev, A.L. Rigorous and compliant approaches to one-class classification. Chemom. Intell. Lab. Syst. 2016, 159, 89–96. [Google Scholar] [CrossRef]
  47. Lalkhen, A.G.; McCluskey, A. Clinical tests: Sensitivity and specificity. Contin. Educ. Anaesth. Crit. Care Pain 2008, 8, 221–223. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The idea of solving the multi-class classification problem using class-specific features by one-against-all (OVA) strategy.
Figure 1. The idea of solving the multi-class classification problem using class-specific features by one-against-all (OVA) strategy.
Sensors 20 01528 g001
Figure 2. The flowchart of the proposed method.
Figure 2. The flowchart of the proposed method.
Sensors 20 01528 g002
Figure 3. The average gene expression levels of each class in dataset 15.
Figure 3. The average gene expression levels of each class in dataset 15.
Sensors 20 01528 g003
Figure 4. The selected class-specific features of dataset 1 by EPSVM: (a) treating class 1 as the positive class; (b) treating class 2 as the positive class; (c) treating class 3 as the positive class. Indicates the selected feature and indicates the average feature vector of one class.
Figure 4. The selected class-specific features of dataset 1 by EPSVM: (a) treating class 1 as the positive class; (b) treating class 2 as the positive class; (c) treating class 3 as the positive class. Indicates the selected feature and indicates the average feature vector of one class.
Sensors 20 01528 g004
Figure 5. The select global features: (a) support vector machine recursive feature elimination (SVM-RFE); (b) Relieff. Indicates the selected feature.
Figure 5. The select global features: (a) support vector machine recursive feature elimination (SVM-RFE); (b) Relieff. Indicates the selected feature.
Sensors 20 01528 g005
Figure 6. Principal component scores scatter plot (PC 1 × PC 2) of Dataset 11 with different features: (a) raw features; (b) global features selected by SVM-RFE; (c) global features selected by Relieff; (d) class-specific features treating healthy donors (HD) class as positive class; (e) class-specific features treating BrCa class as positive class; (f) class-specific features treating colorectal cancer (CRC) class as positive class.
Figure 6. Principal component scores scatter plot (PC 1 × PC 2) of Dataset 11 with different features: (a) raw features; (b) global features selected by SVM-RFE; (c) global features selected by Relieff; (d) class-specific features treating healthy donors (HD) class as positive class; (e) class-specific features treating BrCa class as positive class; (f) class-specific features treating colorectal cancer (CRC) class as positive class.
Sensors 20 01528 g006
Figure 7. Feature selection results of proposed method with different α in one experiment for Dataset 11 (only the top 500 features are shown), where ×, , indicate the selected class-specific features respectively, and indicates the gene expression levels of one sample: (a) α = 0.1 , the total number of selected feature is 477; (b) α = 0.5 , the total number of selected feature is 272; (c) α = 0.9 , the total number of selected feature is 253.
Figure 7. Feature selection results of proposed method with different α in one experiment for Dataset 11 (only the top 500 features are shown), where ×, , indicate the selected class-specific features respectively, and indicates the gene expression levels of one sample: (a) α = 0.1 , the total number of selected feature is 477; (b) α = 0.5 , the total number of selected feature is 272; (c) α = 0.9 , the total number of selected feature is 253.
Sensors 20 01528 g007
Table 1. Structure of dataset 1.
Table 1. Structure of dataset 1.
ClassesFeatures from [1] to [60]Features from [61] to [120]
1 to 2021 to 4041 to 6061 to 8081 to 100101 to 120
Class 1 F 1 + F 3 F 2 + F 3 F 2 + F 3 F 1 + F 3 F 2 + F 3 F 2 + F 3
Class 2 F 2 + F 3 F 1 + F 3 F 2 + F 3 F 2 + F 3 F 1 + F 3 F 2 + F 3
Class 3 F 2 + F 3 F 2 + F 3 F 1 + F 3 F 2 + F 3 F 2 + F 3 F 1 + F 3
Table 2. Descriptions of the platelets data.
Table 2. Descriptions of the platelets data.
DataNo. of ClassesNo. of FeaturesNo. of Samples
Data 10HD+BrCa180994
Data 11HD+BrCa+CRC3200136
Data 12HD+BrCa+CRC+NSCLC3398196
Data 13HD+BrCa+CRC+NSCLC+GBM3575236
Data 14HD+BrCa+CRC+NSCLC+GBM+PAAD3876271
Data 15HD+BrCa+CRC+NSCLC+GBM+PAAD+HBC3728285
Note: The sizes for each class, including HD, BrCa, CRC, NSCLC, GBM, PAAD, HBC are 55, 39, 42, 60, 40, 35, 14, respectively.
Table 3. Descriptions of microarray data for cancer detection.
Table 3. Descriptions of microarray data for cancer detection.
DataNo. of ClassNo. of FeaturesNo. of Samples
Leukemia13532772
Leukemia2311,22572
Small Round Blue Cells Tumor (SRBCT)4832308
Brain_Tumor245010,367
Brain_Tumor15905920
Tumors_99605726
Tumors111112,534174
Table 4. NER of simulation data.
Table 4. NER of simulation data.
Simulation DataEPSVMSVM-REF+SVMRelieff + SVM
Data 11.00000.93440.9750
Data 20.98960.97920.9854
Data 30.98590.97340.9594
Data 40.98250.97380.9362
Data 50.98960.96460.9333
Data 60.99200.92680.9196
Data 70.98830.91480.8914
Data 80.98060.88960.8799
Data 90.97310.85440.8456
Table 5. NER and feature selection results for dataset 10 (HD+BrCa).
Table 5. NER and feature selection results for dataset 10 (HD+BrCa).
Ranking FeaturesEPSVM ( α = 0.5)Relieff + SVMSVM-RFE + SVM
No. of Selected FeaturesMean NERNo. of Selected FeaturesMean NERNo. of Selected FeaturesMean NER
10046.10000.929438.75000.926815.00000.9222
20052.30000.946467.50000.951922.50000.9335
40068.80000.9461140.00000.928340.00000.9200
60064.60000.9409187.50000.947360.00000.9488
80081.30000.9463290.00000.941590.00000.9498
100078.00000.9414162.50000.9524100.00000.9312
120080.30000.9646240.00000.9424120.00000.9623
140085.40000.9422332.50000.9243140.00000.9681
160079.40000.9385200.00000.9472160.00000.9372
180084.20000.9556292.50000.9280180.00000.9488
Table 6. The mean NER of different methods for platelet data.
Table 6. The mean NER of different methods for platelet data.
MethodsDataset 11Dataset 12Dataset 13Dataset 14Dataset 15
EPSVM( α = 0.1)0.89280.82160.81530.80670.8133
( α = 0.5)0.87580.82030.81840.79430.8014
( α = 0.9)0.90350.83910.82690.80300.7904
SVM-RFE+SVM0.86240.79960.79470.74320.7599
Relieff+SVM0.87080.73860.79800.73460.7603
Table 7. SN and SP of dataset 10 for each class ( α = 0.5).
Table 7. SN and SP of dataset 10 for each class ( α = 0.5).
ClassSNSP
EPSVMSVM-REF + SVMRelieff + SVMEPSVMSVM-REF + SVMRelieff + SVM
HD0.90740.83330.90741.00000.98770.9753
BrCa0.79490.82050.87180.89580.84380.8333
CRC0.78570.78570.69050.87100.90320.9462
Table 8. SN and SP of dataset 14 for each class ( α = 0.5).
Table 8. SN and SP of dataset 14 for each class ( α = 0.5).
ClassSNSP
EPSVMSVM-REF + SVMRelieff + SVMEPSVMSVM-REF + SVMRelieff + SVM
HD0.88890.94440.86670.96880.94550.9610
BrCa0.56920.44620.47690.92440.90980.9098
CRC0.75000.59000.64000.90400.89870.8693
NSCLC0.84620.75380.69230.98050.97070.9659
GBM0.48330.45000.50000.96870.94940.9422
PAAD0.64290.54290.55710.89880.88640.9136
HBC0.40000.36000.36000.99330.99110.9889
Table 9. NER of microarray data for tumor detection.
Table 9. NER of microarray data for tumor detection.
DataEPSVMSVM-RFE+SVMRelieff+SVM
Leukemia10.95930.93330.9445
Leukemia20.98520.89870.9589
SRBCT0.99580.96440.9844
Brain_Tumor20.75810.75900.7979
Brain_Tumor10.82930.72440.7862
Tumors_90.82930.72440.7862
Tumors110.95950.90160.9182

Share and Cite

MDPI and ACS Style

Yuan, L.-m.; Sun, Y.; Huang, G. Using Class-Specific Feature Selection for Cancer Detection with Gene Expression Profile Data of Platelets. Sensors 2020, 20, 1528. https://doi.org/10.3390/s20051528

AMA Style

Yuan L-m, Sun Y, Huang G. Using Class-Specific Feature Selection for Cancer Detection with Gene Expression Profile Data of Platelets. Sensors. 2020; 20(5):1528. https://doi.org/10.3390/s20051528

Chicago/Turabian Style

Yuan, Lei-ming, Yiye Sun, and Guangzao Huang. 2020. "Using Class-Specific Feature Selection for Cancer Detection with Gene Expression Profile Data of Platelets" Sensors 20, no. 5: 1528. https://doi.org/10.3390/s20051528

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop