Next Article in Journal
Effect of Water Content Variation on the Tensile Characteristic of Clayey Loess in Ili Valley, China
Next Article in Special Issue
A Radiomics Approach Based on Follow-Up CT for Pathological Subtypes Classification of Pulmonary Ground Glass Nodules
Previous Article in Journal
A Kernel Extreme Learning Machine-Grey Wolf Optimizer (KELM-GWO) Model to Predict Uniaxial Compressive Strength of Rock
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Brief Report

Minimum Sample Size Estimate for Classifying Invasive Lung Adenocarcinoma

School of Electrical and Information Engineering, Tianjin University, Tianjin 300192, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(17), 8469; https://doi.org/10.3390/app12178469
Submission received: 12 July 2022 / Revised: 20 August 2022 / Accepted: 20 August 2022 / Published: 24 August 2022
(This article belongs to the Special Issue Advanced Medical Imaging Technologies and Applications)

Abstract

:
Statistical Learning Theory (SLT) plays an important role in prediction estimation and machine learning when only limited samples are available. At present, determining how many samples are necessary under given circumstances for prediction accuracy is still an unknown. In this paper, the medical diagnosis on lung cancer is taken as an example to solve the problem. Invasive adenocarcinoma (IA) is a main type of lung cancer, often presented as ground glass nodules (GGNs) in patient’s CT images. Accurately discriminating IA from non-IA based on GGNs has important implications for taking the right approach to treatment and cure. Support Vector Machine (SVM) is an SLT application and is used to classify GGNs, wherein the interrelation between the generalization and the lower bound of necessary sampling numbers can be effectively recovered. In this research, to validate the interrelation, 436 GGNs were collected and labeled using surgical pathology. Then, a feature vector was constructed for each GGN sample through the fully connected layer of AlexNet. A 10-dimensional feature subset was then selected with the p-value calculated using Analysis of Variance (ANOVA). Finally, four sets with different sample sizes were used to construct an SVM classifier. Experiments show that a theoretical estimate of minimum sample size is consistent with actual values, and the lower bound on sample size can be solved under various generalization requirements.

1. Introduction

An important task in machine learning and artificial intelligence is to build a classifier for prediction and classification given available training samples or other prior information [1]. Statistical Learning Theory (SLT) plays an important role in prediction estimation and machine learning in that it provides a unified framework for solving finite-sample learning problems [2,3,4]. In the past few decades, Support Vector Machine (SVM), a method based on SLT, has been widely applied for the construction of a classifier when limited samples are available [5,6,7]. Unlike other classifiers, which assume that the number of samples can tend to infinity, SVM offers the best generalization and prediction capability after it is trained using limited samples. SVM is one among many classification methods capable of handling large or small samples, nonlinear, or those with high dimensional issues [8]. A vital question, however, in using SVM is still unresolved, namely, how many samples are necessary for a certain standard of prediction accuracy. To address this question, we devise here a new idea that not only pays attention to the performance of the model on the training set, but also puts additional focus on the model’s prediction accuracy of unknown new data. Typically, a model overfits when it is trained so well on the training data that it fails to generalize. Accordingly, the ability of the model to adapt to new data decreases. It is generally accepted that the primary factor in generalization performance is sample size. However, very little study has focused on studying the quantitative relationship between sample size and generalization. In this paper, we use the medical diagnosis of lung cancer based on computed tomography (CT) images as an example to solve the relationship problem; wherein, it is difficult to obtain large-scale samples in practice.
The construction of any classifier must include training and testing processes according to available samples. In this context, training error refers to the learning ability whereas testing error reflects the prediction and generalization ability [9]. In SLT, a learning error is called an empirical risk (EMP) while a generalization error is called an expected risk (EXP) [10]. Minimizing EMP is a necessary condition for building a classifier with generalization ability or an EXP, but it is not a sufficient condition [11]. In fact, a low EMP may lead to a high EXP, wherein the prediction and generalization ability involves both the number of samples and choice of the classifier type. Under the same EMP, a classifier with simple structure can often lead to a high generalization ability [12]. Nevertheless, SLT can act as the theoretical basis that connects EMP and EXP [13].
Recently, a number of methods have been proposed to diagnose lung cancer using CT images [14,15]. Invasive adenocarcinoma (IA) is a primary type of lung cancer and often presents as ground glass nodules (GGNs) in patient’s CT images [16]. Accurately discriminating IA from non-IA based on GGNs has critical implications for providing an effective treatment plan. Existing studies on classifying IA and non-IA have mostly focused on how minimizing EMP can improve classification accuracy and with very little consideration given to EXP. Moreover, the generalization and prediction ability of the classifier is not guaranteed in practice. To date, little is known about the interrelation between EXP and EMP, and more specifically, it remains unclear as to how many samples are necessary under a certain requirement of prediction accuracy. In this paper, the effect of different sample sizes on EXP is explored, and the minimum sample size is estimated to achieve the desired generalization requirements.

2. Materials and Methods

For this study, we retrospectively collected 436 early GGNs samples from the Tianjin General Hospital, in which the sample rate between IA and non-IA was 216:220. The segmented CT images of GGNs are shown in Figure 1. All subjects gave their informed consent for inclusion before they participated in the study. The study was conducted in accordance with the Declaration of Helsinki, and all experiments were approved by the ethics committee of General Hospital of Tianjin Medical University.
After obtaining a sample set of GGNs, we extracted features from each GGN sample for classifying IA and non-IA. Previous studies have demonstrated how the powerful modeling and feature learning capabilities of deep learning lead to remarkable results [17,18,19,20,21], especially in biomedical image processing applications based on quantitative CT images [22,23,24].
Transfer learning (TL) is a new research area in deep learning. TL transfer the knowledge obtained by a pre-trained network to other image processing tasks [25]. TL methods, when combined with a convolutional neural network (CNN), have been shown as effective solutions to problems associated with relatively small datasets [26]. Specifically, the deep feature extraction capability of TL methods can activate layers of a CNN model to construct the related feature vectors [27]. In the medical field, use of mainstream networks as feature extractors has become a typical and general method to deal with classification problems [28,29,30].
In this work, we use the AlexNet model as the feature extractor to construct a 4096-dimensional feature vector for each GGN sample. The AlexNet model is the first deep CNN model that was trained on the ImageNet competition. It has shown efficacy in image processing [31,32,33] and is convenient in terms of implementation and the ability to fine-tune its processes for new tasks. AlexNet has no complex or deep architectures, which means reduced computation time and the need for robust hardware. In addition, our previous research [34] has shown the effectiveness of features extracted by the fully connected (FC) layer of AlexNet for GGNs classification.
To find the key features in the sample set, we first use the p-value calculated by the Analysis of Variance (ANOVA) as a criterion for feature selection [35]. Typically, lower p-values correspond to larger between-class significant differences. The top ten features with the lowest p-value are selected to form a feature subset.
According to the general principle in SLT, EXP and EMP obey the following relationship with at least (1 − η) probability for a limited number of samples n [36]:
R e x p ( w ) R e m p ( w ) + h ( ln ( 2 n / h ) + 1 ) ln ( η / 4 ) n ,
where the value of EXP is denoted as Rexp(w) and the value of EMP as Remp(w).
Equation (1) can be simplified as
R e x p ( w ) R e m p ( w ) + Φ ( n / h ) .
where Φ(n/h) is the confidence range; η is confidence limits, expressed as the probability that the classifier predicts the outcome to occur; n is the number of training samples; and h is VC dimension, indicating the maximum number of samples that a selected type of classifier can distinguish all possible 2h sample combinations [37]. Only few types of classifiers have the known VC dimensions. Essentially, for the most used linear classifiers, the VC dimension is (d + 1), and d is the number of used features. The widely used SVM is just a linear classifier. When dealing with multi-classification problems using the linear classifier, the actual VC dimension in a one-to-rest manner is (d + 1) × c, where c is the number of classes [38]. From Equation (1), the EXP of the classifier consists of two parts: one part is the EMP; the other part is the confidence range, which is related to the VC dimension and the number of training samples. As the number of samples n increases, the confidence interval decreases monotonically, and the probability that Equation (1) holds increases monotonically.
We build four subsets of different sample sizes by random sampling. The SVM, which is the applicable form of SLT, is used for classifying IA and non-IA. We quantitatively estimate the effect of sample size to the generalization ability and compare the results between the computed EXP using Equation (2) and the actual test from the experiments. The schematic diagram of the experimental design is shown as Figure 2.

3. Results

First, 36 samples in the sample set were randomly selected from the overall 436 GGNs samples as a fixed test set. Four training subsets with different but increasing sample numbers were drawn from the remaining 400 samples as training sets; the proportion between positive (IA) and negative (non-IA) samples was 1:1. Therefore, the number of samples in the four group of training sets were 100, 200, 300, and 400, respectively. We repeated four sets of experiments with different training sets of four sample sizes and a fixed test set. With four sets of sample sizes, the 10-dimensional feature subsets obtained from the AlexNet were inputted into SVM for classifying IA and non-IA. The EXP error of GGNs in the test set and the EMP error in the training set under different sample sizes were calculated separately. We compared the actual value from the test set with the theoretical value computed by Equation (1) at 90% (i.e., η = 0.10) and 95% (i.e., η = 0.05) confidence levels. The experimental results are shown in Table 1.
It can be seen from Table 1 that the actual values from the testing samples and the computed theoretical values of Rexp(w) in Equation (1) show a nearly consistent trend as the sample size increases. Their differences gradually decrease with increase in sample size n, while not exceeding the upper limit of the theoretical value of Rexp(w). The value of confidence range Φ(n/h) is inversely proportional to the sample size, given the same number of features used. Experimental results demonstrate that the generalization ability of the invasive classification model is proportional to the sample size.
From the above conclusions, we can see that it is a necessary condition to minimize only EMP, but it is not a sufficient condition when the number of samples is limited. To increase the generalization and prediction capability of the model, the selected types of classifiers must have as small a VC dimension as is possible. Moreover, there is a requirement to minimize the confidence range as well. This process of selecting the classifier model and learning algorithm can be referred to as adjusting the confidence range. However, due to a lack of theoretical guidance, currently this process in most existing methods is based on prior knowledge and experience, resulting in an excessive dependence on researchers.
In practice, another critical challenge for building an effective classifier is the question of the number of samples necessary under a certain requirement of prediction accuracy. According to Equation (1), we continue to explore the lower bound on sample size under various generalization requirements. The experimental results are shown in Table 2.
From Table 1, we see that the value of Remp(w) is inversely proportional to the sample size when the VC dimension and confidence limits are unchanged, and the values are all lower than 1%. Therefore, we suppose Remp(w) is 1% taking into consideration a degree of classification error, which means that the classification accuracy of the training set reaches 99%. Then, we compute the classification task using only 2 features, meaning the value of d is 2. We take into account that under the binary classification task and the multi-classification task by the linear classifiers, the corresponding VC dimensions are (d + 1) and (d + 1) × c, respectively. When dealing with classification tasks of 2, 4, 6, 8, 10 classes, VC dimension is calculated as 3, 12, 18, 24, 30, respectively. Finally, by plugging the VC dimensions, confidence limits η (η = 0.05, 0.1) and supposed Remp(w) to Equation (1), the lower limit of the achievable generalization requirements can be calculated under different sample sizes as shown in Table 2. Table 2 also shows the minimum sample size required to meet different generalization requirements under various classification tasks. It also shows that if GGNs are divided into two classes, four classes, and six classes, it respectively requires at least 6000, 12,000, and 18,000 samples to achieve 90% generalization ability when using two features.
Furthermore, we compare the sample sizes between the actually used and the computed theoretical values from Equation (1) under a certain expected prediction accuracy and η = 0.05, where the actually used cases are taken from the research of Diego et al. [39], Mohammad et al. [40], and Dai [41], as shown in Table 3. We can see that all the actual sample sizes are greater than the lower bound of the computed sample sizes. This validates the estimated bound of sample size by Equation (1).

4. Conclusions

SLT plays an important role in building a classifier for prediction and classification in the medical diagnosis of lung cancer under available training samples or other prior information. In this paper, the SLT tool was used to analyze two key issues for design classification, i.e., one regarding the computation of the generalization ability of a classifier based on available samples, and its inverse problem, i.e., how many samples are necessary under a certain requirement of prediction accuracy. The research has guiding significance for establishing a necessary sample size under a certain requirement of prediction accuracy, such as the study of the medical diagnosis of lung cancer. Our experimental results validate our method in the process for classifying IA and non-IA. In the future, we will continue exploring the connection between the type of classifiers and generalization ability with SLT.

Author Contributions

Conceptualization, C.M. and S.Y.; methodology, C.M. and S.Y.; software, C.M.; validation, C.M.; writing—original draft preparation, C.M. and S.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant No. 61573251, 61973232.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the ethics committee of General Hospital of Tianjin Medical University (IRB2020-YX-145-01, 2020/12/31).

Informed Consent Statement

Patient consent was waived due to the secondary use of medical record specimens by the ethics committee of General Hospital of Tianjin Medical University.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jordan, M.I.; Mitchell, T.M. Machine learning: Trends, perspectives, fand prospects. Science 2015, 349, 255–260. [Google Scholar] [CrossRef] [PubMed]
  2. Golino, H.F.; Epskamp, S. Exploratory graph analysis: A new approach for estimating the number of dimensions in psychological research. PLoS ONE 2017, 12, e0174035. [Google Scholar] [CrossRef] [PubMed]
  3. Ha, M.H.; Tian, J. The theoretical foundations of statistical learning theory based on fuzzy number samples. Inf. Sci. 2008, 178, 3240–3246. [Google Scholar] [CrossRef]
  4. Loh, P.L. On lower bounds for statistical learning theory. Entropy 2018, 19, 667. [Google Scholar] [CrossRef]
  5. Jain, S. Computer-aided detection system for the classification of non-small cell lung lesions using SVM. Curr. Comput. Aided Drug Des. 2020, 16, 833–840. [Google Scholar] [CrossRef]
  6. Wang, J.; Liu, X.; Dong, D.; Song, J.; Xu, M.; Zang, Y.; Tian, J. Prediction of malignant and benign Lung tumors using a quantitative radiomic method. Acta Autom. Sin. 2017, 43, 2109–2114. [Google Scholar]
  7. Ye, Y.; Tian, M.; Liu, Q.; Tai, H.-M. Pulmonary nodule detection using v-net and high-level descriptor based SVM classifier. IEEE Access 2020, 8, 176033–176041. [Google Scholar] [CrossRef]
  8. Ladayya, F.; Purnami, S.W.; Irhamah. Fuzzy Support Vector Machine for Microarray Imbalanced Data Classification. In Proceedings of the 13th IMT-GT International Conference on Mathematics, Statistics and their Applications (ICMSA), Universiti Utara Malaysia, Kedah, Malaysia, 4–7 December 2017. [Google Scholar]
  9. Ardila, D.; Kiraly, A.P.; Bharadwaj, S.; Choi, B.; Reicher, J.J.; Peng, L.; Tse, D.; Etemadi, M.; Ye, W.; Corrado, G.; et al. End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nat. Med. 2019, 25, 954–961. [Google Scholar] [CrossRef]
  10. Land, W.H.; Margolis, D.; Gottlieb, R.; Yang, J.Y.; Krupinski, E.A. Improving CT prediction of treatment response in patients with metastatic colorectal carcinoma using statistical learning. Int. J. Comput. Biol. Drug Des. 2010, 3, 8–15. [Google Scholar] [CrossRef]
  11. Muselli, M.; Ruffino, F. Consistency of Empirical Risk Minimization for Unbounded Loss Functions. In Proceedings of the 15th Italian Workshop on Neural Nets, Perugia, Italy, 14–17 September 2004. [Google Scholar]
  12. Zhang, X.; Ha, M.; Wu, J.; Wang, C. The Bounds on the Rate of Uniform Convergence of learning Process on Uncertainty Space. In Proceedings of the 6th International Symposium on Neural Networks, Wuhan, China, 26–29 May 2009. [Google Scholar]
  13. Chen, D.R.; Sun, T. Consistency of multiclass empirical risk minimization methods based on convex loss. J. Mach. Learn. Res. 2006, 7, 2435–2447. [Google Scholar]
  14. Naik, A.; Edla, D.R. Lung Nodule Classification on computed tomography images using deep learning. Wirel. Pers. Commun. Vol. 2021, 116, 655–690. [Google Scholar] [CrossRef]
  15. Alamgeer, M.; Mengash, H.A.; Marzouk, R.; Nour, M.K.; Hilal, A.M.; Motwakel, A.; Zamani, A.S.; Rizwanullah, M. Deep learning enables computer aided diagnosis model for lung cancer using biomedical CT images. Comput. Mater. Contin. 2022, 73, 1437–1448. [Google Scholar]
  16. Nemec, U.; Heidinger, B.; Anderson, K.R.; Westmore, M.S.; VanderLaan, P.; Bankier, A.A. Software-based risk stratification of pulmonary adenocarcinomas manifesting as pure ground glass nodules on computed tomography. Eur. Radiol. 2018, 28, 235–242. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Dimitri, G.M.; Spasov, S.; Duggento, A.; Passamonti, L.; Toschi, N. Unsupervised Stratification in Neuroimaging through Deep Latent Embeddings. In Proceedings of the 42nd Annual International Conference of the IEEE-Engineering-in-Medicine-and-Biology-Society, Montreal, Canada, 20–24 July 2020. [Google Scholar]
  18. Shin, H.-C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef] [PubMed]
  19. Kim, D.H.; MacKinnon, T. Artificial intelligence in fracture detection: Transfer learning from deep convolutional neural networks. Clin. Radiol. 2018, 73, 439–445. [Google Scholar] [CrossRef] [PubMed]
  20. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  21. Bengio, Y.; Lecun, Y.; Hinton, G. Deep learning for AI. Commun. ACM 2021, 64, 58–65. [Google Scholar] [CrossRef]
  22. Affonso, C.; Rossi, A.L.D.; Vieira, F.H.A.; de Leon Ferreira, A.C.P. D eep learning for biological image classification. Expert Syst. Appl. 2017, 85, 114–122. [Google Scholar] [CrossRef]
  23. Rachapudi, V.; Devi, G.L. Improved convolutional neural network based histopathological image classification. Evol. Intell. 2021, 14, 1337–1343. [Google Scholar] [CrossRef]
  24. Da Nóbrega, R.V.M.; Rebouças Filho, P.P.; Rodrigues, M.B.; da Silva, S.P.; Dourado Júnior, C.M.; de Albuquerque, V.H.C. Lung nodule malignancy classification in chest computed tomography images using transfer learning and convolutional neural networks. Neural Comput. Appl. 2020, 32, 11065–11082. [Google Scholar] [CrossRef]
  25. Krizhevsky, A.; Sutskever, I.; Hinton, G. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Processing Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  26. Pan, S.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  27. Demir, F.; Bajaj, V.; Ince, M.C.; Taran, S.; Şengür, A. Surface EMG signals and deep transfer learning-based physical action classification. Neural Comput. Appl. 2019, 31, 8455–8462. [Google Scholar] [CrossRef]
  28. Polat, H.; Mehr, H.D. Classification of pulmonary CT images by using hybrid 3D-deep convolutional neural network architecture. Appl. Sci. 2019, 9, 940. [Google Scholar] [CrossRef]
  29. Nalik, A.; Edla, D.R. Lung nodules classification using combination of CNN, second and higher order texture features. J. Intell. Fuzzy Syst. 2021, 41, 5243–5251. [Google Scholar]
  30. Dutta, A.K. Detecting lung cancer using machine learning techniques. Intell. Autom. Soft Comput. 2021, 31, 1007–1023. [Google Scholar] [CrossRef]
  31. Shayma’a, A.H.; Sayed, M.S.; I Abdalla, M.I.; Rashwan, M.A. Breast cancer masses classification using deep convolutional neural networks and transfer learning. Multimed. Tools Appl. 2020, 79, 30735–30768. [Google Scholar]
  32. Cengil, E.; Nar, A. The effect of deep feature concatenation in the classification problem: An approach on COVID disease detection. Int. J. Imaging Syst. Technol. 2021, 32, 26–40. [Google Scholar] [CrossRef]
  33. Hou, X.X.; Xu, X.Z.; Zhu, J.; Guo, Y. Computer aided diagnosis method for breast cancer based on AlexNet and ensemble classifiers. J. Shandong Univ. Eng. Sci. 2019, 49, 74–79. [Google Scholar]
  34. Ma, C.C.; Yue, S.H.; Li, Q. Deep Transfer Learning Strategy for Invasive Lung Adenocarcinoma Classification Appearing as Ground Glass Nodules. In Proceedings of the 2021 IEEE International Instrumentation and Measurement Technology Conference, Glasgow, UK, 17–20 May 2021. [Google Scholar]
  35. Monica, L.; Gaddis, P.D. Statistical methodology: Analysis of variance, analysis of covariance, and multivariate analysis of variance. Acad. Emerg. Med. 1998, 5, 258–265. [Google Scholar]
  36. Song, X.S.; Jiang, X.Y.; Luo, J.H.; Wang, X. SVM parameter selection based on the bound of structure risk. Sci. Technol. Rev. 2011, 29, 72–74. [Google Scholar]
  37. Vaart, V.D.; Wellner, J.A. A note on bounds for VC dimensions. Inst. Math. Stat. Collect. 2009, 5, 103–107. [Google Scholar] [PubMed]
  38. Jayadeva. Learning a hyperplane classifier by minimizing an exact bound on the VC dimension. Neurocomputing 2015, 149, 683–689. [Google Scholar] [CrossRef]
  39. Diego, M.; Rodrigo, M. Improvement of patient classification using feature selection applied to Bi-Directional Axial Transmission. IEEE Trans. Ultrason. Ferroelectr. Freq. Control. 2022. [Google Scholar] [CrossRef]
  40. Mohammad, A.A.; Mwaffaq, O.; Hamza, J. Comprehensive and comparative global and local feature extraction framework for lung cancer detection using CT scan images. IEEE Access 2021, 9, 158140–158154. [Google Scholar]
  41. Dai, C. SVM Visual Classification Based on Weighted Feature of Genetic Algorithm. In Proceedings of the 2015 Sixth International Conference on Intelligent Systems Design and Engineering Applications, Guiyang, China, 18–19 August 2015. [Google Scholar]
Figure 1. Examples of IA and non-IA nodules images. (ad) Segmented IA nodules, (eh) segmented non-IA nodules. (a,e) Larger volume; (b,f) smaller volume; (c,g) more solid components; (d,h) less solid components.
Figure 1. Examples of IA and non-IA nodules images. (ad) Segmented IA nodules, (eh) segmented non-IA nodules. (a,e) Larger volume; (b,f) smaller volume; (c,g) more solid components; (d,h) less solid components.
Applsci 12 08469 g001
Figure 2. Schematic diagram for classifying IA and non-IA based on SVM and GGNs.
Figure 2. Schematic diagram for classifying IA and non-IA based on SVM and GGNs.
Applsci 12 08469 g002
Table 1. Comparison of theoretical and actual values of Rexp(w).
Table 1. Comparison of theoretical and actual values of Rexp(w).
Number of SamplesRemp(w)EXP (η = 0.05)EXP (η = 0.10)Rexp(w) in Test Set
Φ(n/h) in Equation (2)Theoretical ValueΦ(n/h) in Equation (2)Theoretical Value
n = 10000.68770.68770.68260.68260.361
n = 2000.0050.5240.5290.52070.52570.222
n = 3000.0030.44490.44790.44230.44530.194
n = 4000.0020.39540.39740.39320.39520.139
Table 2. The relationship between EXP and sample size.
Table 2. The relationship between EXP and sample size.
Number of Classesn = 3000n = 6000n = 9000n = 12,000n = 15,000n = 18,000n = 21,000n = 24,000
95% confidence (d = 2)c = 20.11030.08330.07100.06350.05830.05440.05140.0490
c = 40.18410.13860.11760.10470.09570.08900.08380.0795
c = 60.21570.16240.13770.12250.11200.10400.09780.0928
c = 80.24160.18200.15420.13720.12530.11640.10940.1037
c = 100.26390.19890.16860.14990.13690.12720.11950.1132
90% confidence (d = 2)c = 20.10910.08250.07040.06300.05780.05400.05100.0486
c = 40.18350.13820.11720.10440.09550.08880.08360.0793
c = 60.21510.16210.13740.12230.11170.10380.09760.0926
c = 80.24110.18170.15400.13700.12510.11630.10930.1036
c = 100.26340.19860.16830.14970.13670.12700.11930.1131
Note: The numbers in orange refer to the achievement of 90% generalization ability.
Table 3. The comparison results between the actual and the computed theoretical values on the lower bound of the sample size estimated in Equation (1).
Table 3. The comparison results between the actual and the computed theoretical values on the lower bound of the sample size estimated in Equation (1).
CaseNumber of ClassesNumber of FeaturesEXPNumber of Samples (η = 0.05)
Actual ValueTheoretical Value
Diego et al. [39]c = 2d = 20.32301235
Mohammad et al. [40]c = 2d = 100.311000776
Dai [41]c = 2d = 400.471000958
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ma, C.; Yue, S. Minimum Sample Size Estimate for Classifying Invasive Lung Adenocarcinoma. Appl. Sci. 2022, 12, 8469. https://doi.org/10.3390/app12178469

AMA Style

Ma C, Yue S. Minimum Sample Size Estimate for Classifying Invasive Lung Adenocarcinoma. Applied Sciences. 2022; 12(17):8469. https://doi.org/10.3390/app12178469

Chicago/Turabian Style

Ma, Chenchen, and Shihong Yue. 2022. "Minimum Sample Size Estimate for Classifying Invasive Lung Adenocarcinoma" Applied Sciences 12, no. 17: 8469. https://doi.org/10.3390/app12178469

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop