Next Article in Journal
Revocable-Attribute-Based Encryption with En-DKER from Lattices
Previous Article in Journal
Comparative Sensitivity Analysis of Some Fuzzy AHP Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of the Health Status of Older Adults Using Oversampling and Neural Network

Institute of Population Research, Peking University, Beijing 100871, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(24), 4985; https://doi.org/10.3390/math11244985
Submission received: 17 November 2023 / Revised: 10 December 2023 / Accepted: 15 December 2023 / Published: 17 December 2023

Abstract

:
Self-rated health (SRH) serves as an important indicator for measuring the physical and mental well-being of older adults, holding significance for their health management and disease prevention. In this paper, we introduce a novel classification method based on oversampling and neural network with the objective of enhancing the accuracy of predict the SRH of older adults. Utilizing data from the 2020 China Family Panel Studies (CFPS), we included a total of 6596 participants aged 60 years and above in our analysis. To mitigate the impact of imbalanced data, an improved oversampling was proposed, known as weighted Tomek-links adaptive semi-unsupervised weighted oversampling (WTASUWO). It firstly removes the features that are not relevant to the classification by ReliefF. Consequently, it combines undersampling and oversampling. To improve the prediction accuracy of the classifier, an improved multi-layer perception (IMLP) for predicting the SRH was constructed based on bagging and adjusted learning rate. Referring to the experimental results, WTASUWO can effectively improve the prediction performance of a classifier when being applied on an imbalanced dataset, and the IMLP using WTASUWO achieves a higher accuracy. This method can more objectively and accurately assess the health status and identify factors affecting the SRH of older adults. By mining relevant information related the health status of older adults and constructing the prediction model, we can provide policymakers and healthcare professionals with targeted intervention techniques to focus on the health needs of older adults. Meanwhile, this method provides a practical research basis for improving the health level of older adults in China.

1. Introduction

As China transitions into an aging society, the depth of ageing is on the rise. According to the seventh national census data, individuals aged 60 and above in China constitute now 18.7% of the total population, with those aged 65 and above accounting for 13.5%. Projections indicate that the peak will reach 34.9% around 2053. Notably, China’s aging rate is 1.5 times that of the global population [1]. In the wave of silver hair, health issues have garnered widespread attention, particularly for the elderly population. With the rapid development of the social economy and the increasing trend of population aging, health problems in old age are anticipated to become increasingly prominent [2].
In this context, utilizing artificial intelligence (AI) information technology to enhance the quality of health and aging services is beneficial to meet their needs for a happy life and may address the challenge of social aging [3]. Furthermore, the health status of older adults is not only related to the basic survival state, but also depends on the long-term development of society. Gaining insights into the health status of older adults and predicting their health needs holds significant importance in formulating targeted health policies and services. The content of the study addresses a crucial topic in the pursuit of achieving healthy and active aging in China. It not only broadens the research scope in this field but also carries substantial theoretical and practical significance for improving the health level of older adults and facilitating the realization of healthy aging.
At present, the assessment of the health status of older adults primarily relies on objective physiological indicators and medical data. However, these often fall short in reflecting the subjective health of older adults [4]. Self-rated health (SRH) emerges as a comprehensive health assessment index, allowing older adults to provide a self-assessment based on their physiological, psychological, and social suitability. It has increasingly been employed by researchers to measure the health status of older adults [5]. The prediction of SRH among older adults is beneficial to prevent potential health problems in advance and improve their overall quality of life. Therefore, the accurate prediction of the SRH of older adults, enabling the provision of more personalized and effective health services, has emerged as a research hotspot. Simultaneously, this method can offer valuable insights for health prediction and evaluation in various fields, thereby fostering the development of health science research.
In the research on predicting SRH in older adults, the imbalance in the number of older adults with various health states can lead to sample imbalance during the training process of commonly used classification methods. This imbalance significantly affects the practicability of the classifier. Effectively improving the accuracy of minority classes and the overall performance of classifiers has thus become a hot topic in the field of prediction [6,7]. Imbalanced datasets, characterized by a significant difference in the number of samples among different classes [8,9], poses a challenge. Currently, addressing the issue of data imbalance in classification involves existing research carried out in two aspects: data equalization processing and classification optimization [10].
At the data level, enhancing the accuracy of minority classes primarily involves altering the distribution of samples and selecting features in the training dataset. Resampling techniques are commonly employed for this purpose, with random undersampling and oversampling being the most utilized. In more detail, random undersampling refers to the random elimination of majority samples; however, this method may lose some useful information, resulting in affecting the classifier performance. As for the random oversampling, it mainly consists of randomly copying minority samples to make the class distribution of data balanced. However, this method increases the possibility of overfitting the model due to repeated replication of minority samples [11,12]. At the algorithm level, enhancing classification algorithms involves improving existing ones or designing new algorithms based on the inherent characteristics of imbalanced datasets. This is primarily achieved through ensemble and cost-sensitive learning to optimize the models [13]. As a powerful machine learning technique, neural networks (NNs) have demonstrated remarkable results in several fields in recent years [14]. Particularly in health prediction, NNs excel at accurately predicting individual health status by learning complex features and nonlinear relationships within large datasets [15].
To sum up, the primary objective of the current study is to develop a SRH prediction model for older adults with high accuracy, aiming to improve the refinement of health management. The main contributions include:
-
A data processing method, based on oversampling technique, is proposed to address class imbalance in datasets related to the health status of older adults;
-
An improved NN model is introduced to predict the health status of older adults, providing higher accuracy and stability in predicting the health status of older adults compared to traditional machine learning algorithms;
-
Based on the empirical and publicly available datasets, the proposed method further demonstrates its effectiveness;
-
By predicting the health status of older adults, potential health problems can be identified in advance and personalized health advice as well as interventions for the elderly, therefore improving their quality of life.

2. Literature Review

An imbalanced dataset refers to the number of samples in one class that is significantly larger than those in other classes. Due to sampling disparity, the traditional methods often fail to acquire the accurate prediction results. Nowadays, the methods applied to solve the classification problem for imbalanced dataset can be divided into two categories: the first one starting with the dataset, it transfers the imbalanced data into a balanced data by changing the data distribution or selecting the features that can better express the data information. As for the second category, it starts from the classification algorithm, where it enhances the accuracy according to the limitations resulting in the application of the algorithm on the imbalanced dataset.

2.1. Research on Oversampling Techniques

Regarding the oversampling techniques, Chawla et al. [16] proposed a novel technique, known as the synthetic minority oversampling technique (SMOTE). This technique achieves the class distribution balance by adding artificial minority oversampling samples to the data. Moreover, SMOTE has provided a new direction for addressing the imbalanced data and has become an effective oversampling technique that has been applied in several fields [17]. However, there are still some shortcomings regarding this technique: the fuzzy boundary problem, caused by SMOTE, ignores the imbalance in minority classes, and it is easily affected by noise. Therefore, it is difficult to get the best performance of a classifier by empirically selecting the different parameters.
Regarding these problems, some researchers have made improvements to SMOTE to adapt to different data characteristics. For instance, Han et al. [18] considered the minority class samples distributed near the class boundary and introduced the Borderline-SMOTE algorithm to improve the learning ability of the boundary samples. Moreover, He et al. [19] proposed the adaptive synthetic (ADASYN) sampling method based on sample distribution, effectively improving the prediction performance of the classifier [20]. In addition, Safe-level SMOTE [21] focused on the class overlap problem caused by SMOTE. It assigned a safety factor to each minority class sample before synthesizing the new samples to ensure that the new samples were distributed in the safe area. Furthermore, Douzas et al. [22] proposed a cluster-based oversampling method, called K-means SMOTE, which effectively addresses the problem of inter- and intra-class imbalance [23]. However, the oversampling results were contingent on the effectiveness of K-means, potentially leading to errors in minority class samples. In a different approach, Tao et al. [24] proposed a weighted oversampling method based on the support vector data description (SVDD) to tackle inter-class overlap and internal imbalanced problems.
In addition, Nekooeimehr et al. [25] proposed the adaptive semi-unsupervised weighted oversampling (ASUWO) algorithm. This method adopted a semi-supervised learning and adaptively adjusted the weights of samples to achieve a balance in the number of samples between classes, thereby enhancing the generalization performance of the model. However, attention should be given to computational complexity and parameter sensitivity. In case of dealing with severely imbalanced datasets, combining it with other methods may be necessary. In practical applications, variations in base models may lead to significant differences, and appropriate adjustment and optimization can be applied based on specific problems [26].
In terms of hybrid sampling, Batista et al. [27] proposed SMOTE Tomek and SMOTE edited nearest neighbor (SMOTE ENN) by combining both common methods. SMOTE Tomek first oversamples the minority class samples to balance the data distribution. Consequently, it is combines with the Tomek-links technique to remove the blurred samples between the majority and minority classes. Moreover, Seiffert et al. [28] combined Adaboost with oversampling and undersampling, and proposed SMOTEBoost and RUSBoost algorithms to increase the robustness of the classifier and improve its accuracy.

2.2. Research on Neural Network Models

At the algorithm level, modification primarily involves adjusting the bias of the algorithm on the dataset level, directing decisions to be biased towards the minority class. Techniques, such as ensemble learning and feature selection, play a crucial role in achieving the bias. In the realm of machine learning applications, approaches include logistic regression (LR), K-nearest neighbor (KNN), decision tree (DT), random forest (RF), support vector machine (SVM), and other single algorithms. On the other hand, some researchers use the voting methods, adaptive boosting (AdaBoost), gradient boosting decision tree (GBDT), and extreme gradient boosting (XGBoost) for prediction and classification [29]. As a widely used method in machine learning, NNs have been widely used in the healthcare domain, addressing challenges such as classification, regression, and clustering.
Moreover, artificial neural network (ANN) is a nonlinear mapping and adaptive dynamic technique formed by several simple neurons widely interconnected [30]. Added to that, back propagation (BP) and radial basis function (RBF) networks have the disadvantages of slow learning speed, being easy to fall into local extremes, and having low accuracy of prediction results. As for the self-organizing map (SOM) network, it adopts unsupervised learning rules and lacks classification information. For instance, Rumelhart et al. proposed multiple-layers perceptron (MLP), a kind of neural network using forward propagation and error backward propagation; it is mainly applied to solve the multi-class prediction problem of nonlinear data [31,32]. The generalization ability and processing efficiency of MLP have several advantages compared to BP, RBF, and SOM networks. If MLP is applied to fit more complex mapping functions with higher accuracy, the main challenges become the structural design and parameter optimization of the model.
The model structure refers to the depth of the model and the width of each layer as well as the activation function. MLP can automatically extract the features and abstraction levels from the data. Generally, the deeper the depth is, the fewer nodes could be used in each layer. However, due to the gradient problems, it is difficult to train such a model. However, the accuracy can be enhanced by increasing the model width even if it is shallow. The wider the width of the model is, the better the fitting ability will be, but the easier the overfitted will appear. The advantages and limitations of various activation functions are not conclusive, and it is required to be selected according to the application scenarios and the datasets. Therefore, to get a suitable model structure for the specific application, it is necessary to adjust it after computing several experiments.
Concerning the model training, the optimization algorithms mainly improve the learning factors and the training methods. Among the first-order optimization algorithms, stochastic gradient descent (SGD) and its variants are among the most widely used in deep learning (DL). SGD is simple and has low computational complexity, but its convergence rate is slow. Given the learning rate significantly influences model performance, researchers have proposed several adaptive learning rate algorithms based on SGD. Among these, the adaptive moment (Adam) estimation algorithm has gained popularity. For example, Khan et al. [33] used the genetic algorithm (GA) to optimize the BP network, comparing the optimized prediction results obtained using some technical indicators and original BP. The results indicated that the prediction accuracy of the GA-based BP was higher. Moreover, Kar et al. [34] applied the evolutionary algorithm (EA) to train the ANN model. The synergy of NNs with optimization algorithms presents an avenue for further exploration. Studying how to apply improved algorithms to enhance the effectiveness and accuracy of NNs for prediction is a worthwhile pursuit.
As a crucial parameter in NNs, a fixed learning rate has no adaptability to the objective function. Setting the learning rate too high may result in training shock and divergence, while setting it too low may lead to slow convergence. In practical applications, adjusting the learning rate based on the specific task and dataset becomes necessary. A common approach is to employ dynamic methods that adjust this parameter, such as gradually reducing the learning rate during the learning process. This helps prevent the model from learning too quickly at the beginning of training and potentially falling into a local optimal solution [35].
In addition, deep neural networks (DNNs) have a deeper network structure compared to shallow neural networks and can learn more complex features and higher level abstractions. For instance, Lin et al. [36] trained transformer protein language models with up to 15 billion parameters using experimental and high-quality predicted structures. They converted protein sequences into textual forms processable by the language model, leveraging the model’s learning ability to predict protein structure. This approach significantly improves the accuracy and efficiency of predictions, particularly when dealing with large datasets, showcasing robust parallel processing capabilities. However, language models face limitations in certain fields with limited protein sequence data due to their dependency on large training datasets. To address this challenge, Lin et al. [37] proposes an approach called ghost convolution with BottleneckCSP and a tiny target prediction head incorporating YOLOv5 for PV panel defect detection. The approach enhances the network representation by introducing ghost convolution and BottleneckCSP. Furthermore, the prediction of small targets is improved by tiny target prediction head. These improvements are combined with YOLOv5, resulting in a new and enhanced detection model. While this method improves the prediction of small targets, enabling the detection of smaller defects, it is important to note that the YOLOv5 model has a large computational load and may require longer processing time for large-scale detection tasks. Overall, DNNs represent a very promising approach, but the structure of such models is more complex and they may pose a risk of overfitting for some simple detection tasks. Therefore, the specific applicability still requires evaluation based on the actual problem and data situation.

2.3. Major Findings

In recent years, classification methods based on oversampling and NNs have been widely applied in the field of health prediction. In practical applications, NN prediction methods also face some challenges. Assessing the health status of older adults involves multiple dimensions, including physiological, psychological, and social aspects. The effective extraction and integration of information becomes pivotal for the prediction model. Notably, there is a limited discussion on this issue in the context of health assessment for older adults in China.
(1) Data quality
For the SRH prediction of older adults, the selection of the data features is crucial to the prediction effect of the model. In more detail, undersampling may lead to lose information of majority classes, and oversampling poses a challenge in ensuring the generated samples conformity with the real data distribution. It is also robust to outliers, mitigating the impact of large deviations in predictions caused by outlier data points. Although hybrid sampling combines undersampling and oversampling methods, it also yields limitations. To improve the prediction rate of a classifier, it is necessary to optimize the oversampling technique.
(2) Model optimization
The traditional statistical methods are difficult to cope with the task regarding feature extraction and classification of complex data. Meanwhile, due to the large heterogeneity of elderly population, the health status of distinct individuals is obviously different. Therefore, NN has good nonlinear fitting ability and high fault tolerance. Even for small sample size data, it can generate a good prediction effect. Meanwhile, MLP has a strong learning ability. However, there are limited studies deploying MLP to predict the health status of older adults. Nevertheless, there is certain research significance of improving the application of MLP in the SRH prediction for older adults.
This paper focused on using oversampling and NNs to predict the health status of older adults. Although there are many research studies in this area, we believe that our work is of innovative value in the following ways: The oversampling technique is applied to balance the data and improve the generalization ability of the model. Moreover, the NN structure is applied to grasp the complex features of the health status of older adults, further improving the prediction accuracy. This method is applied to train the model with existing datasets and can effectively identify health risk factors of older adults. Theoretically, the results can enrich the data processing methods as well as the theoretical framework of the health status prediction research. In practice, it serves as a basis for the implementation and improving the health status of older adults.

3. Methods

In this section, we mainly explore the way to use oversampling and NNs to improve the prediction accuracy of older adults’ health status, as well as explore the influencing factors on the health status, in order to provide a scientific basis for health management and policy formulation. The basic principles of the study include the following:
(1) Description and analysis of experimental data;
(2) Proposition of a new oversampling technique and exploration of the main influencing factors;
(3) Development of a new high-accuracy NN prediction model.

3.1. Data Sources and Analysis

To study the SRH prediction of older adults in China, the experimental data were gathered from the 2020 China Family Panel Studies (CFPS). The reasons include the following:
(1) Representativeness: CFPS is a nationwide comprehensive social longitudinal survey, with high reliability, the changes in the society, economy, population, and health in China.
(2) Richness: The data not only contain the basic information of older adults, but also cover their health status, lifestyle, and other aspects; this provides rich data support for studying the influencing factors on the SRH of older adults.
(3) Timeliness: The data reflect the current health status and the factors affecting it and can provide policymakers and relevant departments with the latest reference basis [38].
In addition, it is relevant to the problem solved in this paper, summarized by the following two points:
(1) It helps to reveal the current situation and characteristics of older adults’ health status and can explore the influencing factors and provide a theoretical basis for enhancing their health level;
(2) The data have high representativeness, reliability, and richness, helping in promoting and applying the results of the study and providing empirical support for solving the aging problem in China.
In this paper, the health status of older adults aged 60 and above was selected as the research object. First, the original data were preprocessed, including missing data handling and normalization. For each feature, samples containing missing values were removed, and a total of 6596 samples with 36 features were included for modeling. Normalization is a commonly used data preprocessing method to remove the effect of data magnitude and accelerate model converge. The feature attributes are normalized to between [0, 1], as follows:
x i j = x i j x i j min x i j max x i j min
where xij is the jth feature of the ith sample, and x i j max and x i j min are the maximum and minimum values of the jth feature, respectively. This dataset contains detailed information of older adults, such as basic information, working status, health status, lifestyle, network services, network importance, and importance of access to information. Table 1 describes the features of the data.
According to the question “How healthy do you think you are?”, the answers are classified into unhealthy, average, relatively healthy, very healthy, and super healthy. The numbers of each class were 1607, 1081, 2448, 722, and 738, respectively. This paper combined the classes “relatively healthy”, “very healthy”, and “super healthy” into “healthy”; thus, the final categories were “unhealthy”, “average”, and “healthy”, whose data amounts were 1607, 1081, 3908, having the values of 1, 2, and 3, respectively. Currently, there are no unified criteria for defining the classification of various health levels, which are generally based on the purpose and problem of the study [39]. Therefore, the SRH of older adults tends to be good. Among them, the “healthy” older adults’ category accounts for the highest proportion, reaching 59.25% of the total number of participants, indicating that most older adults can maintain good health in daily life. This class was followed by “unhealthy” and “average” older adults, accounting for 24.36% and 16.39%, respectively, indicating that some older adults face health problems.

3.2. Oversampling Technique Based on WTASUWO

Oversampling is a technique yielding to data augmentation. It consists of increasing the number of minority class samples to balance the number of samples among the other classes, reducing, therefore, the influence of imbalanced class problem on model performance. In this paper, oversampling was selected due to several reasons:
(1) Oversampling is an operation based on using the raw data and does not lose information;
(2) Oversampling facilitates the model in more effective learning features from the samples belonging to minority classes during the training;
(3) Oversampling is easier to implement compared to other data preprocessing techniques;
(4) Oversampling has been widely applied to the prediction problems in several unbalanced data scenarios.
For example, some studies have used the oversampling techniques to solve the imbalanced problem in the diabetes risk prediction, and experiments have proved that it can effectively enhance the prediction performance of the model [40]. Moreover, the ASUWO algorithm was proposed in 2016 based on the SMOTE oversampling; this algorithm balances the number of samples by sampling with retraction in minority class samples as well as the application of weights to improve the performance of the model on the imbalanced dataset. However, some problems still reside in this algorithm.
(1) Calculate sample weights: Due to the noise and outliers in the samples, this can lead to uneven weight distribution of ASUWO, affecting the model’s prediction performance.
(2) Synthesize sample outliers: The synthesized samples by the boundary between the minority and majority classes may contain both data characteristics, resulting in outliers in the synthesized samples. Therefore, the imbalanced distribution of samples may affect the accuracy of the outlier detection based on clustering [41].
Thus, in this paper, we proposed a ReliefF-based of feature weighted Tomek-links ASUWO (WTASUWO) oversampling, which consists of combining Tomek-links undersampling and ASUWO oversampling. The improvements of this techniques can be summarized as follows:
(1) A weighting strategy that was used to improve the generalization ability and accuracy of the model.
Problems like sparse data distribution, feature redundancy, and irrelevance in high-dimensional imbalanced data pose difficulties for traditional learning algorithms in identifying minority class samples. A prevalent approach is to address this by reducing the dimensionality of the data before classification, operating in a new dimension space. Feature selection aims to identify an “optimal” subset of relevant features. This not only reduce the complexity of the problem and enhances the processing efficiency but also effectively streamlines the design of the classifier [42].
The Relief algorithm, initially proposed by Kira, was limited to the classifying of two types of data. ReliefF, an improved feature selection method building upon Relief, introduces the concept of positive and negative samples. It calculates feature correlation coefficients, utilizing the square root to measure the distance between samples. ReliefF is versatile, applicable to both classification and regression problems, enhancing model performance by selecting features with higher weights. In practice applications, ReliefF can also be combined with various machine learning algorithms. Notably, it assigns substantial weights to features that significantly contribute to classification, offering a simple and fast approach for feature selection. This approach has no restriction regarding the type of data [43,44]. Therefore, ReliefF is employed to sort the importance of the features to reduce its dimension. The algorithm is described as follows:
Step 1: Initialize the weights of all features to 0.
Step 2: For i = 1 to m do (m is the number of sampling times)
(1)
Randomly select a sample R.
(2)
Finds k nearest neighbor samples Hj from the samples of the same class R, and finds k nearest neighbor samples Mj from the samples of different class R. (j = 1,2,…,k).
(3)
For i = 1 to N do (N is the number of runs)
W ( A ) = W ( A ) j = 1 k d i f f ( A , R , H j ) / ( m k ) + C c l a s s ( R ) [ p ( C ) 1 p ( c l a s s ( R ) ) j = 1 k d i f f ( A , R , M j ( C ) ) ] / ( m k )
where diff(A, R1, R2) represents the distance between sample R1 and sample R2 on the feature A; Mj(C) represents the jth nearest neighbor sample of class C; and p(C) represents the probability of the class C.
(4)
Sort the weight W values.
Step 3: Calculate the weight value for each feature.
The feature weights obtained by ReliefF all belong to the interval [−1, 1]. The greater the weight, the more important the feature is. Figure 1 shows the importance and ranking of each factor. This study reduced the complexity of the model by keeping important variables and removing variables that contribute less or are not useful. The number of final features was 15, including x12, x13, x9, x5, x1, x8, x35, x34, x20, x11, x33, x31, x6, x14, and x17.
(2) A resampling technique combining undersampling and oversampling was presented.
Hybrid sampling is a means to remove the noise samples in the imbalanced problems, but the combination of noise filtering technology can also eliminate the wrong samples [45]. For the problem of synthesizing sample errors, Tomek-links was used to eliminate outliers, and it calculated that the samples with the closest distance between classes form a Tomek chain. Then, according to the Tomek chain, the minority class samples located at the boundary between classes were identified, and the overlapping samples were deleted. Next, the adaptive semi-unsupervised clustering in ASUWO was used for oversampling, so that the data distribution was basically balanced while reducing the number of noisy samples [46]. By combining labeled and unlabeled data for sample selection, ASUWO can learn more characteristics of the data, and it has better balance compared with SMOTE.
Tomek-links refers to the nearest neighbors of each other in two samples from different classes, providing a better decision boundary for the classifier. If there are two samples x and y belonging to different classes, the distance d(x, y) is calculated. At this time, if another sample z cannot be found, so that the distance from any sample to z is smaller than d(x, y), it is called a Tomek pair. For each Tomek pair, one of the samples is removed.
The WTASUWO algorithm (Algorithm 1) is a combination of ReliefF, Tomek-links, and ASUWO, aiming to improve the prediction performance through feature weighting, denoising undersampling, and oversampling. The ReliefF feature weights reflect the importance of each feature for classification. The Tomek-links algorithm removes samples that are more similar to the nearest neighbor samples of its own category by comparing the similarity between the samples, which reduces the effect of noisy samples on classification. The ASUWO algorithm removes those samples with low affiliation, which allows the classifier to be trained with more samples, thus improving the prediction performance.
Algorithm 1. WTASUWO
Inputs: training dataset, category labels, number of nearest neighbor samples.
Output: filtered and weighted balanced dataset.
1: Initialization: number of iterations, feature weights, undersampling rate, etc.
2: For each sample
  a. Feature weight calculation: ReliefF is used to calculate the feature weights.
  b. Undersampling processing: based on the weighted data, the Tomek-links algorithm is used for undersampling to remove noisy samples.
  c. Oversampling processing: based on the weighted and noise denoised feature data, the ASUWO algorithm is used for oversampling to increase the number of samples in minority classes.
3: Output the filtered and weighted training dataset.

3.3. SRH Prediction Model of Older Adults Based on IMLP

In general, traditional statistical methods rely on specific data distribution assumptions and have poor prediction performance when considering high-dimensional and large-scale unknown data. As for MLP, it is a type of NN that describes a complex mapping between a set of input variables and an output variable. This technique is suitable for solving the multi-classification problem of nonlinear fractional data. This feature makes the MLP model highly valuable and promising in the prediction studies. Moreover, Figure 2 illustrates the main structure of the network model, divided into input, output, and hidden layers, where the layers of MLP are fully connected. The numbers below Figure 2 represent the number of nodes in the different layers. As for the number of selected features, it represents the number of nodes in the input layer. Moreover, the determination of the number of nodes in the hidden layer is described in Section 3.3, whereas the number of nodes in the output layer represents the classified health status category.
The construction process of MLP can be divided into forward propagation and back propagation. Forward propagation calculates the outputs of the nodes through the weights and thresholds of the previous iteration. Back propagation was used for parameter training. It is a process of calculating the influence of each weight and threshold on the total error from the output layer, and then adjusting the weights and thresholds to minimize the error.
In the forward propagation, the feature vector of the input layer is set as xij. The weight and threshold values from the input layer to the hidden layer are represented by wij and bj. The output value of the nodes in the hidden layer is Sj. M, N, and K are the number of nodes in the input, hidden, and output layers, respectively. L is the number of samples.
S j = f ( i = 1 M w i j x i j + b j ) , j = 1 , 2 , , N
where f(·) represents the activation function, and the sigmoid function is generally used, as in Equation (4). After obtaining the output values of the hidden layer, they are calculated as the input values of the output layer. The network results are shown in Equation (5).
f ( x ) = 1 1 + e x
y ^ k = f ( j = 1 N w j k x j k + b k ) , k = 1 , 2 , , K
In back propagation learning, y ^ k is the output signal of the network, y ^ l is the predicted value of the network, and yl is the true value. The purpose of NN learning is to reduce the error between y ^ l and yl to the set accuracy, and the commonly used error function E is the mean square error (MSE), expressed as follows:
E = 1 L l = 1 L ( y ^ l y l ) 2 + λ | w |
where λ is a regularized coefficient that represents the degree of constraint on the function space. When λ is large, it limits the function space to be small; when λ is small, it allows the function space to become large. λ | w | , also known as structural risk, is the sum of the absolute values of the weights w in the model. The commonly used | w | include L1 and L2 norms. With this item, the function space corresponding to the MLP can be controlled more flexibly, thus better avoiding the problem of overfitting.
The gradient descent method was used to update the weight parameters of the hidden and output layers. If the gradient of w is 0, that is, = 0 , then the next weight w is adjusted, as in Equation (7).
{ w w + Δ w Δ w = η E w
where η is the learning rate, and ∆w represents the magnitude of the gradient. If the output value y ^ is equal to the expected output y, w does not change. Otherwise, adjust w until the error is less than a certain value to end the training.
However, there still exists some problems with MLP, such as the possibility of disappearing gradients or gradient explosions when dealing with the complex tasks. Meanwhile, using NNs to solve classification problems requires to set values for the model hyperparameters. These hyperparameters include the training algorithm, the initial learning rate, the number of hidden layers, the number of hidden layer neurons, the activation function of each layer, and the training times. To improve the generalization ability of the network, some regularization methods are also deployed to avoid overtraining. The improved MLP (IMLP) consists of the following:
(1) Design the network structure: number of nodes.
The input layer serves as a buffer, integrating data sources into the network. In this model, 15 feature dimensions are derived through feature selection. When a network with a hidden layer includes enough neurons, it can effectively approximate a complex system with arbitrary precision. However, increasing the number of hidden layers may extend the computing time and lead to potential overfitting problems. Hence, for simplicity and efficiency, the model with only one hidden layer is considered [47]. The utilization of an implicit layer enables a multilayer perceptron with nonlinear mapping capability, feature extraction, and abstract representation; however, this configuration will reduce the model complexity and improving training efficiency. Moreover, a NN with one hidden layer with a sufficient number of hidden neurons holds the capability to approximate any continuous function.
Few nodes in the hidden layer are not able to fully express the nonlinear relationship in the system, and many nodes may cause overfitting. Through the application of different number of nodes, Figure 3 demonstrates the accuracy of the model corresponding to the number of nodes in the hidden layers. Moreover, referring to Figure 3, the model has the optimal performance when the number of nodes in the hidden layer reaches 50. The number of neurons in the output layer represents the number of output variables (SRH), so the number of neurons in the output layer is three.
In this paper, defining the number of hidden nodes is an empirical process in designing the network structure [48]. Initially, the NN structure is evaluated based on the problem and data characteristics. When building a NN, a small number of hidden nodes can be set, ensuring that the network has a certain learning ability but may not be sufficient to solve the complex problem. Therefore, the range is set between 10 and 100, and the number of nodes gradually increases using a step of 10 to observe the performance variations of the network. As the number of nodes in the hidden layer increases, the accuracy of the network progressively improves. When it reaches a certain level, the performance of the network becomes invariant, and thus the optimal network structure will be suitable for the problem.
(2) Setup parameter optimization: ensemble learning and learning factors.
Bagging is a method applied to improve the accuracy of weak classifiers by constructing a series of prediction functions. They are then combined into a single function using a certain way. Bagging is suitably combined with unstable learning algorithms, such as NNs and DTs. By voting on these classifiers, weak classifiers are trained to generate stronger classifiers, as illustrated in Figure 4. Considering the instability of the model, bagging is applied to build a global model, generating multiple model sequences to improve the prediction accuracy of MLP [49]. At the same time, using Dropout to improve the training process helps in solving the overfitting problems. As a result, the probability value of discarded neurons is set to 0.2 and the K-fold cross-verification is applied for training.
For this model, the learning factor is specified according to experience. If the learning factor is too small, the training speed may be slow. If it is too large, it may cause oscillations, and the error function will not converge during training. Therefore, the attenuation factor was introduced to dynamically adjust the learning rate in the training process of the model. When the error tends to the target in a reduced way, the learning rate can be increased by correcting the direction. When the error increases beyond a certain range, it indicates that the previous step correction is incorrect, and the step size should be reduced. The learning rate updating method is shown in Equation (8). The connection weights of different neurons are dynamically adapted to the learning rate to ensure that the optimization in each direction can find the optimal solution, in order to maximize the adjustment rate of weights and speed up the convergence of the model.
η = { a t , t < T b e r t c , t > T
where a, b, and c are constants; r is the attenuation rate; t is the number of iterations; and T’ is the default number of iterations. The learning rate change curve of this method is shown in Figure 5, which shows the changes of fixed, linear, exponential decay, and adaptive learning rates. At the beginning of the iteration, the learning rate increases, while in the later stages of the iteration, it decreases exponentially.

3.4. Overall Flowchart

The proposed method (e.g., WTASUWO-IMLP (Algorithm 2)) mainly includes feature extraction, undersampling, oversampling, and classification algorithms. ReliefF is first applied to obtain the best feature subset to simplify the input values. Moreover, Tomek-links and ASUWO are used to construct the overall distribution of the data covering the learnable data. Strategies are provided to sample both minority and majority classes to reach data balance. Finally, the synthesized samples are applied to train the classifier. In the modeling of IMLP, it is necessary to clean and normalize the original dataset; consequently, these data are divided into the training and test datasets after being preprocessed. The training set is imported, and the network structure and optimized parameters are determined. Finally, the test dataset is input to verify the accuracy of the model. The modeling process of the SRH prediction of older adults is displayed in Figure 6.
Algorithm 2. WTASUWO-IMLP
Input: original dataset.
Output: Prediction results.
1: Preprocess the original dataset, including data cleaning and normalization.
2: Using the ReliefF algorithm to obtain the optimal feature subset and weights.
3: Using Tomek-links and ASUWO algorithms to undersample and oversample the training set, generating a new training set.
4: Divide the new dataset into training and testing sets.
5: Use the IMLP algorithm to train the model using the new training set.
  a. Determine the network structure and optimization parameters.
  b. Using the new training set to train models: forward propagation, computational loss, and backpropagation.
6: Verify the accuracy of the model using the test set.
7: Generate the prediction results.
The steps of the WTASWO-IMLP algorithm are as follows:
Step 1: Data preprocessing: data cleaning and normalization is performed on the original dataset to remove noise and outliers and to maintain the data at the same scale;
Step 2: Feature selection: the ReliefF algorithm is used to select the optimal feature subset from the original features based on the weights to simplify the input values;
Step 3: Data sampling: the Tomek-links and ASUWO algorithms are used to undersample and oversample the training set for data balancing;
Step 4: Data partitioning: the preprocessed dataset is partitioned into training set and test set, where the former is used to train the model and the latter is applied to verify the accuracy of the model;
Step 5: Model training: on a new training set, the IMLP algorithm is applied to model training, including determining network structure and optimizing the parameters;
Step 6: Model validation: use the test set as input to verify the accuracy of the model;
Step 7: Output results: output the trained classifier to predict SRH for older adults.

4. Results

In this section, from the perspective of algorithm performance analysis, including the determination of the base classifier and feature selection, performance analysis of the oversampling and NNs, and significance analysis of the proposed method are verified based on the empirical 2020 CFPS dataset. In addition, we also conducted a robustness analysis of the proposed method on the public datasets to further validate its generalization ability and stability.

4.1. Evaluation Indicators

When dealing with the imbalanced data, it is necessary to select the evaluation indicators, which are usually based on the confusion matrix. In this paper, accuracy, precision, recall, specificity, area under the receiver operating characteristic (AUC), and running time (RT) were used to evaluate and compare the performances of different classifiers [50,51]. Due to the research on the health status prediction, it is a labeled classification problem based on the characteristics of the dataset. These metrics can evaluate the performances of classifiers from different perspectives and provide a reference basis for model optimization. For example, accuracy, precision, and recall are concerned with the ability of the classifier to recognize the positive class samples, specificity is concerned with the ability of the classifier to recognize the non-positive class samples, and AUC and RT are concerned with the overall performance and efficiency of the classifier.
(1) Accuracy: the ratio of the number of correctly classified samples to the total number of samples.
Accuracy = T P + T N T P + T N + F P + F N
(2) Precision: the proportion of samples classified as positive samples that are true positive samples.
Precision = T P T P + F P
(3) Recall: the proportion of positive samples in the total sample that are predicted correctly.
Recall = T P T P + F N
(4) Specificity: the proportion of negative samples in the total sample that are predicted correctly.
Specificity = T N T N + F P
(5) The receiver operating characteristic (ROC) curve is usually presented in the case of cross, where the advantages and disadvantages of the classifiers cannot be directly distinguished [52]. AUC can be used as an extension of the judging basis of ROC. The larger the AUC value, the higher the accuracy and the better the classifier.
AUC = i P C r a n k i M ( 1 + M ) 2 M × N
where M is the number of positive samples; N is the number of negative samples; ranki is the sequence number of sample i; PC indicates the positive samples; TP is the positive samples that are predicted to be positive; TN is the negative samples that are predicted to be negative; FP is the negative samples that are predicted to be positive; and FN is the positive samples that are predicted to be negative. The values of the above five indicators range from 0 to 1. The higher the value, the better the performance of the model.
Experimental environment: Windows 10 64-bit operating system, processor AMD Ryzen 7 5800H with Radeon Graphics 3.20 GHz, RAM 16.0 GB (15.9 GB available). The programming languages were Matlab 2022a and Python 3.8.8. To objectively compare various methods and improve the generalization ability of the model, the fivefold cross-validation was used to obtain the results. A total of 70% of samples of the original dataset were used as the training dataset, and these samples and the synthesized samples were used to train the classifier, and the remaining 30% samples were used as the test dataset for model testing.
The concept definitions of the key terms are shown in Table 2.
The abbreviations of terms are shown in Table 3.

4.2. Determination of The Base Classifier and Feature Selection

In this paper, some base classifiers were used to choose the algorithm that was suitable for the studied problem, including LR, KNN, DT, RF, SVM, AdaBoost, GBDT, XGBoost, BP, extreme learning machine (ELM), and MLP. The reasons for selecting these classifiers are as follows: (1) Different machine learning algorithms have their own characteristics and show different performances when dealing with the problem in this paper. (2) In practice, there is no absolute algorithm. The only way to find a suitable algorithm as a base classifier is for a specific problem and dataset. Meanwhile, the prediction of the health status involves complex nonlinear relationships and multidimensional features, which requires strong generalization ability and the ability to process the complex data.
NNs have an advantage over other traditional machine learning algorithms in handling nonlinear problems. Comparison of multiple algorithms can help us to find a more suitable model for the current problem to meet the application requirements. For each classifier, we trained and tested the same datasets to ensure the objectivity of the results, and calculated the indicators of accuracy, precision, recall, specificity, AUC, and RT. The evaluation results of different classifiers are shown in Table 4, where the bold values are the relative optimal values. The visual effect is shown in Figure 7.
From Table 4 and Figure 7, by comparing and evaluating these classifiers, the MLP model with the best performance to solve the studied problem was selected. LR is easy to implement and interpret, and it performs well for linear problems, but it is susceptible to overfitting. KNN does not need to transform the feature of the data, which is good for the nonlinear problems, but it is sensitive to the noise samples. DT is sensitive to feature selection, it can process the missing values and class features. RF is robust to noise and missing values, but it can lead to feature selection bias. SVM performs well for linear problems and has certain advantages in solving samples with small sample size, as well as nonlinear and high-dimensional datasets. AdaBoost has a strong generalization ability, but it is easily affected by noise samples. GBDT can combine multiple weak learners, which has strong robustness to noise, but it has high computational complexity. XGBoost has strong robustness to missing values and noise and high computational efficiency, but it is easily affected by overfitting. BP can deal with nonlinear problems and has strong generalization ability, but it is sensitive to the initial weights. ELM has high computational efficiency and is not sensitive to the initial weights, but it is easily affected by noise and has poor performance for linear problems. MLP is a deep learning model with multiple hidden layers. By learning the nonlinear network structure, potential feature rules were obtained from the dataset, nonlinear activation functions were used to improve the expression ability of the model, and the model parameters were updated by the optimizer. It can learn the essential features of a dataset from the minority class samples.
However, too many features will increase the burden of the model. On the other hand, the invalid features will also decline the model performance, so some feature selection methods were used to select features. Feature selection methods can be divided into filter, wrapper, and embedded [53]. The difference lies in whether the feature selection process is combined with the specific classification learning methods. In this paper, different feature selection algorithms were compared, including variance selection and the chi-squared test in Filter, GA in Wrapper, tree-based model in Embedded, and Lasso and ReliefF machine learning algorithms. Based on the MLP classifier, the prediction performances of the feature selection algorithms were compared, which are shown in Table 5 and Figure 8, with the bold values being the relative optimal values.
Filter is a preprocessing method that selects features based on their variance or correlation. The variance filtering method filters the features by the variance of the features. The chi-squared test is to determine whether there is a dependency between features by calculating chi-squared statistics. Wrapper selects features in the process of model training. It selects features according to the output of the model. GA is a common Wrapper method. It selects the optimal feature subset by simulating the process of natural selection and genetic evolution. Embedded is conducted by embedding feature selection into the process of model training. The tree-based model is conducted by constructing a decision tree for feature selection. Both Lasso and ReliefF methods conduct feature selection by limiting the complexity of the model. Lasso construes the complexity of the model by adding L1 regularization terms to select the relevant features. ReliefF selects features by calculating the relative importance of each feature, employing a new distance metric that is better able to process the high-dimensional data. As shown in Table 5 and Figure 8, ReliefF can select the most relevant features, thus improving the performance of the model.

4.3. Performance Analysis of The Oversampling Technique

Oversampling is a method to deal with the imbalance problem, which can reduce the imbalance among classes and improve the prediction performance of a classifier. In this paper, different oversampling techniques were compared, including random undersampling, random oversampling, SMOTE, BorderLine SMOTE, SVM SMOTE, Safe-level SMOTE, ASUWO, SMOTE Tomek, WCNN ASUWO, and WTASUWO. Based on the base classifier and feature selection, the evaluation results of different oversampling algorithms are shown in Table 6 and Figure 9, with the bold values being the relative optimal values.
Random undersampling selects samples from the majority classes and deletes them. Random oversampling selects samples from the minority classes and copies them. These two methods are simple but can lead to data loss and redundancy. SMOTE is a common oversampling method, with good effect and high efficiency. BorderLine SMOTE finds the boundaries of minority class samples and then generates synthetic samples. SVM SMOTE uses the characteristics of SVM to oversample the minority class samples, which can improve the performance of the model. Safe-level SMOTE generates samples by introducing a safe distance. ASUWO adjusts the proportion of undersampling and oversampling according to the degree of imbalance to produce the synthetic samples. SMOTE Tomek is the combination of SMOTE and Tomek-links, which uses Tomek to oversample the minority class samples. WCNN ASUWO calculates the distance between samples by weighting and combines ASUWO to find the minority class samples. WTASUWO improves on ASUWO by first using weighted Tomek-links for undersampling, and then using ASUWO for oversampling to generate the most representative minority class samples. It should be noted that oversampling may increase the computational time, so it is necessary to choose the appropriate oversampling method according to the specific situation. After oversampling, the data distribution of each category will be more balanced, as shown in Figure 10.

4.4. Performance Analysis of The Neural Network Model

ELM and MLP have relatively strong representations that can fit complex nonlinear relationships, but they are also prone to overfitting and tuning parameters. The emergence of intelligent algorithms can be used to optimize the weights and bias parameters of NNs with better global search capabilities, which helps to improve the network performance. To verify the prediction performance of IMLP, it was compared with cuckoo search (CS)-based ELM, simulated annealing (SA)-based ELM, firefly algorithm (FA)-based ELM, particle swarm optimization (PSO)-based MLP, MLP using WTASUWO, and IMLP using WTASUWO. It is mainly based on the ELM and MLP classifiers, which are used for the improved algorithm analysis. The prediction results are shown in Table 7 and Figure 11, with the bold values being the relative optimal values.
MLP is a feedforward NN in which each neuron is connected to all neurons in the previous layer and nonlinearly transforms the input through an activation function. As shown in Table 7 and Figure 11, IMLP using WTASUWO had the best performance in various evaluation indicators, and the prediction accuracy was improved from 0.602 to 0.855. Followed by MLP using WTASUWO, it shows that the prediction performance of MLP was significantly improved. For the ELM and MLP algorithms optimized by intelligent algorithms, their prediction results were not as good as that of IMLP in this dataset. The reason may be because IMLP has a stronger generalization ability when dealing with such problems, or the WTASUWO oversampling strategy improves the performance of IMLP more significantly. At the same time, the RT is acceptable to the research problem in this paper.
CS, SA, FA, and PSO are optimization algorithms that find the optimal solution by training NNs such as ELM and MLP. These algorithms can optimize the model parameters. MLP is a supervised learning that improves the model performance by updating the parameters through back-propagation. In terms of performance, ELM is a fast-learning algorithm, while MLP requires many iterations to converge. In summary, the comparison between this method and other algorithms depends not only on the characteristics of the algorithm itself, but also on the specific application scenarios and problems.
Figure 12 shows the effect of the test process of IMLP using WTASUWO. ROC was used to evaluate the prediction performance of the model under different thresholds. The horizontal coordinate represents the false positive rate (FPR), and the vertical coordinate represents the true positive rate (TPR). The ideal ROC curve should be TPR increasing with FPR until it reached the maximum.

4.5. Significance and Robustness Analysis of WTASUWO-IMLP

(1) Statistical significance tests of IMLP using WTASUWO.
Due to the test errors between different algorithms, it is necessary to conduct the statistical significance tests on the accuracy of the six algorithms. In this paper, we used the Friedman rank test, which is a nonparametric post method that can be used to test whether there are significant differences between the algorithms. This method does not require the assumption that the data has a normal distribution or variance-aligned [54].
Null hypothesis: CS-based ELM, SA-based ELM, FA-based ELM, PSO-based MLP, MLP using WTASUWO, and IMLP using WTASUWO have the same distribution, meaning there is no significant difference between the different algorithms. These algorithms were tested in 10 experiments, with each training and test dataset randomly distributed. The Friedman test results are shown in Table 8.
The Friedman test results (p = 0.000) indicate that there were at least two of the different algorithms are significantly different from each other (p < 0.001), and the multiple comparison test (Nemenyi) can be further used to determine which algorithms were significantly different. Moreover, it can be seen from Table 8 that IMLP using WTASUWO was significant compared to CS-, SA-, FA-based ELM, and PSO-based MLP. The Nemenyi test was used to calculate the critical domain for the difference in the average order value. CD = q α k ( k + 1 ) 6 × N = 2.38, where k is the number of algorithms, N is the number of experiments, and q α is a constant under the significance α . CD plots were used to represent the performance differences between the algorithms. Friedman average order values of different algorithms are shown in Figure 13.
In Figure 13, the vertical axis shows the comparison algorithms, and the horizontal axis shows the average order values. For each algorithm, a dot is used to show its average order value, and a horizontal line segment centered in the dot represents the size of the critical value domain. If the segments of the two algorithms overlap, it indicates that there is no significant difference in performance between the two algorithms within the confidence interval. Otherwise, it means that the performance is significantly different. From this figure, there is shown to be no significant difference between IMLP using WTASUWO and MLP using WTASUWO because their horizontal segments have overlapping areas. It also illustrates the significant improvement in the prediction performance of the classifier with the WTASUWO oversampling technique. In addition, IMLP using WTASUWO was found to be significantly superior to the CS-, SA-, FA-based ELM, and PSO-based MLP algorithms because their horizontal segments did not overlap. Therefore, from the results of the statistical significance tests, the method of this paper is significantly better than other algorithms.
(2) Robustness analysis of WTASUWO-IMLP.
To evaluate the robustness of the proposed method (WTASUWO-IMLP), and considering the limited number of samples, we used four imbalanced datasets with different ratios in the University of California Irvine (UCI) machine learning repository (http://archive.ics.uci.edu/, accessed on 12 October 2022). The characteristics of these datasets are summarized in Table 9. For each dataset, the data were divided into the training and test datasets, and they were preprocessed. Based on the data before and after oversampling, the proposed method was compared with the existing methods. The results are shown in Table 10, Table 11, Table 12 and Table 13, and in Figure 14, Figure 15, Figure 16 and Figure 17.
Firstly, for the oversampling methods, it can be seen from the results that compared with that without oversampling, the prediction performances of the classifiers based on each dataset were significantly improved after WTASUWO. The application of WTASUWO made the RTs of the classifiers relatively reduced, which shows that the algorithm not only improved the performances of the classifiers, but also had high computational efficiency. Secondly, for the comparison of different classification algorithms, IMLP had high prediction accuracy and strong generalization ability for solving the classification problems, and it was able to effectively identify different categories of samples. Based on the above results, the WTASUWO oversampling technique and IMLP model had better effects in improving the performances and efficiencies of the classifiers. In practical applications, WTASUWO can be used for oversampling, and then IMLP can be used for classification and prediction.

5. Discussion

In this section, we focus on exploring the SRH of older adults in China based on the CFPS dataset from the perspective of the key factors affecting their health status, as well as suggesting useful policy recommendations.

5.1. Discussion of Experimental Results

This study indicated that most older people rate their health as good, accounting for 59.25%. The older adults with unhealthy and average health status accounted for 24.36% and 16.39%, respectively. Through feature selection, the final 15 features were used as the influencing factors, containing health changes, chronic diseases, after-tax wage income, living area, and so on. Taking these four key factors as examples, we conducted the chi-squared analysis on the SRH of older adults. The results are shown in Table 14.
The difference analysis employs the hypothesis testing method to assess whether influencing factors can account for changes in the dependent variable. Given that the SRH and influencing factors are categorical variables, the chi-squared test was employed to analyze differences in the impact of key factors on the SRH of older adults. Referring to Table 14, it is evident that these features significantly affect the SRH of older adults. Health changes directly influence the SRH of older adults, with those experiencing chronic diseases exhibiting poorer health status. Additionally, a higher after-tax wage income correlates with better health status of older adults. Moreover, older adults living in urban areas have a higher level of SRH than those living in rural areas. The findings of this study conform with previous studies.
The second data analysis of the Population Health Index Survey (PHIS) verified that chronic diseases are associated with lower SRH [55]. The lower SRH may not solely be attributed to the severity of chronic diseases but rather to the disparity between actual and expected health status. Additionally, the disruptive perception of pain, dysfunction, and inconvenience of daily life stemming from chronic diseases exacerbates the SRH of older adults.
A study conducted in the UK identified the positive effects of income increasing on SRH, emphasizing that income stability is strongly correlated with SRH and overall well-being [56]. Individuals with higher income levels typically enjoy better access to health knowledge and healthcare services, and they adopt healthier lifestyles compared to those with lower incomes. This, in turn, contributes to better SRH among individuals with high incomes.
Consistent with prior evidence, it has been established that the SRH of individuals in urban areas tends to be better than those in rural areas. The discrepancy is attributed to the varying availability of sanitation, preventive medicine, and healthcare services [57]. Urban areas typically boast sanitary conditions, abundant medical resources, and advanced facilities, fostering a conducive environment for the health status of older adults. However, these relatively poor sanitary conditions and limited resources in rural areas contribute to lower SRH among older adults residing in those regions.
The Canadian Longitudinal Study on aging, which enrolled community-dwelling adults aged 45 to 85, investigated the intersectional impact of sex and social factors on SRH. The findings revealed that women were less likely to report poor SRH [58]. The sex difference can be clarified by the different consideration between men and women. In the health assessment, men often concentrate on physical health, whereas women adopt a more comprehensive approach, considering factors such as physical health, function, and mental health.

5.2. Policy Implications

There are several factors affecting the SRH of older adults in China. Special attention should be paid to older adults with lower income, with chronic diseases, in rural areas, and in worse health status. Comprehensive and consecutive efforts based on different aspects are still necessary to improve the health status of older adults.
(1) Reinforce publicity and education to guide older adults to develop good living habits.
Healthcare institutions should establish health archives for older adults, especially for those with hypertension, diabetes, and other chronic diseases, helping to manage the health status of sensitive population and slow the development of chronic diseases. Meanwhile, older adults should keep healthy lifestyles, including regular physical exercise, balanced diet patterns, and positive social connections. It is important to keep good attitudes and strengthen the prevention of serious diseases to improve the SRH of older adults [59]. Through health promotion, the occurrence of chronic diseases for older adults can be effectively prevented and controlled.
(2) Establish a sound old-age security system to improve the income level of older adults with financial difficulties.
The economic living standard has a significant positive effect on older adults’ SRH. The health status of such a population, with more adequate financial sources, is better. Therefore, the social pension security system should be established and improved in line with the national conditions, and financial assistance to older adults with economic difficulties should be provided. The government can increase the transfer income of older adults with poor family status based on income redistribution or encourage social organizations to provide help for poor older adults.
(3) Improve the level of medical and health services through the government’s investment and the joint efforts of the society.
Medical conditions can affect the health status of older adults, as the SRH of older adults in urban areas is significantly higher. The primary medical institutions in rural areas should make full use of the existing health resources by improving the community care and health network and helping in the enhancement of the health status of older adults in rural areas. Therefore, the investment in medical resources should increase and their distribution should expand. Moreover, a multi-level and rationally distributed medical and health service system should be developed to ensure that older adults with diseases can receive timely and effective treatment.

6. Limitations

The proposed method has advantages in the accuracy and stability of prediction. However, it still has some limitations that are described here below:
(1) Sample size and research perspective.
Research on the prediction of the health status of older adults necessitates a large amount of sample data to ensure the stability and accuracy of the model. Furthermore, the SRH of older adults is influenced by various factors, such as physical, psychological, and social aspects. We applied seven aspects and fail to cover the whole data information to explore the interrelationship between these factors. The diversity of sample data and the generalization ability of the model still requires to be further improved.
(2) Model optimization perspective.
Although NNs have certain advantages in predicting the health status of older adults, AI techniques still have limitations, such as overfitting and the inability to explain the model decision-making process. The difficulty in NN modeling lies in the determination of the number and threshold of hidden neuron layers.
(3) Discuss strategies
In the discussion section, the impact of certain factors on the health status of older adults is analyzed, but it is necessary to consider multiple factors to develop a comprehensive management program.

7. Conclusions

With the aging of the population in China, health issues for older adults have become an increasing concern. Therefore, this study aimed to develop a classification method based on oversampling and NNs to accurately predict the health status of older adults. The main findings are as follows:
(1) WTASUWO was proposed by combining the Weighted Tomek-links strategy and ASUWO oversampling. Compared to different data processing methods, the results show that WTASUWO is very important for improving the accuracy of a classifier.
(2) IMLP was proposed by setting the network structure and optimizing parameters. Based on the oversampled data, IMLP was compared to CS-, SA-, FA-based ELM; PSO-based MLP; and MLP using WTASUWO, and the accuracy was improved from 0.602 to 0.855, showing that this method achieved remarkable improvement in the accuracy of prediction.
To sum up, the experimental results show that the proposed method based on oversampling and NNs has high feasibility in practical applications. Finally, we discuss the experimental results and provide some policy implications.

8. Future Work

In future works, we can discuss the already listed limitations and the way to make targeted improvements in future research. Therefore, we will continue to optimize the model structure, expand the sample size, and explore influencing factors to improve the universality and practicability of the prediction problem. In more detail, here are the main future ideas:
(1) The effect of sample size on model performance can be analyzed using the CFPS dataset, and more large-scale and diverse data can be collected. Thus, we will explore the factors influencing the health status of older adults from other perspectives and develop personalized interventions.
(2) NNs have great potential for SRH prediction, and in the future, we can try to optimize the model structure, such as using more advanced network architectures and optimization algorithms to improve the prediction accuracy. Moreover, we will consider DNNs and deeper layers. In addition, it is possible to combine other machine learning and traditional statistical methods with NNs to achieve more efficient and accurate prediction.
(3) Promoting the attention of the government and all society sectors to the health problems of older adults and increasing policy support related to health benefits, achieving an effective health management for older adults in the future.

Author Contributions

Conceptualization, Y.L.; methodology, Y.L.; software, Y.L.; formal analysis, Y.L.; resources, G.C.; writing—original draft preparation, Y.L., Q.H., and G.X.; writing—review and editing, Y.L.; visualization, Y.L.; supervision, G.C.; funding acquisition, G.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Wuhan East Lake High-Tech Development Zone (also known as the Optics Valley of China, or OVC) National Comprehensive Experimental Base for Governance of Intelligent Society (grant #2023003437).

Data Availability Statement

Data employed in this study are public data with sources provided in the article.

Acknowledgments

We are grateful to the Institute for Artificial Intelligence, Peking University, and PKU-WUHAN Institute for Artificial Intelligence, Dingsheng Luo, for providing funding acquisition. The authors are grateful to the editor and the anonymous referees for their many helpful comments on earlier version of our paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bai, C.; Lei, X. New trends in population aging and challenges for China’s sustainable development. China Econ. J. 2020, 13, 3–23. [Google Scholar] [CrossRef]
  2. Alanazi, H.; Daim, T. Health technology diffusion: Case of remote patient monitoring (RPM) for the care of senior population. Technol. Soc. 2021, 66, 101662. [Google Scholar] [CrossRef]
  3. Elreedy, D.; Atiya, A.F. A comprehensive analysis of synthetic minority oversampling technique (SMOTE) for handling class imbalance. Inf. Sci. 2019, 505, 32–64. [Google Scholar] [CrossRef]
  4. Liu, H.; Kong, L.; Sun, Q.; Ma, X. The effects of mindfulness-based interventions on nurses’ anxiety and depression: A meta-analysis. Nurs. Open 2023, 10, 3622–3634. [Google Scholar] [CrossRef] [PubMed]
  5. Chen, Y.; While, A.E.; Hicks, A. Self-rated health and associated factors among older people living alone in Shanghai: SRH of older people living alone. Geriatr. Gerontol. Int. 2015, 15, 457–464. [Google Scholar] [CrossRef] [PubMed]
  6. Bahnsen, A.C.; Aouada, D.; Stojanovic, A.; Ottersten, B. Feature engineering strategies for credit card frand detection. Expert Syst. Appl. 2016, 51, 134–142. [Google Scholar] [CrossRef]
  7. Prati, R.C.; Batista, G.E.A.P.A.; Silva, D.F. Class imbalance revisited: A new experimental setup to access the performance of treatment methods. Knowl. Inf. Syst. 2015, 45, 247–270. [Google Scholar] [CrossRef]
  8. Manokaran, J.; Vairavel, G. GIWRF-SMOTE: Gini impurity-based weighted random forest with SMOTE for effective malware attack and anomaly detection in IoT-Edge. Smart Sci. 2023, 11, 276–292. [Google Scholar] [CrossRef]
  9. Shilaskar, S.; Ghatol, A.; Chatur, P. Medical decision support system for extremely imbalanced datasets. Inf. Sci. 2017, 384, 205–219. [Google Scholar] [CrossRef]
  10. He, H.; Garcia, E.A. Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar] [CrossRef]
  11. Dai, Q.; Liu, J.W.; Liu, Y. Multi-granularity relabeled under-sampling algorithm for imbalanced data. Appl. Soft Comput. 2022, 124, 109083. [Google Scholar] [CrossRef]
  12. Zhao, J.; Liu, N. Semi-supervised classification based mixed sampling for imbalanced data. Open Phys. 2019, 17, 975–983. [Google Scholar] [CrossRef]
  13. Li, Y.; Qi, J.; Jin, H.; Tian, D.; Mu, W.; Feng, J. An improved genetic-XGBoost classifier for customer consumption behavior prediction. Comput. J. 2023, bxad041. [Google Scholar] [CrossRef]
  14. Wang, W.; Wu, Y. Risk analysis of the Chinese financial market with the application of a novel hybrid volatility prediction model. Mathematics 2023, 11, 3937. [Google Scholar] [CrossRef]
  15. He, B.; Zhang, Y.; Zhou, Z.; Wang, B.; Liang, Y.; Lang, J.; Lin, H.; Bing, P.; Yu, L.; Sun, D.; et al. A neural network framework for predicting the tissue-of-origin of 15 common cancer types based on RNA-seq data. Front. Bioeng. Biotechnol. 2020, 8, 737. [Google Scholar] [CrossRef]
  16. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  17. Espinosa, E.; Figueira, A. On the quality of synthetic generated tabular data. Mathematics 2023, 11, 3278. [Google Scholar] [CrossRef]
  18. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Advances in Intelligent Computing, Proceedings of the International Conference on Intelligent Computing (ICIC), Hefei, China, 23–26 August 2005; Lecture Notes in Computer Science, Huang, D.S., Zhang, X.P., Huang, G.B., Eds.; Springer: Berlin/Heidelberg, Germany, 2005. [Google Scholar] [CrossRef]
  19. He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; IEEE: Piscataway, NJ, USA, 2008. [Google Scholar] [CrossRef]
  20. Yang, L.; Zhang, J.; Wang, X.; Li, Z.; Li, Z.; He, Y. An improved ELM-based and data preprocessing integrated approach for phishing detection considering comprehensive features. Expert Syst. Appl. 2021, 165, 113863. [Google Scholar] [CrossRef]
  21. Bunkhumpornpat, C.; Sinapiromsaran, K.; Lursinsap, C. Safe-level-SMOTE: Safe-level-synthetic minority over-sampling Technique for handling the class imbalanced problem. In Advances in Knowledge Discovery and Data Mining, Proceedings of the Lecture Notes in Computer Science, Pacific-Asia Conference on Knowledge Discovery and Data Mining (PAKDD), Bangkok, Thailand, 27–30 April 2009; Theeramunkong, T., Kijsirikul, B., Cercone, N., Ho, T.B., Eds.; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar] [CrossRef]
  22. Douzas, G.; Bacao, F.; Last, F. Improving imbalanced learning through a heuristic oversampling method based on k-means and SMOTE. Inf. Sci. 2018, 465, 1–20. [Google Scholar] [CrossRef]
  23. Zhu, T.; Lin, Y.; Liu, Y. Improving interpolation-based oversampling for imbalanced data learning. Knowl.-Based Syst. 2020, 187, 104826. [Google Scholar] [CrossRef]
  24. Tao, X.; Zheng, Y.; Chen, W.; Zhang, X.; Qi, L.; Fan, Z.; Huang, S. SVDD-based weighted oversampling technique for imbalanced and overlapped dataset learning. Inf. Sci. 2022, 588, 13–51. [Google Scholar] [CrossRef]
  25. Nekooeimehr, I.; Lai-Yuen, S.K. Adaptive semi-unsupervised weighted oversampling (A-SUWO) for imbalanced datasets. Expert Syst. Appl. 2016, 46, 405–416. [Google Scholar] [CrossRef]
  26. Li, Y.; Jia, X.; Wang, R.; Qi, J.; Jin, H.; Chu, X.; Mu, W. A new oversampling method and improved radial basis function classifier for customer consumption behavior prediction. Expert Syst. Appl. 2022, 199, 116982. [Google Scholar] [CrossRef]
  27. Batista, G.E.A.P.A.; Prati, R.C.; Monard, M.C. A study of the behavior of several methods for balancing machine learning training data. ACM SIGKDD Explor. Newsl. 2004, 6, 20–29. [Google Scholar] [CrossRef]
  28. Seiffert, C.; Khoshgoftaar, T.M.; Hulse, J.V.; Napolitano, A. RUSBoost: A hybrid approach to alleviating class imbalance. IEEE Trans. Syst. Man Cybern. 2010, 40, 185–197. [Google Scholar] [CrossRef]
  29. Xie, X.; Xie, B.; Xiong, D.; Hou, M.; Zuo, J.; Wei, G.; Chevallier, J. New theoretical ISM-K2 Bayesian network model for evaluating vaccination effectiveness. J. Ambient. Intell. Humaniz. Comput. 2023, 14, 12789–12805. [Google Scholar] [CrossRef] [PubMed]
  30. Hamidzadeh, J.; Moradi, M. Improved one-class classification using filled function. Appl. Intell. 2018, 48, 3263–3279. [Google Scholar] [CrossRef]
  31. Aljarah, I.; Faris, H.; Mirjalili, S. Optimizing connection weights in neural networks using the whale optimization algorithm. Soft Comput. 2018, 22, 1–15. [Google Scholar] [CrossRef]
  32. Bi, C.; Tian, Q.; Chen, H.; Meng, X.; Wang, H.; Liu, W.; Jiang, H. Optimizing a multi-layer perceptron based on an improved gray wolf algorithm to identify plant diseases. Mathematics 2023, 11, 3312. [Google Scholar] [CrossRef]
  33. Khan, A.U.; Bandopadhyaya, T.K.; Sharma, S. Comparisons of stock rates prediction accuracy using different technical indicators with backpropagation neural network and genetic algorithm based backpropagation neural network. In Proceedings of the International Conference on Emerging Trends in Engineering and Technology, Nagpur, India, 16–18 July 2008; IEEE: Piscataway, NJ, USA, 2008. [Google Scholar] [CrossRef]
  34. Kar, B.P.; Nayak, S.K.; Nayak, S.C. Opposition-based GA learning of artificial neural networks for financial time series forecasting. In Advances in Intelligent Systems and Computing, Proceedings of the International Conference on Computational Intelligence in Data Mining (ICCIDM), Bhubaneswar, India, 5–6 December 2015; Behera, H., Mohapatra, D., Eds.; Springer: New Delhi, India, 2015. [Google Scholar] [CrossRef]
  35. Malalur, S.S.; Manry, M.T. Multiple optimal learning factors for feed forward networks. In Proceedings of the Conference on Independent Component Analyses, Wavelets, Neural Networks, Biosystems, and Nanoengineering VIII, Orlando, FL, USA, 7–9 April 2010. [Google Scholar] [CrossRef]
  36. Lin, Z.; Akin, H.; Rao, R.; Hie, B.; Zhu, Z.; Lu, W.; Smetanin, N.; Verkuil, R.; Kabeli, O.; Shmueli, Y.; et al. Evolutionary-scale prediction of atomic-level protein structure with a language model. Science 2023, 379, 1123–1130. [Google Scholar] [CrossRef]
  37. Li, L.; Wang, Z.; Zhang, T. GBH-YOLOv5: Ghost convolution with BottleneckCSP and tiny target prediction head incorporating YOLOv5 for PV panel defect detection. Electronics 2023, 12, 561. [Google Scholar] [CrossRef]
  38. Xin, Y.; Ren, X. Predicting depression among rural and urban disabled elderly in China using a random forest classifier. BMC Psychiatry 2022, 22, 118. [Google Scholar] [CrossRef] [PubMed]
  39. Damaskinos, P.; Koletsi-Kounari, C.; Mamai-Homata, H.; Papaioannou, W. Social, clinical and psychometric factors affecting self-rated oral health, self-rated health and wellbeing in adults: A cross-sectional survey. Health 2022, 14, 104–124. [Google Scholar] [CrossRef]
  40. Thamrin, S.A.; Sidik, D.; Kuswanto, H.; Lawi, A.; Ansariadi, A. Exploration of obesity status of Indonesia basic health research 2013 with synthetic minority over-sampling techniques. Indones. J. Stat. Its Appl. 2021, 5, 75–91. [Google Scholar] [CrossRef]
  41. Xia, S.; Zheng, Y.; Wang, G.; He, P.; Li, H.; Chen, Z. Random space division sampling for label-noisy classification or imbalanced classification. IEEE Trans. Cybern. 2022, 52, 10444–10457. [Google Scholar] [CrossRef]
  42. Liu, Y.; Zheng, Y.F. FS_SFS: A novel feature selection method for support vector machines. Pattern Recognit. 2006, 39, 1333–1345. [Google Scholar] [CrossRef]
  43. Sun, L.; Kong, X.; Xu, J.; Xue, Z.; Zhai, R.; Zhang, S. A hybrid gene selection method based on ReliefF and ant colony optimization algorithm for tumor classification. Sci. Rep. 2019, 9, 8978. [Google Scholar] [CrossRef]
  44. Vasilyeva, M.; Tyrylgin, A. Machine learning for accelerating macroscopic parameters prediction for poroelasticity problem in stochastic media. Comput. Math. Appl. 2021, 84, 185–202. [Google Scholar] [CrossRef]
  45. Seiffert, C.; Khoshgoftaar, T.M.; Hulse, J.V. Hybrid sampling for imbalanced data. In Proceedings of the IEEE International Conference on Information Reuse and Integration, Las Vegas, NV, USA, 13–15 July 2008; IEEE: Piscataway, NJ, USA, 2008. [Google Scholar] [CrossRef]
  46. Pereira, R.M.; Costa, Y.M.G.; Silla, C.N., Jr. MLTL: A multi-label approach for the Tomek Link undersampling algorithm. Neurocomputing 2020, 383, 95–105. [Google Scholar] [CrossRef]
  47. Kaastra, I.; Boyd, M. Designing a neural network for forecasting financial and economic time series. Neurocomputing 1996, 10, 215–236. [Google Scholar] [CrossRef]
  48. Li, Y.; Chu, X.; Fu, Z.; Feng, J.; Mu, W. Shelf life prediction model of postharvest table grape using optimized radial basis function (RBF) neural network. Br. Food J. 2019, 121, 2919–2936. [Google Scholar] [CrossRef]
  49. Schwenk, H.; Beng, Y. Boosting neural networks. Neural Comput. 2000, 12, 1869–1887. [Google Scholar] [CrossRef] [PubMed]
  50. Fernandes, E.R.Q.; de Carvalho, A.C.P.L.F. Evolutionary inversion of class distribution in overlapping areas for multi-class imbalanced learning. Inf. Sci. 2019, 494, 141–154. [Google Scholar] [CrossRef]
  51. Luque, A.; Carrasco, A.; Martín, A.; Heras, A.D.L. The impact of class imbalance in classification performance metrics based on the binary confusion matrix. Pattern Recognit. 2019, 91, 216–231. [Google Scholar] [CrossRef]
  52. Krupinski, E.A. Receiver operating characteristic (ROC) analysis. Frontline Learn. Res. 2017, 5, 31–42. [Google Scholar] [CrossRef]
  53. Apolloni, J.; Leguizamón, G.; Alba, E. Two hybrid wrapper-filter feature selection algorithms applied to high-dimensional microarray experiments. Appl. Soft Comput. 2016, 38, 922–932. [Google Scholar] [CrossRef]
  54. Li, Y.; Chu, X.; Tian, D.; Feng, J.; Mu, W. Customer segmentation using K-means clustering and the adaptive particle swarm optimization algorithm. Appl. Soft Comput. 2021, 113, 107924. [Google Scholar] [CrossRef]
  55. Ge, L.; Ong, R.; Yap, C.W.; Heng, B.H. Effects of chronic diseases on health-related quality of life and self-rated health among three adult age groups. Nurs. Health Sci. 2019, 21, 214–222. [Google Scholar] [CrossRef]
  56. Akanni, L.; Lenhart, O.; Morton, A. Income trajectories and self-rated health status in the UK. SSM Popul. Health 2022, 17, 101035. [Google Scholar] [CrossRef]
  57. Duboz, P.; Boëtsch, G.; Gueye, L.; Macia, E. Self-rated health in Senegal: A comparison between urban and rural areas. PLoS ONE 2017, 12, e0184416. [Google Scholar] [CrossRef]
  58. Vafaei, A.; Yu, J.; Phillips, S.P. The intersectional impact of sex and social factors on subjective health: Analysis of the Canadian longitudinal study on aging (CLSA). BMC Geriatr. 2021, 21, 473. [Google Scholar] [CrossRef] [PubMed]
  59. Yang, F.; Zhang, J. Traditional Chinese sports under China’s health strategy. J. Environ. Public Health 2022, 2022, 1381464. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Importance of feature variables after ReliefF.
Figure 1. Importance of feature variables after ReliefF.
Mathematics 11 04985 g001
Figure 2. Basic structure of the MLP model.
Figure 2. Basic structure of the MLP model.
Mathematics 11 04985 g002
Figure 3. Accuracy of the model corresponding to the number of nodes in the hidden layers.
Figure 3. Accuracy of the model corresponding to the number of nodes in the hidden layers.
Mathematics 11 04985 g003
Figure 4. Bagging-based network model training architecture.
Figure 4. Bagging-based network model training architecture.
Mathematics 11 04985 g004
Figure 5. Change curves for different learning rate methods.
Figure 5. Change curves for different learning rate methods.
Mathematics 11 04985 g005
Figure 6. Flowchart of the SRH prediction model of older adults.
Figure 6. Flowchart of the SRH prediction model of older adults.
Mathematics 11 04985 g006
Figure 7. Comparison of the prediction results based on different classifiers.
Figure 7. Comparison of the prediction results based on different classifiers.
Mathematics 11 04985 g007
Figure 8. Prediction performances of different feature selection algorithms.
Figure 8. Prediction performances of different feature selection algorithms.
Mathematics 11 04985 g008
Figure 9. Comparison of the prediction performances of different oversampling algorithms based on MLP and ReliefF.
Figure 9. Comparison of the prediction performances of different oversampling algorithms based on MLP and ReliefF.
Mathematics 11 04985 g009
Figure 10. Distribution of data categories after oversampling based on WTASUWO.
Figure 10. Distribution of data categories after oversampling based on WTASUWO.
Mathematics 11 04985 g010
Figure 11. Comparison of the prediction performances of the improved classification algorithms based on WTASUWO.
Figure 11. Comparison of the prediction performances of the improved classification algorithms based on WTASUWO.
Mathematics 11 04985 g011
Figure 12. ROC curves with the adaptive learning rate based on IMLP using WTASUWO.
Figure 12. ROC curves with the adaptive learning rate based on IMLP using WTASUWO.
Mathematics 11 04985 g012
Figure 13. Friedman average order values of different algorithms.
Figure 13. Friedman average order values of different algorithms.
Mathematics 11 04985 g013
Figure 14. Comparison of the prediction performances of the proposed classification method based on the HCV dataset.
Figure 14. Comparison of the prediction performances of the proposed classification method based on the HCV dataset.
Mathematics 11 04985 g014
Figure 15. Comparison of the prediction performances of the proposed classification method based on the CM dataset.
Figure 15. Comparison of the prediction performances of the proposed classification method based on the CM dataset.
Mathematics 11 04985 g015
Figure 16. Comparison of the prediction performances of the proposed classification method based on the Rice dataset.
Figure 16. Comparison of the prediction performances of the proposed classification method based on the Rice dataset.
Mathematics 11 04985 g016
Figure 17. Comparison of the prediction performances of the proposed classification method based on the PS dataset.
Figure 17. Comparison of the prediction performances of the proposed classification method based on the PS dataset.
Mathematics 11 04985 g017
Table 1. Description of experimental data based on the 2020 CFPS.
Table 1. Description of experimental data based on the 2020 CFPS.
ModulesFeaturesDescriptions
Basic informationx1: Sex1 = male; 2 = female
x2: Age1 = 60–70; 2 = 70–80; 3 = 80 or above
x3: Education1 = no education; 2 = illiterate/semi-illiterate; 3 = primary school; 4 = junior high school; 5 = senior high school; 6 = junior college; 7 = bachelor’s degree or above
x4: Marriage status1 = unmarried; 2 = remarried; 3 = cohabiting; 4 = divorced; 5 = widowed
x5: Living area1 = urban; 2 = rural
x6: Family numbers1 = 1–3; 2 = 4–6; 3 = 7–9; 4 = 10 or above
x7: Household registration status1 = agricultural; 2 = non-agricultural; 3 = no household registration; 4 = resident; 5 = unknown
Working statusx8: Working condition1 = employed; 2 = unemployed; 3 = no longer labor force
x9: After-tax wage income in the past year1 = 0–1000; 2 = 1000–2000; 3 = 2000–3000; 4 = 3000–4000; 5 = 4000–5000; 6 = 5000 or above (RMB)
x10: Whether receiving pensions and retirement pay1 = yes; 2 = no; 3 = unknown
x11: Whether receiving old-age insurance1 = yes; 2 = no; 3 = unknown
Health statusx12: Health changes1 = worse; 2 = unchanged; 3 = better; 4 = unknown
x13: Any chronic diseases within six months1 = yes; 2 = no; 3 = unknown
x14: Where do you usually go to see a doctor1 = general hospitals; 2 = specialized hospitals; 3 = service centers; 4= service stations; 5 = clinics; 6 = other
Lifestylex15: Frequency of physical exercise (times)1 = less than 1 time per month; 2 = more than 1 time per month, but less than 1 time per week; 3 = 1–2 times per week; 4 = 3–4 times per week; 5 = 5 or more times per week; 6 = 1 time per day; 7 = 2 or more times per day; 8 = never attend; 9 = unknown
x16: Have you smoked in the past month1 = yes; 2 = no; 3 = unknown
x17: Have you had a drink three times a week in the past month1 = yes; 2 = no; 3 = unknown
x18: Whether to read1 = yes; 2 = no; 3 = unknown
Network servicesx19: Mobile Internet access1 = yes; 2 = no
x20: Whether surfing online1 = yes; 2 = no
x21: Whether playing video games1 = yes; 2 = no; 3 = unknown
x22: Whether doing shopping online1 = yes; 2 = no; 3 = unknown
x23: Whether watching short videos1 = yes; 2 = no; 3 = unknown
x24: Whether learning online1 = yes; 2 = no; 3 = unknown
x25: Whether using WeChat1 = yes; 2 = no; 3 = unknown
Network importancex26: Importance of network to work1~5, 1 = very unimportant, 5 = very important
x27: Importance of network to entertainment1~5, 1 = very unimportant, 5 = very important
x28: Importance of network to keeping in touch with family and friends1~5, 1 = very unimportant, 5 = very important
x29: Importance of network to studying1~5, 1 = very unimportant, 5 = very important
x30: Importance of network to daily life1~5, 1 = very unimportant, 5 = very important
Importance of access to informationx31: Television1~5, 1 = very unimportant, 5 = very important
x32: Internet1~5, 1 = very unimportant, 5 = very important
x33: Newspapers, periodicals, and magazines1~5, 1 = very unimportant, 5 = very important
x34: Broadcast1~5, 1 = very unimportant, 5 = very important
x35: Text messages1~5, 1 = very unimportant, 5 = very important
x36: Told by others1~5, 1 = very unimportant, 5 = very important
Table 2. Concept definitions of the key terms in this paper.
Table 2. Concept definitions of the key terms in this paper.
TermsConcepts
Older adultsRefers to adults aged 60 and above.
SRHRefers to an individual’s SRH status, including “unhealthy”, “average”, and “healthy”.
Feature selectionThe process of selecting a feature subset from the original features that are relevant and valuable to the current learning task.
UndersamplingA technique used to handle the imbalanced data by reducing the number of samples in the majority classes.
OversamplingA technique used to handle the imbalanced data by increasing the sample size of the minority classes.
Neural networkA computational model that simulates the neurons for processing and analyzing large amounts of complex data.
Class imbalanceRefers to the fact that some categories in a dataset have far more samples than others.
PredictionAn assessment of unknown health status based on existing data or information.
Machine learningRefers to the algorithms by which machines learn from large amounts of historical data and thus generate empirical models.
CFPSA long-term and large-scale survey project to collect socio-economic information on Chinese households and individuals.
ReliefFIt is a feature selection algorithm used to identify features related to a specific classification.
Tomek-linksIt is an undersampling method used to remove noise or duplicate samples by calculating the distance between samples.
ASUWOAn oversampling method for handling the imbalanced dataset by adaptively generating new samples in minority classes according to their weights.
MLPIt is a feed-forward ANN model that can simulate complex nonlinear mapping relationships.
BaggingAn ensemble learning method that constructs multiple base learners by randomly extracting different subsets of training data.
ClassifierAn algorithm or model used to classify data into two or more categories.
Table 3. Abbreviations of the key terms in this paper.
Table 3. Abbreviations of the key terms in this paper.
TermsAbbreviationsTermsAbbreviationsTermsAbbreviationsTermsAbbreviations
Self-rated healthSRHLogistic regressionLRRadial basis functionRBFRunning timeRT
China Family Panel StudiesCFPSK-nearest neighborKNNSelf-organizing mapSOMReceiver operating characteristicROC
Adaptive semi-unsupervised weighted oversamplingASUWODecision treeDTMultiple-layers perceptronMLPArea under ROCAUC
Weighted Tomek-links ASUWOWTASUWORandom forestRFStochastic gradient descentSGDExtreme learning machineELM
Improved multi-layer perceptionIMLPSupport vector machineSVMDeep learningDLCuckoo searchCS
Artificial intelligenceAIAdaptive boostingAdaBoostAdaptive momentAdamSimulated annealingSA
Neural networksNNsGradient boosting decision treeGBDTGenetic algorithmGAFirefly algorithmFA
Synthetic minority oversampling techniqueSMOTEExtreme gradient boostingXGBoostEvolutionary algorithmEAParticle swarm optimizationPSO
Adaptive syntheticADASYNArtificial neural networkANNDeep neural networksDNNsFalse positive rateFPR
Support vector data descriptionSVDDBack propagationBPMean square errorMSETrue positive rateTPR
Table 4. Prediction results of different base classifiers.
Table 4. Prediction results of different base classifiers.
ClassifiersAccuracyPrecisionRecallSpecificityAUCRT
LR0.5710.5040.5530.4320.5380.082
KNN0.5270.4470.4770.5640.6190.131
DT0.5490.4910.4560.5230.6410.042
RF0.4280.4280.4230.5650.6841.021
SVM0.5520.4880.5810.3550.7372.552
AdaBoost0.5780.5100.7570.4400.7420.504
GBDT0.5910.5250.6610.5360.72837.544
XGBoost0.5810.5160.6700.5110.7220.529
BP0.5750.5140.5640.5830.690128.548
ELM0.5880.5240.5990.5150.6690.051
MLP0.6020.5380.6600.6060.70110.178
Table 5. Prediction results of different feature selection algorithms based on MLP.
Table 5. Prediction results of different feature selection algorithms based on MLP.
Feature Selection MethodsAccuracyPrecisionRecallSpecificityAUCRT
Variance0.4740.4490.4870.1610.6982.165
Chi-squared0.4390.4390.4350.3210.6961.020
GA0.5540.4940.7160.4270.7202.818
Tree-based0.5740.5100.7440.4410.7341.856
Lasso0.5850.5220.6590.5280.7239.937
ReliefF0.5960.5270.7310.5670.7355.509
Table 6. Prediction results of different oversampling algorithms based on MLP and ReliefF.
Table 6. Prediction results of different oversampling algorithms based on MLP and ReliefF.
Sampling TechniquesAccuracyPrecisionRecallSpecificityAUCRT
Random undersampling0.5760.4650.5650.6820.5889.108
Random oversampling0.6480.5730.6810.7320.72619.355
SMOTE0.6420.5630.6760.7240.72415.540
BorderLine SMOTE0.6500.5750.6760.7370.72912.246
SVM SMOTE0.6550.5820.6870.7390.7358.351
Safe-level SMOTE0.7140.6700.7800.7820.79425.781
ASUWO0.7730.6670.6740.8130.78523.337
SMOTE Tomek0.6460.5700.6790.7300.72610.434
WCNN ASUWO0.6260.4390.4390.7190.69011.996
WTASUWO0.7770.6620.6750.8280.85218.543
Table 7. Prediction results of the improved classification algorithms based on WTASUWO.
Table 7. Prediction results of the improved classification algorithms based on WTASUWO.
Improvement MethodsAccuracyPrecisionRecallSpecificityAUCRT
CS-based ELM0.6710.5070.5110.7520.75011.580
SA-based ELM0.6430.4650.4740.7270.71323.901
FA-based ELM0.6510.4760.4820.7350.7274.785
PSO-based MLP0.6650.4980.5510.7230.75225.654
MLP using WTASUWO0.7770.6620.6750.8280.85218.543
IMLP using WTASUWO0.8550.7420.7450.8710.89229.970
Table 8. Friedman results between different algorithms and IMLP using WTASUWO.
Table 8. Friedman results between different algorithms and IMLP using WTASUWO.
CS-Based ELMSA-Based ELMFA-Based ELMPSO-Based MLPMLP Using WTASUWO
Significance0.012<0.001<0.0010.0280.252
Table 9. Information of the UCI datasets.
Table 9. Information of the UCI datasets.
Hepatitis C Virus (HCV) for Egyptian PatientsContraceptive Method (CM) ChoiceRice (Cammeo and Osmancik)Predict Students’ (PS) Dropout and Academic Success
Instances1385147338104424
Attributes289736
Classes4323
Distribution336/332/355/362629/333/5111630/21801421/794/2209
Imbalance ratio (IR)1.01:1:1.07:1.091.89:1:1.531:1.341.79:1:2.78
Subject areaLife scienceLife scienceComputer scienceOther
Distribution after WTASUWO210/219/207/207336/328/3131331/13481253/1282/1286
Attributes after WTASUWO108724
Table 10. Prediction results of the oversampling and classification algorithms based on the HCV dataset.
Table 10. Prediction results of the oversampling and classification algorithms based on the HCV dataset.
OversamplingMethodsAccuracyPrecisionRecallSpecificityAUCRT
No resamplingCS-based ELM0.6170.2460.2610.7350.5082.840
SA-based ELM0.6270.2520.2550.7500.49914.218
FA-based ELM0.6210.2480.2590.7400.4941.437
PSO-based MLP0.5290.2490.2920.5080.49843.911
MLP0.5100.2490.3790.5210.4840.162
IMLP0.6380.2840.3950.7520.60434.324
WTASUWOCS-based ELM0.6450.2850.2890.7620.6132.142
SA-based ELM0.6500.2930.2960.7660.6134.026
FA-based ELM0.6320.2630.2710.7500.5570.954
PSO-based MLP0.6370.2790.2920.7510.56832.699
MLP0.7040.4020.4050.8020.7132.848
IMLP0.7320.4640.4770.8170.7666.267
Table 11. Prediction results of the oversampling and classification algorithms based on the CM dataset.
Table 11. Prediction results of the oversampling and classification algorithms based on the CM dataset.
OversamplingMethodsAccuracyPrecisionRecallSpecificityAUCRT
No resamplingCS-based ELM0.5600.3960.4730.6080.6472.854
SA-based ELM0.5710.3980.4250.6500.6375.755
FA-based ELM0.5920.4290.4730.6570.6831.312
PSO-based MLP0.4700.3540.6180.3910.57433.743
MLP0.5760.4140.4960.6190.6690.719
IMLP0.6130.4550.4880.6810.7035.981
WTASUWOCS-based ELM0.6360.4570.5180.6950.7242.218
SA-based ELM0.6360.4520.4670.7200.7114.304
FA-based ELM0.6450.4660.4850.7250.7131.047
PSO-based MLP0.6250.4430.4900.6930.70721.489
MLP0.6720.5050.5050.7550.7533.870
IMLP0.7270.5880.6010.7900.8104.515
Table 12. Prediction results of the oversampling and classification algorithms based on the Rice dataset.
Table 12. Prediction results of the oversampling and classification algorithms based on the Rice dataset.
OversamplingMethodsAccuracyPrecisionRecallSpecificityAUCRT
No resamplingCS-based ELM0.8710.8670.8750.8670.9487.630
SA-based ELM0.8610.8620.8660.8560.94415.299
FA-based ELM0.8580.8620.8590.8560.9422.761
PSO-based MLP0.5070.5070.8790.8200.67057.994
MLP0.5160.5160.8280.8330.5000.114
IMLP0.8700.8710.8750.8690.9487.424
WTASUWOCS-based ELM0.8790.8780.8790.8790.9354.539
SA-based ELM0.8960.8960.8960.8960.9458.694
FA-based ELM0.8530.8520.8540.8520.9202.220
PSO-based MLP0.8430.8410.8460.8400.91439.383
MLP0.9210.9210.9020.9210.9605.647
IMLP0.9270.9270.9270.9270.9624.717
Table 13. Prediction results of the oversampling and classification algorithms based on the PS dataset.
Table 13. Prediction results of the oversampling and classification algorithms based on the PS dataset.
OversamplingMethodsAccuracyPrecisionRecallSpecificityAUCRT
No resamplingCS-based ELM0.7380.6200.5180.6870.6259.171
SA-based ELM0.7060.5890.5780.6610.60423.256
FA-based ELM0.7050.5850.5800.6450.6063.037
PSO-based MLP0.4020.3890.5520.5530.639107.070
MLP0.4020.3870.5370.5660.6282.373
IMLP0.7690.6740.6220.7670.74658.603
WTASUWOCS-based ELM0.7230.6580.5860.7910.7077.629
SA-based ELM0.7100.5960.5670.7810.70816.845
FA-based ELM0.7230.5820.5900.7880.6992.898
PSO-based MLP0.7160.5730.5760.7860.712114.901
MLP0.7440.6120.6300.8000.6976.406
IMLP0.7520.6820.6440.8060.77335.923
Table 14. Chi-squared analysis of key factors on the SRH of older adults (%).
Table 14. Chi-squared analysis of key factors on the SRH of older adults (%).
FeaturesGroupsUnhealthyAverageHealthy χ 2 p Values
Health changesworse42.6620.0137.331060.718<0.05
unchanged10.7813.8975.33
better15.0913.4871.43
unknown50.0050.000.00
Chronic diseasesyes45.4816.7837.75665.251<0.05
no16.2116.2567.54
unknown33.330.0066.67
After-tax wage income0–100027.2016.6156.19170.695<0.05
1000–200016.0024.0060.00
2000–300016.6720.8362.50
3000–400012.9022.5864.52
4000–500020.0015.0065.00
5000 or above9.4714.7075.83
Living areaurban20.3718.5261.1162.794<0.05
rural28.2114.3457.45
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Hu, Q.; Xie, G.; Chen, G. Prediction of the Health Status of Older Adults Using Oversampling and Neural Network. Mathematics 2023, 11, 4985. https://doi.org/10.3390/math11244985

AMA Style

Li Y, Hu Q, Xie G, Chen G. Prediction of the Health Status of Older Adults Using Oversampling and Neural Network. Mathematics. 2023; 11(24):4985. https://doi.org/10.3390/math11244985

Chicago/Turabian Style

Li, Yue, Qingyu Hu, Guilan Xie, and Gong Chen. 2023. "Prediction of the Health Status of Older Adults Using Oversampling and Neural Network" Mathematics 11, no. 24: 4985. https://doi.org/10.3390/math11244985

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop