Next Article in Journal
Shack-Hartmann Wavefront Sensing of Ultrashort Optical Vortices
Next Article in Special Issue
External Auditory Stimulation as a Non-Pharmacological Sleep Aid
Previous Article in Journal
Non-Linear Regression Models with Vibration Amplitude Optimization Algorithms in a Microturbine
Previous Article in Special Issue
Enhancing System Acceptance through User-Centred Design: Integrating Patient Generated Wellness Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Measuring Patient Similarity on Multiple Diseases by Joint Learning via a Convolutional Neural Network

1
Research Center of Electrical and Information Technology, Seoul National University of Science and Technology, Seoul 01811, Korea
2
Department of Industrial Engineering, Yonsei University, Seoul 03722, Korea
3
Department of Applied Artificial Intelligence, Seoul National University of Science and Technology, Seoul 01811, Korea
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(1), 131; https://doi.org/10.3390/s22010131
Submission received: 30 October 2021 / Revised: 22 December 2021 / Accepted: 23 December 2021 / Published: 25 December 2021

Abstract

:
Patient similarity research is one of the most fundamental tasks in healthcare, helping to make decisions without incurring additional time and costs in clinical practices. Patient similarity can also apply to various medical fields, such as cohort analysis and personalized treatment recommendations. Because of this importance, patient similarity measurement studies are actively being conducted. However, medical data have complex, irregular, and sequential characteristics, making it challenging to measure similarity. Therefore, measuring accurate similarity is a significant problem. Existing similarity measurement studies use supervised learning to calculate the similarity between patients, with similarity measurement studies conducted only on one specific disease. However, it is not realistic to consider only one kind of disease, because other conditions usually accompany it; a study to measure similarity with multiple diseases is needed. This research proposes a convolution neural network-based model that jointly combines feature learning and similarity learning to define similarity in patients with multiple diseases. We used the cohort data from the National Health Insurance Sharing Service of Korea for the experiment. Experimental results verify that the proposed model has outstanding performance when compared to other existing models for measuring multiple-disease patient similarity.

1. Introduction

With the ever-growing size and availability of electronic health records (EHRs), many applications on health-related studies become possible to analyze. Patient similarity analysis is of significant interest to researchers because it can be applied to various health research fields, such as precision medicine [1]. It establishes a group of similar patients by measuring the distance between patients and predicts the target patient’s condition by the common phenotype of the similar cohort. With the right patient similarity network built from large-scale data in place, physicians can retrieve a cohort of similar patients for a target patient based on the case of study, make medical comparisons, and, thereafter, make effective personalized treatment plans [2]. Patient similarity analysis is one way to utilize EHRs and facilitate integration from the whole database for data-driven medical decision-making.
EHRs are a complex collection of various medical concepts that can be broadly divided into structured and unstructured data. Unstructured data have no fixed structure, and are usually composed of text, images, and signals. At the same time, structural data are stored in a table, which is more applicable for computer processing than unstructured data. EHRs are also hierarchical, heterogeneous, and sequential. ‘Hierarchical’ means that there may be multiple medical concepts under each observation. For example, multiple medications and multiple procedures may occur under a visit of a patient’s profile. ‘Heterogeneous’ means that the difference between each patient’s profile can be huge, as the frequency of each patient’s visit to the hospital and the medical events under each visit may vary widely. Lastly, sequential means that the patient’s condition changes with time. That is, the observation results are different at different times. These challenges described above often lead to difficulties in utilizing EHRs.
Converting complex, irregular, and sequential EHRs into the appropriate form is the first step in the health-related research by patient data representation. Due to insufficient domain knowledge, statistics are often used to evaluate patient information. For example, the frequency-inverse document frequency (TF-IDF) method can evaluate the importance of clinical concepts in medical records. The importance of the concept increases proportionally with the number of times it occurs in the patient’s medical records, but at the same time decreases, as opposed to the frequency it occurs in other patients’ medical records. Reference [3] used a vector consisting of the international classification of diseases (ICD)-10 codes to represent individual patients, and assigned weights to ICD-10 codes using the TF-IDF method. References [4,5] features were extracted from diagnostic codes, drug components, and lab tests with TF-IDF methods, and similarity was measured using Euclidean distances. However, the above studies did not take into account the sequential properties of obtaining the patient presentation. With the development of deep learning, which is good at longitudinal data processing, several studies have conducted patient presentations using neural networks. Reference [6] used auto-encoder as a deep learning framework for the patient representation. References [7,8] established a vector representation for medical concepts from longitudinal EHRs, and a vector representation of each visit is obtained from the sum of the vectors of medical concepts arising from that visit. This vector representation also considers two levels of embeddings: the medical concept level and the visit level. Not only does this work enhance the effectiveness of the prediction task, but it is also meant to apply the learned interpretable representation to other problems. Reference [9] proposes a model that uses a new time-aware long short-term memory (LSTM) module to process irregularly spaced longitudinal EHRs and obtain effective patient representation by capturing independencies in the sequence.
The purpose of patient similarity is to derive appropriate similarity between patient pairs based on EHRs. Well-measured similarity can be applied to various medical fields, including personalized medicine, disease prediction, and cohort analysis [10,11,12,13,14]. Recently, patient similarity studies based on supervised learning have been conducted to use feedback information from medical professionals. The purpose of supervised patient similarity learning is to learn patient similarity more accurately using expertise as feedback. Reference [15] applies the large-scale machine learning (LSML) algorithm to measure patient similarity and proposes a local supervised similarity learning algorithm. The algorithm identifies homogeneous and heterogeneous neighbors simultaneously, using expert-given labels for each pair of patients as supervised information. Reference [16] proposed a patient similarity framework that introduces group lasso into the objective function to obtain a low-dimensional sparse similarity matrix. This low-ranking mapping avoids the drawbacks of similarity learning algorithms, which have relatively high computational costs. Reference [17] proposed the uncorrelated patient similarity learning framework using maximum likelihood estimation to obtain parameters of similarity functions. In addition, two regular entries are introduced to implement feature selection independent of sparse feature selection. Reference [18] proposed multi-task sparse metric learning to measure patient similarity, a multi-task triple constraint sparse metric learning framework. Optimization is based on triple constraints, which bring similar patients closer and distance other speakers to achieve metric learning.
Recently, studies have been conducted to learn patient similarity measurements and obtain patient presentation considering the temporal properties of longitudinal EHRs. Reference [19] represented the longitudinal EHRs of each patient by a matrix constructed by stacking equal-length embedded vector representations of each visit as the input to the network, which preserves the expression of the temporal characteristics of the longitudinal data. Based on the structure composed of bi-directional convolutional neural networks (CNN) and supervised metric learning, a matching matrix is used to calculate the similarity between the two embedded vectors derived from the deep network. Reference [20] proposed a time-fusion CNN framework that preserves the local temporal relationship and obtains a global contribution of different time intervals for patient similarity measurement. The work also made predictions for target patients according to the phenotype of the k-nearest patients. Reference [21] proposed a CNN-based triple patient similarity learning framework to learn a margin that separates similar and dissimilar patients better.
However, most similarity measurement studies have only studied single or specific diseases. It is not realistic to consider only one specific disease, and it ignores the characteristics of the accompanying disease. Therefore, a similarity measurement study for patients with multiple diseases is needed. Recently, there are studies dealing with multiple diseases patient similarity [22,23], but still need more participation from the researchers. To measure the similarity of patients with multiple diseases, we proposed a CNN-based model that jointly learns identity representation and measures the similarity of patients with multiple diseases. The proposed model consists of two parts. The former is feature learning, which classifies the input data as having any disease and gives them an identity presentation. The latter part is similarity learning, which measures similarity between patients. We trained these two parts simultaneously to measure the similarity of patients with multiple diseases. Furthermore, we compared the similarity score between our proposed ‘joint learning’ model with feature learning and similarity-based learning models to validate the performance. Moreover, we verified the performance by classification evaluation metrics to test the performance of proposed joint learning.

2. Materials and Methods

This section explains how the patient’s demographic information, diseases, and prescription was represented and how to measure the similarity between patients with multiple diseases. Lastly, a CNN-based model, which applies feature learning and similarity learning, was introduced. All clinical processes were discussed and validated by a doctor from Seoul Red Cross hospital in Seoul, South Korea.

2.1. Patient Record Representation

The purpose of this section is to define the input to the model. As shown in Figure 1, raw EHRs are sequential and irregular. We need to transform the raw EHRs to numeric form to process through our proposed model. The most basic method is to use one-hot encoding. Since EHRs are longitudinal data, the temporal property should be considered when obtaining patient representation; therefore, we used matrix to represent the patient’s medical record, which was obtained by stacking the same length of embedded vector for each visit. In our research, each visit contains demographic information, such as gender and age group, diagnosis codes, and prescription codes.
Through Doc2Vec, we can obtain the visit level embedding vector and same dimensional vector representations for patients’ historical records [24]. The patient p can be represented by a matrix P with size e × v , which e is the embedding size and v is the number of visits. We used paragraph vector with the distributed memory (PV-DM) model of Doc2vec to obtain patient representations, used demographics, disease, and treatment information from EHRs, and compared the embedding dimensions by setting 128 and 256. We confirmed that the patient’s representation performance was better at 128 dimensions and that each patient’s visit was converted to 128 dimensions vector. Furthermore, to equalize the number of visits to every patient, we set the maximum number of visits to 200, and applied zero padding. Consequently, a patient representation matrix of 200 × 128 were obtained.

2.2. Self-Attention Module

The objective of the attention mechanism is to select more important information from that task, and it is derived from the human visual attention mechanism. Restricted attention to capturing useful information significantly improves efficiency and accuracy. Attention mechanisms combined with CNN and recurrent neural networks (RNN) structures have achieved superior performance in natural language processing and computer vision.
In our research, 200 × 128 size matrix representation of the patient history records can be converted into a vector by convolution and the pooling layer through the above method. However, although the basic convolution operation can capture the temporal information, there is a limitation that the information is equally treated during this process. To further improve the ability of patient vector representation, we introduce an attention mechanism in the convolutional layer in the network.

2.3. Patient Labeling for Similarity Learning

2.3.1. Feature Learning Label

We considered three diseases—diabetes, cerebrovascular disease, and ischemic heart disease, which may cause cascaded disease occurrence to each other. We used feature learning to obtain the identity representation of patients embedded with a matrix, and feature learning needs a label for supervised learning. We considered patients’ age, gender, blood pressure (BP), body mass index (BMI), and disease codes, and then those factors are represented as one-hot encoding. For continuous measurements, such as age, blood pressure, and BMI, we divide them into three groups, to express with one-hot encoding, as shown in Table 1.

2.3.2. Similarity Learning Label

Similarity learning measures similarity based on distance and needs a label as supervised learning. We use a patient’s age, gender, BP, BMI, and disease information as the patient’s label to measure the similarity. We generate similarity labels by comparing feature categories among seven categories, (age, gender, BP, BMI, diabetes, cerebrovascular disease, ischemic heart disease) if there are more than equal to four categories (median of 7) match between patients, we label them “1”, as similar patients, otherwise “0”, as shown in Table 2.

2.4. Patient Similarity Measurement

2.4.1. CNN Based Framework

CNN has superior advantages in feature extraction due to its strong representation capabilities. Feature extraction is performed on local information via convolutional operations. Complete information can be obtained by integrating local information. The feature detection layer of CNN avoids explicit feature extraction, but implicitly learns from the training data. Furthermore, the complexity of the network is reduced due to the sharing of local weights, which is also a significant advantage of convolutional networks for fully connected networks.
This section will show how to obtain the embedded vector representation of a patient’s history matrix via a CNN structure. Feature extraction is first performed via convolutional layers. Our goal is to explore relationships between accesses in the patient matrix, so we use 1D filters to perform convolutional tasks on the access dimensions of patient matrix representations. More specifically, information from matrix history maps is extracted using different filters, and the parameters of each filter are obtained through network optimization.
For each filter, w h × e , where h represents the size of the filter, which is the number of consecutive visits used to generate features in the convolution operation. Where e is the embedded dimension for each access defined in the previous work. By performing a feature extraction with temporal meaning through a convolution operation, we have Equation (1), where * represents the convolution operation, b represents the error term, P i : i + h 1 represents the concatenation of the i t h to the i + h 1 t h visit vectors, and the result c i represents a feature obtained after the convolution operation of P i : i + h 1 .
c i = R e L U w * P i : i + h 1 + b
Applying the filter to the entire patient matrix representation with a stride of 1, we can obtain v + h 1 features, where v is the number of visits. Through these features, we can get a feature map c as Equation (2).
c = c 1 , c 2 , c 3 , , c v h + 1
The pooling layer reduces the obtained feature map. Here, we use max pooling, and the purpose is to extract the most essential information in the feature map to get a reduced feature map, which can reduce the computational complexity and make the feature representation more robust and avoid overfitting. We obtained a feature map through convolution operations of the data of embedded patients represented by matrix. We used Conv1D because our data have sequential properties. In addition, Maxpooling1D was used to obtain the largest value while reducing dimensions on the obtained feature map. Since our EHRs has sequential features, we added a self-attention module to effectively extract information based on its sequential features and patient historical records. A CNN block computes Conv1D, Maxpooling1D, and self-attention, in order. Finally, three CNN blocks result in vector presentation of the patient.
In each CNN block, as shown in Figure 2, three convolutional operations and pooling operations are performed on the input feature map in a parallel manner. All of the output is joined into a relative deep feature map. Because different convolution operations and pooling operations can obtain additional information of the input feature map, parallel, conducting these operations, and combining all the outputs, will get better feature representation. The different convolutional layers aim to extract various features, and max-pooling layers are used to reduce the intermediate representation. Then, the feature maps derived from different branches are concatenated together as a feature map for the following step processing.

2.4.2. Feature Learning

We employ a sigmoid classifier on top of the base network for feature learning to learn patient identity representation [25]. With a patient embedded matrix into the CNN network, it extracts patient information. Moreover, identity presentation is obtained through feature learning. Feature learning labels are used to supervise the categorization training. We use sigmoid to obtain probability values for each label’s output to classify the patient’s disease and then use binary cross-entropy as a loss function as Equation (3).
L c l a s s i f y y , y ˜ = i y i * log y i ˜ + 1 y i * log 1 y i ˜
where y represents the true label and y ˜ represents the output of the network. The y i represents the i t h label in the multiple labels and the y ˜ i indicates the i t h output probability that the patients pairs share the i t h disease. The feature-learning model is shown in Figure 3.

2.4.3. Similarity-Based Learning

Appropriate metrics derive the patient similarity. With a patient embedded matrix into the CNN network, it extracts the patient’s information. Thus, two vectors are obtained. Both vectors are 800-dimensional vectors, which obtain 800-dimensional vectors through vector-based Euclidean distance calculations. To determine whether the patient is similar to another patient, we measure the similarity by the contrastive loss function. This loss function learns to be similar if the values from the distance calculation of both data are small (contrarily, if the values are large). Thus, ultimately, if we calculate contrastive loss, each disease results in a similarity score. We compare similarity measurement performances with this similarity score.
We adopt contrastive loss in [26] as a loss function to measure similarity after computing the vector-based Euclidean distance. This loss function Equation (4) was applied for each disease. Y is a binary label with 1 if the two inputs X i a , X i b are similar, otherwise 0. e i is a vector-based Euclidean calculation of two feature representations extracted from input data via CNN. This loss function learns that if the inputs are similar, the following formulations are removed to make the values smaller, and if they are not similar, the preceding formulations are removed to make the values bigger.
L s i m i l a r i t y e i , Y , X a , X b i = Y 1 2 e i 2 + 1 Y 1 2 max 0 , m a r g i n e i 2

2.5. Joint Similarity Learning Architecture

Figure 4 shows joint learning structure transforms embedded matrix representations containing temporal information from patient records into vector representations and learns two methods simultaneously through transformed vector representations, which we draw ideas from [27]. The feature extraction part of this CNN-based network is learned by sharing weights. For each CNN block, we perform convolutional operations to extract the input features and use a max-pooling layer to reduce the intermediate representation. We then introduce a self-attention module for each CNN block to improve the ability of the representation. Finally, we classify what diseases the patient has, and then learn the similarity for each disease after calculating the distance of the two vectors, and optimize them by summing their respective loss functions Equation (5).
L t o t a l = L c l a s s i f y a + L c l a s s i f y b + L s i m i l a r i t y
The algorithm of joint learning is shown as Algorithm 1.
Algorithm 1 Joint Learning
Input: P: patient data, D: Doc2vec matrix, L: dataset true labels
Output: patient similarity score on test dataset
for i in dataset do
  let P i be the patient matrix of i
  for j in i do
    V j <- vectorize(j,D)
   append V j   to   P i
P t r a i n ,   P t e s t ,   L t r a i n ,   L t e s t <- split train subset and test subset
  S <- CNN Joint Learning ( [ P t r a i n _ a ,   P t r a i n _ b ] , [ L t r a i n _ a ,   L t r a i n _ b ])
 score <- evaluate ( i ,   [ L t e s t _ a , L t e s t _ b ], S)
return score

2.6. Data Descriptions

The data are provided by National Health Insurance Service of Korea and covers 12 years of EHRs. We filter the visit history related to diabetes (ICD-10 codes of E10-E14), cerebrovascular disease (ICD-10 codes of I61-I69), and ischemic heart disease (ICD-10 codes of I20-I25). The description of the collected dataset is shown in Table 3 and Figure 5.
According to the patient’s diagnostic information, classify them in the order of visits with the same patient identification number, and then extract diagnostic information, demographics, to form an EHR set. Since extreme data can bias, the 99th quartile of each medical event is checked, and the data are removed if they have more than 12 diagnostics, and 200 visits.

2.7. Handling Data Imbalance

To deal with the data imbalance issue, we applied the data augmentation for imbalance multi-label data, which is known as multi-label synthetic minority over-sampling (MLSMOTE) [28]. MLSMOTE is one of the most popular and effective data augmentation techniques in the case of multi-label classification. As the name suggests, it is an extension or variant of the synthetic minority over-sampling technique (SMOTE) [29]. In SMOTE, we provide data and augment it to generate more samples of the same class from which the reference point has been chosen, but it failed in a multi-label situation because the instances of the data had many labels attached with it. Therefore, there is a chance that a sample with a minority label could also have a majority label, so we will have to construct labels for the synthetic data as well. Labels in the majority were referred to as head labels, while labels in the minority were referred to as tail labels in multi-label contexts [30].

2.8. Model Parameters

First, model training allows us to obtain optimized parameters in our network. Our deep learning parameters are set as follows: the width of the convolution filter is 3, and the number of convolution filters takes on 128, 64, and 32. Max-pooling and self-attention modules are added after each convolution operation. We added a dropout normalization with a dropout rate of 0.3 after the max-pooling layer to overcome the overfitting problem. In the feature extraction part, the last feature extraction layer, where the vector representation size was 800. The batch size was 128 and epochs with 100, and the early stop was used to prevent overfitting. Moreover, 100,000 patient pairs as experimental data with the ratio of training and test sets was 8:2. The performance of the test set proves that our model can improve the ability of patient similarity learning. The model was implemented by TensorFlow-Keras and optimized by Adam [31].

3. Results and Discussions

3.1. Validation of Proposed Joint Learning Performance

To validate the performance of joint learning, we compared the performance according to each learning approach: similarity-based learning and joint learning. Both models were trained based on CNN. To test the performance of similarity-based learning, we measured similarity, except feature learning. We measured the similarity between every combination of two diseases and three diseases. The results of similarity-based learning are shown in Table 4.
Similarity-based learning measures similarity without feature learning and learns without identity presentation. The maximum similarity scores of each combination of diseases seem to be high, but the mean of similarity scores are below 85%. To validate the impact of joint learning, we showed the similarity score results in Table 5.
The similarity score of joint learning increased by 0.084 on average than similarity-based learning. Moreover, the mean of the similarity score of every combination exceeded 90%, except the diabetes and cerebrovascular disease combination, but almost reached 90%. These results verified that feature learning effectively expresses patient information. Therefore, our proposed joint learning improved measuring patient similarity with multiple diseases.

3.2. Validation of Joint Learning Model by Classification Evaluation Metrics

This section compares the performance of three different learning approaches: feature learning, similarity-based learning, and joint learning by classification evaluation metrics. In feature learning, we follow the procedure of base CNN network training as a categorization classifier. Similarity-based learning is a model that removes the categorization classifier. Finally, joint learning is a method of combining the two models. Since we transform the similarity learning problem into a supervised binary classification problem, we compare the proposed model with feature learning and similarity learning performance in binary classification tasks. We selected accuracy, precision, recall, and F1 scores as performance metrics from classification evaluation. Classification metrics are calculated from true positives, false positives, false negatives, and true negatives. Accuracy is one metric, which gives the fraction of predictions our model got right. Precision gives the fraction of correctly identified as positive out of all predicted as positives. Recall gives the fraction that correctly identified as positive out of all positives. Lastly, the F1 score is the harmonic mean of the model’s precision and recall because it is not sensitive to extremely large values, unlike simple average [32]. The results are shown in Figure 6.
In Figure 6, as we can observe, our proposed joint learning model performed better in accuracy, precision, recall, and F1 score, about 20%, than applying feature learning and similarity-based learning. With these results, we validated that using feature learning alone does not measure similarity properly. Moreover, by using similarity-based learning alone, it poorly classifies embedded data for each patient.

3.3. Comparison between Existing Similarity Measurement Algorithms

To validate the performance of the proposed joint learning algorithm, we include another four existing distance metric algorithms for comparison purposes. The considered existing algorithms are Euclidean distance, Cosine similarity, Mahalanobis distance [33], and localized supervised metric learning (LSML) [34]. The performance comparison results by classification evaluation metrics of five algorithms are shown in Table 6.
The Euclidean distance and Cosine similarity algorithms are basic metrics for measuring patients’ similarity. Those algorithms directly measure similarity without learning parameters and, therefore, the results are low. Moreover, Mahalanobis and LSML results are higher than Euclidean and Cosine, but as we can observe, the proposed joint learning algorithm has the highest scores among all algorithms.

4. Conclusions, Limitations, and Future Works

Patient similarity measurement research is a fundamental study for application in various medical fields. Appropriate similarity measurements are required for application to other healthcare areas. Patients’ historical records, particularly those of patients with chronic diseases, rely heavily on temporal information. It is important to obtain a vector representation of the patient’s historical records, with temporal properties to capture information in the historical patient records, and measure the similarity between patients. Patient similarity learning is based on learning a metric matrix that may effectively quantify the similarity between patients using paired constraints. Compared to the primary distance metric, patient similarity learning can change the patient representation into a new characteristic space based on the learned matching matrix, resulting in closer distances between similar patients and further distances between different patient samples.
Existing studies measure the similarity of single diseases well, but measuring similarity with other diseases still needs more research. To address the limitations of existing similarity studies that consider only one disease, we propose a model to measure multiple disease similarities by jointly combining feature learning and basic similarity learning. This research verified that our proposed joint learning model improves the similarity measurement performance of multiple diseases by over-performing among other existing similarity measurement algorithms. Therefore, we validated that the proposed model is suitable for measuring the patient similarity of multiple diseases.
The limitation of our research is the population of the cohort database because we only dealt with one kind of medical database, which the National Health Institute of Korea provides. As shown in Figure 5, there are only a few records of diabetes, cerebrovascular disease, and ischemic heart diseases. For those three diseases, as well as other diseases, there are an unsatisfying number of records available, unless the disease is general, such as hypertension. It was challenging to find insight and to measure similarities between patients with the small dataset.
In future works, we would like to search and acquire other electronic health records by collaborating with the research center of hospitals in Korea or databases from other countries, such as Medical Information Mart for Intensive Care (MIMIC)-III to expand the population number and verify our proposed model. Moreover, we will consider more appropriate and adequate variables, such as more specific body check measurements to strengthen our concept of the similarity measurement. Furthermore, we will extend our research to disease prediction and precision medicine by applying our similarity measurement model to broaden the research on healthcare applications.

Author Contributions

S.H.O. and S.B. conceived and conducted the experiments. J.P. verified and advised the experimental outcome as a medical doctor. All authors have read and agreed to the published version of the manuscript.

Funding

This research is result of a study on the “Convergence and Open Sharing System” Project grant number [COSS-2021-A1-01], supported by the Ministry of Education and National Research Foundation of Korea and supported by the Research Program funded by the SeoulTech (Seoul National University of Science and Technology).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board (IRB) of Seoul National University of Science and Technology (IRB no. 2021-0030-01).

Informed Consent Statement

Patient consent was waived as the research type only dealt with gathered data from the National Health Insurance Sharing Service. The authors did not collect any additional patient information.

Data Availability Statement

The datasets generated and/or analyzed in the current study are not publicly available due to patient information collected by the National Health Insurance Sharing Service, which requires payment for access. However, sample data are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Sharafoddini, A.; Dubin, J.; Lee, J. Patient Similarity in Prediction Models Based on Health Data: A Scoping Review. JMIR Med. Inform. 2017, 5, e7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Dai, L.; Zhu, H.; Liu, D. Patient similarity: Methods and applications. arXiv 2020, arXiv:2012.01976. [Google Scholar]
  3. Roque, F.; Jensen, P.; Schmock, H.; Dalgaard, M.; Andreatta, M.; Hansen, T.; Søeby, K.; Bredkjær, S.; Juul, A.; Werge, T.; et al. Using Electronic Patient Records to Discover Disease Correlations and Stratify Patient Cohorts. PLoS Comput. Biol. 2011, 7, e1002141. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Wang, F. Medical prognosis based on patient similarity and expert feedback. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012), Tsukuba, Japan, 11–15 November 2012. [Google Scholar]
  5. Wang, F.; Sun, J. PSF: A Unified Patient Similarity Evaluation Framework through Metric Learning with Weak Supervision. IEEE J. Biomed. Health Inform. 2015, 19, 1053–1060. [Google Scholar] [CrossRef] [PubMed]
  6. Miotto, R.; Li, L.; Kidd, B.; Dudley, J. Deep Patient: An Unsupervised Representation to Predict the Future of Patients from the Electronic Health Records. Sci. Rep. 2016, 6, 26094. [Google Scholar] [CrossRef] [PubMed]
  7. Choi, E.; Bahadori, M.; Searles, E.; Coffey, C.; Thompson, M.; Bost, J.; Tejedor-Sojo, J.; Sun, J. Multi-layer Representation Learning for Medical Concepts. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2016, San Francisco, CA, USA, 13–17 August 2016. [Google Scholar]
  8. Bajor, J.; Mesa, D.; Osterman, T.; Lasko, T. Embedding Complexity in the Data Representation instead of in the Model: A Case Study Using Heterogeneous Medical Data. arXiv 2018, arXiv:1802.04233v1. [Google Scholar]
  9. Baytas, I.; Xiao, C.; Zhang, X.; Wang, F.; Jain, A.; Zhou, J. Patient Subtyping via Time-Aware LSTM Networks. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2017, Halifax, NS, Canada, 13–17 August 2017. [Google Scholar]
  10. Gupta, V.; Sachdeva, S.; Bhalla, S. A Novel Deep Similarity Learning Approach to Electronic Health Records Data. IEEE Access 2020, 8, 209278–209295. [Google Scholar] [CrossRef]
  11. Cheng, Q.; Gu, X. Deep attentional fine-grained similarity network with adversarial learning for cross-modal retrieval. Multimed. Tools Appl. 2020, 79, 31401–31428. [Google Scholar] [CrossRef]
  12. Ahmad, A.; Fröhlich, H. Towards clinically more relevant dissection of patient heterogeneity via survival-based Bayesian clustering. Bioinformatics 2017, 33, 3558–3566. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Planey, C.; Gevaert, O. CoINcIDE: A framework for discovery of patient subtypes across multiple datasets. Genome Med. 2016, 8, 27. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Parimbelli, E.; Marini, S.; Sacchi, L.; Bellazzi, R. Patient similarity for precision medicine: A systematic review. J. Biomed. Inform. 2018, 83, 87–96. [Google Scholar] [CrossRef] [PubMed]
  15. Sun, J.; Wang, F.; Hu, J.; Edabollahi, S. Supervised patient similarity measure of heterogeneous patient records. ACM SIGKDD Explor. Newsl. 2012, 14, 16–24. [Google Scholar] [CrossRef]
  16. Zhan, M.; Cao, S.; Qian, B.; Chang, S.; Wei, J. Low-Rank Sparse Feature Selection for Patient Similarity Learning. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining (ICDM), Barcelona, Spain, 12–15 December 2016. [Google Scholar]
  17. Huai, M.; Miao, C.; Suo, Q.; Li, Y.; Gao, J.; Zhang, A. Uncorrelated Patient Similarity Learning. In Proceedings of the 2018 SIAM International Conference on Data Mining; Society of Industrial and Applied Mathematics: Philadelphia, PA, USA, 2018; pp. 270–278. [Google Scholar]
  18. Suo, Q.; Zhong, W.; Ma, F.; Ye, Y.; Huai, M.; Zhang, A. Multi-task Sparse Metric Learning for Monitoring Patient Similarity Progression. In Proceedings of the 2018 IEEE International Conference on Data Mining (ICDM), Singapore, 17–20 November 2018. [Google Scholar]
  19. Suo, Q.; Ma, F.; Yuan, Y.; Huai, M.; Zhong, W.; Zhang, A.; Gao, J. Personalized disease prediction using a CNN-based similarity learning method. In Proceedings of the 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Kansas City, MO, USA, 13–16 November 2017. [Google Scholar]
  20. Suo, Q.; Ma, F.; Yuan, Y.; Huai, M.; Zhong, W.; Gao, J.; Zhang, A. Deep Patient Similarity Learning for Personalized Healthcare. IEEE Trans. NanoBioscience 2018, 17, 219–227. [Google Scholar] [CrossRef] [PubMed]
  21. Zhu, Z.; Yin, C.; Qian, B.; Cheng, Y.; Wei, J.; Wang, F. Measuring Patient Similarities via a Deep Architecture with Medical Concept Embedding. In Proceedings of the 2016 IEEE 16th International Conference on Data Mining (ICDM), Barcelona, Spain, 12–15 December 2016. [Google Scholar]
  22. Zhao, F.; Xu, J.; Lin, Y. Similarity Measure for Patients via a Siamese CNN Network. In Algorithms and Architectures for Parallel Processing; Springer Nature: Basingstoke, UK, 2018; pp. 319–328. [Google Scholar]
  23. Brown, S. Patient Similarity: Emerging Concepts in Systems and Precision Medicine. Front. Physiol. 2016, 7, 561. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Le, Q.; Mikolov, T. Distributed Representations of Sentences and Documents. In Proceedings of the 31st International Conference on Machine Learning, Beijing, China, 21–26 June 2014; pp. 1188–1196. [Google Scholar]
  25. Woo, S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–19. [Google Scholar]
  26. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality Reduction by Learning an Invariant Mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2. [Google Scholar]
  27. Alfasly, S.; Hu, Y.; Li, H.; Liang, T.; Jin, X.; Liu, B.; Zhao, Q. Multi-Label-Based Similarity Learning for Vehicle Re-Identification. IEEE Access 2019, 7, 162605–162616. [Google Scholar] [CrossRef]
  28. Charte, F.; Rivera, A.; del Jesus, M.; Herrera, F. MLSMOTE: Approaching imbalanced multilabel learning through synthetic instance generation. Knowl. -Based Syst. 2015, 89, 385–397. [Google Scholar] [CrossRef]
  29. Chawla, N.; Bowyer, K.; Hall, L.; Kegelmeyer, W. SMOTE: Synthetic Minority Over-sampling Technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  30. Sukhwani, N. Handling Data Imbalance in Multi-label Classification (MLSMOTE). Available online: https://medium.com/thecyphy/handling-data-imbalance-in-multi-label-classification-mlsmote-531155416b87 (accessed on 1 December 2021).
  31. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. Available online: https://arxiv.org/abs/1412.6980 (accessed on 12 August 2021).
  32. Gulati, M. How to Choose Evaluation Metrics for Classification Models; Analytics Vidhya: Indore, India, 2020. [Google Scholar]
  33. De Maesschalck, R.; Jouan-Rimbaud, D.; Massart, D. The Mahalanobis distance. Chemom. Intell. Lab. Syst. 2000, 50, 1–18. [Google Scholar] [CrossRef]
  34. Sun, J.; Sow, D.; Hu, J.; Ebadollahi, S. Localized Supervised Metric Learning on Temporal Physiological Data. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010. [Google Scholar]
Figure 1. An Illustration of EHRs.
Figure 1. An Illustration of EHRs.
Sensors 22 00131 g001
Figure 2. CNN block.
Figure 2. CNN block.
Sensors 22 00131 g002
Figure 3. An Illustration of Feature Learning.
Figure 3. An Illustration of Feature Learning.
Sensors 22 00131 g003
Figure 4. An illustration of proposed similarity learning.
Figure 4. An illustration of proposed similarity learning.
Sensors 22 00131 g004
Figure 5. Population distribution of diabetes, cerebrovascular disease, and ischemic heart disease.
Figure 5. Population distribution of diabetes, cerebrovascular disease, and ischemic heart disease.
Sensors 22 00131 g005
Figure 6. Performance of feature learning, similarity learning joint learning.
Figure 6. Performance of feature learning, similarity learning joint learning.
Sensors 22 00131 g006
Table 1. Group division for continuous measurement.
Table 1. Group division for continuous measurement.
AgeBPBMI
Group 10–30PrehypertensionUnderweight
Group 231–60Stage 1Normal
Group 3Over 61Stage 2Overweight
In BP, the prehypertension range is 120–139 mmHg of systolic blood pressure (SBP) and 80–89 mmHg of diastolic blood pressure (DBP). Stage 1 range is SBP of 140–159 mmHg and 90–99 mmHg of DBP. Lastly, the range of stage 2 is SBP over 160 mmHg and DBP over 100 mmHg. In BMI, the underweight group range is below 18.5 kg/m2. Normal group range is between 18.5 kg/m2 to 22.9 kg/m2. The overweight group is composed of BMI more than 23 kg/m2.
Table 2. Example of similarity labeling.
Table 2. Example of similarity labeling.
FeaturesPatient APatient BMatchNo. of Match
Age0–300004
31–60111
Over 61000
GenderMale111
Female000
BPPrehypertension000
Stage 1100
Stage 2010
BMIUnderweight000
Normal000
Overweight111
DiseaseDiabetes100
Cerebrovascular disease111
Ischemic heart disease000
OutputSimilarity Label1
If the patient’s record belongs to a certain feature, we write 1, otherwise 0. In the match column, if the features of patients A and B match, we write 1, otherwise 0. Similarity label will be 1 if the number of matches is more than equal to 4, otherwise it is 0.
Table 3. Dataset descriptions.
Table 3. Dataset descriptions.
DescriptionsMaleFemale
Sex (%)5644
Age, mean(SD)57.8 (23.4)60.5 (21.5)
SBP (mmHg), mean (SD)126.4 (27.1)131.8 (24.8)
DBP (mmHg), mean (SD)78.8 (12.6)81.8 (14.2)
BMI (kg/m2), mean (SD)22.3 (5.8)23.1 (4.2)
No. of visits1,052,085648,066
Table 4. Similarity-based learning results.
Table 4. Similarity-based learning results.
Diabetes and
Cerebrovascular Disease
Diabetes and
Ischemic Heart
Disease
Cerebrovascular Disease and
Ischemic Heart Disease
Diabetes and
Cerebrovascular
Disease and Ischemic Heart Disease
Average
Mean0.78590.80270.84290.84740.8197
Max0.87570.86070.89520.92740.8897
Min0.55710.74700.80350.81370.7304
The metric presented in Table 4 is the similarity score calculated by Equation (4). The range of the score is between 0 and 1. If the score is closer to 1, it means higher similarity; otherwise, it is closer to 0. The mean, max, and min came from the score of patient pairs.
Table 5. Joint learning results.
Table 5. Joint learning results.
Diabetes and
Cerebrovascular Disease
Diabetes and
Ischemic Heart
Disease
Cerebrovascular Disease and
Ischemic Heart Disease
Diabetes and
Cerebrovascular
Disease and Ischemic Heart Disease
Average
Mean0.84460.91480.92270.93330.9039
Max0.95890.97890.98490.97820.9752
Min0.64390.81250.85260.86450.7934
The metric presented in Table 5 is the similarity score calculated by Equation (5). The range of the score is between 0 and 1. If the score is closer to 1, it means higher similarity; otherwise, it is closer to 0. The mean, max, and min came from the score of patient pairs.
Table 6. Performance comparison between existing algorithms.
Table 6. Performance comparison between existing algorithms.
AlgorithmsAccuracyPrecisionRecallF1 Score
Euclidean0.54230.52180.62850.5457
Cosine0.57510.56840.60710.5895
Mahalanobis0.65730.68170.78250.7782
LSML0.80150.81480.85330.8759
Joint Learning0.85720.85110.89250.9227
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Oh, S.H.; Back, S.; Park, J. Measuring Patient Similarity on Multiple Diseases by Joint Learning via a Convolutional Neural Network. Sensors 2022, 22, 131. https://doi.org/10.3390/s22010131

AMA Style

Oh SH, Back S, Park J. Measuring Patient Similarity on Multiple Diseases by Joint Learning via a Convolutional Neural Network. Sensors. 2022; 22(1):131. https://doi.org/10.3390/s22010131

Chicago/Turabian Style

Oh, Sang Ho, Seunghwa Back, and Jongyoul Park. 2022. "Measuring Patient Similarity on Multiple Diseases by Joint Learning via a Convolutional Neural Network" Sensors 22, no. 1: 131. https://doi.org/10.3390/s22010131

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop