Next Article in Journal
Association between Subjective Body Image, Body Mass Index and Psychological Symptoms in Chinese Adolescents: A Nationwide Cross-Sectional Study
Next Article in Special Issue
Patient Satisfaction and Hospital Quality of Care Evaluation in Malaysia Using SERVQUAL and Facebook
Previous Article in Journal
Psychological Aspects of Media Communication during the COVID-19 Pandemic: Insights from Healthcare and Pharmacy Specialists in Lithuania
Previous Article in Special Issue
Forecast of the COVID-19 Epidemic Based on RF-BOA-LightGBM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Application of Projection Word Embeddings on Medical Records Scoring System

1
School of Medicine, National Defense Medical Center, Taipei 114, Taiwan
2
School of Public Health, National Defense Medical Center, Taipei 114, Taiwan
3
Graduate Institute of Life Sciences, National Defense Medical Center, Taipei 114, Taiwan
4
Artificial Intelligence of Things Center, Tri-Service General Hospital, National Defense Medical Center, Taipei 114, Taiwan
5
Division of Cardiovascular Surgery, Cheng Hsin Rehabilitation and Medical Center, Taipei 112, Taiwan
6
Department of Informatics, Taoyuan Armed Forces General Hospital, Taoyuan 325, Taiwan
7
Department of Medical Informatics, Tri-Service General Hospital, National Defense Medical Center, Taipei 114, Taiwan
8
Division of Colorectal Surgery, Department of Surgery, Tri-Service General Hospital, National Defense Medical Center, Taipei 114, Taiwan
9
Department of Family and Community Medicine, Department of Internal Medicine, Tri-Service General Hospital, National Defense Medical Center, Taipei 114, Taiwan
*
Authors to whom correspondence should be addressed.
C.L., Y.-T.L. and F.-J.W. contribute equally in the article.
Healthcare 2021, 9(10), 1298; https://doi.org/10.3390/healthcare9101298
Submission received: 24 August 2021 / Revised: 24 September 2021 / Accepted: 28 September 2021 / Published: 29 September 2021

Abstract

:
Medical records scoring is important in a health care system. Artificial intelligence (AI) with projection word embeddings has been validated in its performance disease coding tasks, which maintain the vocabulary diversity of open internet databases and the medical terminology understanding of electronic health records (EHRs). We considered that an AI-enhanced system might be also applied to automatically score medical records. This study aimed to develop a series of deep learning models (DLMs) and validated their performance in medical records scoring task. We also analyzed the practical value of the best model. We used the admission medical records from the Tri-Services General Hospital during January 2016 to May 2020, which were scored by our visiting staffs with different levels from different departments. The medical records were scored ranged 0 to 10. All samples were divided into a training set (n = 74,959) and testing set (n = 152,730) based on time, which were used to train and validate the DLMs, respectively. The mean absolute error (MAE) was used to evaluate each DLM performance. In original AI medical record scoring, the predicted score by BERT architecture is closer to the actual reviewer score than the projection word embedding and LSTM architecture. The original MAE is 0.84 ± 0.27 using the BERT model, and the MAE is 1.00 ± 0.32 using the LSTM model. Linear mixed model can be used to improve the model performance, and the adjusted predicted score was closer compared to the original score. However, the project word embedding with the LSTM model (0.66 ± 0.39) provided better performance compared to BERT (0.70 ± 0.33) after linear mixed model enhancement (p < 0.001). In addition to comparing different architectures to score the medical records, this study further uses a mixed linear model to successfully adjust the AI medical record score to make it closer to the actual physician’s score.

Graphical Abstract

1. Introduction

With the increasing advancement of technology, the data amount generated by humans is growing explosively [1]. Effectively taking advantage of these growing data may bring valuable information, which many successful cases from different industries [2] have already proved. However, the majority of these data are not structured [3], which cannot be directly used by traditional analytical methods. At the same time, it is expected to employ new algorithms to use these data to allow for stronger decision-making capacity [4,5]. In recent years, with the breakthrough developments of the deep neural network in diverse fields, we are already capable of directly analyzing data in the forms of videos, texts, and voices. Hence, the focus of researches is now to develop applications to solve practical problems.
The medical system is an important field that is very suitable to develop the abovementioned applications. Medical knowledge is accumulating quickly, making it more and more possible for doctors to have knowledge gaps [6], which may cause misdiagnoses and, thus, urgently need to be solved [7]. Computer-aided diagnosis systems have been greatly developed in recent years, aiming to solve this problem, yet unsuccessfully so far [8]. This is probably because the majority of medical data are non-structural data [9]; take cancer, for example, where about 96% of cancer diagnoses are made from pathological section reports, the data of which, however, are recorded in text descriptions and videos [10]. Thus, it is difficult for traditional models to link these original non-structural data with diagnosis information directly. With the advancement of artificial intelligence (AI) technology, the new generation of computer-aided diagnosis systems is expected to make great contributions to the intellectualization of medical systems. It can further eliminate human errors to increase the quality of medical care [11]. In 2012, AlexNet was the ILSVRC champion, leading the 3rd AI revolution [12]. Since then, more powerful deep learning models have been developed, such as VGGNet [13], Inception Net [14], ResNet [15], DenseNet [16], etc. This revolution led by deep learning has made enormous progress in image recognition tasks, driving breakthroughs in related research. Computer-aided diagnosis tools built based on deep learning technology have led to an increase in medical care quality [11]. Examples include lymph node metastasis detection [17], diabetic retinopathy detection [18], skin cancer classification [19], pneumonia detection [20], bleeding identification [21], etc. There have been over 300 studies (mostly in the last 2 years) using such technologies in medical image analysis [22]. It is worth mentioning that the most impressive capacity of deep learning technology is automatic feature extraction. With the precondition of a large database for annotation, it has been proven to reach, or even surpass, the level of human experts [15,23,24].
The current method to use a large amount of information from medical records is to code through recognition by experts and according to ICD (The International Statistical Classification of Diseases and Related Health Problems). This work is not only necessary for our national health insurance declaration system but may also be used in disease monitoring, hospital management, clinical studies, and policy planning. However, artificial classification is not only expensive but is also time-inefficient, which is the most important. For example, in disease monitoring, since the outbreak of infectious disease will cause large casualties [25], many countries have developed their disease monitoring systems specifically aiming at contagious diseases, such as the Real-time Outbreak and Disease Surveillance (RODS) system [26]. To ensure time efficiency, this system stipulates emergency physicians to input data within required time limits when identifying notifiable diseases, making it hard to be promoted to other diseases. With the advancement of data science, it has been universally expected that an automatic disease interpretation model can be developed to solve the high-cost and time-inefficient problems of artificial interpretation.
Due to the popularization of medical records electronization, a great number of studies have attempted to use this information for text mining and ICD code classification. The current technology primarily uses a bag-of-words model to standardize text medical records, then uses a support vector machine (SVM), random forest tree, and other classifiers for diagnosis classification [27,28,29,30,31]. However, previous studies have found that these methods were incapable of accurate diagnosis classification because of the particularity and diversity of clinical terms, where synonyms need to be properly processed before data preprocessing [10]. A complete medical dictionary integrates the currently recommended forms of clinical terms; yet, it is almost impossible due to the complexity of clinical terms. Therefore, traditional automatic classification programs can hardly make significant progress. In addition, the bag-of-words model treats different characters as different features and counts the number of features in one article. Although this makes it possible to use a dictionary to handle the synonym problem, similar characters would be considered two different features. Thus, the number of features integrated by the bag-of-words model will be strikingly huge, causing a curse of dimensionality when classified by subsequent classifiers, leading to inefficiency and slow progress of traditional algorithms.
Other than classification efficiency, the greatest challenge for traditional algorithms is new diseases. For instance, there was an H1N1 outbreak in 2009, with related cases that had never been recorded before 2008. Traditional classification algorithms are completely unable to perform proper classification of newly emerged words [27,28,29,30,31]. This disadvantage makes it absolutely impossible for traditional methods to reach full automation. Regarding this issue, we proposed word embedding as a technical breakthrough in disease classification. Since the 20th century, word embedding has been an important technology to allow computers to understand the semantic meaning further. Its core logic is hoping to characterize every single word into a vector in high-dimensional space and expecting similar vectors for similar characters/words to express semantic meaning [32,33]. The word2vec published by the Google team in 2013 is considered the most important breakthrough in recent word embedding studies. It has been verified to allow similar characters to have very high cosine similarity and very close Euclidean distance in vector space [34]. However, this technology has a disadvantage that, once applied, it converts an article into an unequal matrix, making it inapplicable for traditional classifiers, such as SVM and random forest trees. A general solution is to average or weighted average the word vector of all characters in an article as semanteme [35]. However, from the MultiGenre NLI (MultiNLI) Corpus competition release by the natural language research team of Stanford (https://nlp.stanford.edu/projects/snli/), we can still see that combining modern AI technology gives better efficiency to models. Language processing conducts analysis mostly based on Recurrent Neural Network (RNN) or Convolutional Neural Network (CNN). Its core principle is to use convolutional layer (does not have memory but can gradually integrate surrounding single-character information in higher-order features, requires more layers) or Long Short-Term Memory Unit (has short- and long-term memory, thus needing fewer layers) for feature extraction and is able to process information in matrix form [36]. CNN has become the primary method in all computer vision competitions. Its reason for success is a fuzzy matching technique of convolutional layer, allowing for integrating similar image features. We will be able to change the convolutional layer from recognizing similar image features to recognizing similar vocabularies through certain designs. Hence, CNN has been applied in text mining, such as semantic classification [37], short sentence searching [38], and chapter analysis [39], and has shown considerably good efficiency. In the most recent study, Bidirectional Encoder Representations from Transformers (BERT), developed by Google, has swept all kinds of natural language process competitions [40]. Yet, its core is still good work/sentence/paragraph embedding. Generally speaking, combining good embedding technology with modern deep learning neural networks is undoubtedly the best option for current natural language processing tasks.
Our team has already applied it in disease classification of discharge record summaries and proved that it compared with traditional models. AI model with combined word embedding model and CNN reduces 30% error rate in disease classification tasks, makes modeling easier by avoiding troublesome text integration preprocessing, and learns external language resources through unmonitored learning to integrate similarity among clinical clauses [41]. However, although the combination of word embedding and CNN is better in disease classification tasks than traditional methods, its accuracy still cannot be compared with humans. One of the reasons is the error in understanding the semantic meaning. Therefore, improving the word embedding model’s understanding of the meaning of medical terms might increase its subsequent analytical efficiency [42]. There are two studies that have evaluated the application of word embedding models trained by different resources on biomedical NLP and found EHR-trained word embedding could better capture semantic property [43,44]. On the other hand, external data resources have a neglected advantage in that the vocabulary diversity of external internet data resources is far more than that of internal task database. This advantage will greatly affect real disease coding tasks. Hence, an embedded training process needs to be developed to maintain the vocabulary diversity of internet resources and medical terms’ understanding of the internal task database. A recent word embedding comparison study showed that EHR-trained work embedding could usually better capture medical semantic meaning [43]. Even the research team of abroad Mayo Clinic uses an EHR with a large amount of data. The total number of words is only about 100,000, the vocabulary diversity of which is still far less than the external database [43,44]. This is due to the lack of some rare diseases and periodic diseases, such as the 2003 SARS outbreak and the 2009 H1N1 outbreak. Therefore, EHR-trained word embedding models are unable to include enough vocabulary. For this reason, our team developed a projection word embedding model that has the vocabulary diversity of Wikipedia/PubMed, as well as an understanding of medical terms in EHR [45].
A medical record is a historical record and also the foundation of a patient’s medical care. It records the patient’s conditions, reasons, results of examinations/tests, treatment methods, and results during care processes. It integrates and analyzes patients’ related information, presents the executive ground of medical decisions, and even affects national health policy. The basic purpose of medical records is to remind oneself or other medical care colleagues of a patient’s daily conditions and attending physician’s current thoughts. When medical treatment is being performed, the medical record serves as the communication tool among physicians and means for continuous treatment. In other words, the medical record is the only text material that records a patient’s conditions and focuses on all medical care personnel. A medical record is an index of medical care quality reflecting a physician’s clinical thinking and diagnostic basis. It serves as the reference for learning, research, and education. Meanwhile, it also serves as the evidence for medical disputes to clarify the attribution of liabilities. The medical record is the foundation of patient care as it records the contents of patient care provided by medical personnel. Thus, all results obtained from observation or examination can be found on the medical record. Therefore, any change in the patient’s condition can be found from the medical record so that the patient’s current condition can be evaluated for suitable treatments. Moreover, communication with a patient should also be included in the medical record so that medical personnel can learn the patient’s expectations on the treatment, resulting in a closer doctor-patient relationship. For other professionals, a detailed medical record saves a lot of communication time and avoids misunderstanding or missing the patient’s previous conditions that may lead to mistreatment.
The content of medical records also has legal effects. It is the basis of insurance benefits and even affects national health policy. For example, public health studies usually need to include case information under national health insurance, and, through studying a large number of medical records, such studies can help public health researchers and medical officials to establish more suitable public health decisions and administrative rules that protect the rights and interests of both doctors and patients. Clinical decision-making guides formulated by many specialized medical associations also used information from medical records. The implicit demographic information from these medical records is also collected at the national level and published as national health demographic information to compare with other countries so as to serve as a way to communicate and learn from each other for mutual benefits.
In this study, as shown in the graphical abstract, a scoring database was established by experts performing scoring on medical records. An AI model was trained to learn experts’ scoring logics so as to screen high-quality medical record summaries. In contrast, the database made up of which will have the chance to promote the establishment of other subsequent AI models, improve model accuracy, and serve as a teaching example to improve medical education efficiency.

2. Method

2.1. Data Source

In this study, inpatient medical records from Tri-Service General Hospital from 1 January 2016 to 31 December 2019 were used as the basic database, which was ethically approved by institutional review board (IRB NO. A202005104). Physicians of different levels from different departments were invited for medical records summary scoring. Scoring dimensions include different indexes, based on clinical writing standards, it contains 12 scoring items from each detailed structure of the QNOTE scale’s inpatient record, including chief of complaint, history of the present illness, problem list, past medical history, medications, adverse drug reactions and allergies, social and family history, review of systems, physical findings, assessment, plan of care, and follow-up information. The completeness of each item’s record, as well as the 5 structures (completeness, correctness, concordance, plausibility, and currency) of electronic medical records’ examination information, are evaluated in 5 levels of the Likert scale: strongly disagree, disagree, no comment (not agree nor disagree), agree, and strongly agree. Specialists from different departments were required to review 227,689 medical records and preliminarily score them on a 10-point Likert scale based on the average of above 5 structures. These scores were then used as the training target of the AI model to represent medical record writing quality. All samples were divided into a training set (n = 74,959) and testing set (n = 152,730) based on time, and then they were evaluated by different departments. Data of the testing set was compared with the actual scores for analysis, and MAE from the Likert scale was used as the evaluation index for model performance. In the end, the aforementioned model was applied in Tri-Service General Hospital. A medical record auto-scoring system was established in the hospital so as to screen high-quality medical records for future teaching and research studies.

2.2. AI Algorithm

The collected medical records and various writing quality indicators can be used for artificial intelligence model training. The model architecture uses the word embedding and LSTM model developed by our team. The word embedding also uses the projection word embedding comparison table to perform single-character conversion mathematical vectors and uses the entire input article as the input matrix. We used projection word embedding to construct a deep convolutional network model to enable the network to integrate the transformed semantic vectors and extract written medical records based on different word combinations. First, we used the word embedding comparison table trained by Wikipedia and PubMed library, and then we used EHR to perform projection word embedding training. Next, we connected the converted text matrix in parallel so that the network can refer to two different word embedding sources simultaneously. In addition, we used different word embeddings separately as conversion sources to compare their effects on prediction performance.

2.2.1. Long Short-Term Memory (LSTM)

In RNN, the output can be given back to the network as input, thereby creating a loop structure. RNNs are trained through backpropagation. In the process of backpropagation, RNN will encounter the problem of vanishing gradient. We use the gradient to update the weight of the neural network. The problem of vanishing gradient is when the gradient shrinks as it propagates backwards in time. Therefore, the layers that obtain small gradients will not learn but will, instead, cause the network to have short-term memory.
The LSTM architecture was introduced by Hochreiter and Schmidhuber [46] to alleviate the problem of vanishing gradients. LSTMs can use a mechanism called gates to learn long-term dependencies. These gates can learn which information in the sequence is important to keep or discard. LSTMs have three gates: input, forget, and output. This is the core of the LSTM model, where pointwise addition and multiplication are performed to add or delete information from the memory. These operations are performed using the input and forget gate of the LSTM block, which also contains the output “tanh” activation function. In addition to using the original architecture and model parameters, the other settings are Epochs = 20, Batch size = 300, and Learning rate = 0.001.

2.2.2. Bidirectional Encoder Representation from Transformers (BERT)

Other than the original word embedding and LSTM architecture, BERT architecture was also used for feature extraction. BERT is a recent attention-based model with a bidirectional Transformer network that was pre-trained on a large corpus. This pre-trained model is then effectively used to solve various language tasks with fine-tuning [40,47]. In brief terms, the task-specific BERT architecture represents input text as sequential tokens. The input representation is generated with the sum of the token embeddings, the segmentation embeddings and the position embeddings [40]. For a classification task, the first word in the sequence is a unique token which is denoted with [CLS]. An encoder layer is followed with a fully-connected layer at the [CLS] position. Finally, a softmax layer is used as the aggregator for classification purposes [47]. If the NLP task has pair of sentences as in question-answer case, the sentence pairs may be separated with another special token [SEP]. BERT multilingual base model (cased) is used as transfer feature learning, and other parameters are set to Epochs = 30, Batch size = 32, and Learning rate = 0.00001.
Through these two methods, we can enable the network to learn the semantic meanings of different individual characters. We can also let the network learn from different texts, such as from Wikipedia and PubMed. Then, through EHR for Fine-tune retraining, the BERT architecture that has finished learning only needs to change from predicting its context output to predicting the categories of multiple medical record quality dimensions; then, it can be trained with medical record information.

2.3. Linear Mixed Model Function for Medical Records Scoring Prediction

Suppose data are collected from m independent groups of observations (called clusters or subjects in longitudinal data).
Y m = X m B m + e m .
Here, Ym is an n × 1 vector of the dependent variable for patient m, and Xi is an n × q matrix of all the independent variables for patient m. Bm is a q × 1 unknown vector of regression coefficients, and em is an n × 1 vector of residuals. This results in a multi-level mixed model with random effects for all samples, which is expressed as
Y = X B + Z u + e ,
where Z is a matrix of known constants included in the information of the independent variables with random effects, and u is a matrix of random effects for all patients.
The best linear unbiased prediction (BLUP) is important for predicting the medical record score in each patient, and it can be calculated by following the steps in [48].
Ym is an n × 1 vector of the dependent variable for patient m, and Xi is an n × q matrix of all independent variables for patient m. Moreover, Zm is an n × p matrix of independent variables with random effects for patient m. These matrices contain the observed data and are defined as
Y m = [ y 1 , m y 2 , m y n , m ] ,   X m = [ 1 x 1 , 1 , m x 1 , q 1 , m 1 x 2 , 1 , m x 2 , q 1 , m x 1 x n , 1 , m x n , q 1 , m ] ,   Z m = [ 1 x 1 , 1 , m x 1 , p 1 , m 1 x 2 , 1 , m x 2 , p 1 , m x 1 x n , 1 , m x n , p 1 , m ] .
After building the prediction tool, we have the G matrix, B vector and σ2. G is a variance co-variance matrix of the random effects (p × p), and B is the fixed effect coefficients vector (q × 1). σ2 is the variance of the residuals. We can calculate a matrix R (n × n) using
G = [ τ 1 2 τ 12 τ 1 p τ 12 τ 2 2 τ 2 p τ 1 p τ 2 p τ p 2 ] ,   B = [ b 0 b 1 b q 1 ] ,   R = σ 2 I n × n = [ σ 2 0 0 0 σ 2 0 0 0 σ 2 ] .
If the independence assumption holds (i.e., [ u e ] ~ N ( [ 0 0 ] , [ G 0 0 R ] ) ), then we can calculate the variance co-variance matrix (Σm) of Ym using
Σ m = Z m G Z m T + R .
Finally, the BLUP of the random effect in patient m can be estimated using
B L U P m = G Z m T Σ m 1 ( Y m X m B ) .
We can estimate the regression coefficients (Bm) in patient m based on the above result, and Bm can be used to predict the disease progression. Bm can be calculated using
B m = B + B L U P m
Note that this calculation cannot make direct forecasts without the co-variable values. Thus, the co-variables information at the time of interest must be generated. We propose two methods for generating this information: (1) assume consistency between the last time and the time of interest and (2) predict the linear expectations. We will assess these methods in our analysis. Unquestionably, clinicians can use the most reasonable values based on their judgment to predict the co-variables at the time of interest. In summary, we can combine this method with population information to predict the medical record score.

2.4. Evaluation Criteria

We evaluated the generalization performance of each model in the training and testing samples. Mean absolute error (MAE) were used to compare the performance of the models, as follows:
M A E = i = 1 N | y i y ^ i | N   .

3. Results

The research scheme is shown in Figure 1, where a total of 227,689 medical records were scored by experts. In AI model training, the medical records were divided into the training set and testing set based on year, where 74,959 records were used to establish BERT and LSTM models, and 152,730 records were used to test record scoring. LMM was then employed to modify BERT and LSTM to establish another two models. In the end, MAE was used to compare the four models’ efficiencies in predicting medical record scores.
Table 1 shows the distribution of medical records in different departments. It can be seen that 74,959 records were included for modeling, and then 152,730 records were used for prediction. The average score from experts was 7.24 ± 1.02 for the training set and 7.67 ± 0.84 for the testing set; after BERT and LSTM modeling of medical record scoring, the average score of BERT prediction in the testing set was 7.47 ± 0.89, and 7.15 ± 1.05 for LSTM. After training through the BERT and LSTM models, the artificial intelligence model had already scored the medical records.
Our team’s projection word embedding model allowed the model to have both the vocabulary diversity of Wikipedia/PubMed and an understanding of medical terms in EHR. The concept of projection word embedding used the results of our previous studies, a concept in linear algebra that projects through matrix multiplication to allow all coordinates to convert into a new coordinate system. Such conversion changes the correlation of certain points while at the same time maintaining all current coordinates. In addition to the original projection word embedding and LSTM architecture, we attempted to use BERT architecture for feature extraction. BERT stands for Bidirectional Encoder Representations from Transformers, the elementary unit of BERT architecture is the encoder’s Multi-Head Self-Attention Layer in the transformer. In contrast, the overall architecture of BERT is stacked by a bidirectional Transformer Encoder Layer. As shown in Table 2, in general, on the ground of experts’ scoring, the trained scoring model BERT had a prediction score of 7.49 ± 0.28. In contrast, LSTM had 7.17 ± 0.31; after modification by the linear mixed model (LMM), BERT’s and LSTM’s prediction scores were 7.36 ± 0.56 and 7.33 ± 0.65, respectively. After layering different departments, such as internal medicine, surgery, obstetrics, and pediatrics, it can be learned that BERT all had higher prediction scores than LSTM, while, after LMM modification, all LSTM prediction scores increased. Through further looking into different departments, it was found that most departments’ BERT prediction scores were higher than that of LSTM, and the latter increased after LMM modification.
It can be learned from Table 3 that, when reviewer physicians’ scores and AI scores were calculated using mean absolute error (MAE), both BERT and LSTM AI scores were 0.6~1.3 points lower than reviewer physicians’ scores; thus, the linear mixed model (LMM) was introduced for modification, thereby reducing the score difference to 0.3~1 points, showing a significant reduction (p < 0.001) in score difference. The reason for the modification using LMM is that an ordinary linear regression contains only two influencing factors: fixed effect and noise. The latter is a random factor not considered in our model, while the former are those predictable factors that can also be completely divided. The AI scoring of medical records after modification by LMM is also more realistic. After department layering, it was found that, in some departments, LMM-modified MAE was not significantly reduced comparing with the original MAE. Hence, experts’ scores were made into a heat map (Figure 2), where it was found that some groups of scoring physicians and scored physicians had closer scores, and were separately analyzed. In Table 4, medical record prediction scores and MAE are analyzed from Block A to H, respectively, and, except for block F, most blocks had similar record scores with previous results, and the MAE of LSTM prediction scores significantly reduced (p < 0.05) after LMM modification.
In spite of this, we were still unable to identify the reason why the MAE of certain departments had no significant reduction after LMM modification. Thus, heat map analysis was performed on LMM-modified LSTM prediction scores. Figure 3 shows that some reviewers’ LMM-modified LSTM prediction scores had relatively greater MAE. After grouping using LMM modified MAE (Grade-LMM modified LSTM), experts’ scores were close among groups, but BERT and LSTM prediction scores were lower than the original experts’ scores. In Figure 4, We further using MAE to evaluate model efficiency, and then comparing MAE (|Grade—LMM modified BERT|, |Grade—LMM modified BERT|) of LMM-modified BERT or LSTM with the MAE (|Grade—BERT|, |Grade—LSTM|) of the original BERT or LSTM, it was found MAE was effectively reduced through LMM modification in Q1~Q3, but not in Q4. Thus, it is suspected that some scoring physicians in Q4 may have scored incorrectly.

4. Discussion

In this study, the projection word embedding model was used to develop an AI system to evaluate the writing quality of inpatient medical records. The AI system is already capable of accurate classification to level 3 ICD-10 coding, combined with results from previous studies. Since level 3 coding is already at the disease level, subsequent coding will all just be remarks (such as location), and reaching such a level will allow for the possibility of full automation of common disease classification tasks, as well as extraction of disease features from other medical descriptions, through this algorithm. In addition to the original word embedding and LSTM architecture, BERT architecture was also employed to extract disease features for medical record scoring. LMM was further used for modification to get AI scores closer to actual reviewer physicians’ scores. Moreover, it was also identified that some physicians over-scored medical records. If these scoring standards can be improved in the future, a better medical writing quality could be expected.
In addition, why is the quality of medical record writing so important? Because the medical record is the historical record of the patient’s health care; it is also the basis of care, and its content records the patient’s condition during the care process, the reason and result of the inspection, and the treatment method and result. In recent studies, it is feasible to use electronic health records (EHR) to predict disease risk, such as atrial fibrillation (AF) [49], coronary heart disease in patients with hypertension [50], fall risk [51], multiple sclerosis disease [52], and cervical cancer [53]. Over the past two decades, the investigation of genetic variation underlying disease susceptibility has increased considerably. Most notably, genome-wide association studies (GWAS) have investigated tens of millions of single-nucleotide polymorphisms (SNPs) for associations with complex diseases. However, results from numerous GWAS have revealed that the majority of statistically significantly associated genetic variants have small effects [54] and may not be predictive of disease risks [55], and many diseases are associated with tens of thousands of genetic variants [56]. These findings have led to the resurgence of the polygenic risk score (PRS), an aggregate measure of many genetic variants weighted by their individual effects on a given phenotype. However, epidemiologic studies are expensive and complex to run, which raises the question of whether a PRS could be developed and applied in a clinical setting using genetic data that are more readily available. Recently, some scholars proposed new ideas for developing and implementing PRS predictions using biobank-linked EHR data [57].
For the medical records scoring system, this not only saves doctors the time for scoring medical records but also can get feedback immediately after the writing is completed to improve the quality of medical record writing. In the past research, clinicians spent 3.7 h per day, or 37% of their work day, on EHR [58]. There was a marked reduction in EHR time with both clinician and resident seniority. Despite this improvement, the total time spent on EHR remained exceedingly high amongst even the most experienced physicians [58]. The significance of an increasing shift towards EHR is a growing paradigm that cannot be understated, particularly in the current era of healthcare, and there is increasing scrutiny on documentation [59,60]. These increased demands can lead to EHR fatigue and physician burnout. In a survey of a general internal medicine group, 38% reported feeling burnt out, with 60% citing high documentation pressure and 50% describing too much EHR time at home [61]. Burnout has been linked to an increased risk of resident’s wellbeing [62].
There are still some limitations for electronic medical records. First, this scoring system can only be used in our hospital because the medical record system of different hospitals do not talk to each other. Second, entering data into an EHR requires a doctor to spend a lot of time doing so, leading to most physicians experiencing burnout symptoms due to EMR-related workloads. Third, cyber-attacks are a perennial concern for EHRs. It is, therefore, imperative that cybersecurity is continually enhanced. Fourth, timing discrepancies occur in EHRs, and they can lead to serious clinical consequences.
In summary, combining projection word embedding and LSTM with LMM can give better prediction scores. This system can be used to assist medical record scoring so that young physicians can get immediate writing feedback, so as to improve the quality of medical record writing in my country and let the public, Medical units, and insurance units can all get better help. In the future, it may be possible to actively introduce such technologies into hospitals to achieve personalized precision medicine.

Author Contributions

Data curation, C.L., C.-J.H., and C.-C.L.; Project administration, W.-H.F.; Resources, Y.-T.L., F.-J.W., and W.-H.F.; Software, S.-A.L.; Supervision, C.L.; Visualization, D.-J.T.; Writing—original draft, Y.-T.L., F.-J.W., and D.-J.T.; Writing—review & editing, C.L. and D.-J.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was approved by the institutional review board in Tri-Service General Hospital, Taipei, Taiwan (IRB NO. A202005104).

Informed Consent Statement

Patients’ consent was waived because data were collected retrospectively and in anonymized files and encrypted from the hospital to the data controller.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

This study was supported by funding from the Ministry of Science and Technology, Taiwan (MOST109-2314-B-016-021, MOST110-2314-B-016-008 to W.H. Fang and MOST110-2314-B-016-010-MY3 to C. Lin), the National Science and Technology Development Fund Management Association, Taiwan (MOST109-3111-Y-016-002 to C. Lin and MOST110-3111-Y-016-005 to C. Lin), Cheng Hsin General Hospital (CHNDMC-109-19 to Y.T. Lee and C. Lin), and Taoyuan Armed Forces General Hospital (TYAFGH-D-109043 to F.J. Wu and C. Lin).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hilbert, M.; Lopez, P. The world’s technological capacity to store, communicate, and compute information. Science 2011, 332, 60–65. [Google Scholar] [CrossRef] [Green Version]
  2. McAfee, A.; Brynjolfsson, E. Big data: The management revolution. Harv. Bus. Rev. 2012, 90, 60–68. [Google Scholar]
  3. Unstructured Data and the 80 Percent Rule; Clarabridge Bridgepoints: Reston, VA, USA, 2008; Volume Q3.
  4. Jeong, S.R.; Ghani, I. Semantic Computing for Big Data: Approaches, Tools, and Emerging Directions (2011–2014). TIIS 2014, 8, 2022–2042. [Google Scholar]
  5. Cox, M.; Ellsworth, D. Application-controlled demand paging for out-of-core visualization. In Proceedings of the 8th Conference on Visualization’97, Phoenix, AZ, USA, 24 October 1997. [Google Scholar]
  6. McDonald, C.J. Medical heuristics: The silent adjudicators of clinical practice. Ann. Intern. Med. 1996, 124, 56–62. [Google Scholar] [CrossRef]
  7. National Academies of Sciences E, Medicine. Improving Diagnosis in Health Care; National Academies Press: Washington, DC, USA, 2016. [Google Scholar]
  8. El-Kareh, R.; Hasan, O.; Schiff, G.D. Use of health information technology to reduce diagnostic errors. BMJ Qual. Saf. 2013, 22 (Suppl 2), ii40–ii51. [Google Scholar] [CrossRef] [Green Version]
  9. Murdoch, T.B.; Detsky, A.S. The inevitable application of big data to health care. JAMA 2013, 309, 1351–1352. [Google Scholar] [CrossRef]
  10. Spasic, I.; Livsey, J.; Keane, J.A.; Nenadic, G. Text mining of cancer-related information: Review of current status and future directions. Int. J. Med. Inform. 2014, 83, 605–623. [Google Scholar] [CrossRef]
  11. Cahan, A.; Cimino, J.J. A Learning Health Care System Using Computer-Aided Diagnosis. J. Med. Internet 2017, 19, e54. [Google Scholar] [CrossRef]
  12. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  13. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  14. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  16. Huang, G.; Liu, Z.; Weinberger, K.Q.; van der Maaten, L. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  17. Bejnordi, B.E.; Veta, M.; van Diest, P.J.; van Ginneken, B.; Karssemeijer, N.; Litjens, G.; van der Laak, J.A.W.M.; the CAMELYON16 Consortium. Diagnostic Assessment of Deep Learning Algorithms for Detection of Lymph Node Metastases in Women with Breast Cancer. JAMA 2017, 318, 2199–2210. [Google Scholar] [CrossRef]
  18. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J.; et al. Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs. JAMA 2016, 316, 2402–2410. [Google Scholar] [CrossRef]
  19. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  20. Rajpurkar, P.; Irvin, J.; Zhu, K.; Yang, B.; Mehta, H.; Duan, T.; Ding, D.; Bagul, A.; Langlotz, C.; Shpanskaya, K.; et al. CheXNet: Radiologist-Level Pneumonia Detection on Chest X-Rays with Deep Learning. arXiv 2017, arXiv:1711.05225. [Google Scholar]
  21. Grewal, M.; Srivastava, M.M.; Kumar, P.; Varadarajan, S. RADNET: Radiologist Level Accuracy using Deep Learning for HEMORRHAGE detection in CT Scans. arXiv 2017, arXiv:1710.04934. [Google Scholar]
  22. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio AA, A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  23. Xiong, W.; Droppo, J.; Huang, X.; Seide, F.; Seltzer, M.; Stolcke, A.; Yu, D.; Zweig, G. Achieving human parity in conversational speech recognition. arXiv 2016, arXiv:1610.05256. [Google Scholar]
  24. Amodei, D.; Ananthanarayanan, S.; Anubhai, R.; Bai, J.; Battenberg, E.; Case, C.; Casper, J.; Catanzaro, B.; Cheng, Q.; Chen, G.; et al. Deep speech 2: End-to-end speech recognition in english and mandarin. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 20–22 June 2016. [Google Scholar]
  25. Dembek, Z.F.; Kortepeter, M.G.; Pavlin, J.A. Discernment between deliberate and natural infectious disease outbreaks. Epidemiol. Infect. 2007, 135, 353–371. [Google Scholar] [CrossRef]
  26. Tsui, F.C.; Espino, J.U.; Dato, V.M.; Gesteland, P.H.; Hutman, J.; Wagner, M.M. Technical description of RODS: A real-time public health surveillance system. J. Am. Med. Inform. Assoc. JAMIA 2003, 10, 399–408. [Google Scholar] [CrossRef] [Green Version]
  27. Koopman, B.; Zuccon, G.; Nguyen, A.; Bergheim, A.; Grayson, N. Automatic ICD-10 classification of cancers from free-text death certificates. Int. J. Med. Inform. 2015, 84, 956–965. [Google Scholar] [CrossRef] [Green Version]
  28. Koopman, B.; Karimi, S.; Nguyen, A.; McGuire, R.; Muscatello, D.; Kemp, M.; Truran, D.; Zhang, M.; Thackway, S. Automatic classification of diseases from free-text death certificates for real-time surveillance. BMC Med. Inform. Decis. Mak. 2015, 15, 53. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Koopman, B.; Zuccon, G.; Wagholikar, A.; Chu, K.; O’Dwyer, J.; Nguyen, A.; Keijzers, G. Automated Reconciliation of Radiology Reports and Discharge Summaries. In AMIA Annual Symposium Proceedings; American Medical Informatics Association: Bethesda, MD, USA, 2015; Volume 2015, pp. 775–784. [Google Scholar]
  30. Khachidze, M.; Tsintsadze, M.; Archuadze, M. Natural Language Processing Based Instrument for Classification of Free Text Medical Records. BioMed Res. Int. 2016, 2016, 8313454. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Mujtaba, G.; Shuib, L.; Raj, R.G.; Rajandram, R.; Shaikh, K.; Al-Garadi, M.A. Automatic ICD-10 multi-class classification of cause of death from plaintext autopsy reports through expert-driven feature selection. PLoS ONE 2017, 12, e0170242. [Google Scholar] [CrossRef] [PubMed]
  32. Bengio, Y.; Ducharme, R.; Vincent, P.; Jauvin, C. A neural probabilistic language model. J. Mach. Learn. Res. 2003, 3, 1137–1155. [Google Scholar]
  33. Yih, W.T.; Toutanova, K.; Platt, J.C.; Meek, C. Learning discriminative projections for text similarity measures. In Proceedings of the Fifteenth Conference on Computational Natural Language Learning, Portland, OR, USA, 23 June 2011. [Google Scholar]
  34. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.S.; Dean, J. Distributed representations of words and phrases and their compositionality. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, Nevada, 5–10 December 2013. [Google Scholar]
  35. Arora, S.; Liang, Y.; Ma, T. A Simple but Tough-to-Beat Baseline for Sentence Embeddings. 2016. Available online: https://openreview.net/forum?id=SyK00v5xx (accessed on 27 September 2021).
  36. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE. 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  37. Yih, W.T.; He, X.; Meek, C. Semantic Parsing for Single-Relation Question Answering. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Baltimore, MD, USA, 22–27 June 2014. [Google Scholar]
  38. Shen, Y.; He, X.; Gao, J.; Deng, L.; Mesnil, G. Learning semantic representations using convolutional neural networks for web search. In Proceedings of the 23rd International Conference on World Wide Web, Seoul, Korea, 7–11 April 2014. [Google Scholar]
  39. Kim, Y. Convolutional neural networks for sentence classification. arXiv 2014, arXiv:1408.5882. [Google Scholar]
  40. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  41. Lin, C.; Hsu, C.J.; Lou, Y.S.; Yeh, S.J.; Lee, C.C.; Su, S.L.; Chen, H.C. Artificial Intelligence Learning Semantics via External Resources for Classifying Diagnosis Codes in Discharge Notes. J. Med. Internet Res. 2017, 19, e380. [Google Scholar] [CrossRef] [Green Version]
  42. Choi, Y.; Chiu, C.Y.; Sontag, D. Learning Low-Dimensional Representations of Medical Concepts. AMIA Jt. Summits Transl. Sci. Proceedings. AMIA Jt. Summits Transl. Sci. 2016, 2016, 41–50. [Google Scholar]
  43. Wang, Y.; Liu, S.; Afzal, N.; Rastegar-Mojarad, M.; Wang, L.; Shen, F.; Kingsbury, P.; Liu, H. A comparison of word embeddings for the biomedical natural language processing. J. Biomed. Inform. 2018, 87, 12–20. [Google Scholar]
  44. Pakhomov, S.V.; Finley, G.; McEwan, R.; Wang, Y.; Melton, G.B. Corpus domain effects on distributional semantic modeling of medical terms. Bioinformatics 2016, 32, 3635–3644. [Google Scholar] [CrossRef] [Green Version]
  45. Lin, C.; Lou, Y.S.; Tsai, D.J.; Lee, C.C.; Hsu, C.J.; Wu, D.C.; Wang, M.C.; Fang, W.H. Projection Word Embedding Model With Hybrid Sampling Training for Classifying ICD-10-CM Codes: Longitudinal Observational Study. JMIR Med. Inform. 2019, 7, e14499. [Google Scholar] [CrossRef] [PubMed]
  46. Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  47. Gao, Z.; Feng, A.; Song, X.; Wu, X. Target-dependent sentiment classification with BERT. IEEE Access 2019, 7, 154290–154299. [Google Scholar] [CrossRef]
  48. Robinson, G.K. That BLUP is a good thing: The estimation of random effects. Stat. Sci. 1991, 15–32. [Google Scholar] [CrossRef]
  49. Hulme, O.L.; Khurshid, S.; Weng, L.C.; Anderson, C.D.; Wang, E.Y.; Ashburner, J.M.; Ko, D.; McManus, D.D.; Benjamin, E.J.; Ellinor, P.T.; et al. Development and Validation of a Prediction Model for Atrial Fibrillation Using Electronic Health Records. JACC. Clin. Electrophysiol. 2019, 5, 1331–1341. [Google Scholar] [CrossRef] [PubMed]
  50. Du, Z.; Yang, Y. Accurate Prediction of Coronary Heart Disease for Patients with Hypertension from Electronic Health Records with Big Data and Machine-Learning Methods: Model Development and Performance Evaluation. JMIR Med. Inform. 2020, 8, e17257. [Google Scholar] [CrossRef] [PubMed]
  51. Ye, C.; Li, J.; Hao, S.; Liu, M.; Jin, H.; Zheng, L.; Xia, M.; Jin, B.; Zhu, C.; Alfreds, S.T.; et al. Identification of elders at higher risk for fall with statewide electronic health records and a machine learning algorithm. Int. J. Med. Inform. 2020, 137, 104105. [Google Scholar] [CrossRef]
  52. Ahuja, Y.; Kim, N.; Liang, L.; Cai, T.; Dahal, K.; Seyok, T.; Lin, C.; Finan, S.; Liao, K.; Savovoa, G.; et al. Leveraging electronic health records data to predict multiple sclerosis disease activity. Ann. Clin. Transl. Neurol. 2021, 8, 800–810. [Google Scholar] [CrossRef]
  53. Weegar, R.; Sundström, K. Using machine learning for predicting cervical cancer from Swedish electronic health records by mining hierarchical representations. PLoS ONE. 2020, 15, e0237911. [Google Scholar] [CrossRef]
  54. Manolio, T.A.; Collins, F.S.; Cox, N.J.; Goldstein, D.B.; Hindorff, L.A.; Hunter, D.J.; McCarthy, M.I.; Ramos, E.M.; Cardon, L.R.; Chakravarti, A.; et al. Finding the missing heritability of complex diseases. Nature 2009, 461, 747–753. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Lo, A.; Chernoff, H.; Zheng, T.; Lo, S.-H. Why significant variables aren’t automatically good predictors. Proc. Natl. Acad. Sci. USA 2015, 112, 13892–13897. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Visscher, P.M.; Wray, N.R.; Zhang, Q.; Sklar, P.; McCarthy, M.I.; Brown, M.A.; Yang, J. 10 years of GWAS discovery: Biology, function, and translation. Am. J. Hum. Genet. 2017, 101, 5–22. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Li, R.; Chen, Y.; Ritchie, M.D.; Moore, J.H. Electronic health records and polygenic risk scores for predicting disease risk. Nat. Rev. Genet. 2020, 21, 493–502. [Google Scholar] [CrossRef] [PubMed]
  58. Verma, G.; Ivanov, A. Analyses of electronic health records utilization in a large community hospital. PLoS ONE 2020, 15, e0233004. [Google Scholar] [CrossRef]
  59. Pizziferri, L.; Kittler, A.F.; Volk, L.A.; Honour, M.M.; Gupta, S.; Wang, S.; Wang, T.; Lippincott, M.; Li, Q.; Bates, D.W. Primary care physician time utilization before and after implementation of an electronic health record: A time-motion study. J. Biomed. Inform. 2005, 38, 176–188. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  60. Pizziferri, L.; Kittler, A.F.; Volk, L.A.; Shulman, L.N.; Kessler, J.; Carlson, G.; Michaelidis, T.; Bates, D.W. Impact of an Electronic Health Record on oncologists’ clinic time. In Proceedings of the AMIA Annual Symposium Proceedings, Washington, DC, USA, 22–26 October 2005; p. 1083. [Google Scholar]
  61. Linzer, M.; Poplau, S.; Babbott, S.; Collins, T.; Guzman-Corrales, L.; Menk, J.; Murphy, M.L.; Ovington, K. Worklife and wellness in academic general internal medicine: Results from a national survey. J. Gen. Intern. Med. 2016, 31, 1004–1010. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  62. Van der Heijden, F.; Dillingh, G.; Bakker, A.; Prins, J. Suicidal thoughts among medical residents with burnout. Arch. Suicide Res. 2008, 12, 344–346. [Google Scholar] [CrossRef]
Figure 1. Training and testing sets generation. Schematic of the data set creation and analysis strategy, which was devised to assure a robust and reliable data set for training and testing of the network. Once a medical records data were placed in one of the data sets, that individual’s data were used only in that set, avoiding ‘cross-contamination’ among the training and testing sets. The details of the flow chart and how each of the data sets was used are described in the Methods.
Figure 1. Training and testing sets generation. Schematic of the data set creation and analysis strategy, which was devised to assure a robust and reliable data set for training and testing of the network. Once a medical records data were placed in one of the data sets, that individual’s data were used only in that set, avoiding ‘cross-contamination’ among the training and testing sets. The details of the flow chart and how each of the data sets was used are described in the Methods.
Healthcare 09 01298 g001
Figure 2. Heat map of medical record scores from scoring and scored physicians. X-axis: physicians who wrote the medical records; Y-axis: scoring physicians and their departments. A redder grid means record scoring physicians give a higher score to record writing physicians. There are clusters in some areas; thus, we put out some blocks and observe the block (A to H) characteristics in Table 4.
Figure 2. Heat map of medical record scores from scoring and scored physicians. X-axis: physicians who wrote the medical records; Y-axis: scoring physicians and their departments. A redder grid means record scoring physicians give a higher score to record writing physicians. There are clusters in some areas; thus, we put out some blocks and observe the block (A to H) characteristics in Table 4.
Healthcare 09 01298 g002
Figure 3. MAE heat map of LMM-modified LSTM prediction scores from scoring and scored physicians. X-axis: physicians who wrote the medical records; Y-axis: scoring physicians and their departments. By subtracting the MAE of the original score from the LMM modified LSTM prediction score, and using the MAE and coring physicians to conduct a heat map analysis, it can be found that some reviewer scores are on the high side.
Figure 3. MAE heat map of LMM-modified LSTM prediction scores from scoring and scored physicians. X-axis: physicians who wrote the medical records; Y-axis: scoring physicians and their departments. By subtracting the MAE of the original score from the LMM modified LSTM prediction score, and using the MAE and coring physicians to conduct a heat map analysis, it can be found that some reviewer scores are on the high side.
Healthcare 09 01298 g003
Figure 4. Using LMM modified MAE (Grade-LMM modified LSTM) for interquartile range grouping. (A): Compare the scores of Experts, BERT and LSTM. Y-axis: medical record scores, X-axis: Experts’ score, BERT prediction score, LSTM prediction score. (B): Compare the original MAE with the LMM modified MAE. Y-axis: mean absolute error (MAE), X-axis: |Grade—BERT|, |Grade—LMM modified BERT|, |Grade—LSTM|, |Grade—LMM modified BERT| for model efficiency evaluation. The LMM modified MAE (Grade-LMM modified LSTM) is grouped by interquartile range and divided into Q1, Q2, Q3, and Q4.
Figure 4. Using LMM modified MAE (Grade-LMM modified LSTM) for interquartile range grouping. (A): Compare the scores of Experts, BERT and LSTM. Y-axis: medical record scores, X-axis: Experts’ score, BERT prediction score, LSTM prediction score. (B): Compare the original MAE with the LMM modified MAE. Y-axis: mean absolute error (MAE), X-axis: |Grade—BERT|, |Grade—LMM modified BERT|, |Grade—LSTM|, |Grade—LMM modified BERT| for model efficiency evaluation. The LMM modified MAE (Grade-LMM modified LSTM) is grouped by interquartile range and divided into Q1, Q2, Q3, and Q4.
Healthcare 09 01298 g004
Table 1. Medical records distribution and scoring in the training set and testing set of different departments.
Table 1. Medical records distribution and scoring in the training set and testing set of different departments.
Training Set (n = 74,959)Testing Set (n = 152,730)p-Value
Department <0.001 *
General surgery4843 (6.5%)10,504 (6.9%)
Pleural surgery1932 (2.6%)3472 (2.3%)
Cardiovascular surgery3904 (5.2%)8319 (5.4%)
Colorectal & rectal surgery491 (0.7%)3479 (2.3%)
Urology surgery1330 (1.8%)3313 (2.2%)
Pediatric Surgery99 (0.1%)85 (0.1%)
Plastic surgery1748 (2.3%)4009 (2.6%)
Pulmonary Medicine10,268 (13.7%)19,065 (12.5%)
Cardiology2723 (3.6%)4765 (3.1%)
Nephrology2473 (3.3%)3749 (2.5%)
Blood Oncology9257 (12.3%)17,110 (11.2%)
Endocrine and metabolic839 (1.1%)1477 (1.0%)
Gastroenterology3861 (5.2%)7372 (4.8%)
Rheumatism, immunology and allergy1247 (1.7%)2624 (1.7%)
Trauma756 (1.0%)940 (0.6%)
Infection and Tropical Medicine3701 (4.9%)8488 (5.6%)
Psychiatric department6531 (8.7%)14,331 (9.4%)
Neurological department3159 (4.2%)7374 (4.8%)
Pediatric department1138 (1.5%)2474 (1.6%)
Dental department1223 (1.6%)2483 (1.6%)
Surgery department607 (0.8%)817 (0.5%)
Dermatology department5 (0.0%)109 (0.1%)
ENT department2388 (3.2%)3907 (2.6%)
Radiology40 (0.1%)175 (0.1%)
Emergency department0 (0.0%)300 (0.2%)
Family and Community Medicine188 (0.3%)655 (0.4%)
Nuclear Medicine Department144 (0.2%)153 (0.1%)
Neurosurgery3219 (4.3%)6937 (4.5%)
Orthopedic department3482 (4.6%)7876 (5.2%)
Obstetrics and Gynecology1766 (2.4%)3222 (2.1%)
Ophthalmology department607 (0.8%)903 (0.6%)
Rehabilitation department990 (1.3%)2243 (1.5%)
Experts’ scores7.24 ± 1.027.67 ± 0.84<0.001 *
BERT prediction score 7.47 ± 0.89
LSTM prediction score 7.15 ± 1.05
*: p-value < 0.05.
Table 2. BERT and LSTM original prediction scores and LMM-modified scores.
Table 2. BERT and LSTM original prediction scores and LMM-modified scores.
Experts’ ScoresBERT Prediction ScoresLSTM Prediction ScoresLMM-Modified BERT Prediction ScoresLMM-Modified LSTM Prediction Scores
Overall7.69 ± 0.647.49 ± 0.287.17 ± 0.317.36 ± 0.567.33 ± 0.65
Internal medicine7.49 ± 0.667.37 ± 0.217.01 ± 0.207.14 ± 0.567.08 ± 0.65
Surgery7.78 ± 0.557.49 ± 0.227.16 ± 0.177.54 ± 0.437.54 ± 0.51
Obstetrics and pediatrics8.08 ± 0.697.68 ± 0.317.37 ± 0.317.70 ± 0.617.68 ± 0.79
Other departments7.76 ± 0.607.57 ± 0.337.32 ± 0.407.39 ± 0.537.37 ± 0.61
Department
General surgery7.69 ± 0.747.48 ± 0.537.26 ± 0.287.45 ± 0.567.45 ± 0.57
Pleural surgery7.87 ± 0.257.55 ± 0.357.22 ± 0.167.55 ± 0.437.64 ± 0.48
Cardiovascular surgery7.73 ± 0.567.38 ± 0.377.01 ± 0.057.34 ± 0.177.35 ± 0.34
Colorectal & rectal surgery7.92 ± 0.187.73 ± 0.377.22 ± 0.167.87 ± 0.357.97 ± 0.40
Urology surgery7.76 ± 0.187.48 ± 0.297.14 ± 0.097.54 ± 0.257.48 ± 0.37
Pediatric Surgery6.16 ± NA6.86 ± 0.507.09 ± NA6.86 ± NA6.65 ± NA
Plastic surgery7.98 ± 0.087.58 ± 0.327.20 ± 0.157.65 ± 0.237.65 ± 0.29
Pulmonary Medicine7.58 ± 0.837.30 ± 0.576.98 ± 0.197.26 ± 0.587.22 ± 0.65
Cardiology7.19 ± 0.977.02 ± 0.646.99 ± 0.086.83 ± 0.686.75 ± 0.73
Nephrology8.13 ± 0.697.54 ± 0.557.12 ± 0.067.42 ± 0.477.39 ± 0.60
Blood Oncology7.21 ± 0.556.89 ± 0.506.71 ± 0.166.77 ± 0.526.71 ± 0.74
Endocrine and metabolic7.64 ± 0.267.38 ± 0.357.17 ± 0.047.35 ± 0.447.25 ± 0.55
Gastroenterology7.19 ± 0.257.15 ± 0.266.96 ± 0.127.16 ± 0.307.09 ± 0.33
Rheumatism, immunology and allergy7.79 ± 0.217.33 ± 0.326.98 ± 0.147.29 ± 0.177.19 ± 0.22
Trauma7.84 ± 1.327.39 ± 0.577.18 ± 0.027.21 ± 0.357.14 ± 0.47
Infection and Tropical Medicine7.33 ± 0.537.09 ± 0.576.98 ± 0.076.94 ± 0.746.89 ± 0.87
Psychiatric department8.41 ± 0.488.08 ± 0.478.00 ± 0.167.94 ± 0.597.94 ± 0.67
Neurological department7.89 ± 0.247.62 ± 0.237.39 ± 0.067.60 ± 0.187.63 ± 0.25
Pediatric department7.91 ± 0.857.51 ± 0.667.14 ± 0.107.52 ± 0.667.48 ± 0.93
Dental department7.95 ± 0.257.05 ± 0.526.53 ± 0.096.89 ± 0.046.76 ± 0.04
Surgery department7.81 ± NA7.40 ± 0.267.14 ± NA7.33 ± NA7.25 ± NA
Dermatology department8.58 ± NA7.67 ± 0.646.83 ± NA7.73 ± NA7.85 ± NA
ENT department7.37 ± 0.497.36 ± 0.387.29 ± 0.157.32 ± 0.477.37 ± 0.54
Radiology6.85 ± NA6.70 ± 0.176.67 ± NA6.51 ± NA6.57 ± NA
Family and Community Medicine7.37 ± 0.417.19 ± 0.617.29 ± 0.096.91 ± 0.806.90 ± 1.15
Nuclear Medicine Department8.76 ± NA8.01 ± 0.457.54 ± NA7.83 ± NA8.02 ± NA
Neurosurgery7.95 ± 0.497.59 ± 0.567.12 ± 0.077.78 ± 0.637.78 ± 0.75
Orthopedic department7.38 ± 0.407.21 ± 0.347.09 ± 0.097.14 ± 0.387.11 ± 0.44
Obstetrics and Gynecology8.31 ± 0.347.96 ± 0.417.67 ± 0.237.95 ± 0.497.96 ± 0.51
Ophthalmology department7.86 ± 0.197.65 ± 0.267.56 ± 0.067.54 ± 0.277.53 ± 0.33
Rehabilitation department8.06 ± 0.597.63 ± 0.417.29 ± 0.167.61 ± 0.257.51 ± 0.37
Table 3. The difference between the original AI/LMM-modified score and the expert score.
Table 3. The difference between the original AI/LMM-modified score and the expert score.
Original MAE aLMM-modified MAE bp-Value
Overall
BERT0.84 ± 0.270.70 ± 0.33<0.001 *
LSTM1.00 ± 0.320.66 ± 0.39<0.001 *
Internal medicineBERT0.82 ± 0.270.66 ± 0.370.007 *
LSTM0.96 ± 0.320.63 ± 0.41<0.001 *
SurgeryBERT0.86 ± 0.240.72 ± 0.250.011 *
LSTM1.04 ± 0.250.67 ± 0.30<0.001 *
Obstetrics and pediatricsBERT1.05 ± 0.300.82 ± 0.320.069
LSTM1.21 ± 0.310.74 ± 0.44<0.001 *
Other departmentsBERT0.79 ± 0.260.70 ± 0.350.142
LSTM0.96 ± 0.350.67 ± 0.41<0.001 *
Department
General surgeryBERT0.80 ± 0.210.75 ± 0.150.645
LSTM1.03 ± 0.200.72 ± 0.120.003 *
Pleural surgeryBERT0.72 ± 0.100.49 ± 0.260.200
LSTM0.91 ± 0.200.38 ± 0.270.100
Cardiovascular surgeryBERT0.88 ± 0.260.86 ± 0.420.589
LSTM1.09 ± 0.390.79 ± 0.510.065
Colorectal & rectal surgeryBERT0.74 ± 0.120.61 ± 0.250.686
LSTM0.97 ± 0.100.57 ± 0.340.057
Urology surgeryBERT0.73 ± 0.060.67 ± 0.100.318
LSTM0.93 ± 0.100.63 ± 0.200.002 *
Plastic surgeryBERT0.76 ± 0.050.59 ± 0.150.057
LSTM0.97 ± 0.080.52 ± 0.220.029 *
Pulmonary MedicineBERT0.94 ± 0.320.69 ± 0.290.040 *
LSTM1.14 ± 0.360.65 ± 0.270.002 *
CardiologyBERT1.01 ± 0.410.75 ± 0.330.136
LSTM1.12 ± 0.340.74 ± 0.340.024 *
NephrologyBERT0.89 ± 0.290.89 ± 0.410.841
LSTM1.22 ± 0.470.82 ± 0.490.222
Blood OncologyBERT0.85 ± 0.210.66 ± 0.230.130
LSTM0.91 ± 0.220.72 ± 0.280.195
Endocrine and metabolicBERT0.82 ± 0.030.68 ± 0.160.343
LSTM0.95 ± 0.090.63 ± 0.230.114
GastroenterologyBERT0.60 ± 0.110.42 ± 0.200.050 *
LSTM0.66 ± 0.170.37 ± 0.230.015 *
Rheumatism, immunology and allergyBERT0.74 ± 0.110.69 ± 0.130.548
LSTM1.02 ± 0.150.70 ± 0.160.032 *
TraumaBERT1.08 ± 0.220.88 ± 0.701.000
LSTM1.19 ± 0.630.84 ± 0.720.667
Infection and Tropical MedicineBERT0.69 ± 0.170.66 ± 0.810.028 *
LSTM0.78 ± 0.260.63 ± 0.910.028 *
Psychiatric departmentBERT0.73 ± 0.260.59 ± 0.470.328
LSTM1.03 ± 0.290.52 ± 0.540.028 *
Neurological departmentBERT0.72 ± 0.060.56 ± 0.060.002 *
LSTM0.82 ± 0.090.44 ± 0.110.002 *
Pediatric departmentBERT1.18 ± 0.350.95 ± 0.330.328
LSTM1.36 ± 0.300.90 ± 0.490.007 *
Dental departmentBERT0.96 ± 0.101.12 ± 0.240.400
LSTM1.52 ± 0.191.23 ± 0.230.400
ENT departmentBERT0.73 ± 0.130.53 ± 0.170.024 *
LSTM0.78 ± 0.150.46 ± 0.20<0.001 *
Family and Community MedicineBERT0.75 ± 0.060.74 ± 0.430.700
LSTM0.80 ± 0.050.81 ± 0.620.700
NeurosurgeryBERT1.12 ± 0.280.80 ± 0.100.002 *
LSTM1.21 ± 0.300.77 ± 0.140.002 *
Orthopedic departmentBERT0.78 ± 0.340.71 ± 0.380.630
LSTM0.92 ± 0.280.68 ± 0.420.089
Obstetrics and GynecologyBERT0.88 ± 0.030.64 ± 0.240.009 *
LSTM1.02 ± 0.190.53 ± 0.280.004 *
Ophthalmology departmentBERT0.56 ± 0.170.55 ± 0.260.690
LSTM0.60 ± 0.090.57 ± 0.300.222
Rehabilitation departmentBERT0.88 ± 0.120.77 ± 0.220.180
LSTM1.06 ± 0.380.77 ± 0.380.180
a Original MAE: Expert’s score—BERT/LSTM prediction score. b LMM-modified MAE: Expert’s score—LMM-modified BERT/LSTM prediction score. *: p-value < 0.05.
Table 4. Experts’ scores, BERT and LSTM prediction scores, and MAE of different blocks.
Table 4. Experts’ scores, BERT and LSTM prediction scores, and MAE of different blocks.
BlockExperts’ Score (a)BERT Score (b)LSTM Score (c)p-ValueLMM-Modified BERT Score (d)LMM-Modified LSTM Score (e)p-Value|a-b| #|a-d| #p-Value|a-c| #|a-e| #p-Value
A7.44 ± 0.667.35 ± 0.176.99 ± 0.17<0.001 *7.08 ± 0.567.02 ± 0.660.6260.83 ± 0.270.66 ± 0.380.008 *0.97 ± 0.330.63 ± 0.43<0.001 *
B7.35 ± 0.517.43 ± 0.067.32 ± 0.170.0877.32 ± 0.477.38 ± 0.540.8240.7 ± 0.130.51 ± 0.170.013 *0.76 ± 0.160.45 ± 0.20.002 *
C7.88 ± 0.147.56 ± 0.097.4 ± 0.10.016 *7.59 ± 0.187.63 ± 0.240.7400.69 ± 0.030.54 ± 0.10.005 *0.77 ± 0.080.41 ± 0.14<0.001 *
D7.94 ± 17.43 ± 0.197.13 ± 0.080.005 *7.57 ± 0.67.61 ± 0.840.9321.29 ± 0.290.88 ± 0.310.042 *1.44 ± 0.30.74 ± 0.350.004 *
E7.74 ± 0.917.51 ± 0.086.98 ± 0.18<0.001 *7.19 ± 0.457.12 ± 0.560.7721.05 ± 0.330.85 ± 0.40.2271.25 ± 0.410.8 ± 0.350.016 *
F7.3 ± 0.636.97 ± 0.246.61 ± 0.170.004 *6.75 ± 0.546.69 ± 0.740.8740.88 ± 0.220.73 ± 0.280.2581 ± 0.270.78 ± 0.30.154
G7.76 ± 0.157.46 ± 0.087.08 ± 0.11<0.001 *7.52 ± 0.257.46 ± 0.370.7070.69 ± 0.080.63 ± 0.120.2380.93 ± 0.120.6 ± 0.20.003 *
H8.41 ± 0.478.1 ± 0.068.05 ± 0.180.4367.93 ± 0.597.95 ± 0.670.9620.71 ± 0.230.58 ± 0.480.4891 ± 0.30.51 ± 0.550.045 *
#: The mean absolute error (MAE), the absolute value of the original score minus the predicted score. *: p-value < 0.05.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, C.; Lee, Y.-T.; Wu, F.-J.; Lin, S.-A.; Hsu, C.-J.; Lee, C.-C.; Tsai, D.-J.; Fang, W.-H. The Application of Projection Word Embeddings on Medical Records Scoring System. Healthcare 2021, 9, 1298. https://doi.org/10.3390/healthcare9101298

AMA Style

Lin C, Lee Y-T, Wu F-J, Lin S-A, Hsu C-J, Lee C-C, Tsai D-J, Fang W-H. The Application of Projection Word Embeddings on Medical Records Scoring System. Healthcare. 2021; 9(10):1298. https://doi.org/10.3390/healthcare9101298

Chicago/Turabian Style

Lin, Chin, Yung-Tsai Lee, Feng-Jen Wu, Shing-An Lin, Chia-Jung Hsu, Chia-Cheng Lee, Dung-Jang Tsai, and Wen-Hui Fang. 2021. "The Application of Projection Word Embeddings on Medical Records Scoring System" Healthcare 9, no. 10: 1298. https://doi.org/10.3390/healthcare9101298

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop