Next Article in Journal
Lessons Learned from the COVID-19 Outbreak as Experienced by ICU Nurses: Manifest Qualitative Content Analysis
Next Article in Special Issue
Transforming Healthcare Analytics with FHIR: A Framework for Standardizing and Analyzing Clinical Data
Previous Article in Journal
Research Mapping of Trauma Experiences in Autism Spectrum Disorders: A Bibliometric Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Review on Electronic Health Record Text-Mining for Biomedical Name Entity Recognition in Healthcare Domain

1
School of Computer Science, Harbin Institute of Technology, Harbin 150001, China
2
Department of Computer Engineering, Gachon University, Seongnam 13120, Republic of Korea
*
Author to whom correspondence should be addressed.
Healthcare 2023, 11(9), 1268; https://doi.org/10.3390/healthcare11091268
Submission received: 28 February 2023 / Revised: 24 April 2023 / Accepted: 26 April 2023 / Published: 28 April 2023
(This article belongs to the Special Issue Data Mining and Sentiment Analysis in Healthcare)

Abstract

:
Biomedical-named entity recognition (bNER) is critical in biomedical informatics. It identifies biomedical entities with special meanings, such as people, places, and organizations, as predefined semantic types in electronic health records (EHR). bNER is essential for discovering novel knowledge using computational methods and Information Technology. Early bNER systems were configured manually to include domain-specific features and rules. However, these systems were limited in handling the complexity of the biomedical text. Recent advances in deep learning (DL) have led to the development of more powerful bNER systems. DL-based bNER systems can learn the patterns of biomedical text automatically, making them more robust and efficient than traditional rule-based systems. This paper reviews the healthcare domain of bNER, using DL techniques and artificial intelligence in clinical records, for mining treatment prediction. bNER-based tools are categorized systematically and represent the distribution of input, context, and tag (encoder/decoder). Furthermore, to create a labeled dataset for our machine learning sentiment analyzer to analyze the sentiment of a set of tweets, we used a manual coding approach and the multi-task learning method to bias the training signals with domain knowledge inductively. To conclude, we discuss the challenges facing bNER systems and future directions in the healthcare field.

1. Introduction

1.1. Background of Biomedical Name Entity Recognition (bNER)

Information extraction in biomedical informatics is essential for handling medical literature, which grows exponentially and extracts information in medical research [1]. Biomedical informatics analyze and identify biomedical name entity recognition (bNER) and healthcare as biomedical entities in electronic health records (EHRs), such as drugs (medicines), proteins from compound genes, and diseases in unstructured medical texts [2,3,4,5,6]. With the unprecedented expansion of electronic medical records (EMR) worldwide, millions of pieces of data have been collected since the publication of the basic norms of EMR [7,8,9,10]. The correlated features, temporal dynamics, and uncertainty propose a variational recurrent imputation network that combines imputation and the prediction of treatment information. Herbal treatment in Chinese medicine has evolved, over thousands of years, into a unique and complicated system [11,12]. Western countries are getting more into it as an alternative and complementary medicine. Unfortunately, this unstructured data cannot be directly used in the clinical practice setting to improve patient care [13]. Based on 4 years of EHR (2018–2022), covering the entire region or country, an accurate and transparent Bayesian model is constructed on knowledge of how to test the model [14]. Data-driven studies, health outcomes management, and clinical decision-making are increasingly made possible with the improvement of Information Technology’s positive impact [15].
In addition, name entity recognition (NER) aims to identify text belonging to predefined semantic types such as a person, location, organization (P-L-O) [16], etc. NER is an information extraction tool and plays a variety of roles in natural language processing (NLP), question answering (Q/A), knowledge-based (KB) [17] automatic text summarization (ATS), sentiment analysis [18], machine translation (MT) [19], and information retrieval (IR) [20]. The named entity (NE) is identified by the names P-L-O (medical code, amounts, period, and percentage expression) into predefined classes in the EHR [21,22]. Recently, NER has gained interest and is being discussed at several scientific conferences (e.g., CoNLL03 [23], SemEval [24], and IREX). NE describes something or someone such as a place, person, or location [25]. Semantic (sense) ambiguity, as well as the model’s stability and generalizability, is one of the most common problems with NER. Moreover, proper nouns may have multiple meanings depending on the context. Mansouri et al. [26] limited natural kind terms and proper names such as biological species and substances. However, Kripke et al. [27] confined the word-named entities to a rigid designator. NEs are domain-specific (proteins, enzymes, and genes) and general (person/location). Specifically, this study examines generic and domain-specific bNER in unilingual (English language) healthcare text-mining propaganda (https://corenlp.run/ (accessed on 03 January 2023)). Furthermore, this study is beyond the scope of NER’s work on multilingualism.
Moreover, the development of appropriate models in NLP has made it possible for recent advances in propaganda detection to be achieved. Social media, fake news, and various propaganda techniques have contributed to increased misinformation on the Internet in the last ten years [28,29,30,31]. In addition to threatening the social order, broadcasting, economy, climate change, health, and democracy, these techniques are linked to terrorism [32,33]. An individual or group may commit a propaganda act when they deliberately intend to influence the actions or opinions of other individuals or groups in a way that will lead them toward achieving a specific goal [34]. According to O’Donnell et al. [35], propaganda is a deliberate and systematic process that influences a person’s opinions, thoughts, and behaviors to achieve the campaign’s intent. It is important to distinguish propaganda from lies, distortions, and deception, where distorted messages are aimed at misleading or deceiving the public, either intentionally or unintentionally [36,37]. The concept of propaganda, at the level of social media, should be well understood since it affects the harmony of the social network. The short-text propaganda stream classification is extended using a large-scale semantic network developed from a Web corpus derived from a large collection of short-text propaganda streams [38]. Currently, a significant challenge is facing one of the most challenging and significant tasks in EHR: the classification of short-text streams.

1.2. Impact of Biomedical EHRs Analysis

In biomedical analysis, NLP is becoming more common, which helps process and analyze EHRs automatically. The pre-trained model enables representations, a method of computing the contributions of different diseases and admissions, to maximize EHR data utilization and fully utilize healthcare domain information [39]. For cross-domain medical applications, a missingness-aware feature extraction network is an additional feature for medical analysis by incorporating healthcare conditions into the feature extraction [40]. EHR mining integration methods are applied to real-time practice parameters using meta-learning methods that automatically adjust sample weights based on real-time parameters [41]. Random noise in the samples, as well as sample bias, may lead to issues with overfitting and poor generalization. NE corpora that humans annotate are essential for training and testing automated bNER systems. Unified clinical terminology systems (UCTS), clinical oncology, and medical databases, such as DrugBank, contribute to bNER in clinical records in English, including the unified medical language system and others [42]. Certain entity types in the EHR are annotated, including drugs, treatments, symptoms, tests, temporary words, anatomy, and operations [43,44,45,46,47]. In COVID-19, more organizations use healthcare propaganda to spread awareness and promote prevention and control. Organizations responsible for public health should monitor social media communication to discover what rumors and myths are circulating or actively promoted in the community, what misconceptions exist about various treatments and diseases, and what ideas, problems, and rumors concern the public the most. This study reviews healthcare entries based on clinical records used to treat various diseases.

1.3. Motivations of the Study

This study presents an overview of the shift in techniques from machine learning to handcrafted rules. Kundeti et al. [48] focused on NER from the perspective of the clinical domain in 2016. Particularly, we focus on the fallacies, challenges, and analyses of the NE approach to clinical reports. The recent review mentions new domains bNER [49] and clinical NER (cNER) [50]. Accordingly, the existing review focuses on feature-based ML models rather than DL-based and bNER systems in healthcare. However, deep-learning (DL) techniques are not included in their analysis. DL has been applied to bNER over the last few years, significantly advancing the state-of-the-art performance for tagging chunking, name entities, and semantic labels [51,52,53]. As a result of this trend, this review is focused on mining text from electronic health records to recognize the names of biomedical entities within the healthcare domain.
This paper discusses techniques that influence bNER in EHR corpora, addressing issues of both approaches, which improve the annotation process and quality. This study aims to identify factors influencing bNER performance, healthcare issues, and associated challenges by comparing their DL architectures. Although bNER studies have thrived for a few decades now, the field of reviews on bNER studies has developed more, and to our knowledge, few reviews in this field have been published to date. The following are our main contributions:
  • This study examines DL techniques in bNER, as well as their applications in propagandist text in electronic health records, and it provides researchers and practitioners with the current understanding of these techniques. A major focus of this study is to consolidate bNER and provide the research community with relevant corpora for bNER.
  • This study presents a broad review of DL techniques for bNER entries in clinical records, for online or electronic treatment, for the diagnosis and treatment of various diseases.
  • The study proposes a new classification that categorizes DL-based bNER approaches, input representations, and encoders and decoders. The proposed classification receives context for tag decoding and predicts labels for words based on given sequences.
  • Additionally, this study examines the most representative DL methods for solving bNER problems and challenges. As a result, we summarize the challenges and future directions facing healthcare propaganda and bNER systems in the coming years.
  • Information extraction (IE) and NLP are two powerful tools that can be used to process and analyze text. IE is the process of identifying and classifying named entities in text. At the same time, NLP is a field of computer science that deals with the interaction between computers and human (natural) languages.
The paper is structured as follows: Section 1 introduces the topic of text mining for healthcare. Section 2 provides a detailed analysis of the proposed healthcare solution and data collection procedures. Section 3 discusses the previous research on this topic and the gaps and limitations of this research. Section 4 of the paper discusses the in-depth healthcare analysis. It covers clinical decision support, healthcare administration, and public healthcare. Section 5 discusses the findings, challenges, limitations, and future work. Section 6 concludes the paper.

2. Material and Methods

In order to help scientific authors report a wide range of systematic reviews and meta-analyses, PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) is an evidence-based minimum set of items that are designed to help them assess the benefits and harms of a variety of healthcare interventions [54]. Database searches yielded 3128 records, and gray literature searches yielded 13. After removing duplicate articles and reviewing abstracts, the full text of 409 articles is reviewed. A flow chart illustrating the PRISMA process (identification, screening, and inclusion) [55] is shown in Figure 1.
In the identification process, each database is searched individually, including all search terms and truncations. Each database’s final number of records or articles is determined once all search terms are combined and all applicable limits are applied. Additionally, many researchers add notations in the box to indicate how many results there were for each database search, which could vary from PubMed (n = 1186) to healthcare (n = 84), or vice versa. The articles appearing more than once in search results are removed to avoid reviewing duplicate articles. A screening process is performed to determine whether the records are relevant to the research query, which is included in the full-text, according to certain criteria. To the right, note the number of articles that can find the full text that were unable to find it during text screening, and list any articles that cannot find the full text obtained in preparation for the full text screening (i.e., Screening Full-Text Records). Adding the record that is not retrieved to the number of reports sought for retrieval should give you this number. Review their full texts to determine whether these articles qualify for inclusion in the “Include and Exclude Report” systematic review. Enter the number of articles removed from the full-text screening process in the “Reports excluded” box, along with a description of the exclusions’ causes and the number of records removed for each reason. This study identified several shortcomings, such as incorrect settings, populations, interventions, and doses. To calculate the number of excluded articles, subtract the number of records excluded during the eligibility review of full-text articles from the total number of articles reviewed for eligibility. This method counts excluded articles once in the list, even if they meet multiple exclusion criteria. If necessary, combine the number with the literature search results in the box labeled “Studies included in the review.”

2.1. Datasets and Tools

Annotations of high quality are crucial to both the learning process and the evaluation process of a model. The following summarizes widely used datasets in unilingual tools (English) in NER. Tags are created using a corpus of documents containing annotations for multiple entities [56]. Data sources and the number of tags or entities are widely used for some datasets, as NER tasks are annotated, mainly, on news articles with a few entity types. The tag types used in MUC-2 to MUC-7 become significantly similar [57]. Along with these domain-specific datasets, a number of text-based datasets were developed with OntoNote information, as shown in Table 1.
Several recent NER studies used the English-based Conference on Natural Language Learning 2003 (CoNLL03) and OntoNotes Release 5.0 (OntoNotes05 (https://catalog.ldc.upenn.edu/LDC2013T19 (accessed on 8 January 2023)) datasets. However, with OntoNotes, 2005 Wikipedia, a huge corpus containing multiple genres is annotated with shallow semantics and structural information. A total of 18 tag types for entities are annotated in the texts. The English dataset contains a considerable percentage of news with annotations in three entity types (organization, place, and location). Table 1 summarizes the popular ones for multilingual (non-English) NER by academia and industry. Reuters News annotations in two languages are included in CoNLL03: German and Dutch. The main content contains five terms: clinical manifestation, syndrome, disease, treatment law, and herbs. Other side content information for Traditional Chinese Medicine (TCM) clinical treatment records includes a patient’s name and age. The datasets are generally very complex, large, hierarchical, heterogeneous, and differ in features [58,59]. Data preprocessing and transformation are required even before mining and discovery are applied.
Formal techniques can be used to check the strict correctness of AI-based solutions, as AI terms broadly include closely related areas (such as machine learning) [60,61]. Mathematical proofs, simulations, and testing are some methods that can be used. However, mathematical proofs provide the highest level of assurance, while simulations and testing are less time-consuming but do not provide the same level of assurance. In addition to formal techniques, a number of other things can be done to check the correctness of AI-based solutions, especially when it comes to gathering and processing data. These include:
  • Data provenance: This involves tracking the origin of data and how it has been processed. This can help to identify any errors or biases in the data.
  • Data quality: This involves checking data quality to ensure it is accurate, complete, and consistent.
  • Data auditing: This involves reviewing the data to identify errors or biases.
  • Data visualization: This involves using graphs, charts, and other visuals to represent data in a way that makes it easy to understand. This can help to identify any errors or biases in the data.
  • Expert review: This involves having an expert review the AI solution to ensure it is correct.
These steps make it possible to check the correctness of AI-based solutions, especially when gathering and processing data.

2.2. Deep Learning Role in bNER

DL identifies the data representations at multiple levels by combining multiple processing layers with learning representations of data [62]. Traditional machine-learning techniques had limitations in processing raw natural data. In the forward pass, a weighted quantity of the inputs from the earlier layer is computed (via a non-linear function), and the result is passed through a nonlinear function. This study provided a comprehensive overview of a DL-based recommendation, which is analyzed within four dimensions: utilizing a system, remedies for challenges, awareness, purposeful properties, and prevalence recommendation system domains [63,64]. Machines are used for representation learning, i.e., to learn how to identify or classify data from raw data by feeding them raw data and automatically discovering the necessary representations. In DL, representations are addressed at multiple levels, each of which is derived by combining simple but non-linear modules to transform the original representation into an advanced abstract representation. DL provides representation learning and semantic composition capabilities as a function of both vector representation and neural processing. It allows raw data to be fed into a machine that automatically discovers latent representations, as well as processing that is required for detection or classification. In DL-based models, non-linear activation functions are used to learn complex and intricate features by comparing them to linear models.
In addition, NER features are designed significantly more efficiently using DL. It requires considerable engineering expertise and domain knowledge to use feature-based approaches. Meanwhile, DL-based models automatically extract useful representations from raw data. In addition, end-to-end bNER models undergo end-to-end gradient descent (GD) training [11]. In order to design potentially complex bNER systems, it takes advantage of this property. The existing taxonomy describes encoders at the character and word levels, as well as tag decoders [8,65,66]. We argue that the encoder at the word level in a DL-based NER model uses inaccurate word-level information. The EHR word-level and character-level representations are used to gather raw features and context dependencies for tag encoders and decoders. This study summarizes recent improvements in bNER, presenting the general architecture for the treatment and the journey of cancer patients alongside healthcare professionals, as shown in Figure 2.
Healthcare professionals detect the patients’ symptoms through diagnosis, which is treated through chemotherapy (in cancer) and other scientific procedures [67]. Each patient’s data is recorded through electronic information storage devices for future usage, treatments, and research. In addition to embedding at the word and character levels, distributed representations also incorporate features such as POS tags and gazetteers that are proven to be effective in feature-based representations. In context encoding, Convolutional neural networks (CNNs) [68], recurrent neural networks (RNN), or other networks are used to capture context dependencies. The decoder predicts tags in the input sequence based on the input sequence data. It is predicted that each token will have a tag of BIOES, indicating the B-(begin), I-(inside), O-(outside), E-(end), and S-(singleton) of the named entity [69]. In addition to detecting entity boundaries, the tag decoder is trained to classify and detect text spans to entity types by detecting entity boundaries. Note that the tag sequence or tag notations are not controlled during the training process when implementing the transformers-based (RoBERTa) model [70]. Conditional random fields (CRF) is used to control the sequence of each tag with tokens. An analysis of word shift plots is performed to determine what creates the dip in note scores after 21 days, as shown in Figure 3.
The ‘words’ associated with lung cancer treatment, such as ‘mouth’, ‘chemotherapy’, ‘dose’, ‘lung’, ‘treatment’, and ‘cancer’ perform to drive the score down. While more broad words, such as ‘support’, ‘care’, ‘independent’, ‘patient’, and ‘activity’, associated with patient care, inspire the note scores on peak days [18].

2.3. Deep Learning-Based Healthcare System

DL is an emerging technique of artificial intelligence used in several applications, including healthcare monitoring [71], speech recognition [72], computer vision [73], portfolio optimization [74], NLP [75], social network analysis [76], visual data processing, etc. Data analysis is becoming increasingly important in creating patterns and assisting in decision-making today. Moreover, this state-of-the-art method is capable of improving learning execution, expanding the scope of the research, and simplifying the measurement process. Health propaganda detection significantly deviates from other clarifications and, thereby, arouses doubt that a different mechanism in treatment causes it. Several detection techniques have been developed to solve similar problems. DL is one of the most effective methods for detecting misinformation within social networks and electronic data. In this case, DL techniques are applied to the same type of misinformation and the healthcare propaganda problem. Moreover, these DL techniques are dependent on a variety of data features, and they detect propaganda automatically.
DL algorithms classify tweet data into non-propaganda and propaganda classes [77]. Together with CNNs, long-short term memory (LSTM) [4], and Transformers-based Bidirectional Encoder Representations from Transformers (BERT) [78] language models in bNER, numerous researchers are driving the development of DL networks. Several studies have demonstrated that deep neural networks have improved NER in the biomedical field, such as BioNER [79], PubTator [80], Hunner. [81], GeneView [82], GENIA [83], and BioCreative [84]. Furthermore, word2vec models are compared [85], including FastText models [86], ELMo models [87], and a well-tuned BERT model [88], on a closed biomedical domain corpus to see how they contribute to the deep neural network (DNN). Persuasion is categorized as disinformation, which defines propaganda as a deliberate and systematic process to influence a person’s opinions, thoughts, and behavior in order to attain the intended aim [35]. Beside healthcare, propaganda plays an important role in politics, as it is used, primarily, to gain people’s faith in a person or a group. Governments and other organizations generally use propaganda for fast influence, which is either true or false [89]. Propaganda comes in several forms: text, images, and videos. In conclusion, researchers previously found that propaganda is related to healthcare (COVID-19), political discussions, and sectarianism.
Approximately 60% of the population is affected by Diabetes Mellitus, which is a disease with a high mortality rate. This study uses a multi-expert system ensemble model to diagnose diabetes type II, according to expert systems [90], to analyze various DL techniques for binary classification regarding illness, i.e., to diagnose whether a patient is suffering from disease or not [91]. Public authorities and governments urge consumers to verify the authenticity of circulating news on social media presenting misinformation about health, resulting in anxiety and panic among patients diagnosed with COVID-19 [92]. A data-mining analysis of information distribution on major social media platforms exposed COVID-19 rumors on a large scale by using information diffusion on COVID-19 [93]. Millions of people are affected by the COVID-19 spread worldwide, mostly in Europe, the United Kingdom, Spain, the United States, and the Asian subcontinent, and COVID-19 is the subject of extensive research for developing drugs to treat it. In order to cure this deadly virus, misinformation about how to treat it is spread extensively. Some of this misinformation claims that drinking alcohol or drinking cow urine will cure the virus. However, none of these claims is medically proven to exist. The World Health Organization suggested that the public take precautions against COVID-19 and has expressed concern about public health. Many hashtags are used to spread information about COVID-19 through social networks, such as Facebook and Twitter. Furthermore, the next section is an overview on the development of DL-based sentiment analysis.

2.4. Development of DL Sentiment Analyzer

To create a labeled data set where our machine learning sentiment analyzer can analyze the sentiment of a set of tweets, a manual coding approach is used, as with the topic classification. The reliability of the intercoder is measured by selecting a subsample of 300 Facebook reviews at random. The neutral or unidentified reviews are removed since we prefer binary sentiment classification. A stratified sampling technique is used to select training data for sentiment analysis, in which 80% of reviews within a class are selected for training. In order to train machine learning models, the Python packages (nltk, spacy, and scikit-learn) are utilized, which employ three basic classifiers: Naïve Bayes (NB), support vector machines (SVM), and logistic regression (LR). Several techniques are applied in this study to develop a sentiment analyzer [94,95,96,97,98], which analyzes patient online reviews to classify emerging and fading themes and sentiment trends during the early stage of the COVID-19 outbreak [99].

2.4.1. Multi-Task Healthcare Analysis

The multi-task learning method uses the domain knowledge contained in training to inductively bias the training signals in an enhanced simplification by using domain information. The shared representation system learns tasks in parallel, which helps other tasks, in better ways, in parallel [100,101,102]. The multi-task algorithm can find internal representations used for various tasks, which learn regularities in a specific language and achieve POS, Chunk, NER, and SRL together. Lin et al. [103] combined transfer modules using two sharing parameters. A multilingual architecture for multi-task low-resource conditions effectively transfers dissimilar types of knowledge (auxiliary) to advance the main model. It is possible to use a multi-task learning framework in addition to NER, along with other tasks, to extract entities and relationships jointly [104]. bNER has developed four steps (data collection, annotation, model training, and evaluation) as two associated subtasks involving biomedical records and entity segmentation corpus likelihood [8,105,106,107].
Data collection: The first step is collecting a biomedical text corpus. This corpus should be large and diverse, and it should include a variety of biomedical entities. Subtasks are:
  • Corpus cleaning: The corpus may need to be cleaned before it can be annotated. This may involve removing stop words, punctuation, and other noise from the corpus.
  • Feature extraction: Features can be extracted from the corpus to help the machine learning model learn to recognize biomedical entities. These features can be lexical, syntactic, or semantic in nature.
  • Annotation: The next step is to annotate the corpus with biomedical entities. This can be done manually or automatically. Subtasks are:
  • Entity segmentation: The corpus is first segmented into entities. This can be done using a variety of techniques, such as rule-based systems or machine-learning models.
  • Entity labeling: Once the entities are segmented, they are labeled with the appropriate biomedical entity type. This can also be done using a variety of techniques, such as rule-based systems or machine-learning models.
Model training: Once the corpus is annotated, a machine-learning model can be trained to recognize biomedical entities. This model can be trained using a variety of machine-learning techniques, such as support vector machines or neural networks. Subtasks are:
  • Model selection: The first step is to select a machine-learning model. This can be done by considering the corpus size, the entities’ complexity, and the model’s desired performance.
  • Model training: The model is then trained on the annotated corpus. This can be done using various techniques, such as batch training or online training.
  • Model evaluation: A held-out test set evaluates the model’s performance. This ensures that the model is not overfitting the training data.
Evaluation: The final step is to evaluate the performance of the model. This can be done using a held-out test set or a cross-validation approach. Subtasks are:
  • Precision: Precision is the fraction of entities that the model correctly identifies.
  • Recall: Recall is the fraction of entities that are actually in the corpus and that are correctly identified by the model.
  • F1-score: The F1-score is a weighted harmonic mean of precision and recall.
As NER on each dataset is considered a separate task in the biomedical domain, multi-task tuning is necessary to account for the differences between datasets [108,109]. In this case, the characters and words in the different datasets are assumed to be the same. As a result of this multi-task learning, the data is analyzed more efficiently, and more generalized representations are learned with the models.

2.4.2. Healthcare Mining from EHR

Researchers in the mining community are giving considerable attention to healthcare. Healthcare methods are “a series of activities aimed at diagnosing, treating and preventing diseases and improving a patient’s health” [110]. Medical treatment and healthcare organization processes can be discovered and analyzed with the help of process mining [111]. Analyzing the behavior of processes between healthcare organizations is another application of process mining. In Elhaj et al. [112], supervised machine learning models are compared to determine which model is better at evaluating patient triage outcomes. In addition, Pandya et al.’s [113] InfusedHeart Framework outperformed RNNs, LSTMs, Bi-LSTMs, CNNs, K-means Clustering, and SVM-based classification approaches to extract more information from the recorded heartbeat sound signals.

2.4.3. Healthcare Context in AI

Medical AI research presents unique challenges compared to research in other technical fields. A significant problem in medical diagnostics (e.g., the exact relationship between diseases and their causes) is the lack of such computable models. Even when it comes to the same clinical case, a doctor’s response may differ greatly from the response of another doctor. Consequently, it would be very challenging to train AI-based tools with biased responses, which are contaminated with discrete biases from clinicians who are ignorant of the truth of the reality in which they operate. The researchers are presented with the grand challenges of generating human-centered AI technologies that are ethically fair and enhance the human condition [114].
The healthcare industry is full of medical challenges with various characteristics, making AI research difficult. An AI model for cancer may not be able to be trained using the mathematical model developed for cardiovascular applications, for example, because the process is not generalized. Consequently, AI models tailored to cardiovascular applications cannot be trained. Moreover, the results may be inaccurate if underserved individuals and communities with infirmities are not considered during the AI system’s development. Depending on the application, specialized AI processes might be necessary. Several factors influence this, including healthcare decisions, target populations, data variability, and the quantity of helpful information contained within the data. Expert opinions are at the bottom of the hierarchy, followed by randomized controlled trials, systematic reviews, and meta-analyses, as shown in Table 2.
In recent years, expert opinions are outdated decision-making methods that have become obsolete. There is the possibility that an expert opinion may be unfitting, and often, there is disagreement among experts. Therefore, it occupies a low position in the hierarchy due to its low ranking. In clinical medicine, a series of cases refers to patients who share specific clinical, pathophysiological, or surgical characteristics used to identify particular disease features, treatments or diagnoses. According to the literature, high-quality evidence is needed for clinical practice, which is achieved traditionally with randomized controlled trials (RCT) [127]. Therefore, this is the decisive preview and exercise for patients. New drugs undergo evaluation phases, starting with pre-clinical, usually leading to the gold standard, which tests a larger population.

2.5. Tag Encoder-Decoder Architecture

The CRF is a statistical tool that evaluates the probability of a given observation sequence based on a global set of variables, and it is widely used in feature-based supervised learning approaches. CRF is used as a tag decoder in many DL models of neural networks, such as LSTMs and CNNs, which are built on top of each other and take advantage of the special properties of the CRF [128]. As shown in Table 3, the CRF tag decoder is the most commonly chosen tag decoder for CoNLL03 and OntoNotes5.0, and with the CRF tag decoder, state-of-the-art performance is obtained in these applications. As a result, CRFs, which are designed to fetch the information of a segment at the word level, cannot take full advantage of the information contained at the segment level because segment-level information is not encoded at the word level.
Among the many types of machine learning algorithms, the RNN is an increasingly popular one. In addition to solving or approximating increasingly difficult tasks, this RNN’s output is passed into a series of layers of RNNs; as explained below, short-term memory is a characteristic of RNNs. LSTM-Gates, coupled with a network, can generate long-term memory. This model updates the hidden unit at each time step, and the sequence length is not restricted. GRU-TN (gated recurrent unit tensor network) and LSTM-based models with bidirectional LSTM are proposed when dealing with a variety of sequence data types; they facilitate the generation of sequence data and improve the model’s flexibility [133,134]. Several RNN models, including BiRNN, divide the neurons into forward and backward computation with regard to their function [135,136]. Among the achievements of RNN are the recurrent gated units (GRU), along with the LSTM, which have shown great promise in modeling sequential data.
A biRNN effectively extracts input and future (positive and negative) information simultaneously for a specific timeframe, which is not the case with regular RNNs [137]. NER performance is enhanced with limited data by adding knowledge and data augmentation via a multi-task bidirectional RNN model combined with deep transfer learning [138,139]. The RNN encoder–decoder encodes a sequence of symbols into a static length vector representation of multi-RNN neural networks and symbols [140]. BiLSTM and CRF are applied as the basic RNN and DL models to encode and decode the context information [141,142,143,144,145,146,147,148]. According to Santos et al. [149], POS tagging is performed using DNN that learns character-level word representations and associates them with regular word representations. Recently, some studies [150] have integrated hypergraph representation, for nested entities, using linguistic features with designed LSTM-based neural networks [151]. Tan et al. [152] proposed the construction of a novel sequence-to-set neural network for nested NERs. A non-autoregressive decoder predicts the complete set of entities in one pass while capturing their interdependencies.
Pre-trained language models (PLMs) have become increasingly popular as NLP enters a new era. When a language model (LM) encounters a specific text set, it tries to reduce the amount of confusion that results [153]. A language model encoder is used for text production, and there are numerous different classifiers within a domain, so only one needs to be trained per domain [154]. PLMs are a very popular topic in NLP, and rapid development has led to the achievements of natural language today. Over time, as DL evolved, so did the number of model parameters. A significantly larger dataset is required to prevent overfitting, as shown in Figure 4.

3. In-Depth Healthcare Analytics

There are three application areas included in this review: clinical decision support, healthcare administration, and public health. It provides a detailed description of the operational definitions for these three areas.

3.1. Clinical Decision Support

An analysis of clinical decision support is aimed, mainly, at patients with diabetes, cancer, and cardiovascular disease, as well as patients in the internal care unit (ICU). Data-mining algorithms developed by some studies are reviewed in this article. Clinical decision-making papers are organized by major disease categories, with topics investigated and data sources used, as shown in Table 4.

3.2. Healthcare Administration

The majority of articles reviewed used data mining for administrative purposes in healthcare. Data mining is applied to various areas, including data warehousing and cloud computing, quality improvement, cost reduction, resource utilization, patient management, etc. Table 5 shows the number of articles, their focus areas, associated problems, and data sources.

3.3. Public Healthcare

Data mining is increasingly used to inform public healthcare decisions. Using data mining, researchers developed automated analytics tools for non-experts, identified resource utilization and patient satisfaction, and developed healthcare programs and emergency responses. It may be possible to develop a patient-centered, robust healthcare system if this effort is continued, as shown in Table 6.

4. Results

The most recent publications are listed first, so a summary of the 60 articles is in the qualitative analysis. There were 162 facilitators in the 60 articles and 97 barriers in relation to job satisfaction in long-term care facilities. The SVM model for binary relevance, as well as the classifier chain, had the highest F1-score (0.58) when examining the Tangible dimension. According to the F1-score for reliability and assurance, LR with binary relevance has the highest score, while NB with the label power set receives the highest score for responsiveness (0.63), and LR with the label power set receives the highest score for empathy (0.88). A consistently high F1 score is achieved only by SVMs with classifier chains in all service quality (SERVQUAL) dimensions (tangibility, reliability, responsiveness, assurance, and empathy). The five-fold cross-validation performance metrics are listed in Table 7.
In spite of the lower average accuracy compared to prior supervised machine learning studies, we also observed high predictive accuracy across SERVQUAL dimensions and a high F1 score. There was a 0.63–0.74 accuracy range for tangible dimension, a 0.65–0.71 accuracy range for reliability, a 0.53–0.71 accuracy range for responsiveness, a 0.57–0.69 accuracy range for assurance, and a 0.71–0.78 accuracy range for empathy. A range of F1-scores is found, for tangible dimensions, ranging from 0.38 to 0.62; for dependability dimensions, it ranges from 0.76 to 0.81; for responsiveness, it ranges from 0.40 to 0.65; for assurance, it ranges from 0.64 to 0.70; for empathy, it ranges from 0.82 to 0.87.
The prediction performance using supervised machine learning is summarized in Table 8. The best-performing multi-label classifier and classification model, prediction performance ratings, ranged from 0.13 to 0.25, with an F1 value of 0.68 to 0.75, and the models correctly classified the reviews. Based on the SVM model and classifier chain method, overall, the SVM model with the F1-score of 0.75 has the highest accuracy (0.21). However, for the topic classification model, hamming loss is more significant because it measures how many class labels are mispredicted.
As compared to all other models, the SVM model with a classifier chain is the lowest hamming loss (0.27). After SVM with the classifier chain, SVM with binary relevance is second, and five-fold cross-validation is used to evaluate all models.

5. Discussion

This review is focused on the unique role of AI in the treatment decisions of healthcare treatment (cancer).

5.1. Summary of Finding

Extracting information from medical literature is a crucial aspect of biomedical informatics, which grows exponentially. bNER architectures are designed with DL-BiLSTM-CRF, using 1104 entities and 67,799 tokens for each annotator tag, as well as fine-grained biomedical corpus [169]. Specifically, 13 entity types are used in the language expressions in clinical records, revealing some interesting points from the corpus analysis. It is challenging to uncover the rules governing clinical records’ construction without analyzing the data. This analysis is used to identify the rules of healthcare and biomedical clinical expressions (entities) to generate questions for future research and summarize, as shown in Table 9.

5.2. Challenges of Text-Mining in Healthcare

Data mining has shown some promise in assisting healthcare researchers in modeling the healthcare system and improving healthcare delivery using its predictive techniques. However, mining healthcare data soon revealed many challenges attached to the validity of healthcare data and the limitations of predictive modeling, leading to the failure of data mining projects. Data-mining techniques and methods are gaining interest again, as the Big Data movement has gained momentum over the past few years.
DL techniques have been shown to be effective in detecting misinformation and healthcare propaganda [171]. However, a number of challenges must be addressed in order to improve the accuracy and reliability of these techniques. One challenge is the need for large and high-quality datasets. DL models require a large amount of data to train on, and this data must be of high quality in order to produce accurate results. However, it can be difficult to obtain large and high-quality datasets of misinformation and healthcare propaganda. Another challenge is the need to account for the diversity of misinformation and healthcare propaganda. Misinformation and healthcare propaganda can take many different forms, and it can be difficult to develop a single DL model that can effectively detect all forms of misinformation and healthcare propaganda. Finally, DL models can be computationally expensive to train and deploy. This can be a challenge for organizations that do not have the resources to invest in DL technology. Despite these disease challenges, DL techniques have the potential to be a powerful tool for detecting misinformation and healthcare propaganda. With continued research and development, DL models have improved to the point where they can be used effectively to protect people from the harmful effects of misinformation and healthcare propaganda.
In addition, to handle language ambiguity, we consider the annotations’ quality and consistency. Historical diagnoses are seeds that are incorporated into a patient’s comorbidity network as the patient’s comorbidities are expanded, and the disease risk is explored iteratively. The process of remote diagnosis is challenging since it is difficult to detect relevant data when observed remotely [14]. Patients may have different diseases based on the same historical diagnosis because they are different. In view of the sensitive nature of the data, it is impossible to differentiate between patients with similar historical diagnoses. An approach that uses data-driven approaches, based on the characteristics of their EHRs, is applied to predict disease, readmission times, and diagnoses of patients. EHR data is not utilized fully by most existing predictive models because some supervised events lack labels. Existing methods provide generic interpretability while being complex at the same time [39]. Data integration with multi-omics EHR data with low noise is challenging because it contributes to overfitting, poor generalization performance, and underfitting [41].
The clinical observations recorded in data mining are sparse, irregular, and highly dimensional, which creates significant challenges in estimating downstream clinical outcomes [42]. DL-based bNER systems are supervised with significant annotated data in training. Annotating data is an expensive and time-consuming process despite technological advances [56]. Annotation schemes commonly handle nested entities, fine-grained entities, and entities assigned multiple types. For many languages and domains with limited resources and skills, it is a big challenge to perform annotation tasks in a resource-poor environment since domain experts are necessary to accomplish these tasks.
Secondary data analysis is essential to ensure data quality and preprocess clinical data [59]. The treatments and diagnoses of diseases contain the needed information (for example, symptoms, diagnoses, prescriptions, and laboratory results). Since clinical data from multiple sources has different terms, big data management with lexicon sources is essential. These terms allow clinical data to be integrated, uniformly arranged, analyzed, and retrieved quickly. Developing a unified health information model is necessary for the efficient data analysis of large-scale clinical data. In addition, controlled vocabulary and terminology systems are used to enhance clinical data functionality and usability. Electronic medical records are used as clinical record structures in bNER. High data quality, preprocessing, and managing large-scale data, including multi-relational analyses based on domain knowledge, are key challenges.

5.3. Implications

This study provides useful managerial implications for healthcare data scientists, practitioners, and data analysts. Big Data has emerged recently, as biomedical data has grown explosively, and medical technologies and computers have advanced rapidly. Among all of the research articles presented in this study, one specific feature stands out: It discusses how clinical decision-making and ML approaches are used in a wide range of applications in healthcare. They include bNER using EHR data for treatment prediction, benefit package design based on healthcare needs, and research on the effects of health scheme design on healthcare utilization.
This section examines healthcare and sentiment analysis in bNER areas that have effectively used DL tools to create models to solve specific tasks using EHRs. Hence, healthcare data scientists and analysts can take advantage of the present study and the research articles in the analysis to gain a comprehensive and synthesized view of how this gap can be filled, with successful case studies, across all healthcare problems and all the issues involved in each healthcare treatment prediction.
Applying treatment prediction methodology, in practice, requires a thoughtful implementation study with public healthcare in the lead. Data science also needs a high-quality, up-to-date technical infrastructure.

5.4. Limitations and Future Work

This study discusses several limitations and research directions for future healthcare and text-mining researchers. First, this study did not analyze post-COVID-19 healthcare literature; therefore, future researchers may collect more data after the pandemic, as text-mining methods are more efficient with larger datasets. A possible future work to address this limitation could be to collect and analyze post-COVID-19 healthcare literature using a text-mining approach. This could involve identifying relevant healthcare topics in post-COVID-19 literature and quantifying the frequency and sentiment of those topics. In addition, NLP techniques were used to identify relationships between topics within the literature [99,172,173]. This could provide insights into how healthcare topics have changed after the COVID-19 pandemic.
Second, this study employed nested named entities, which can be difficult to identify and extract because they are not explicitly labeled in the text. Hence, nested entities pose a limitation to current NER tasks in the healthcare domain. To address this limitation, future research should focus on the development of improved annotation schemes to incorporate nested entities, more robust models to detect nested entities, and experimentation with different feature and model architectures. Moreover, it should be investigated how to better detect nested entities by including contextual information in the models.
Third, in this study, we expect to pay attention to bNER as modeling languages advance and real-world applications increase. bNER, on the other hand, is applied before downstream processing. Therefore, various clinical datasets with disparate characteristics of cancer patients permit further investigation to find the behavior of imputation in healthcare studies [42]. The sensitivity of the data makes it impossible to access other information, such as medical diagnoses, except for personal information.
Fourth, the model in this study cannot differentiate between patients with the same diagnosis. It uses co-occurring frequency as a form of external knowledge, but disease relationships in the EHR are not limited to this type of knowledge [14,39]. The length of the time series is also constrained, which limits the model’s ability to learn complex patterns. The study proposes a future direction of research to address these limitations by developing a model that can differentiate between patients with the same diagnosis, using additional forms of external knowledge and removing the length constraint on time series. As a result, EHR predictions would be more accurate and comprehensive [41].
Fifth, this study proposes a bNER system for EHRs in English. However, the system is limited to English text. To address this limitation, future work could develop a multilingual bNER system. This could be done by training a new bNER system on a dataset of biomedical text in multiple languages. Alternatively, future work could develop a bNER system that can automatically translate biomedical text into a single language before performing NER. This would allow the system to handle multilingual text without having to be trained on a dataset of biomedical text in each language.

6. Conclusions

A biomedical informatics system analyzes and identifies bNER and healthcare propaganda, as biomedical data moves from cellular to electronic form. Information Technology and computational techniques are needed to uncover new information about researchers and patients in the biomedical field. EHRs contain basic and relevant medical information in their text, identifying medical entities with special meanings, such as people, places, and organizations. The expansion of EHR and clinical data has been unprecedented, as well as the spread of misinformation, fakes, and propaganda through social media and the free-text internet. This study aims to review recent studies on DL-based bNER and health propaganda solutions to help potential researchers build a comprehensive understanding of this field. Additionally, we present some preliminary findings regarding classifying healthcare tasks, traditional approaches for data mining, basic concepts, and evaluation metrics in DL.

Author Contributions

Conceptualization, P.N.A. and A.M.S.; methodology, P.N.A.; software, P.N.A.; validation, P.N.A., A.M.S. and K.L.; formal analysis, P.N.A.; investigation, A.M.S.; resources, K.L.; data curation, A.M.S.; writing—original draft preparation, P.N.A.; writing—review and editing, A.M.S.; visualization, A.M.S.; supervision, K.L.; project administration, K.L.; funding acquisition, K.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant of the Korea Health Technology R&D Project through the Korea Health Industry Development Institute (KHIDI), funded by the Ministry of Health & Welfare, Republic of Korea (grant number: HI22C1569) and was supported by the National Research Foundation of Korea (NRF-2022R1F1A1069069).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Savova, G.K.; Masanz, J.J.; Ogren, P.V.; Zheng, J.; Sohn, S.; Kipper-Schuler, K.C.; Chute, C.G. Mayo Clinical Text Analysis and Knowledge Extraction System (CTAKES): Architecture, Component Evaluation and Applications. J. Am. Med. Inform. Assoc. 2010, 17, 507–513. [Google Scholar] [CrossRef] [PubMed]
  2. Shickel, B.; Tighe, P.J.; Bihorac, A.; Rashidi, P. Deep EHR: A Survey of Recent Advances in Deep Learning Techniques for Electronic Health Record (EHR) Analysis. IEEE J. Biomed. Health Inform. 2017, 22, 1589–1604. [Google Scholar] [CrossRef] [PubMed]
  3. Han, S.; Zhang, R.F.; Shi, L.; Richie, R.; Liu, H.; Tseng, A.; Quan, W.; Ryan, N.; Brent, D.; Tsui, F.R. Classifying Social Determinants of Health from Unstructured Electronic Health Records Using Deep Learning-Based Natural Language Processing. J. Biomed. Inform. 2022, 127, 103984. [Google Scholar] [CrossRef]
  4. Park, J.; Artin, M.G.; Lee, K.E.; Pumpalova, Y.S.; Ingram, M.A.; May, B.L.; Park, M.; Hur, C.; Tatonetti, N.P. Deep Learning on Time Series Laboratory Test Results from Electronic Health Records for Early Detection of Pancreatic Cancer. J. Biomed. Inform. 2022, 131, 104095. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, X.; Wang, H.; He, T.; Liao, Y.; Jian, C. Recent Advances in Representation Learning for Electronic Health Records: A Systematic Review. J. Phys. Conf. Ser. 2022, 2188, 012007. [Google Scholar] [CrossRef]
  6. Li, I.; Pan, J.; Goldwasser, J.; Verma, N.; Wong, W.P.; Nuzumlalı, M.Y.; Rosand, B.; Li, Y.; Zhang, M.; Chang, D. Neural Natural Language Processing for Unstructured Data in Electronic Health Records: A Review. Comput. Sci. Rev. 2022, 46, 100511. [Google Scholar] [CrossRef]
  7. An, Y.; Xia, X.; Chen, X.; Wu, F.-X.; Wang, J. Chinese Clinical Named Entity Recognition via Multi-Head Self-Attention Based BiLSTM-CRF. Artif. Intell. Med. 2022, 127, 102282. [Google Scholar] [CrossRef]
  8. Chai, Z.; Jin, H.; Shi, S.; Zhan, S.; Zhuo, L.; Yang, Y. Hierarchical Shared Transfer Learning for Biomedical Named Entity Recognition. BMC Bioinform. 2022, 23, 8. [Google Scholar] [CrossRef]
  9. Chang, L.; Zhang, R.; Lv, J.; Zhou, W.; Bai, Y. A Review of Biomedical Named Entity Recognition. J. Comput. Methods Sci. Eng. 2022, 22, 893–900. [Google Scholar] [CrossRef]
  10. Dai, S.; Ding, Y.; Zhang, Z.; Zuo, W.; Huang, X.; Zhu, S. GrantExtractor: Accurate Grant Support Information Extraction from Biomedical Fulltext Based on Bi-LSTM-CRF. IEEE/ACM Trans. Comput. Biol. Bioinform. 2021, 18, 205–215. [Google Scholar] [CrossRef]
  11. Zhao, A.; Li, Z.; Zhao, G.; Cai, Y.; Xu, Y.; Zhu, X.; Cao, N.; Yang, J. A Pilot Study of WCA (a Chinese Jianpi Herbal Formula) Integrated to Systemic Chemotherapy (CT) in the First-Line Treatment for Advanced Gastric Cancer (AGC). J. Clin. Oncol. 2012, 30, 143. [Google Scholar] [CrossRef]
  12. Kajiura, S.; Hosokawa, A.; Yoshita, H.; Nakada, N.; Itaya, Y.; Ueda, A.; Ando, T.; Fujinami, H.; Ogawa, K.; Sugiyama, T. Preventive Effect of Rikkunshito, Traditional Japanese Medicine, for Chemotherapy-Induced Nausea and Vomiting. J. Clin. Oncol. 2012, 30, 135. [Google Scholar] [CrossRef]
  13. Festen, S.; Van der Wal-Huisman, H.; van der Leest, A.H.; Reyners, A.K.; de Bock, G.H.; de Graeff, P.; van Leeuwen, B.L. The Effect of Treatment Modifications by an Onco-Geriatric MDT on One-Year Mortality, Days Spent at Home and Postoperative Complications. J. Geriatr. Oncol. 2021, 12, 779–785. [Google Scholar] [CrossRef]
  14. Xu, Z.; Zhang, J.; Zhang, Q.; Xuan, Q.; Yip, P.S.F. A Comorbidity Knowledge-Aware Model for Disease Prognostic Prediction. IEEE Trans. Cybern. 2022, 52, 9809–9819. [Google Scholar] [CrossRef] [PubMed]
  15. Quinn, S.M.; Fernandez, H.; McCorkle, T.; Rogers, R.; Hussain, S.; Ford, C.A.; Barg, F.K.; Ginsburg, K.R.; Amaral, S. The Role of Resilience in Healthcare Transitions among Adolescent Kidney Transplant Recipients. Pediatr. Transplant. 2019, 23, e13559. [Google Scholar] [CrossRef]
  16. Eligüzel, N.; Çetinkaya, C.; Dereli, T. Application of Named Entity Recognition on Tweets during Earthquake Disaster: A Deep Learning-Based Approach. Soft Comput. 2022, 26, 395–421. [Google Scholar] [CrossRef]
  17. Nadeau, D.; Sekine, S. A Survey of Named Entity Recognition and Classification. Lingvisticae Investig. 2007, 30, 3–26. [Google Scholar] [CrossRef]
  18. Elbers, D.C.; La, J.; Minot, J.R.; Gramling, R.; Brophy, M.T.; Do, N.V.; Fillmore, N.R.; Dodds, P.S.; Danforth, C.M. Sentiment Analysis of Medical Record Notes for Lung Cancer Patients at the Department of Veterans Affairs. PLoS ONE 2023, 18, e0280931. [Google Scholar] [CrossRef]
  19. Liu, Y.; Gu, J.; Goyal, N.; Li, X.; Edunov, S.; Ghazvininejad, M.; Lewis, M.; Zettlemoyer, L. Multilingual Denoising Pre-Training for Neural Machine Translation. Trans. Assoc. Comput. Linguist. 2020, 8, 726–742. [Google Scholar] [CrossRef]
  20. Guo, Y.; Ma, Z.; Mao, J.; Qian, H.; Zhang, X.; Jiang, H.; Cao, Z.; Dou, Z. Webformer: Pre-Training with Web Pages for Information Retrieval. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 1502–1512. [Google Scholar]
  21. Wang, C.; Gao, J.; Rao, H.; Chen, A.; He, J.; Jiao, J.; Zou, N.; Gu, L. Named Entity Recognition (NER) for Chinese Agricultural Diseases and Pests Based on Discourse Topic and Attention Mechanism. Evol. Intell. 2022, 1–10. [Google Scholar] [CrossRef]
  22. Grishman, R.; Sundheim, B. Message Understanding Conference 6: A Brief History. In Proceedings of the COLING 1996 Volume 1: The 16th International Conference on Computational Linguistics, Copenhagen, Denmark, 5–9 August 1996. [Google Scholar]
  23. Kim, J.; Kim, Y.; Kang, S. Weakly Labeled Data Augmentation for Social Media Named Entity Recognition. Expert Syst. Appl. 2022, 209, 118217. [Google Scholar] [CrossRef]
  24. Litkowski, K.C.; Hargraves, O. SemEval-2007 Task 06: Word-Sense Disambiguation of Prepositions. In Proceedings of the Proceedings of the Fourth International Workshop on Semantic Evaluations (SemEval-2007), Prague, Czech Republic, 23–24 June 2007; pp. 24–29. [Google Scholar]
  25. Wacholder, N.; Ravin, Y.; Choi, M. Disambiguation of Proper Names in Text. In Proceedings of the Fifth Conference on Applied Natural Language Processing, Washington, DC, USA, 31 March–3 April 1997; pp. 202–208. [Google Scholar]
  26. Mansouri, A.; Affendey, L.S.; Mamat, A. Named Entity Recognition Approaches. Int. J. Comput. Sci. Netw. Secur. 2008, 8, 339–344. [Google Scholar]
  27. Kripke, S. Naming and Necessity; Semantics of Natural Language; Davidson, D., Harman, G., Eds.; D. Reidel: Dordrecht, The Netherlands, 1972. [Google Scholar]
  28. Balakrishnan, V.; Shi, Z.; Law, C.L.; Lim, R.; Teh, L.L.; Fan, Y. A Deep Learning Approach in Predicting Products’ Sentiment Ratings: A Comparative Analysis. J. Supercomput. 2022, 78, 7206–7226. [Google Scholar] [CrossRef] [PubMed]
  29. Severin, W.J.; Tankard, J.W. Communication Theories: Origins, Methods, and Uses in the Mass Media; Longman: New York, NY, USA, 1997. [Google Scholar]
  30. Vosoughi, S.; Roy, D.; Aral, S. The Spread of True and False News Online. Science 2018, 359, 1146–1151. [Google Scholar] [CrossRef] [PubMed]
  31. Wang, X.; Wang, X.; Min, G.; Hao, F.; Chen, C.P. An Efficient Feedback Control Mechanism for Positive/Negative Information Spread in Online Social Networks. IEEE Trans. Cybern. 2020, 52, 87–100. [Google Scholar] [CrossRef] [PubMed]
  32. Dobber, T.; Metoui, N.; Trilling, D.; Helberger, N.; de Vreese, C. Do (Microtargeted) Deepfakes Have Real Effects on Political Attitudes? Int. J. Press/Politics 2021, 26, 69–91. [Google Scholar] [CrossRef]
  33. House of Commons. Disinformation and ‘Fake News’: Final Report; House of Commons: London, UK, 2019. [Google Scholar]
  34. Wardle, C.; Derakhshan, H. Information Disorder: Toward an Interdisciplinary Framework for Research and Policymaking; Council of Europe: Strasbourg, France, 2017. [Google Scholar]
  35. Jowett, G.S.; O’donnell, V. Propaganda & Persuasion; Sage Publications: Thousand Oaks, CA, USA, 2018. [Google Scholar]
  36. Alhindi, T.; Pfeiffer, J.; Muresan, S. Fine-Tuned Neural Models for Propaganda Detection at the Sentence and Fragment Levels. arXiv 2019, arXiv:1910.09702. [Google Scholar]
  37. Barrón-Cedeno, A.; Jaradat, I.; Da San Martino, G.; Nakov, P. Proppy: Organizing the News Based on Their Propagandistic Content. Inf. Process. Manag. 2019, 56, 1849–1864. [Google Scholar] [CrossRef]
  38. Huang, J.; Li, G.; Huang, Q.; Wu, X. Joint Feature Selection and Classification for Multilabel Learning. IEEE Trans. Cybern. 2017, 48, 876–889. [Google Scholar] [CrossRef]
  39. Lu, C.; Reddy, C.K.; Ning, Y. Self-Supervised Graph Learning With Hyperbolic Embedding for Temporal Health Event Prediction. IEEE Trans. Cybern. 2023, 53, 2124–2136. [Google Scholar] [CrossRef]
  40. Yang, B.; Ye, M.; Tan, Q.; Yuen, P.C. Cross-Domain Missingness-Aware Time-Series Adaptation With Similarity Distillation in Medical Applications. IEEE Trans. Cybern. 2022, 52, 3394–3407. [Google Scholar] [CrossRef] [PubMed]
  41. Yang, Z.; Wu, N.; Liang, Y.; Zhang, H.; Ren, Y. SMSPL: Robust Multimodal Approach to Integrative Analysis of Multiomics Data. IEEE Trans. Cybern. 2022, 52, 2082–2095. [Google Scholar] [CrossRef] [PubMed]
  42. Mulyadi, A.W.; Jun, E.; Suk, H.-I. Uncertainty-Aware Variational-Recurrent Imputation Network for Clinical Time Series. IEEE Trans. Cybern. 2022, 52, 9684–9694. [Google Scholar] [CrossRef] [PubMed]
  43. Chiu, J.P.; Nichols, E. Named Entity Recognition with Bidirectional LSTM-CNNs. Trans. Assoc. Comput. Linguist. 2016, 4, 357–370. [Google Scholar] [CrossRef]
  44. Li, J.; Sun, A.; Han, J.; Li, C. A Survey on Deep Learning for Named Entity Recognition. IEEE Trans. Knowl. Data Eng. 2020, 34, 50–70. [Google Scholar] [CrossRef]
  45. Yin, X.; Huang, J.X.; Li, Z.; Zhou, X. A Survival Modeling Approach to Biomedical Search Result Diversification Using Wikipedia. IEEE Trans. Knowl. Data Eng. 2013, 25, 1201–1212. [Google Scholar] [CrossRef]
  46. Yin, D.; Cheng, S.; Pan, B.; Qiao, Y.; Zhao, W.; Wang, D. Chinese Named Entity Recognition Based on Knowledge Based Question Answering System. Appl. Sci. 2022, 12, 5373. [Google Scholar] [CrossRef]
  47. Tanabe, L.; Xie, N.; Thom, L.H.; Matten, W.; Wilbur, W.J. GENETAG: A Tagged Corpus for Gene/Protein Named Entity Recognition. BMC Bioinform. 2005, 6, S3. [Google Scholar] [CrossRef]
  48. Kundeti, S.R.; Vijayananda, J.; Mujjiga, S.; Kalyan, M. Clinical Named Entity Recognition: Challenges and Opportunities. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1937–1945. [Google Scholar]
  49. Song, B.; Li, F.; Liu, Y.; Zeng, X. Deep Learning Methods for Biomedical Named Entity Recognition: A Survey and Qualitative Comparison. Brief. Bioinform. 2021, 22, bbab282. [Google Scholar] [CrossRef]
  50. Liu, P.; Guo, Y.; Wang, F.; Li, G. Chinese Named Entity Recognition: The State of the Art. Neurocomputing 2022, 473, 37–53. [Google Scholar] [CrossRef]
  51. Collobert, R.; Weston, J.; Bottou, L.; Karlen, M.; Kavukcuoglu, K.; Kuksa, P. Natural Language Processing (Almost) from Scratch. J. Mach. Learn. Res. 2011, 12, 2493–2537. [Google Scholar]
  52. Huang, Z.; Xu, W.; Yu, K. Bidirectional LSTM-CRF Models for Sequence Tagging. arXiv 2015, arXiv:1508.01991. [Google Scholar]
  53. Kim, N.H.; Kim, J.M.; Park, D.M.; Ji, S.R.; Kim, J.W. Analysis of Depression in Social Media Texts through the Patient Health Questionnaire-9 and Natural Language Processing. Digit. Health 2022, 8, 20552076221114204. [Google Scholar] [CrossRef] [PubMed]
  54. Moher, D.; Liberati, A.; Tetzlaff, J.; Altman, D.G.; PRISMA Group. The Preferred Reporting Items for Systematic Reviews and Meta-Analyses: The PRISMA Statement. Ann. Intern. Med. 2009, 151, 264–269. [Google Scholar] [CrossRef]
  55. Haddaway, N.R.; Page, M.J.; Pritchard, C.C.; McGuinness, L.A. PRISMA2020: An R Package and Shiny App for Producing PRISMA 2020-Compliant Flow Diagrams, with Interactivity for Optimised Digital Transparency and Open Synthesis. Campbell Syst. Rev. 2022, 18, e1230. [Google Scholar] [CrossRef]
  56. Pustejovsky, J.; Ide, N.; Verhagen, M.; Suderman, K. Enhancing Access to Media Collections and Archives Using Computational Linguistic Tools. In CDH@ TLT; CEUR-WS: Bloomington, UK, 2017; pp. 19–28. [Google Scholar]
  57. Sundheim, B.M. Overview of Results of the MUC-6 Evaluation; Defence Technical Information Center: Fort Belvoir, VA, USA, 1995. [Google Scholar]
  58. Raza Abidi, S.S.; Goh, A. Applying Knowledge Discovery to Predict Infectious Disease Epidemics. In Proceedings of the PRICAI’98: Topics in Artificial Intelligence: 5th Pacific Rim International Conference on Artificial Intelligence, Singapore, 22–27 November 1998; Springer: Berlin/Heidelberg, Germany, 1998; pp. 170–181. [Google Scholar]
  59. Cios, K.J.; Moore, G.W. Uniqueness of Medical Data Mining. Artif. Intell. Med. 2002, 26, 1–24. [Google Scholar] [CrossRef]
  60. Seshia, S.A.; Sadigh, D.; Sastry, S.S. Toward Verified Artificial Intelligence. Commun. ACM 2022, 65, 46–55. [Google Scholar] [CrossRef]
  61. Krichen, M.; Mihoub, A.; Alzahrani, M.Y.; Adoni, W.Y.H.; Nahhal, T. Are Formal Methods Applicable To Machine Learning And Artificial Intelligence? In Proceedings of the 2022 2nd International Conference of Smart Systems and Emerging Technologies (SMARTTECH), Riyadh, Saudi Arabia, 9–11 May 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 48–53. [Google Scholar]
  62. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  63. Batmaz, Z.; Yurekli, A.; Bilge, A.; Kaleli, C. A Review on Deep Learning for Recommender Systems: Challenges and Remedies. Artif. Intell. Rev. 2019, 52, 1–37. [Google Scholar] [CrossRef]
  64. Khan, A.H.; Siddqui, J.; Sohail, S.S. A Survey of Recommender Systems Based on Semi-Supervised Learning. In Proceedings of the International Conference on Innovative Computing and Communications; Springer: Singapore, 2022; pp. 319–327. [Google Scholar]
  65. Wang, R.; Hou, F.; Cahan, S.; Chen, L.; Jia, X.; Ji, W. Fine-Grained Entity Typing with a Type Taxonomy: A Systematic Review. IEEE Trans. Knowl. Data Eng. 2023, 35, 4794–4812. [Google Scholar] [CrossRef]
  66. Shen, Y.; Yun, H.; Lipton, Z.C.; Kronrod, Y.; Anandkumar, A. Deep Active Learning for Named Entity Recognition. arXiv 2017, arXiv:1707.05928. [Google Scholar]
  67. Bridges, J.; Hughes, J.; Farrington, N.; Richardson, A. Cancer Treatment Decision-Making Processes for Older Patients with Complex Needs: A Qualitative Study. BMJ Open 2015, 5, e009674. [Google Scholar] [CrossRef] [PubMed]
  68. Cho, M.; Ha, J.; Park, C.; Park, S. Combinatorial Feature Embedding Based on CNN and LSTM for Biomedical Named Entity Recognition. J. Biomed. Inform. 2020, 103, 103381. [Google Scholar] [CrossRef] [PubMed]
  69. Zhang, Y.; Chen, H.; Zhao, Y.; Liu, Q.; Yin, D. Learning Tag Dependencies for Sequence Tagging. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; pp. 4581–4587. [Google Scholar]
  70. Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. Roberta: A Robustly Optimized Bert Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
  71. Islam, M.R.; Kamal, A.R.M.; Sultana, N.; Islam, R.; Moni, M.A. Detecting Depression Using K-Nearest Neighbors (Knn) Classification Technique. In Proceedings of the 2018 International Conference on Computer, Communication, Chemical, Material and Electronic Engineering (IC4ME2), Rajshahi, Bangladesh, 8–9 February 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–4. [Google Scholar]
  72. Hinton, G.; Deng, L.; Yu, D.; Dahl, G.E.; Mohamed, A.; Jaitly, N.; Senior, A.; Vanhoucke, V.; Nguyen, P.; Sainath, T.N. Deep Neural Networks for Acoustic Modeling in Speech Recognition: The Shared Views of Four Research Groups. IEEE Signal Process. Mag. 2012, 29, 82–97. [Google Scholar] [CrossRef]
  73. Wang, N.; Yeung, D.-Y. Learning a Deep Compact Image Representation for Visual Tracking. Adv. Neural Inf. Process. Syst. 2013, 26, 809–817. [Google Scholar]
  74. Vo, N.N.; He, X.; Liu, S.; Xu, G. Deep Learning for Decision Making and the Optimization of Socially Responsible Investments and Portfolio. Decis. Support Syst. 2019, 124, 113097. [Google Scholar] [CrossRef]
  75. Young, T.; Hazarika, D.; Poria, S.; Cambria, E. Recent Trends in Deep Learning Based Natural Language Processing. IEEE Comput. Intell. Mag. 2018, 13, 55–75. [Google Scholar] [CrossRef]
  76. Reddy, M.; Nardelli, J.C.; Pereira, Y.L.; Oliveira, L.B.; Silva, T.H.; Vasconcelos, M.; Horowitz, M. Higher Education’s Influence on Social Networks and Entrepreneurship in Brazil. Soc. Netw. Anal. Min. 2023, 13, 1–21. [Google Scholar] [CrossRef]
  77. Khanday, A.M.U.D.; Khan, Q.R.; Rabani, S.T. Identifying Propaganda from Online Social Networks during COVID-19 Using Machine Learning Techniques. Int. J. Inf. Technol. 2021, 13, 115–122. [Google Scholar] [CrossRef]
  78. Nitish, S.; Darsini, R.; Shashank, G.S.; Tejas, V.; Arya, A. Bidirectional Encoder Representation from Transformers (BERT) Variants for Procedural Long-Form Answer Extraction. In Proceedings of the 2022 12th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India, 27–28 January 2022; pp. 71–76. [Google Scholar] [CrossRef]
  79. Habibi, M.; Weber, L.; Neves, M.; Wiegandt, D.L.; Leser, U. Deep Learning with Word Embeddings Improves Biomedical Named Entity Recognition. Bioinformatics 2017, 33, i37–i48. [Google Scholar] [CrossRef] [PubMed]
  80. Weber, L.; Münchmeyer, J.; Rocktäschel, T.; Habibi, M.; Leser, U. HUNER: Improving Biomedical NER with Pretraining. Bioinformatics 2020, 36, 295–302. [Google Scholar] [CrossRef] [PubMed]
  81. Wei, C.-H.; Kao, H.-Y.; Lu, Z. PubTator: A Web-Based Text Mining Tool for Assisting Biocuration. Nucleic Acids Res. 2013, 41, W518–W522. [Google Scholar] [CrossRef]
  82. Thomas, P.; Starlinger, J.; Vowinkel, A.; Arzt, S.; Leser, U. GeneView: A Comprehensive Semantic Search Engine for PubMed. Nucleic Acids Res. 2012, 40, W585–W591. [Google Scholar] [CrossRef]
  83. Kim, J.-D.; Ohta, T.; Tateisi, Y.; Tsujii, J. GENIA Corpus—A Semantically Annotated Corpus for Bio-Textmining. Bioinformatics 2003, 19, i180–i182. [Google Scholar] [CrossRef] [PubMed]
  84. Song, Y.; Kim, E.; Lee, G.G.; Yi, B.-K. POSBIOTM—NER: A Trainable Biomedical Named-Entity Recognition System. Bioinformatics 2005, 21, 2794–2796. [Google Scholar] [CrossRef] [PubMed]
  85. Bojanowski, P.; Grave, E.; Joulin, A.; Mikolov, T. Enriching Word Vectors with Subword Information. Trans. Assoc. Comput. Linguist. 2017, 5, 135–146. [Google Scholar] [CrossRef]
  86. Adewumi, T.P.; Liwicki, F.; Liwicki, M. Exploring Swedish & English FastText Embeddings for NER with the Transformer. arXiv 2020, arXiv:2007.16007. [Google Scholar]
  87. Sun, C.; Yang, Z.; Luo, L.; Wang, L.; Zhang, Y.; Lin, H.; Wang, J. A Deep Learning Approach with Deep Contextualized Word Representations for Chemical–Protein Interaction Extraction from Biomedical Literature. IEEE Access 2019, 7, 151034–151046. [Google Scholar] [CrossRef]
  88. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-Training of Deep Bidirectional Transformers for Language Understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  89. Zhou, Y. Pro-ISIS Fanboys Network Analysis and Attack Detection through Twitter Data. In Proceedings of the 2017 IEEE 2nd International Conference on Big Data Analysis (ICBDA), Beijing, China, 10–12 March 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 386–390. [Google Scholar]
  90. Sarwar, A.; Ali, M.; Manhas, J.; Sharma, V. Diagnosis of Diabetes Type-II Using Hybrid Machine Learning Based Ensemble Model. Int. J. Inf. Technol. 2020, 12, 419–428. [Google Scholar] [CrossRef]
  91. Gao, J.; Zheng, P.; Jia, Y.; Chen, H.; Mao, Y.; Chen, S.; Wang, Y.; Fu, H.; Dai, J. Mental Health Problems and Social Media Exposure during COVID-19 Outbreak. PLoS ONE 2020, 15, e0231924. [Google Scholar]
  92. Al-Zaman, M.S. COVID-19-Related Social Media Fake News in India. J. Media 2021, 2, 100–114. [Google Scholar] [CrossRef]
  93. Gisondi, M.A.; Barber, R.; Faust, J.S.; Raja, A.; Strehlow, M.C.; Westafer, L.M.; Gottlieb, M. A Deadly Infodemic: Social Media and the Power of COVID-19 Misinformation. J. Med. Internet Res. 2022, 24, e35552. [Google Scholar] [CrossRef]
  94. Mane, H.; Yue, X.; Yu, W.; Doig, A.C.; Wei, H.; Delcid, N.; Harris, A.-G.; Nguyen, T.T.; Nguyen, Q.C. Examination of the Public’s Reaction on Twitter to the Over-Turning of Roe v Wade and Abortion Bans. Healthcare 2022, 10, 2390. [Google Scholar] [CrossRef]
  95. Ong, S.-Q.; Pauzi, M.B.M.; Gan, K.H. Text Mining and Determinants of Sentiments towards the COVID-19 Vaccine Booster of Twitter Users in Malaysia. Healthcare 2022, 10, 994. [Google Scholar] [CrossRef]
  96. Reshi, A.A.; Rustam, F.; Aljedaani, W.; Shafi, S.; Alhossan, A.; Alrabiah, Z.; Ahmad, A.; Alsuwailem, H.; Almangour, T.A.; Alshammari, M.A.; et al. COVID-19 Vaccination-Related Sentiments Analysis: A Case Study Using Worldwide Twitter Dataset. Healthcare 2022, 10, 411. [Google Scholar] [CrossRef]
  97. Rahim, A.I.A.; Ibrahim, M.I.; Chua, S.-L.; Musa, K.I. Hospital Facebook Reviews Analysis Using a Machine Learning Sentiment Analyzer and Quality Classifier. Healthcare 2021, 9, 1679. [Google Scholar] [CrossRef]
  98. Lee, K.; Mileski, M.; Fohn, J.; Frye, L.; Brooks, L. Facilitators and Barriers Surrounding the Role of Administration in Employee Job Satisfaction in Long-Term Care Facilities: A Systematic Review. Healthcare 2020, 8, 360. [Google Scholar] [CrossRef]
  99. Shah, A.M.; Yan, X.; Qayyum, A.; Naqvi, R.A.; Shah, S.J. Mining Topic and Sentiment Dynamics in Physician Rating Websites during the Early Wave of the COVID-19 Pandemic: Machine Learning Approach. Int. J. Med. Inform. 2021, 149, 104434. [Google Scholar] [CrossRef]
  100. Hu, Y.; Wen, G.; Liao, H.; Wang, C.; Dai, D.; Yu, Z. Automatic Construction of Chinese Herbal Prescriptions From Tongue Images Using CNNs and Auxiliary Latent Therapy Topics. IEEE Trans. Cybern. 2021, 51, 708–721. [Google Scholar] [CrossRef] [PubMed]
  101. Rodriguez, N.E.; Nguyen, M.; McInnes, B.T. Effects of Data and Entity Ablation on Multitask Learning Models for Biomedical Entity Recognition. J. Biomed. Inform. 2022, 130, 104062. [Google Scholar] [CrossRef] [PubMed]
  102. Caruana, R. Multitask Learning. Mach. Learn. 1997, 28, 41–75. [Google Scholar] [CrossRef]
  103. Lin, Y.; Yang, S.; Stoyanov, V.; Ji, H. A Multi-Lingual Multi-Task Architecture for Low-Resource Sequence Labeling. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, Melbourne, Australia, 15–20 July 2018; Volume 1: Long Papers. pp. 799–809. [Google Scholar]
  104. Wang, P.; Xu, B.; Xu, J.; Tian, G.; Liu, C.-L.; Hao, H. Semantic Expansion Using Word Embedding Clustering and Convolutional Neural Network for Improving Short Text Classification. Neurocomputing 2016, 174, 806–814. [Google Scholar] [CrossRef]
  105. Zhang, Y.; Lu, W.; Ou, W.; Zhang, G.; Zhang, X.; Cheng, J.; Zhang, W. Chinese Medical Question Answer Selection via Hybrid Models Based on CNN and GRU. Multimed. Tools Appl. 2020, 79, 14751–14776. [Google Scholar] [CrossRef]
  106. Lv, J.; Zhang, Z.; Jin, L.; Li, S.; Li, X.; Xu, G.; Sun, X. Trigger Is Non-Central: Jointly Event Extraction via Label-Aware Representations with Multi-Task Learning. Knowl.-Based Syst. 2022, 252, 109480. [Google Scholar] [CrossRef]
  107. Chai, Z.; Jin, H.; Shi, S.; Zhan, S.; Zhuo, L.; Yang, Y.; Lian, Q. Noise Reduction Learning Based on XLNet-CRF for Biomedical Named Entity Recognition. IEEE/ACM Trans. Comput. Biol. Bioinform. 2022, 20, 595–605. [Google Scholar] [CrossRef] [PubMed]
  108. Hanisch, D.; Fundel, K.; Mevissen, H.-T.; Zimmer, R.; Fluck, J. ProMiner: Rule-Based Protein and Gene Entity Recognition. BMC Bioinform. 2005, 6, S14. [Google Scholar] [CrossRef]
  109. Crichton, G.; Pyysalo, S.; Chiu, B.; Korhonen, A. A Neural Network Multi-Task Learning Approach to Biomedical Named Entity Recognition. BMC Bioinform. 2017, 18, 368. [Google Scholar] [CrossRef]
  110. Leemans, S.J.; van Zelst, S.J.; Lu, X. Partial-Order-Based Process Mining: A Survey and Outlook. Knowl. Inf. Syst. 2023, 65, 1–29. [Google Scholar] [CrossRef]
  111. Fahrenkrog-Petersen, S.A.; Kabierski, M.; van der Aa, H.; Weidlich, M. Semantics-Aware Mechanisms for Control-Flow Anonymization in Process Mining. Inf. Syst. 2023, 114, 102169. [Google Scholar] [CrossRef]
  112. Elhaj, H.; Achour, N.; Tania, M.H.; Aciksari, K. A Comparative Study of Supervised Machine Learning Approaches to Predict Patient Triage Outcomes in Hospital Emergency Departments. Array 2023, 17, 100281. [Google Scholar] [CrossRef]
  113. Pandya, S.; Gadekallu, T.R.; Reddy, P.K.; Wang, W.; Alazab, M. InfusedHeart: A Novel Knowledge-Infused Learning Framework for Diagnosis of Cardiovascular Events. IEEE Trans. Comput. Soc. Syst. 2022. [Google Scholar] [CrossRef]
  114. Ozmen Garibay, O.; Winslow, B.; Andolina, S.; Antona, M.; Bodenschatz, A.; Coursaris, C.; Falco, G.; Fiore, S.M.; Garibay, I.; Grieman, K. Six Human-Centered Artificial Intelligence Grand Challenges. Int. J. Hum.-Comput. Interact. 2023, 39, 391–437. [Google Scholar] [CrossRef]
  115. Albawi, S.; Arif, M.H.; Waleed, J. Skin Cancer Classification Dermatologist-Level Based on Deep Learning Model. Acta Scientiarum. Technol. 2023, 45, e61531. [Google Scholar] [CrossRef]
  116. Chia, M.A.; Hersch, F.; Sayres, R.; Bavishi, P.; Tiwari, R.; Keane, P.A.; Turner, A.W. Validation of a Deep Learning System for the Detection of Diabetic Retinopathy in Indigenous Australians. Br. J. Ophthalmol. 2023, 63, 2092-F0081. [Google Scholar] [CrossRef] [PubMed]
  117. Peng, H.; Wu, C.; Nguyen, D.; Schuemann, J.; Mairani, A.; Pu, Y.; Jiang, S. Recent Advancements of Artificial Intelligence in Particle Therapy. IEEE Trans. Radiat. Plasma Med. Sci. 2023, 7, 213–224. [Google Scholar] [CrossRef]
  118. Liu, S.; Wang, X.; Xiang, Y.; Xu, H.; Wang, H.; Tang, B. Multi-Channel Fusion LSTM for Medical Event Prediction Using EHRs. J. Biomed. Inform. 2022, 127, 104011. [Google Scholar] [CrossRef]
  119. Biswas, S.; Dash, S. LSTM-CNN Deep Learning–Based Hybrid System for Real-Time COVID-19 Data Analysis and Prediction Using Twitter Data. In Assessing COVID-19 and Other Pandemics and Epidemics Using Computational Modelling and Data Analysis; Springer: Cham, Switzerland, 2022; pp. 239–257. [Google Scholar]
  120. Amethiya, Y.; Pipariya, P.; Patel, S.; Shah, M. Comparative Analysis of Breast Cancer Detection Using Machine Learning and Biosensors. Intell. Med. 2022, 2, 69–81. [Google Scholar] [CrossRef]
  121. Dubin, J.R.; Simon, S.D.; Norrell, K.; Perera, J.; Gowen, J.; Cil, A. Risk of Recall among Medical Devices Undergoing US Food and Drug Administration 510 (k) Clearance and Premarket Approval, 2008–2017. JAMA Netw. Open 2021, 4, e217274. [Google Scholar] [CrossRef]
  122. Sayres, R.; Taly, A.; Rahimy, E.; Blumer, K.; Coz, D.; Hammel, N.; Krause, J.; Narayanaswamy, A.; Rastegar, Z.; Wu, D. Using a Deep Learning Algorithm and Integrated Gradients Explanation to Assist Grading for Diabetic Retinopathy. Ophthalmology 2019, 126, 552–564. [Google Scholar] [CrossRef] [PubMed]
  123. Wong, K.C.; Moradi, M.; Wu, J.; Syeda-Mahmood, T. Identifying Disease-Free Chest x-Ray Images with Deep Transfer Learning. In Medical Imaging 2019: Computer-Aided Diagnosis; SPIE: Bellingham, WA, USA, 2019; Volume 10950, pp. 179–184. [Google Scholar]
  124. Wu, T.; Sultan, L.R.; Tian, J.; Cary, T.W.; Sehgal, C.M. Machine Learning for Diagnostic Ultrasound of Triple-Negative Breast Cancer. Breast Cancer Res. Treat. 2019, 173, 365–373. [Google Scholar] [CrossRef] [PubMed]
  125. Keren, L.; Bosse, M.; Marquez, D.; Angoshtari, R.; Jain, S.; Varma, S.; Yang, S.-R.; Kurian, A.; Van Valen, D.; West, R. A Structured Tumor-Immune Microenvironment in Triple Negative Breast Cancer Revealed by Multiplexed Ion Beam Imaging. Cell 2018, 174, 1373–1387. [Google Scholar] [CrossRef] [PubMed]
  126. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-Level Classification of Skin Cancer with Deep Neural Networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  127. Subbiah, V. The next Generation of Evidence-Based Medicine. Nat. Med. 2023, 29, 49–58. [Google Scholar] [CrossRef]
  128. Zhuo, J.; Cao, Y.; Zhu, J.; Zhang, B.; Nie, Z. Segment-Level Sequence Modeling Using Gated Recursive Semi-Markov Conditional Random Fields. In Proceedings of the Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016; Volume 1: Long Papers. pp. 1413–1423. [Google Scholar]
  129. Rei, M.; Crichton, G.K.; Pyysalo, S. Attending to Characters in Neural Sequence Labeling Models. arXiv 2016, arXiv:1611.04361. [Google Scholar]
  130. Naim, J.; Hossain, T.; Tasneem, F.; Chy, A.N.; Aono, M. Leveraging Fusion of Sequence Tagging Models for Toxic Spans Detection. Neurocomputing 2022, 500, 688–702. [Google Scholar] [CrossRef]
  131. Luo, Y.; Xiao, F.; Zhao, H. Hierarchical Contextualized Representation for Named Entity Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 8441–8448. [Google Scholar]
  132. Aguilar, G.; Maharjan, S.; López-Monroy, A.P.; Solorio, T. A Multi-Task Approach for Named Entity Recognition in Social Media Data. arXiv 2019, arXiv:1906.04135. [Google Scholar]
  133. Graves, A.; Schmidhuber, J. Framewise Phoneme Classification with Bidirectional LSTM and Other Neural Network Architectures. Neural Netw. 2005, 18, 602–610. [Google Scholar] [CrossRef]
  134. Tjandra, A.; Sakti, S.; Manurung, R.; Adriani, M.; Nakamura, S. Gated Recurrent Neural Tensor Network. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 448–455. [Google Scholar]
  135. Yuan, M.; Ren, J. Numerical Feature Transformation-Based Sequence Generation Model for Multi-Disease Diagnosis. Int. J. Pattern Recognit. Artif. Intell. 2021, 35, 2159034. [Google Scholar] [CrossRef]
  136. Liu, Y.; Gou, X. A Text Classification Method Based on Graph Attention Networks. In Proceedings of the 2021 International Conference on Information Technology and Biomedical Engineering (ICITBE), Nanchang, China, 24–26 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 35–39. [Google Scholar]
  137. Schuster, M.; Paliwal, K.K. Bidirectional Recurrent Neural Networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef]
  138. Zhao, Q.; Xu, D.; Li, J.; Zhao, L.; Rajput, F.A. Knowledge Guided Distance Supervision for Biomedical Relation Extraction in Chinese Electronic Medical Records. Expert Syst. Appl. 2022, 204, 117606. [Google Scholar] [CrossRef]
  139. Chen, Y.; Yao, Z.; Chi, H.; Gabbay, D.; Yuan, B.; Bentzen, B.; Liao, B. BTPK-Based Learning: An Interpretable Method for Named Entity Recognition. arXiv 2022, arXiv:2201.09523. [Google Scholar]
  140. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning Phrase Representations Using RNN Encoder-Decoder for Statistical Machine Translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
  141. Liu, X.; Cheng, M.; Zhang, H.; Hsieh, C.-J. Towards Robust Neural Networks via Random Self-Ensemble. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 369–385. [Google Scholar]
  142. Lin, J.C.-W.; Shao, Y.; Djenouri, Y.; Yun, U. ASRNN: A Recurrent Neural Network with an Attention Model for Sequence Labeling. Knowl.-Based Syst. 2021, 212, 106548. [Google Scholar] [CrossRef]
  143. Luong, M.-T.; Manning, C.D. Achieving Open Vocabulary Neural Machine Translation with Hybrid Word-Character Models. arXiv 2016, arXiv:1604.00788. [Google Scholar]
  144. Xiangyu, X.; Tong, Z.; Wei, Y.; Jinglei, Z.; Rui, X.; Shikun, Z. A Hybrid Character Representation for Chinese Event Detection. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–8. [Google Scholar]
  145. Ho, K.K.; Chan, J.Y.; Chiu, D.K. Fake News and Misinformation During the Pandemic: What We Know and What We Do Not Know. IT Prof. 2022, 24, 19–24. [Google Scholar] [CrossRef]
  146. Zheng, S.; Wang, F.; Bao, H.; Hao, Y.; Zhou, P.; Xu, B. Joint Extraction of Entities and Relations Based on a Novel Tagging Scheme. arXiv 2017, arXiv:1706.05075. [Google Scholar]
  147. Chen, K.; Song, X.; Yuan, H.; Ren, X. Fully Convolutional Encoder-Decoder With an Attention Mechanism for Practical Pedestrian Trajectory Prediction. IEEE Trans. Intell. Transp. Syst. 2022, 23, 20046–20060. [Google Scholar] [CrossRef]
  148. Horiguchi, S.; Fujita, Y.; Watanabe, S.; Xue, Y.; García, P. Encoder-Decoder Based Attractors for End-to-End Neural Diarization. IEEE/ACM Trans. Audio Speech Lang. Process. 2022, 30, 1493–1507. [Google Scholar] [CrossRef]
  149. Dos Santos, C.; Zadrozny, B. Learning Character-Level Representations for Part-of-Speech Tagging. In Proceedings of the International Conference on Machine Learning, PMLR, Beijing, China, 21–26 June 2014; pp. 1818–1826. [Google Scholar]
  150. Nguyen, T.-S.; Nguyen, L.-M. Nested Named Entity Recognition Using Multilayer Recurrent Neural Networks. In Proceedings of the International Conference of the Pacific Association for Computational Linguistics, Yangon, Myanmar, 16–18 August 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 233–246. [Google Scholar]
  151. Katiyar, A.; Cardie, C. Nested Named Entity Recognition Revisited. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, New Orleans, LA, USA, 1–6 June 2018; Volume 1. [Google Scholar]
  152. Jagfeld, G.; Jenne, S.; Vu, N.T. Sequence-to-Sequence Models for Data-to-Text Natural Language Generation: Word-vs. Character-Based Processing and Output Diversity. arXiv 2018, arXiv:1810.04864. [Google Scholar]
  153. Gao, T.; Fisch, A.; Chen, D. Making Pre-Trained Language Models Better Few-Shot Learners. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Online, 1–6 August 2021; Volume 1: Long Papers. Association for Computational Linguistics: Cedarville, OH, USA, 2021; pp. 3816–3830. [Google Scholar]
  154. Qiu, X.; Sun, T.; Xu, Y.; Shao, Y.; Dai, N.; Huang, X. Pre-Trained Models for Natural Language Processing: A Survey. Sci. China Technol. Sci. 2020, 63, 1872–1897. [Google Scholar] [CrossRef]
  155. Tsipouras, M.G.; Exarchos, T.P.; Fotiadis, D.I.; Kotsia, A.P.; Vakalis, K.V.; Naka, K.K.; Michalis, L.K. Automated Diagnosis of Coronary Artery Disease Based on Data Mining and Fuzzy Modeling. IEEE Trans. Inf. Technol. Biomed. 2008, 12, 447–458. [Google Scholar] [CrossRef]
  156. Rakha, E.A.; Reis-Filho, J.S.; Ellis, I.O. Basal-like Breast Cancer: A Critical Review. J. Clin. Oncol. 2008, 26, 2568–2581. [Google Scholar] [CrossRef] [PubMed]
  157. Tapak, L.; Mahjub, H.; Hamidi, O.; Poorolajal, J. Real-Data Comparison of Data Mining Methods in Prediction of Diabetes in Iran. Healthc. Inform. Res. 2013, 19, 177–185. [Google Scholar] [CrossRef] [PubMed]
  158. Iqbal, U.; Hsu, C.-K.; Nguyen, P.A.A.; Clinciu, D.L.; Lu, R.; Syed-Abdul, S.; Yang, H.-C.; Wang, Y.-C.; Huang, C.-Y.; Huang, C.-W. Cancer-Disease Associations: A Visualization and Animation through Medical Big Data. Comput. Methods Programs Biomed. 2016, 127, 44–51. [Google Scholar] [CrossRef]
  159. Agrawal, A.; Misra, S.; Narayanan, R.; Polepeddi, L.; Choudhary, A. Lung Cancer Survival Prediction Using Ensemble Data Mining on SEER Data. Sci. Program. 2012, 20, 29–42. [Google Scholar] [CrossRef]
  160. Yeh, J.-Y.; Wu, T.-H.; Tsao, C.-W. Using Data Mining Techniques to Predict Hospitalization of Hemodialysis Patients. Decis. Support Syst. 2011, 50, 439–448. [Google Scholar] [CrossRef]
  161. Arvisais-Anhalt, S.; Lehmann, C.U.; Bishop, J.A.; Balani, J.; Boutte, L.; Morales, M.; Park, J.Y.; Araj, E. Searching Full-Text Anatomic Pathology Reports Using Business Intelligence Software. J. Pathol. Inform. 2022, 13, 100014. [Google Scholar] [CrossRef]
  162. Pur, A.; Bohanec, M.; Lavrač, N.; Cestnik, B. Primary Health-Care Network Monitoring: A Hierarchical Resource Allocation Modeling Approach. Int. J. Health Plan. Manag. 2010, 25, 119–135. [Google Scholar] [CrossRef]
  163. Koskela, T.-H.; Ryynanen, O.-P.; Soini, E.J. Risk Factors for Persistent Frequent Use of the Primary Health Care Services among Frequent Attenders: A Bayesian Approach. Scand. J. Prim. Health Care 2010, 28, 55–61. [Google Scholar] [CrossRef] [PubMed]
  164. Oliwa, T.; Maron, S.B.; Chase, L.M.; Lomnicki, S.; Catenacci, D.V.; Furner, B.; Volchenboum, S.L. Obtaining Knowledge in Pathology Reports through a Natural Language Processing Approach with Classification, Named-Entity Recognition, and Relation-Extraction Heuristics. JCO Clin. Cancer Inform. 2019, 3, 1–8. [Google Scholar] [CrossRef] [PubMed]
  165. Nimmagadda, S.L.; Dreher, H.V. On Robust Methodologies for Managing Public Health Care Systems. Int. J. Environ. Res. Public Health 2014, 11, 1106–1140. [Google Scholar] [CrossRef] [PubMed]
  166. Kostkova, P.; Fowler, D.; Wiseman, S.; Weinberg, J.R. Major Infection Events over 5 Years: How Is Media Coverage Influencing Online Information Needs of Health Care Professionals and the Public? J. Med. Internet Res. 2013, 15, e107. [Google Scholar] [CrossRef]
  167. Rathore, M.M.; Ahmad, A.; Paul, A.; Wan, J.; Zhang, D. Real-Time Medical Emergency Response System: Exploiting IoT and Big Data for Public Health. J. Med. Syst. 2016, 40, 283. [Google Scholar] [CrossRef]
  168. Lavrač, N.; Bohanec, M.; Pur, A.; Cestnik, B.; Debeljak, M.; Kobler, A. Data Mining and Visualization for Decision Support and Modeling of Public Health-Care Resources. J. Biomed. Inform. 2007, 40, 438–447. [Google Scholar] [CrossRef]
  169. Zhang, T.; Wang, Y.; Wang, X.; Yang, Y.; Ye, Y. Constructing Fine-Grained Entity Recognition Corpora Based on Clinical Records of Traditional Chinese Medicine. BMC Med. Inform. Decis. Mak. 2020, 20, 64. [Google Scholar] [CrossRef]
  170. Jones, R.A.; Hollen, P.J.; Wenzel, J.; Weiss, G.; Song, D.; Sims, T.; Petroni, G. Understanding Advanced Prostate Cancer Decision-Making Utilizing an Interactive Decision Aid. Cancer Nurs. 2018, 41, 2. [Google Scholar] [CrossRef]
  171. Kishwar, A.; Zafar, A. Fake News Detection on Pakistani News Using Machine Learning and Deep Learning. Expert Syst. Appl. 2023, 211, 118558. [Google Scholar] [CrossRef]
  172. Shah, A.; Yan, X.B.; Shah, S.A. Tracking Patients Healthcare Experiences during the COVID-19 Outbreak: Topic Modeling and Sentiment Analysis of Doctor Reviews. J. Eng. Res. 2021, 9, 219–239. [Google Scholar] [CrossRef]
  173. Shah, A.M.; Naqvi, R.A.; Jeong, O.-R. Detecting Topic and Sentiment Trends in Physician Rating Websites: Analysis of Online Reviews Using 3-Wave Datasets. Int. J. Environ. Res. Public Health 2021, 18, 4743. [Google Scholar] [CrossRef] [PubMed]
Figure 1. PRISMA flow diagram illustrating the steps involved in conducting a systematic literature review.
Figure 1. PRISMA flow diagram illustrating the steps involved in conducting a systematic literature review.
Healthcare 11 01268 g001
Figure 2. A general pipeline for the treatment of cancer patients along their journey through the healthcare system.
Figure 2. A general pipeline for the treatment of cancer patients along their journey through the healthcare system.
Healthcare 11 01268 g002
Figure 3. A visual representation of the decrease in positive sentiment between Day 19 and Day 21 of treatment.
Figure 3. A visual representation of the decrease in positive sentiment between Day 19 and Day 21 of treatment.
Healthcare 11 01268 g003
Figure 4. Pre-trained language models (PLMs) with typical examples.
Figure 4. Pre-trained language models (PLMs) with typical examples.
Healthcare 11 01268 g004
Table 1. Uniform Resource Locators (URL) for all monolingual NER datasets (English).
Table 1. Uniform Resource Locators (URL) for all monolingual NER datasets (English).
CorpusDateText-SourceTagData URL 1
MUC-2, MUC-3, MUC-4,1995MUC Data Sets https://www-nlpir.nist.gov/related_projects/ (accessed on 5 January 2023)
MUC-61995Wall Street Journal7https://catalog.ldc.upenn.edu/LDC2003T13 (accessed on 5 January 2023)
MUC-6 Plus1995Additional news7https://catalog.ldc.upenn.edu/LDC96T10 (accessed on 5 January 2023)
MUC-7 1997New York Times news7https://catalog.ldc.upenn.edu/LDC2001T02 (accessed on 5 January 2023)
CoNLL032003Reuters news4https://www.clips.uantwerpen.be/conll2003/ner/ (accessed on 5 January 2023)
ACE 20032003Wikipedia7https://www-nlpir.nist.gov/ (accessed on 5 January 2023)
ACE 20052003Wikipedia7https://catalog.ldc.upenn.edu/LDC2004T09 (accessed on 5 January 2023)
OntoNotes2005Wikipedia18https://catalog.ldc.upenn.edu/LDC2013T19 (accessed on 5 January 2023)
BTC2016Twitter-http://www.i2b2.org (accessed on 5 January 2023)
i2b2-20062007Medical-https://www.i2b2.org/NLP/DataSets/Main.php (accessed on 5 January 2023)
CADEC2015Medical-http://data.csiro.au/ (accessed on 5 January 2023)
CONLL032003News4https://github.com/synalp/NER/tree/master/corpus/CoNLL-2003 (accessed on 5 January 2023)
BiLingual NER
UNER2008SSEALURDUhttp://www.iiu.edu.pk/?page_id=5181 (accessed on 5 January 2023)
MK-PUCIT2019-URDUhttps://www.dropbox.com/sh/1ivw7ykm2tugg94/AAB9t5wnN7FynESpo7TjJW8la (accessed on 5 January 2023)
ACE-032003WikipediaCh/Arhttps://catalog.ldc.upenn.edu/LDC2004T09 (accessed on 5 January 2023)
ACE-042004WikipediaCh/Arhttps://catalog.ldc.upenn.edu/LDC2005T09 (accessed on 5 January 2023)
ACE-072007WikipediaCh/Arhttps://catalog.ldc.upenn.edu/LDC2006T06 (accessed on 5 January 2023)
QUAERO-French MedicalFrhttps://quaerofrenchmed.limsi.fr/ (accessed on 5 January 2023)
KIND--Ithttps://github.com/dhfbk/KIND (accessed on 5 January 2023)
hr500k--Crhttp://hdl.handle.net/11356/1183 (accessed on 5 January 2023)
CoNLL032003GermanGrhttps://www.clips.uantwerpen.be/conll2003/ner/ (accessed on 5 January 2023)
CoNLL032003DutchDuhttps://www.clips.uantwerpen.be/conll2002/ner/ (accessed on 5 January 2023)
1 URL is provided to assist the access of free available data.
Table 2. AI in healthcare summary of recent and relevant articles.
Table 2. AI in healthcare summary of recent and relevant articles.
ReferenceYearMethodApplicationFieldLimitationsKey Features
[115]2023Neural network (NN)Develops and implements a NN-based method for skin cancer predictionDermatologyThis investigation failed to identify the optimal CNN architecture.Image recognition
[116]2023DL systemsComparing the performance of a DL against a retinal specialist vision-threatening all-cause referable Ophthalmology This study does not use: DLS-driven DR screening, a focus on probable validation, and real-world implementation.Image recognition
[117]2022 Neural network Image segmentation review and six optimizations of the radiotherapy dose. Oncology Many-to-one mapping; intrinsic fitting error; patient-specific tissue heterogeneitiesImage recognition
[118]2022Multi-channel LSTMTo signify a medical event that is strongly linked to the prediction task using EHRsMedical event predictionFast knowledge and effectively managing the clinical result from intricate and heterogeneous EHRsNLP
[119]2022LSTM-CNNTwitter analysis and prediction of real-time COVID-19 dataPathologyState-of-the-art results, implementing bidirectional LSTM NLP
[120]2022ResNet-CNN-BiLSTMTo conclude, what kind of cyberbullying is being committed, and classify the individuals involvedPsychological careThe validity and reliability of the data, as well as biosensor and biomarker-based techniques, are affected by a number of factors, such as sample quality, instrument accuracy, and environmental conditions.NLP
[121]2021NNDiagnosing cardiovascular diseases using cardiac MRI imagesRadiologyLack of generalizability of the results, which ends with lesion classification, is not automated here during MRI analysis. Image recognition
[122]2019NN-integrated gradientsGrading diabetic retinopathy more rapidly and preciselyOphthalmologyInsufficient discussion or interpretation of the results. It is likely that the length of the grading process will affect the experience in some way.Image recognition
[123]2019NNTo regulate which of the X-ray images showed a healthy chestRadiologyOptimization of the probability threshold for classification of an Inception-ResNet-v2 model trained on ImageNetImage recognition
[124]2019LR
(Logistic regression)
Ultrasound inspections are used to establish the criteria for classifying breast cancer known as triple negative (TN).RadiologyLacking computerized segmentation of the masses, fewer people are involved in sampling bias.Image recognition
[125]2018-Multi-step processing, including deep-learning-based segmentation, revealed variability in the composition of tumor-immune It offers a unique challenge when it comes to de-noising MIBI-TOF data due to two properties: Low intensity values, low abundant antigens, and signal is sparse and pixelated Image recognition
[126] 2017 Neural networkTo achieve dermatologist-level accuracy when diagnosing skin cancerDermatologyCNNs require large amounts of data to train. Collecting large amount of labeled skin cancer images is difficult and time-consuming. Image recognition
Table 3. CRF/LSTM is the state-of-the-art performance encoder/decoder for CoNLL03 and OntoNotes5.0.
Table 3. CRF/LSTM is the state-of-the-art performance encoder/decoder for CoNLL03 and OntoNotes5.0.
WorkInput RepresentationEncoder/Decoder 1F1-Score
WordCharacter
[129,130]Google Word2vecLSTMLSTM/CRF0.8409
[131]Globel VectorLSTMLSTM/CRF0.9337
[132]Twitter Word2vecCNNLSTM/CRF0.4186
[88]WordPiece-Transformer/Softmax0.9208
1 Decoder is the most normally chosen tag.
Table 4. The major disease categories and topics using clinical decision-making are arranged here, as well as the data sources for the papers.
Table 4. The major disease categories and topics using clinical decision-making are arranged here, as well as the data sources for the papers.
AuthorSymptom 1Data SourceTopic Relation
Tsipouras et al. [155]Coronary diseaseCardiology department University Hospital GreeceCoronary heart disease treatment
Rakha et al. [156]CancerBreast cancer data set, WisconsinBreast cancer classification patients with innovative algorithm
Tapak et al. [157]DiabetesIranian national non-communicable diseases risk factors surveillanceEvaluation of diabetes classification accuracy algorithms
Iqbal et al. [158]CancerTaiwan National Health Insurance DatabaseVisualization tool for cancer
Agrawal et al. [159]Lung cancerSEER Program of the National Cancer Institute, USALung cancer survival prediction
Yeh et al. [160]OtherHemodialysis center in TaiwanHospitalization prediction of Hemodialysis patients
1 Data mining is applied to identify rare forms of diabetes, coronary heart Disease, and cancer.
Table 5. Data sources and problems in healthcare administration with regard to the analysis of problems.
Table 5. Data sources and problems in healthcare administration with regard to the analysis of problems.
AuthorResearch 1Data SourceLimitation Analysis
Arvisais et al. [161] cloud computingNot specifiedReports from radiology repositories
Pur et al. [162]Utilization, quality, and costs of healthcareNational Institute of Public Health; Health Care Institute, Celje; SlovenianFramework and model for monitoring HC networks’ resource allocation
Koskela et al. [163]Management of patientsTampere Health Centre, FinlandPredicting persistent healthcare attendance by examining risk factors
Oliwa et al. [164]Healthcare quality Pathology company in AustraliaEfficient pathology ordering organization
1 Keeping electronic patient records secure and cost-effective requires data warehousing and cloud computing.
Table 6. Data sources and problems in healthcare administration with regards to the analysis of problems.
Table 6. Data sources and problems in healthcare administration with regards to the analysis of problems.
AuthorResearchData SourceLimitation Analysis
Nimmagada et al. [165]Healthcare World Health OrganizationPrograms designed to prevent health problems
Kostkova et al. [166]Healthcare qualityNational electronic Library of InfectionPublic and professional information seeking behavior
Rathore et al. [167]Healthcare dataInternet of Things and big data for real-time emergency responseUCI machine learning depository
Lavrač et al. [168]Healthcare qualitySlovenian national Institute of Public HealthInnovative use of data mining
Table 7. The five-fold cross-validation metrics for each SERVQUAL dimension in MLQC [97].
Table 7. The five-fold cross-validation metrics for each SERVQUAL dimension in MLQC [97].
1 MetricsTangibleReliabilityResponsiveAssuranceEmpathy
NBSVMLRNBSVMLRNBSVMLRNBSVMLRNBSVMLR
Binary relevance
Acc0.6750.70.680.690.710.710.630.640.650.640.710.730.780.780.79
F10.390.580.470.810.820.810.480.550.540.680.720.730.870.870.88
P0.760.690.670.680.760.710.660.610.670.60.710.710.780.80.79
R0.270.510.360.990.880.970.390.510.460.790.730.7510.80.99
Label powerset
Acc0.660.60.640.690.680.70.550.610.610.560.630.670.780.780.8
F10.530.520.480.810.790.810.630.610.620.670.680.670.870.870.88
P0.610.610.570.690.720.710.50.550.550.520.590.560.780.810.8
R0.490.40.430.990.880.920.870.680.730.940.810.820.90.970.9
Classifier chain
Acc0.670.710.680.690.730.710.630.650.640.650.630.710.780.790.79
F10.3990.580.470.810.810.820.470.570.570.680.720.70.870.870.88
P0.760.690.670.680.760.710.670.630.610.610.720.710.780.820.8
R0.270.510.370.90.870.940.370.530.50.70.740.710.920.98
Rak-EL
Acc0.630.710.670.690.70.710.620.630.650.640.680.690.780.780.79
F10.270.580.490.810.80.810.520.550.560.650.680.690.870.80.88
P0.680.70.650.690.730.730.650.590.640.630.660.670.780.870.8
R0.170.490.390.90.90.930.50.520.520.710.710.7210.950.98
1 The main evaluation metrics are Accuracy (Acc), Precision (P), Recall (R), and F1-score; Naïve Bayes (NB), support vector machines (SVM), and logistic regression (LR), as well as Random k-Label sets (Rak-EL) are used.
Table 8. The performance classification of multi-label models is based on the five-fold cross-validation of the ML models [97].
Table 8. The performance classification of multi-label models is based on the five-fold cross-validation of the ML models [97].
ClassifierAccuracyF1-ScorePrecisionRecall
Binary RelevanceNB14.71%0.730.700.76
SVM21.1%0.750.740.76
LR19.3%0.750.730.77
Classifier ChainNB14.9%0.740.630.89
SVM21.5%0.730.670.79
LR19.1%0.730.660.82
Label PowersetNB13.0%0.730.700.75
SVM16.6%0.750.750.76
LR15.8%0.740.720.77
ML-KNN, BR-KNN 1NB18.0%0.720.690.74
SVM14.0%0.740.720.76
Rak-ELNB15.7%0.740.720.76
SVM18.6%0.710.690.73
LR18.0%0.680.730.64
1 Multi-Label K-Nearest Neighbors algorithm (MLKNN); Binary Relevance K-Nearest Neighbors (Br KNN), Random k-Label sets (Rak-EL).
Table 9. Summaries of the clinical expressions in healthcare and biomedicine.
Table 9. Summaries of the clinical expressions in healthcare and biomedicine.
AuthorYearMethodStudy PopulationDesignSummaryObjective
Shickel et al. [2]2018DL to clinical tasksEHR ProjectsReview Data mining for clinical tasks based on EHRs.
Applications include information extraction, representation learning, outcome prediction, and phenotyping.
Model interpretability, data heterogeneity, and the lack of universal benchmarks are among the limitations found in current research.
Festen et al. [13]2021Descriptive statisticsTwo hundred fourteen patients with cancer of 70 year Recruited from a University Medical Center, Netherlands.Quantitative(1) Implementing nurse-driven geriatric assessment.
(2) Evaluate the patient’s preferences regarding treatment outcomes. (3) Considered patient assessments in their decision-making.
The geriatric assessment of older cancer patients is incorporated into a novel care pathway. A multidisciplinary team (MDT) of oncogeriatric experts could modify treatment.
Quinn et al. [15]2019Content analysis.patients 32 (17 stable, 15 unstableSurveyHealthcare team confidence and connection are associated with stable HCT. In contrast to complete independence or autonomy, interdependence involves building connections with healthcare providers and eliciting their support.Researchers determined that resilience constructs play a protective role in helping AYA with KT maintain stable HCT.
Song et al. [49]2021DL-based methods for BioNER and datasetsBiomedical text miningReviewAlgorithms used in deep learning can be categorized into four types: that which is based on one neural network, that which is based on multi-task learning, that which is based on transfer learning, and that which is based on hybrid models.Dataset size and type determine the results of these algorithms when applied to BioNER in a wide variety of domains.
Bridge et al. [67]2015Face-to-face interviews. Framework Analysis.Healthcare professionals
(n = 22; nurse;
n = 11 physician) Recruited NHS hospital in UK.
Qualitative(1) Provide quality, timely, and accurate information to patients.
(2) In the meeting, patients’ complex information is given attention.
(3) Promoting patient autonomy. (4) Participate in multidisciplinary teams.
Identify factors shaping cancer treatment decisions for older people with complex needs, including health and social care.
Jones et al. [170]2018Semi-structured interview.Thirty-five sets of patients and their decision partners, Recruited from Cancer Center, USAQualitative(1) Making the discussion more thorough between patients and decision partners with the use of decision aids.
(2) Facilitating patient-provider relationships via decision aids.
Patient and partner experiences with an interactive decision aid utilized to make informed, shared treatment choices for advanced prostate cancer.
Peng et al. [117]2022 AI utilization in particle therapy Image segmentation review and six optimization of the radiotherapy dose.Oncology reviewResearch on AI-powered particle therapy based on a comprehensive literature review. Creating data, interpreting it, incorporating fundamental physical processes, and rigorously validating it are all as important as the actual AI model development.The aim is to present six aspect included: dose calculation, treatment planning, image guidance, range and dose verification, adaptive replanning and quality assurance
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ahmad, P.N.; Shah, A.M.; Lee, K. A Review on Electronic Health Record Text-Mining for Biomedical Name Entity Recognition in Healthcare Domain. Healthcare 2023, 11, 1268. https://doi.org/10.3390/healthcare11091268

AMA Style

Ahmad PN, Shah AM, Lee K. A Review on Electronic Health Record Text-Mining for Biomedical Name Entity Recognition in Healthcare Domain. Healthcare. 2023; 11(9):1268. https://doi.org/10.3390/healthcare11091268

Chicago/Turabian Style

Ahmad, Pir Noman, Adnan Muhammad Shah, and KangYoon Lee. 2023. "A Review on Electronic Health Record Text-Mining for Biomedical Name Entity Recognition in Healthcare Domain" Healthcare 11, no. 9: 1268. https://doi.org/10.3390/healthcare11091268

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop