Next Issue
Volume 5, March
Previous Issue
Volume 4, September
 
 

Mach. Learn. Knowl. Extr., Volume 4, Issue 4 (December 2022) – 18 articles

Cover Story (view full-size image): Over the past decade, deep learning has shown excellent performance in complex application scenarios. Due to the complexity of deep learning, many applications rely on black-box software packages such as Tensorflow and Pytorch. These frameworks are constantly improving, and new versions are frequently released. However, like any other software, new versions may contain changes that result in side effects and bugs, which can cause the performance of models to degrade or even fail completely. In this study, the impact of version changes on model performance is investigated for the popular frameworks Tensorflow and Pytorch. The results show that several version changes had a significant negative impact and users are therefore advised to perform regression tests for their models on every upgrade to a new framework version. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
12 pages, 2247 KiB  
Article
An Explainable Deep Learning Framework for Detecting and Localising Smoke and Fire Incidents: Evaluation of Grad-CAM++ and LIME
by Ioannis D. Apostolopoulos, Ifigeneia Athanasoula, Mpesi Tzani and Peter P. Groumpos
Mach. Learn. Knowl. Extr. 2022, 4(4), 1124-1135; https://doi.org/10.3390/make4040057 - 06 Dec 2022
Cited by 5 | Viewed by 2613
Abstract
Climate change is expected to increase fire events and activity with multiple impacts on human lives. Large grids of forest and city monitoring devices can assist in incident detection, accelerating human intervention in extinguishing fires before they get out of control. Artificial Intelligence [...] Read more.
Climate change is expected to increase fire events and activity with multiple impacts on human lives. Large grids of forest and city monitoring devices can assist in incident detection, accelerating human intervention in extinguishing fires before they get out of control. Artificial Intelligence promises to automate the detection of fire-related incidents. This study enrols 53,585 fire/smoke and normal images and benchmarks seventeen state-of-the-art Convolutional Neural Networks for distinguishing between the two classes. The Xception network proves to be superior to the rest of the CNNs, obtaining very high accuracy. Grad-CAM++ and LIME algorithms improve the post hoc explainability of Xception and verify that it is learning features found in the critical locations of the image. Both methods agree on the suggested locations, strengthening the abovementioned outcome. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

17 pages, 4418 KiB  
Article
Ontology Completion with Graph-Based Machine Learning: A Comprehensive Evaluation
by Sebastian Mežnar, Matej Bevec, Nada Lavrač and Blaž Škrlj
Mach. Learn. Knowl. Extr. 2022, 4(4), 1107-1123; https://doi.org/10.3390/make4040056 - 01 Dec 2022
Viewed by 3705
Abstract
Increasing quantities of semantic resources offer a wealth of human knowledge, but their growth also increases the probability of wrong knowledge base entries. The development of approaches that identify potentially spurious parts of a given knowledge base is therefore highly relevant. We propose [...] Read more.
Increasing quantities of semantic resources offer a wealth of human knowledge, but their growth also increases the probability of wrong knowledge base entries. The development of approaches that identify potentially spurious parts of a given knowledge base is therefore highly relevant. We propose an approach for ontology completion that transforms an ontology into a graph and recommends missing edges using structure-only link analysis methods. By systematically evaluating thirteen methods (some for knowledge graphs) on eight different semantic resources, including Gene Ontology, Food Ontology, Marine Ontology, and similar ontologies, we demonstrate that a structure-only link analysis can offer a scalable and computationally efficient ontology completion approach for a subset of analyzed data sets. To the best of our knowledge, this is currently the most extensive systematic study of the applicability of different types of link analysis methods across semantic resources from different domains. It demonstrates that by considering symbolic node embeddings, explanations of the predictions (links) can be obtained, making this branch of methods potentially more valuable than black-box methods. Full article
Show Figures

Figure 1

19 pages, 3329 KiB  
Article
SGD-Based Cascade Scheme for Higher Degrees Wiener Polynomial Approximation of Large Biomedical Datasets
by Ivan Izonin, Roman Tkachenko, Rostyslav Holoven, Kyrylo Yemets, Myroslav Havryliuk and Shishir Kumar Shandilya
Mach. Learn. Knowl. Extr. 2022, 4(4), 1088-1106; https://doi.org/10.3390/make4040055 - 21 Nov 2022
Cited by 2 | Viewed by 1847
Abstract
The modern development of the biomedical engineering area is accompanied by the availability of large volumes of data with a non-linear response surface. The effective analysis of such data requires the development of new, more productive machine learning methods. This paper proposes a [...] Read more.
The modern development of the biomedical engineering area is accompanied by the availability of large volumes of data with a non-linear response surface. The effective analysis of such data requires the development of new, more productive machine learning methods. This paper proposes a cascade ensemble that combines the advantages of using a high-order Wiener polynomial and Stochastic Gradient Descent algorithm while eliminating their disadvantages to ensure a high accuracy of the approximation of such data with a satisfactory training time. The work presents flow charts of the learning algorithms and the application of the developed ensemble scheme, and all the steps are described in detail. The simulation was carried out based on a real-world dataset. Procedures for the proposed model tuning have been performed. The high accuracy of the approximation based on the developed ensemble scheme was established experimentally. The possibility of an implicit approximation by high orders of the Wiener polynomial with a slight increase in the number of its members is shown. It ensures a low training time for the proposed method during the analysis of large datasets, which provides the possibility of its practical use in the biomedical engineering area. Full article
Show Figures

Figure 1

23 pages, 3600 KiB  
Article
A New Rough Set Classifier for Numerical Data Based on Reflexive and Antisymmetric Relations
by Yoshie Ishii, Koki Iwao and Tsuguki Kinoshita
Mach. Learn. Knowl. Extr. 2022, 4(4), 1065-1087; https://doi.org/10.3390/make4040054 - 18 Nov 2022
Cited by 1 | Viewed by 2122
Abstract
The grade-added rough set (GRS) approach is an extension of the rough set theory proposed by Pawlak to deal with numerical data. However, the GRS has problems with overtraining, unclassified and unnatural results. In this study, we propose a new approach called the [...] Read more.
The grade-added rough set (GRS) approach is an extension of the rough set theory proposed by Pawlak to deal with numerical data. However, the GRS has problems with overtraining, unclassified and unnatural results. In this study, we propose a new approach called the directional neighborhood rough set (DNRS) approach to solve the problems of the GRS. The information granules in the DNRS are based on reflexive and antisymmetric relations. Following these relations, new lower and upper approximations are defined. Based on these definitions, we developed a classifier with a three-step algorithm, including DN-lower approximation classification, DN-upper approximation classification, and exceptional processing. Three experiments were conducted using the University of California Irvine (UCI)’s machine learning dataset to demonstrate the effect of each step in the DNRS model, overcoming the problems of the GRS, and achieving more accurate classifiers. The results showed that when the number of dimensions is reduced and both the lower and upper approximation algorithms are used, the DNRS model is more efficient than when the number of dimensions is large. Additionally, it was shown that the DNRS solves the problems of the GRS and the DNRS model is as accurate as existing classifiers. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

23 pages, 4081 KiB  
Article
On the Dimensionality and Utility of Convolutional Autoencoder’s Latent Space Trained with Topology-Preserving Spectral EEG Head-Maps
by Arjun Vinayak Chikkankod and Luca Longo
Mach. Learn. Knowl. Extr. 2022, 4(4), 1042-1064; https://doi.org/10.3390/make4040053 - 18 Nov 2022
Cited by 9 | Viewed by 2382
Abstract
Electroencephalography (EEG) signals can be analyzed in the temporal, spatial, or frequency domains. Noise and artifacts during the data acquisition phase contaminate these signals adding difficulties in their analysis. Techniques such as Independent Component Analysis (ICA) require human intervention to remove noise and [...] Read more.
Electroencephalography (EEG) signals can be analyzed in the temporal, spatial, or frequency domains. Noise and artifacts during the data acquisition phase contaminate these signals adding difficulties in their analysis. Techniques such as Independent Component Analysis (ICA) require human intervention to remove noise and artifacts. Autoencoders have automatized artifact detection and removal by representing inputs in a lower dimensional latent space. However, little research is devoted to understanding the minimum dimension of such latent space that allows meaningful input reconstruction. Person-specific convolutional autoencoders are designed by manipulating the size of their latent space. A sliding window technique with overlapping is employed to segment varied-sized windows. Five topographic head-maps are formed in the frequency domain for each window. The latent space of autoencoders is assessed using the input reconstruction capacity and classification utility. Findings indicate that the minimal latent space dimension is 25% of the size of the topographic maps for achieving maximum reconstruction capacity and maximizing classification accuracy, which is achieved with a window length of at least 1 s and a shift of 125 ms, using the 128 Hz sampling rate. This research contributes to the body of knowledge with an architectural pipeline for eliminating redundant EEG data while preserving relevant features with deep autoencoders. Full article
(This article belongs to the Special Issue Advances in Explainable Artificial Intelligence (XAI))
Show Figures

Figure 1

18 pages, 2707 KiB  
Article
A Morphological Post-Processing Approach for Overlapped Segmentation of Bacterial Cell Images
by Dilanga Abeyrathna, Shailabh Rauniyar, Rajesh K. Sani and Pei-Chi Huang
Mach. Learn. Knowl. Extr. 2022, 4(4), 1024-1041; https://doi.org/10.3390/make4040052 - 17 Nov 2022
Cited by 6 | Viewed by 2976
Abstract
Scanning electron microscopy (SEM) techniques have been extensively performed to image and study bacterial cells with high-resolution images. Bacterial image segmentation in SEM images is an essential task to distinguish an object of interest and its specific region. These segmentation results can then [...] Read more.
Scanning electron microscopy (SEM) techniques have been extensively performed to image and study bacterial cells with high-resolution images. Bacterial image segmentation in SEM images is an essential task to distinguish an object of interest and its specific region. These segmentation results can then be used to retrieve quantitative measures (e.g., cell length, area, cell density) for the accurate decision-making process of obtaining cellular objects. However, the complexity of the bacterial segmentation task is a barrier, as the intensity and texture of foreground and background are similar, and also, most clustered bacterial cells in images are partially overlapping with each other. The traditional approaches for identifying cell regions in microscopy images are labor intensive and heavily dependent on the professional knowledge of researchers. To mitigate the aforementioned challenges, in this study, we tested a U-Net-based semantic segmentation architecture followed by a post-processing step of morphological over-segmentation resolution to achieve accurate cell segmentation of SEM-acquired images of bacterial cells grown in a rotary culture system. The approach showed an 89.52% Dice similarity score on bacterial cell segmentation with lower segmentation error rates, validated over several cell overlapping object segmentation approaches with significant performance improvement. Full article
Show Figures

Figure 1

13 pages, 1898 KiB  
Article
Evidence-Based Regularization for Neural Networks
by Giuseppe Nuti, Andreea-Ingrid Cross and Philipp Rindler
Mach. Learn. Knowl. Extr. 2022, 4(4), 1011-1023; https://doi.org/10.3390/make4040051 - 15 Nov 2022
Cited by 1 | Viewed by 2301
Abstract
Numerous approaches address over-fitting in neural networks: by imposing a penalty on the parameters of the network (L1, L2, etc.); by changing the network stochastically (drop-out, Gaussian noise, etc.); or by transforming the input data (batch normalization, etc.). In contrast, we aim to [...] Read more.
Numerous approaches address over-fitting in neural networks: by imposing a penalty on the parameters of the network (L1, L2, etc.); by changing the network stochastically (drop-out, Gaussian noise, etc.); or by transforming the input data (batch normalization, etc.). In contrast, we aim to ensure that a minimum amount of supporting evidence is present when fitting the model parameters to the training data. This, at the single neuron level, is equivalent to ensuring that both sides of the separating hyperplane (for a standard artificial neuron) have a minimum number of data points, noting that these points need not belong to the same class for the inner layers. We firstly benchmark the results of this approach on the standard Fashion-MINST dataset, comparing it to various regularization techniques. Interestingly, we note that by nudging each neuron to divide, at least in part, its input data, the resulting networks make use of each neuron, avoiding a hyperplane completely on one side of its input data (which is equivalent to a constant into the next layers). To illustrate this point, we study the prevalence of saturated nodes throughout training, showing that neurons are activated more frequently and earlier in training when using this regularization approach. A direct consequence of the improved neuron activation is that deep networks are now easier to train. This is crucially important when the network topology is not known a priori and fitting often remains stuck in a suboptimal local minima. We demonstrate this property by training a network of increasing depth (and constant width); most regularization approaches will result in increasingly frequent training failures (over different random seeds), whilst the proposed evidence-based regularization significantly outperforms in its ability to train deep networks. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

17 pages, 1861 KiB  
Article
Semantic Interactive Learning for Text Classification: A Constructive Approach for Contextual Interactions
by Sebastian Kiefer, Mareike Hoffmann and Ute Schmid
Mach. Learn. Knowl. Extr. 2022, 4(4), 994-1010; https://doi.org/10.3390/make4040050 - 13 Nov 2022
Cited by 1 | Viewed by 2109
Abstract
Interactive Machine Learning (IML) can enable intelligent systems to interactively learn from their end-users, and is quickly becoming more and more relevant to many application domains. Although it places the human in the loop, interactions are mostly performed via mutual explanations that miss [...] Read more.
Interactive Machine Learning (IML) can enable intelligent systems to interactively learn from their end-users, and is quickly becoming more and more relevant to many application domains. Although it places the human in the loop, interactions are mostly performed via mutual explanations that miss contextual information. Furthermore, current model-agnostic IML strategies such as CAIPI are limited to ’destructive’ feedback, meaning that they solely allow an expert to prevent a learner from using irrelevant features. In this work, we propose a novel interaction framework called Semantic Interactive Learning for the domain of document classification, located at the intersection between Natural Language Processing (NLP) and Machine Learning (ML). We frame the problem of incorporating constructive and contextual feedback into the learner as a task involving finding an architecture that enables more semantic alignment between humans and machines while at the same time helping to maintain the statistical characteristics of the input domain when generating user-defined counterexamples based on meaningful corrections. Therefore, we introduce a technique called SemanticPush that is effective for translating conceptual corrections of humans to non-extrapolating training examples such that the learner’s reasoning is pushed towards the desired behavior. Through several experiments we show how our method compares to CAIPI, a state of the art IML strategy, in terms of Predictive Performance and Local Explanation Quality in downstream multi-class classification tasks. Especially in the early stages of interactions, our proposed method clearly outperforms CAIPI while allowing for contextual interpretation and intervention. Overall, SemanticPush stands out with regard to data efficiency, as it requires fewer queries from the pool dataset to achieve high accuracy. Full article
Show Figures

Figure 1

26 pages, 1434 KiB  
Article
FeaSel-Net: A Recursive Feature Selection Callback in Neural Networks
by Felix Fischer, Alexander Birk, Peter Somers, Karsten Frenner, Cristina Tarín and Alois Herkommer
Mach. Learn. Knowl. Extr. 2022, 4(4), 968-993; https://doi.org/10.3390/make4040049 - 31 Oct 2022
Cited by 3 | Viewed by 1915
Abstract
Selecting only the relevant subsets from all gathered data has never been as challenging as it is in these times of big data and sensor fusion. Multiple complementary methods have emerged for the observation of similar phenomena; oftentimes, many of these techniques are [...] Read more.
Selecting only the relevant subsets from all gathered data has never been as challenging as it is in these times of big data and sensor fusion. Multiple complementary methods have emerged for the observation of similar phenomena; oftentimes, many of these techniques are superimposed in order to make the best possible decisions. A pathologist, for example, uses microscopic and spectroscopic techniques to discriminate between healthy and cancerous tissue. Especially in the field of spectroscopy in medicine, an immense number of frequencies are recorded and appropriately sized datasets are rarely acquired due to the time-intensive measurements and the lack of patients. In order to cope with the curse of dimensionality in machine learning, it is necessary to reduce the overhead from irrelevant or redundant features. In this article, we propose a feature selection callback algorithm (FeaSel-Net) that can be embedded in deep neural networks. It recursively prunes the input nodes after the optimizer in the neural network achieves satisfying results. We demonstrate the performance of the feature selection algorithm on different publicly available datasets and compare it to existing feature selection methods. Our algorithm combines the advantages of neural networks’ nonlinear learning ability and the embedding of the feature selection algorithm into the actual classifier optimization. Full article
(This article belongs to the Special Issue Recent Advances in Feature Selection)
Show Figures

Figure 1

14 pages, 613 KiB  
Article
Lottery Ticket Structured Node Pruning for Tabular Datasets
by Ryan Bluteau, Robin Gras, Zachary Innes and Mitchel Paulin
Mach. Learn. Knowl. Extr. 2022, 4(4), 954-967; https://doi.org/10.3390/make4040048 - 28 Oct 2022
Cited by 1 | Viewed by 1573
Abstract
This paper experiments with well known pruning approaches, iterative and one-shot, and presents a new approach to lottery ticket pruning applied to tabular neural networks based on iterative pruning. Our contribution is a standard model for comparison in terms of speed and performance [...] Read more.
This paper experiments with well known pruning approaches, iterative and one-shot, and presents a new approach to lottery ticket pruning applied to tabular neural networks based on iterative pruning. Our contribution is a standard model for comparison in terms of speed and performance for tabular datasets that often do not get optimized through research. We show leading results in several tabular datasets that can compete with ensemble approaches. We tested on a wide range of datasets with a general improvement over the original (already leading) model in 6 of 8 datasets tested in terms of F1/RMSE. This includes a total reduction of over 85% of nodes with the additional ability to prune over 98% of nodes with minimal affect to accuracy. The new iterative approach we present will first optimize for lottery ticket quality by selecting an optimal architecture size and weights, then apply the iterative pruning strategy. The new iterative approach shows minimal degradation in accuracy compared to the original iterative approach, but it is capable of pruning models much smaller due to optimal weight pre-selection. Training and inference time improved over 50% and 10%, respectively, and up to 90% and 35%, respectively, for large datasets. Full article
(This article belongs to the Section Learning)
Show Figures

Figure 1

30 pages, 4010 KiB  
Article
Actionable Explainable AI (AxAI): A Practical Example with Aggregation Functions for Adaptive Classification and Textual Explanations for Interpretable Machine Learning
by Anna Saranti, Miroslav Hudec, Erika Mináriková, Zdenko Takáč, Udo Großschedl, Christoph Koch, Bastian Pfeifer, Alessa Angerschmid and Andreas Holzinger
Mach. Learn. Knowl. Extr. 2022, 4(4), 924-953; https://doi.org/10.3390/make4040047 - 27 Oct 2022
Cited by 14 | Viewed by 3400
Abstract
In many domains of our daily life (e.g., agriculture, forestry, health, etc.), both laymen and experts need to classify entities into two binary classes (yes/no, good/bad, sufficient/insufficient, benign/malign, etc.). For many entities, this decision is difficult and we need another class called “maybe”, [...] Read more.
In many domains of our daily life (e.g., agriculture, forestry, health, etc.), both laymen and experts need to classify entities into two binary classes (yes/no, good/bad, sufficient/insufficient, benign/malign, etc.). For many entities, this decision is difficult and we need another class called “maybe”, which contains a corresponding quantifiable tendency toward one of these two opposites. Human domain experts are often able to mark any entity, place it in a different class and adjust the position of the slope in the class. Moreover, they can often explain the classification space linguistically—depending on their individual domain experience and previous knowledge. We consider this human-in-the-loop extremely important and call our approach actionable explainable AI. Consequently, the parameters of the functions are adapted to these requirements and the solution is explained to the domain experts accordingly. Specifically, this paper contains three novelties going beyond the state-of-the-art: (1) A novel method for detecting the appropriate parameter range for the averaging function to treat the slope in the “maybe” class, along with a proposal for a better generalisation than the existing solution. (2) the insight that for a given problem, the family of t-norms and t-conorms covering the whole range of nilpotency is suitable because we need a clear “no” or “yes” not only for the borderline cases. Consequently, we adopted the Schweizer–Sklar family of t-norms or t-conorms in ordinal sums. (3) A new fuzzy quasi-dissimilarity function for classification into three classes: Main difference, irrelevant difference and partial difference. We conducted all of our experiments with real-world datasets. Full article
Show Figures

Figure 1

12 pages, 3878 KiB  
Article
Prospective Neural Network Model for Seismic Precursory Signal Detection in Geomagnetic Field Records
by Laura Petrescu and Iren-Adelina Moldovan
Mach. Learn. Knowl. Extr. 2022, 4(4), 912-923; https://doi.org/10.3390/make4040046 - 07 Oct 2022
Cited by 3 | Viewed by 2401
Abstract
We designed a convolutional neural network application to detect seismic precursors in geomagnetic field records. Earthquakes are among the most destructive natural hazards on Earth, yet their short-term forecasting has not been achieved. Stress loading in dry rocks can generate electric currents that [...] Read more.
We designed a convolutional neural network application to detect seismic precursors in geomagnetic field records. Earthquakes are among the most destructive natural hazards on Earth, yet their short-term forecasting has not been achieved. Stress loading in dry rocks can generate electric currents that cause short-term changes to the geomagnetic field, yielding theoretically detectable pre-earthquake electromagnetic emissions. We propose a CNN model that scans windows of geomagnetic data streams and self-updates using nearby earthquakes as labels, under strict detectability criteria. We show how this model can be applied in three key seismotectonic settings, where geomagnetic observatories are optimally located in high-seismicity-rate epicentral areas. CNNs require large datasets to be able to accurately label seismic precursors, so we expect the model to improve as more data become available with time. At present, there is no synthetic data generator for this kind of application, so artificial data augmentation is not yet possible. However, this deep learning model serves to illustrate its potential usage in earthquake forecasting in a systematic and unbiased way. Our method can be prospectively applied to any kind of three-component dataset that may be physically connected to seismogenic processes at a given depth. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

24 pages, 2020 KiB  
Article
How Do Deep-Learning Framework Versions Affect the Reproducibility of Neural Network Models?
by Mostafa Shahriari, Rudolf Ramler and Lukas Fischer
Mach. Learn. Knowl. Extr. 2022, 4(4), 888-911; https://doi.org/10.3390/make4040045 - 05 Oct 2022
Cited by 3 | Viewed by 3099
Abstract
In the last decade, industry’s demand for deep learning (DL) has increased due to its high performance in complex scenarios. Due to the DL method’s complexity, experts and non-experts rely on blackbox software packages such as Tensorflow and Pytorch. The frameworks are constantly [...] Read more.
In the last decade, industry’s demand for deep learning (DL) has increased due to its high performance in complex scenarios. Due to the DL method’s complexity, experts and non-experts rely on blackbox software packages such as Tensorflow and Pytorch. The frameworks are constantly improving, and new versions are released frequently. As a natural process in software development, the released versions contain improvements/changes in the methods and their implementation. Moreover, versions may be bug-polluted, leading to the model performance decreasing or stopping the model from working. The aforementioned changes in implementation can lead to variance in obtained results. This work investigates the effect of implementation changes in different major releases of these frameworks on the model performance. We perform our study using a variety of standard datasets. Our study shows that users should consider that changing the framework version can affect the model performance. Moreover, they should consider the possibility of a bug-polluted version before starting to debug source code that had an excellent performance before a version change. This also shows the importance of using virtual environments, such as Docker, when delivering a software product to clients. Full article
Show Figures

Figure 1

23 pages, 728 KiB  
Review
Entropic Statistics: Concept, Estimation, and Application in Machine Learning and Knowledge Extraction
by Jialin Zhang
Mach. Learn. Knowl. Extr. 2022, 4(4), 865-887; https://doi.org/10.3390/make4040044 - 30 Sep 2022
Cited by 1 | Viewed by 1780
Abstract
The demands for machine learning and knowledge extraction methods have been booming due to the unprecedented surge in data volume and data quality. Nevertheless, challenges arise amid the emerging data complexity as significant chunks of information and knowledge lie within the non-ordinal realm [...] Read more.
The demands for machine learning and knowledge extraction methods have been booming due to the unprecedented surge in data volume and data quality. Nevertheless, challenges arise amid the emerging data complexity as significant chunks of information and knowledge lie within the non-ordinal realm of data. To address the challenges, researchers developed considerable machine learning and knowledge extraction methods regarding various domain-specific challenges. To characterize and extract information from non-ordinal data, all the developed methods pointed to the subject of Information Theory, established following Shannon’s landmark paper in 1948. This article reviews recent developments in entropic statistics, including estimation of Shannon’s entropy and its functionals (such as mutual information and Kullback–Leibler divergence), concepts of entropic basis, generalized Shannon’s entropy (and its functionals), and their estimations and potential applications in machine learning and knowledge extraction. With the knowledge of recent development in entropic statistics, researchers can customize existing machine learning and knowledge extraction methods for better performance or develop new approaches to address emerging domain-specific challenges. Full article
(This article belongs to the Section Data)
13 pages, 4850 KiB  
Article
Automatic Extraction of Medication Information from Cylindrically Distorted Pill Bottle Labels
by Kseniia Gromova and Vinayak Elangovan
Mach. Learn. Knowl. Extr. 2022, 4(4), 852-864; https://doi.org/10.3390/make4040043 - 27 Sep 2022
Cited by 3 | Viewed by 3907
Abstract
Patient compliance with prescribed medication regimens is critical for maintaining health and managing disease and illness. To encourage patient compliance, multiple aids, like automatic pill dispensers, pill organizers, and various reminder applications, have been developed to help people adhere to their medication regimens. [...] Read more.
Patient compliance with prescribed medication regimens is critical for maintaining health and managing disease and illness. To encourage patient compliance, multiple aids, like automatic pill dispensers, pill organizers, and various reminder applications, have been developed to help people adhere to their medication regimens. However, when utilizing these aids, the user or patient must manually enter their medication information and schedule. This process is time-consuming and often prone to error. For example, elderly patients may have difficulty reading medication information on the bottle due to decreased eyesight, leading them to enter medication information incorrectly. This study explored methods for extracting pertinent information from cylindrically distorted prescription drug labels using Machine Learning and Computer Vision techniques. This study found that Deep Convolutional Neural Networks (DCNN) performed better than other techniques in identifying label key points under different lighting conditions and various backgrounds. This method achieved a percentage of Correct Key points PCK @ 0.03 of 97%. These key points were then used to correct the cylindrical distortion. Next, the multiple dewarped label images were stitched together and processed by an Optical Character Recognition (OCR) engine. Pertinent information, such as patient name, drug name, drug strength, and directions of use, were extracted from the recognized text using Natural Language Processing (NLP) techniques. The system created in this study can be used to improve patient health and compliance by creating an accurate medication schedule. Full article
(This article belongs to the Special Issue Language Processing and Knowledge Extraction)
Show Figures

Figure 1

13 pages, 3633 KiB  
Article
On the Application of Artificial Neural Network for Classification of Incipient Faults in Dissolved Gas Analysis of Power Transformers
by Bonginkosi A. Thango
Mach. Learn. Knowl. Extr. 2022, 4(4), 839-851; https://doi.org/10.3390/make4040042 - 26 Sep 2022
Cited by 4 | Viewed by 1888
Abstract
Oil-submerged transformer is one of the inherent instruments in the South African power system. Transformer malfunction or impairment may interpose the operation of the electric power distribution and transmission system, coupled with liability for high overhaul costs. Hence, recognition of inchoate faults in [...] Read more.
Oil-submerged transformer is one of the inherent instruments in the South African power system. Transformer malfunction or impairment may interpose the operation of the electric power distribution and transmission system, coupled with liability for high overhaul costs. Hence, recognition of inchoate faults in an oil-submerged transformer is indispensable and it has turned into an intriguing subject of interest by utility owners and transformer manufacturers. This work proposes a hybrid implementation of a multi-layer artificial neural network (MLANN) and IEC 60599:2022 gas ratio method in identifying inchoate faults in mineral oil-based submerged transformers by employing the dissolved gas analysis (DGA) method. DGA is a staunch practice to discover inchoate faults as it furnishes comprehensive information in examining the transformer state. In current work, MLANN was established to pigeonhole seven fault types of transformer states predicated on the three International Electrotechnical Commission (IEC) combustible gas ratios. The designs enmesh the development of numerous MLANN algorithms and picking networks with the optimum performance. The gas ratios are in accordance with the IEC 60599:2022 standard whilst an empirical databank comprised of 100 datasets was used in the training and testing activities. The designated MLANN design produces an overall correlation coefficient of 0.998 in the categorization of transformer state with reference to the combustible gas produced. Full article
(This article belongs to the Section Network)
Show Figures

Figure 1

12 pages, 1310 KiB  
Article
Comparison of Imputation Methods for Missing Rate of Perceived Exertion Data in Rugby
by Amarah Epp-Stobbe, Ming-Chang Tsai and Marc Klimstra
Mach. Learn. Knowl. Extr. 2022, 4(4), 827-838; https://doi.org/10.3390/make4040041 - 23 Sep 2022
Cited by 1 | Viewed by 1797
Abstract
Rate of perceived exertion (RPE) is used to calculate athlete load. Incomplete load data, due to missing athlete-reported RPE, can increase injury risk. The current standard for missing RPE imputation is daily team mean substitution. However, RPE reflects an individual’s effort; group mean [...] Read more.
Rate of perceived exertion (RPE) is used to calculate athlete load. Incomplete load data, due to missing athlete-reported RPE, can increase injury risk. The current standard for missing RPE imputation is daily team mean substitution. However, RPE reflects an individual’s effort; group mean substitution may be suboptimal. This investigation assessed an ideal method for imputing RPE. A total of 987 datasets were collected from women’s rugby sevens competitions. Daily team mean substitution, k-nearest neighbours, random forest, support vector machine, neural network, linear, stepwise, lasso, ridge, and elastic net regression models were assessed at different missingness levels. Statistical equivalence of true and imputed scores by model were evaluated. An ANOVA of accuracy by model and missingness was completed. While all models were equivalent to the true RPE, differences by model existed. Daily team mean substitution was the poorest performing model, and random forest, the best. Accuracy was low in all models, affirming RPE as multifaceted and requiring quantification of potentially overlapping factors. While group mean substitution is discouraged, practitioners are recommended to scrutinize any imputation method relating to athlete load. Full article
(This article belongs to the Section Data)
Show Figures

Figure 1

13 pages, 1012 KiB  
Review
Artificial Intelligence Methods for Identifying and Localizing Abnormal Parathyroid Glands: A Review Study
by Ioannis D. Apostolopoulos, Nikolaos I. Papandrianos, Elpiniki I. Papageorgiou and Dimitris J. Apostolopoulos
Mach. Learn. Knowl. Extr. 2022, 4(4), 814-826; https://doi.org/10.3390/make4040040 - 21 Sep 2022
Cited by 4 | Viewed by 2129
Abstract
Background: Recent advances in Artificial Intelligence (AI) algorithms, and specifically Deep Learning (DL) methods, demonstrate substantial performance in detecting and classifying medical images. Recent clinical studies have reported novel optical technologies which enhance the localization or assess the viability of Parathyroid Glands (PG) [...] Read more.
Background: Recent advances in Artificial Intelligence (AI) algorithms, and specifically Deep Learning (DL) methods, demonstrate substantial performance in detecting and classifying medical images. Recent clinical studies have reported novel optical technologies which enhance the localization or assess the viability of Parathyroid Glands (PG) during surgery, or preoperatively. These technologies could become complementary to the surgeon’s eyes and may improve surgical outcomes in thyroidectomy and parathyroidectomy. Methods: The study explores and reports the use of AI methods for identifying and localizing PGs, Primary Hyperparathyroidism (PHPT), Parathyroid Adenoma (PTA), and Multiglandular Disease (MGD). Results: The review identified 13 publications that employ Machine Learning and DL methods for preoperative and operative implementations. Conclusions: AI can aid in PG, PHPT, PTA, and MGD detection, as well as PG abnormality discrimination, both during surgery and non-invasively. Full article
Show Figures

Figure 1

Previous Issue
Next Issue
Back to TopTop