Topic Editors

Faculty of Mechanical and Industrial Engineering, Warsaw University of Technology, 02-524 Warsaw, Poland
Department of Electrical and Computer Engineering, COMSATS University Islamabad, Lahore 54000, Pakistan
Dr. Yong Wang
School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
Center for Automation and Robotics, Politechnic University of Madrid (UPM) and Spanish Council for Scientific Research (CSIC), Ctra. Campo Real km 0,200 La Poveda, 28500 Madrid, Spain

Advances in Artificial Neural Networks

Abstract submission deadline
31 October 2023
Manuscript submission deadline
31 December 2023
Viewed by
33621

Topic Information

Dear Colleagues,

This Topic Issue (TI), interdisciplinary in character, focuses on the dissemination of ideas between the various branches of pure and applied sciences, covering a range of artificial neural networks, including pattern recognition, computer vision, logical reasoning, knowledge engineering, expert systems, artificial intelligence, intelligent control and intelligent systems. It aims to promote the development of information science and technology.

The central interest of this TI is on mathematical and physical modeling, numerical/analytical study and computation, as well as the experimental investigation of artificial neural networks and their applications in order to understand their operation mechanism, potential application value and future development directions.

From the point of view of theoretical aspects, contributions addressing the following topics are welcome:

- Numerical algorithms and procedures;

- Convolutional Neural Network (CNN);

- Recurrent Neural Network (RNN);

- Snapshot ensembles;

- Dropout;

- Bias correction;

- Cyclical learning rates;

- Neuro-Symbolic Hybrid Intelligent Architecture;

- Radical Basis Neural Networks;

- Neural Network for Adaptive Pattern Recognition;

- Neural Network Learning;

- Recent Advances in Neural Network Applications in Process Control;

- Neural Architectures of Fuzzy Petri Nets.

Application engineers, scientists, and research students from all disciplines with an interest in considering neural networks to solve real-world problems are encouraged to contribute to this Topic Issue.

Dr. Krzysztof Ejsmont
Dr. Aamer Bilal Asghar
Dr. Yong Wang
Dr. Rodolfo Haber
Topic Editors

Keywords

  • artificial neural networks (ANNs)
  • neuro-fuzzy systems
  • adaptive neuro-fuzzy inference system (ANFIS)
  • backpropagation
  • training algorithm
  • MPPT algorithm
  • wavelet
  • support vector machines
  • particle swarm optimization (PSO)
  • adaptive cuckoo search optimization (ACSO)
  • partial shading (PS)
  • hybrid optimization

Participating Journals

Journal Name Impact Factor CiteScore Launched Year First Decision (median) APC
AI
ai
- - 2020 21.8 Days CHF 1200 Submit
Algorithms
algorithms
2.3 3.7 2008 19.1 Days CHF 1600 Submit
Applied Sciences
applsci
2.7 4.5 2011 15.8 Days CHF 2300 Submit
Information
information
3.1 5.8 2010 22.6 Days CHF 1600 Submit
Mathematics
mathematics
2.4 3.5 2013 17.7 Days CHF 2600 Submit

Preprints is a platform dedicated to making early versions of research outputs permanently available and citable. MDPI journals allow posting on preprint servers such as Preprints.org prior to publication. For more details about reprints, please visit https://www.preprints.org.

Published Papers (23 papers)

Order results
Result details
Journals
Select all
Export citation of selected articles as:
Article
Comparative Analysis of Deep Learning Architectures and Vision Transformers for Musical Key Estimation
Information 2023, 14(10), 527; https://doi.org/10.3390/info14100527 - 28 Sep 2023
Viewed by 644
Abstract
The musical key serves as a crucial element in a piece, offering vital insights into the tonal center, harmonic structure, and chord progressions while enabling tasks such as transposition and arrangement. Moreover, accurate key estimation finds practical applications in music recommendation systems and [...] Read more.
The musical key serves as a crucial element in a piece, offering vital insights into the tonal center, harmonic structure, and chord progressions while enabling tasks such as transposition and arrangement. Moreover, accurate key estimation finds practical applications in music recommendation systems and automatic music transcription, making it relevant across academic and industrial domains. This paper presents a comprehensive comparison between standard deep learning architectures and emerging vision transformers, leveraging their success in various domains. We evaluate their performance on a specific subset of the GTZAN dataset, analyzing six different deep learning models. Our results demonstrate that DenseNet, a conventional deep learning architecture, achieves remarkable accuracy of 91.64%, outperforming vision transformers. However, we delve deeper into the analysis to shed light on the temporal characteristics of each deep learning model. Notably, the vision transformer and SWIN transformer exhibit a slight decrease in overall performance (1.82% and 2.29%, respectively), yet they demonstrate superior performance in temporal metrics compared to the DenseNet architecture. The significance of our findings lies in their contribution to the field of musical key estimation, where accurate and efficient algorithms play a pivotal role. By examining the strengths and weaknesses of deep learning architectures and vision transformers, we can gain valuable insights for practical implementations, particularly in music recommendation systems and automatic music transcription. Our research provides a foundation for future advancements and encourages further exploration in this area. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Segmentation Head Networks with Harnessing Self-Attention and Transformer for Insulator Surface Defect Detection
Appl. Sci. 2023, 13(16), 9109; https://doi.org/10.3390/app13169109 - 10 Aug 2023
Viewed by 383
Abstract
Current methodologies for insulator defect detection are hindered by limitations in real-world applicability, spatial constraints, high computational demand, and segmentation challenges. Addressing these shortcomings, this paper presents a robust fast detection algorithm combined segmentation head networks with harnessing self-attention and transformer (HST-Net), which [...] Read more.
Current methodologies for insulator defect detection are hindered by limitations in real-world applicability, spatial constraints, high computational demand, and segmentation challenges. Addressing these shortcomings, this paper presents a robust fast detection algorithm combined segmentation head networks with harnessing self-attention and transformer (HST-Net), which is based on the You Only Look Once (YOLO) v5 to recognize and assess the extent and types of damage on the insulator surface. Firstly, the original backbone network is replaced by the transformer cross-stage partial (Transformer-CSP) networks to enrich the network’s ability by capturing information across different depths of network feature maps. Secondly, an insulator defect segmentation head network is presented to handle the segmentation of defect areas such as insulator losses and flashovers. It facilitates instance-level mask prediction for each insulator object, significantly reducing the influence of intricate backgrounds. Finally, comparative experiment results show that the positioning accuracy and defect segmentation accuracy of the proposed both surpass that of other popular models. It can be concluded that the proposed model not only satisfies the requirements for balance between accuracy and speed in power facility inspection, but also provides fresh perspectives for research in other defect detection domains. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Ensemble System of Deep Neural Networks for Single-Channel Audio Separation
Information 2023, 14(7), 352; https://doi.org/10.3390/info14070352 - 21 Jun 2023
Viewed by 627
Abstract
Speech separation is a well-known problem, especially when there is only one sound mixture available. Estimating the Ideal Binary Mask (IBM) is one solution to this problem. Recent research has focused on the supervised classification approach. The challenge of extracting features from the [...] Read more.
Speech separation is a well-known problem, especially when there is only one sound mixture available. Estimating the Ideal Binary Mask (IBM) is one solution to this problem. Recent research has focused on the supervised classification approach. The challenge of extracting features from the sources is critical for this method. Speech separation has been accomplished by using a variety of feature extraction models. The majority of them, however, are concentrated on a single feature. The complementary nature of various features have not been thoroughly investigated. In this paper, we propose a deep neural network (DNN) ensemble architecture to completely explore the complimentary nature of the diverse features obtained from raw acoustic features. We examined the penultimate discriminative representations instead of employing the features acquired from the output layer. The learned representations were also fused to produce a new features vector, which was then classified by using the Extreme Learning Machine (ELM). In addition, a genetic algorithm (GA) was created to optimize the parameters globally. The results of the experiments showed that our proposed system completely considered various features and produced a high-quality IBM under different conditions. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
NSGA-PINN: A Multi-Objective Optimization Method for Physics-Informed Neural Network Training
Algorithms 2023, 16(4), 194; https://doi.org/10.3390/a16040194 - 03 Apr 2023
Cited by 2 | Viewed by 1687
Abstract
This paper presents NSGA-PINN, a multi-objective optimization framework for the effective training of physics-informed neural networks (PINNs). The proposed framework uses the non-dominated sorting genetic algorithm (NSGA-II) to enable traditional stochastic gradient optimization algorithms (e.g., ADAM) to escape local minima effectively. Additionally, the [...] Read more.
This paper presents NSGA-PINN, a multi-objective optimization framework for the effective training of physics-informed neural networks (PINNs). The proposed framework uses the non-dominated sorting genetic algorithm (NSGA-II) to enable traditional stochastic gradient optimization algorithms (e.g., ADAM) to escape local minima effectively. Additionally, the NSGA-II algorithm enables satisfying the initial and boundary conditions encoded into the loss function during physics-informed training precisely. We demonstrate the effectiveness of our framework by applying NSGA-PINN to several ordinary and partial differential equation problems. In particular, we show that the proposed framework can handle challenging inverse problems with noisy data. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Review
Resilience and Resilient Systems of Artificial Intelligence: Taxonomy, Models and Methods
Algorithms 2023, 16(3), 165; https://doi.org/10.3390/a16030165 - 18 Mar 2023
Cited by 1 | Viewed by 2077
Abstract
Artificial intelligence systems are increasingly being used in industrial applications, security and military contexts, disaster response complexes, policing and justice practices, finance, and healthcare systems. However, disruptions to these systems can have negative impacts on health, mortality, human rights, and asset values. The [...] Read more.
Artificial intelligence systems are increasingly being used in industrial applications, security and military contexts, disaster response complexes, policing and justice practices, finance, and healthcare systems. However, disruptions to these systems can have negative impacts on health, mortality, human rights, and asset values. The protection of such systems from various types of destructive influences is thus a relevant area of research. The vast majority of previously published works are aimed at reducing vulnerability to certain types of disturbances or implementing certain resilience properties. At the same time, the authors either do not consider the concept of resilience as such, or their understanding varies greatly. The aim of this study is to present a systematic approach to analyzing the resilience of artificial intelligence systems, along with an analysis of relevant scientific publications. Our methodology involves the formation of a set of resilience factors, organizing and defining taxonomic and ontological relationships for resilience factors of artificial intelligence systems, and analyzing relevant resilience solutions and challenges. This study analyzes the sources of threats and methods to ensure each resilience properties for artificial intelligence systems. As a result, the potential to create a resilient artificial intelligence system by configuring the architecture and learning scenarios is confirmed. The results can serve as a roadmap for establishing technical requirements for forthcoming artificial intelligence systems, as well as a framework for assessing the resilience of already developed artificial intelligence systems. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
GRNN: Graph-Retraining Neural Network for Semi-Supervised Node Classification
Algorithms 2023, 16(3), 126; https://doi.org/10.3390/a16030126 - 22 Feb 2023
Viewed by 1595
Abstract
In recent years, graph neural networks (GNNs) have played an important role in graph representation learning and have successfully achieved excellent results in semi-supervised classification. However, these GNNs often neglect the global smoothing of the graph because the global smoothing of the graph [...] Read more.
In recent years, graph neural networks (GNNs) have played an important role in graph representation learning and have successfully achieved excellent results in semi-supervised classification. However, these GNNs often neglect the global smoothing of the graph because the global smoothing of the graph is incompatible with node classification. Specifically, a cluster of nodes in the graph often has a small number of other classes of nodes. To address this issue, we propose a graph-retraining neural network (GRNN) model that performs smoothing over the graph by alternating between a learning procedure and an inference procedure, based on the key idea of the expectation-maximum algorithm. Moreover, the global smoothing error is combined with the cross-entropy error to form the loss function of GRNN, which effectively solves the problem. The experiments show that GRNN achieves high accuracy in the standard citation network datasets, including Cora, Citeseer, and PubMed, which proves the effectiveness of GRNN in semi-supervised node classification. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Named Entity Recognition Model Based on Feature Fusion
Information 2023, 14(2), 133; https://doi.org/10.3390/info14020133 - 17 Feb 2023
Cited by 1 | Viewed by 1467
Abstract
Named entity recognition can deeply explore semantic features and enhance the ability of vector representation of text data. This paper proposes a named entity recognition method based on multi-head attention to aim at the problem of fuzzy lexical boundary in Chinese named entity [...] Read more.
Named entity recognition can deeply explore semantic features and enhance the ability of vector representation of text data. This paper proposes a named entity recognition method based on multi-head attention to aim at the problem of fuzzy lexical boundary in Chinese named entity recognition. Firstly, Word2vec is used to extract word vectors, HMM is used to extract boundary vectors, ALBERT is used to extract character vectors, the Feedforward-attention mechanism is used to fuse the three vectors, and then the fused vectors representation is used to remove features by BiLSTM. Then multi-head attention is used to mine the potential word information in the text features. Finally, the text label classification results are output after the conditional random field screening. Through the verification of WeiboNER, MSRA, and CLUENER2020 datasets, the results show that the proposed algorithm can effectively improve the performance of named entity recognition. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Hybrid Backstepping Control of a Quadrotor Using a Radial Basis Function Neural Network
Mathematics 2023, 11(4), 991; https://doi.org/10.3390/math11040991 - 15 Feb 2023
Cited by 1 | Viewed by 1037
Abstract
This article presents a hybrid backstepping consisting of two robust controllers utilizing the approximation property of a radial basis function neural network (RBFNN) for a quadrotor with time-varying uncertainties. The quadrotor dynamic system is decoupled into two subsystems: the position and the attitude [...] Read more.
This article presents a hybrid backstepping consisting of two robust controllers utilizing the approximation property of a radial basis function neural network (RBFNN) for a quadrotor with time-varying uncertainties. The quadrotor dynamic system is decoupled into two subsystems: the position and the attitude subsystems. As part of the position subsystem, adaptive RBFNN backstepping control (ANNBC) is developed to eliminate the effects of uncertainties, trace the quadrotor’s position, and provide the desired roll and pitch angles commands for the attitude subsystem. Then, adaptive RBFNN backstepping is integrated with integral fast terminal sliding mode control (ANNBIFTSMC) to track the required Euler angles and improve robustness against external disturbances. The proposed technique is advantageous because the quadrotor states trace the reference states in a short period of time without requiring knowledge of dynamic uncertainties and external disturbances. In addition, because the controller gains are based on the desired trajectories, adaptive algorithms are used to update them online. The stability of a closed loop system is proved by Lyapunov theory. Numerical simulations show acceptable attitude and position tracking performances. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Outcome Prediction for Patients with Bipolar Disorder Using Prodromal and Onset Data
Appl. Sci. 2023, 13(3), 1552; https://doi.org/10.3390/app13031552 - 25 Jan 2023
Viewed by 973
Abstract
Background: Predicting the outcomes of serious mental illnesses including bipolar disorder (BD) is clinically beneficial, yet difficult. Objectives: This study aimed to predict hospitalization and mortality for patients with incident BD using a deep neural network approach. Methods: We randomly sampled 20,000 US [...] Read more.
Background: Predicting the outcomes of serious mental illnesses including bipolar disorder (BD) is clinically beneficial, yet difficult. Objectives: This study aimed to predict hospitalization and mortality for patients with incident BD using a deep neural network approach. Methods: We randomly sampled 20,000 US Veterans with BD. Data on patients’ prior hospitalizations, diagnoses, procedures, medications, note types, vital signs, lab results, and BD symptoms that occurred within 1 year before and at the onset of the incident BD were extracted as features. We then created novel temporal images of patient clinical features both during the prodromal period and at the time of the disease onset. Using each temporal image as a feature, we trained and tested deep neural network learning models to predict the 1-year combined outcome of hospitalization and mortality. Results: The models achieved accuracies of 0.766–0.949 and AUCs of 0.745–0.806 for the combined outcomes. The AUC for predicting mortality was 0.814, while its highest and lowest values for predicting different types of hospitalization were 90.4% and 70.1%, suggesting that some outcomes were more difficult to predict than others. Conclusion: Deep learning using temporal graphics of clinical history is a new and promising analytical approach for mental health outcome prediction. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Approximate Reasoning for Large-Scale ABox in OWL DL Based on Neural-Symbolic Learning
Mathematics 2023, 11(3), 495; https://doi.org/10.3390/math11030495 - 17 Jan 2023
Viewed by 959
Abstract
The ontology knowledge base (KB) can be divided into two parts: TBox and ABox, where the former models schema-level knowledge within the domain, and the latter is a set of statements of assertions or facts about instances. ABox reasoning is a process of [...] Read more.
The ontology knowledge base (KB) can be divided into two parts: TBox and ABox, where the former models schema-level knowledge within the domain, and the latter is a set of statements of assertions or facts about instances. ABox reasoning is a process of discovering implicit knowledge in ABox based on the existing KB, which is of great value in KB applications. ABox reasoning is influenced by both the complexity of TBox and scale of ABox. The traditional logic-based ontology reasoning methods are usually designed to be provably sound and complete but suffer from long algorithm runtimes and do not scale well for ontology KB represented by OWL DL (Description Logic). In some application scenarios, the soundness and completeness of reasoning results are not the key constraints, and it is acceptable to sacrifice them in exchange for the improvement of reasoning efficiency to some extent. Based on this view, an approximate reasoning method for large-scale ABox in OWL DL KBs was proposed, which is named the ChunfyReasoner (CFR). The CFR introduces neural-symbolic learning into ABox reasoning and integrates the advantages of symbolic systems and neural networks (NNs). By training the NN model, the CFR approximately compiles the logic deduction process of ontology reasoning, which can greatly improve the reasoning speed while ensuring higher reasoning quality. In this paper, we state the basic idea, framework, and construction process of the CFR in detail, and we conduct experiments on two open-source ontologies built on OWL DL. The experimental results verify the effectiveness of our method and show that the CFR can support the applications of large-scale ABox reasoning of OWL DL KBs. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Medical QA Oriented Multi-Task Learning Model for Question Intent Classification and Named Entity Recognition
Information 2022, 13(12), 581; https://doi.org/10.3390/info13120581 - 14 Dec 2022
Cited by 1 | Viewed by 1433
Abstract
Intent classification and named entity recognition of medical questions are two key subtasks of the natural language understanding module in the question answering system. Most existing methods usually treat medical queries intent classification and named entity recognition as two separate tasks, ignoring the [...] Read more.
Intent classification and named entity recognition of medical questions are two key subtasks of the natural language understanding module in the question answering system. Most existing methods usually treat medical queries intent classification and named entity recognition as two separate tasks, ignoring the close relationship between the two tasks. In order to optimize the effect of medical queries intent classification and named entity recognition tasks, a multi-task learning model based on ALBERT-BILSTM is proposed for intent classification and named entity recognition of Chinese online medical questions. The multi-task learning model in this paper makes use of encoder parameter sharing, which enables the model’s underlying network to take into account both named entity recognition and intent classification features. The model learns the shared information between the two tasks while maintaining its unique characteristics during the decoding phase. The ALBERT pre-training language model is used to obtain word vectors containing semantic information and the bidirectional LSTM network is used for training. A comparative experiment of different models was conducted on Chinese medical questions dataset. Experimental results show that the proposed multi-task learning method outperforms the benchmark method in terms of precision, recall and F1 value. Compared with the single-task model, the generalization ability of the model has been improved. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
A Mask-Based Adversarial Defense Scheme
Algorithms 2022, 15(12), 461; https://doi.org/10.3390/a15120461 - 06 Dec 2022
Cited by 1 | Viewed by 1108
Abstract
Adversarial attacks hamper the functionality and accuracy of deep neural networks (DNNs) by meddling with subtle perturbations to their inputs. In this work, we propose a new mask-based adversarial defense scheme (MAD) for DNNs to mitigate the negative effect from adversarial attacks. Our [...] Read more.
Adversarial attacks hamper the functionality and accuracy of deep neural networks (DNNs) by meddling with subtle perturbations to their inputs. In this work, we propose a new mask-based adversarial defense scheme (MAD) for DNNs to mitigate the negative effect from adversarial attacks. Our method preprocesses multiple copies of a potential adversarial image by applying random masking, before the outputs of the DNN on all the randomly masked images are combined. As a result, the combined final output becomes more tolerant to minor perturbations on the original input. Compared with existing adversarial defense techniques, our method does not need any additional denoising structure or any change to a DNN’s architectural design. We have tested this approach on a collection of DNN models for a variety of datasets, and the experimental results confirm that the proposed method can effectively improve the defense abilities of the DNNs against all of the tested adversarial attack methods. In certain scenarios, the DNN models trained with MAD can improve classification accuracy by as much as 90% compared to the original models when given adversarial inputs. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Prediction of Friction Coefficient for Ductile Cast Iron Using Artificial Neural Network Methodology Based on Experimental Investigation
Appl. Sci. 2022, 12(23), 11916; https://doi.org/10.3390/app122311916 - 22 Nov 2022
Cited by 2 | Viewed by 921
Abstract
The key objective of the present study is to analyze the friction coefficient and wear rate for ductile cast iron. Three different microstructures were chosen upon which to perform the experimental tests under different sliding time, load, and sliding speed conditions. These specimens [...] Read more.
The key objective of the present study is to analyze the friction coefficient and wear rate for ductile cast iron. Three different microstructures were chosen upon which to perform the experimental tests under different sliding time, load, and sliding speed conditions. These specimens were perlite + ferrite, ferrite, and bainitic. Moreover, an artificial neural network (ANN) model was developed in order to predict the friction coefficient using a set of data collected during the experiments. The ANN model structure was made up of four input parameters (namely time, load, number, and nodule diameter) and one output parameter (friction coefficient). The Levenberg–Marquardt back-propagation algorithm was applied in the ANN model to train the data using feed-forward back propagation (FFBP). The results of the experiments revealed that the coefficient of friction reduced as the sliding speed increased under a constant load. Additionally, it exhibits the same pattern of action when the test is run with a heavy load and constant sliding speed. Additionally, when the sliding speed increased, the wear rate dropped. The results also show that the bainite structure is harder and wears less quickly than the ferrite structure. Additionally, the results pertaining to the ANN structure showed that a single hidden layer model is more accurate than a double hidden layer model. The highest performance in the validation stage, however, was observed at epochs 8 and 20, respectively, for a double hidden layer and at 0.012346 for a single layer at epoch 20. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Self-Organized Fuzzy Neural Network Nonlinear System Modeling Method Based on Clustering Algorithm
Appl. Sci. 2022, 12(22), 11435; https://doi.org/10.3390/app122211435 - 11 Nov 2022
Cited by 1 | Viewed by 1033
Abstract
In this paper, an improved self-organizing fuzzy neural network (SOFNN-CA) based on a clustering algorithm is proposed for nonlinear systems modeling in industrial processes. In order to reduce training time and increase training speed, we combine offline learning and online identification. The unsupervised [...] Read more.
In this paper, an improved self-organizing fuzzy neural network (SOFNN-CA) based on a clustering algorithm is proposed for nonlinear systems modeling in industrial processes. In order to reduce training time and increase training speed, we combine offline learning and online identification. The unsupervised clustering algorithm is used to generate the initial centers of the network in the offline learning phase, and, in the self-organizing phase of the system, the Mahalanobis distance (MD) index and error criterion are adopted to add neurons to learn new features. A new density potential index (DPI) combined with neuron local field potential (LFP) is designed to adjust the neuron width, which further improves the network generalization. The similarity index calculated by the Gaussian error function is used to merge neurons to reduce redundancy. Meanwhile, the convergence of SOFNN-CA in the case of structural self-organization is demonstrated. Simulations and experiments results show that the proposed SOFNN-CA has a more desirable modeling accuracy and convergence speed compared with SOFNN-ALA and SOFNN-AGA. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
EEG-Based Emotion Recognition Using Convolutional Recurrent Neural Network with Multi-Head Self-Attention
Appl. Sci. 2022, 12(21), 11255; https://doi.org/10.3390/app122111255 - 06 Nov 2022
Cited by 10 | Viewed by 2597
Abstract
In recent years, deep learning has been widely used in emotion recognition, but the models and algorithms in practical applications still have much room for improvement. With the development of graph convolutional neural networks, new ideas for emotional recognition based on EEG have [...] Read more.
In recent years, deep learning has been widely used in emotion recognition, but the models and algorithms in practical applications still have much room for improvement. With the development of graph convolutional neural networks, new ideas for emotional recognition based on EEG have arisen. In this paper, we propose a novel deep learning model-based emotion recognition method. First, the EEG signal is spatially filtered by using the common spatial pattern (CSP), and the filtered signal is converted into a time–frequency map by continuous wavelet transform (CWT). This is used as the input data of the network; then the feature extraction and classification are performed by the deep learning model. We called this model CNN-BiLSTM-MHSA, which consists of a convolutional neural network (CNN), bi-directional long and short-term memory network (BiLSTM), and multi-head self-attention (MHSA). This network is capable of learning the time series and spatial information of EEG emotion signals in depth, smoothing EEG signals and extracting deep features with CNN, learning emotion information of future and past time series with BiLSTM, and improving recognition accuracy with MHSA by reassigning weights to emotion features. Finally, we conducted experiments on the DEAP dataset for sentiment classification, and the experimental results showed that the method has better results than the existing classification. The accuracy of high and low valence, arousal, dominance, and liking state recognition is 98.10%, and the accuracy of four classifications of high and low valence-arousal recognition is 89.33%. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Artificial Neural Network as a Tool for Estimation of the Higher Heating Value of Miscanthus Based on Ultimate Analysis
Mathematics 2022, 10(20), 3732; https://doi.org/10.3390/math10203732 - 11 Oct 2022
Cited by 7 | Viewed by 1407
Abstract
Miscanthus is a perennial energy crop that produces high yields and has the potential to be converted into energy. The ultimate analysis determines the composition of the biomass and the energy value in terms of the higher heating value (HHV), which is the [...] Read more.
Miscanthus is a perennial energy crop that produces high yields and has the potential to be converted into energy. The ultimate analysis determines the composition of the biomass and the energy value in terms of the higher heating value (HHV), which is the most important parameter in determining the quality of the fuel. In this study, an artificial neural network (ANN) model based on the principle of supervised learning was developed to predict the HHV of miscanthus biomass. The developed ANN model was compared with the models of predictive regression models (suggested from the literature) and the accuracy of the developed model was determined by the coefficient of determination. The paper presents data from 192 miscanthus biomass samples based on ultimate analysis and HHV. The developed model showed good properties and the possibility of prediction with high accuracy (R2 = 0.77). The paper proves the possibility of using ANN models in practical application in determining fuel properties of biomass energy crops and greater accuracy in predicting HHV than the regression models offered in the literature. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Design of Optical Tweezers Manipulation Control System Based on Novel Self-Organizing Fuzzy Cerebellar Model Neural Network
Appl. Sci. 2022, 12(19), 9655; https://doi.org/10.3390/app12199655 - 26 Sep 2022
Cited by 1 | Viewed by 963
Abstract
Holographic optical tweezers have unique non-physical contact and can manipulate and control single or multiple cells in a non-invasive way. In this paper, the dynamics model of the cells captured by the optical trap is analyzed, and a control system based on a [...] Read more.
Holographic optical tweezers have unique non-physical contact and can manipulate and control single or multiple cells in a non-invasive way. In this paper, the dynamics model of the cells captured by the optical trap is analyzed, and a control system based on a novel self-organizing fuzzy cerebellar model neural network (NSOFCMNN) is proposed and applied to the cell manipulation control of holographic optical tweezers. This control system consists of a main controller using the NSOFCMNN with a new self-organization mechanism, a robust compensation controller, and a higher order sliding mode. It can accurately move the captured cells to the expected position through the optical trap generated by the holographic optical tweezers system. Both the layers and blocks of the proposed NSOFCMNN can be adjusted online according to the new self-organization mechanism. The compensation controller is used to eliminate the approximation errors. The higher order sliding surface can enhance the performance of controllers. The distances between cells are considered in order to further realize multi-cell cooperative control. In addition, the stability and convergence of the proposed NSOFCMNN are proved by the Lyapunov function, and the learning law is updated online by the gradient descent method. The simulation results show that the control system based on the proposed NSOFCMNN can effectively complete the cell manipulation task of optical tweezers and has better control performance than other neural network controllers. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Convolutional Neural Network Model Compression Method for Software—Hardware Co-Design
Information 2022, 13(10), 451; https://doi.org/10.3390/info13100451 - 26 Sep 2022
Cited by 1 | Viewed by 2274
Abstract
Owing to their high accuracy, deep convolutional neural networks (CNNs) are extensively used. However, they are characterized by high complexity. Real-time performance and acceleration are required in current CNN systems. A graphics processing unit (GPU) is one possible solution to improve real-time performance; [...] Read more.
Owing to their high accuracy, deep convolutional neural networks (CNNs) are extensively used. However, they are characterized by high complexity. Real-time performance and acceleration are required in current CNN systems. A graphics processing unit (GPU) is one possible solution to improve real-time performance; however, its power consumption ratio is poor owing to high power consumption. By contrast, field-programmable gate arrays (FPGAs) have lower power consumption and flexible architecture, making them more suitable for CNN implementation. In this study, we propose a method that offers both the speed of CNNs and the power and parallelism of FPGAs. This solution relies on two primary acceleration techniques—parallel processing of layer resources and pipelining within specific layers. Moreover, a new method is introduced for exchanging domain requirements for speed and design time by implementing an automatic parallel hardware–software co-design CNN using the software-defined system-on-chip tool. We evaluated the proposed method using five networks—MobileNetV1, ShuffleNetV2, SqueezeNet, ResNet-50, and VGG-16—and FPGA processors—ZCU102. We experimentally demonstrated that our design has a higher speed-up than the conventional implementation method. The proposed method achieves 2.47×, 1.93×, and 2.16× speed-up on the ZCU102 for MobileNetV1, ShuffleNetV2, and SqueezeNet, respectively. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Training of an Extreme Learning Machine Autoencoder Based on an Iterative Shrinkage-Thresholding Optimization Algorithm
Appl. Sci. 2022, 12(18), 9021; https://doi.org/10.3390/app12189021 - 08 Sep 2022
Cited by 2 | Viewed by 1880
Abstract
Orthogonal transformations, proper decomposition, and the Moore–Penrose inverse are traditional methods of obtaining the output layer weights for an extreme learning machine autoencoder. However, an increase in the number of hidden neurons causes higher convergence times and computational complexity, whereas the generalization capability [...] Read more.
Orthogonal transformations, proper decomposition, and the Moore–Penrose inverse are traditional methods of obtaining the output layer weights for an extreme learning machine autoencoder. However, an increase in the number of hidden neurons causes higher convergence times and computational complexity, whereas the generalization capability is low when the number of neurons is small. One way to address this issue is to use the fast iterative shrinkage-thresholding algorithm (FISTA) to minimize the output weights of the extreme learning machine. In this work, we aim to improve the convergence speed of FISTA by using two fast algorithms of the shrinkage-thresholding class, called greedy FISTA (G-FISTA) and linearly convergent FISTA (LC-FISTA). Our method is an exciting proposal for decision-making involving the resolution of many application problems, especially those requiring longer computational times. In our experiments, we adopt six public datasets that are frequently used in machine learning: MNIST, NORB, CIFAR10, UMist, Caltech256, and Stanford Cars. We apply several metrics to evaluate the performance of our method, and the object of comparison is the FISTA algorithm due to its popularity for neural network training. The experimental results show that G-FISTA and LC-FISTA achieve higher convergence speeds in the autoencoder training process; for example, in the Stanford Cars dataset, G-FISTA and LC-FISTA are faster than FISTA by 48.42% and 47.32%, respectively. Overall, all three algorithms maintain good values of the performance metrics on all databases. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Fixed-Time Convergent Gradient Neural Network for Solving Online Sylvester Equation
Mathematics 2022, 10(17), 3090; https://doi.org/10.3390/math10173090 - 28 Aug 2022
Cited by 2 | Viewed by 1069
Abstract
This paper aims at finding a fixed-time solution to the Sylvester equation by using a gradient neural network (GNN). To reach this goal, a modified sign-bi-power (msbp) function is presented and applied on a linear GNN as an activation function. Accordingly, a fixed-time [...] Read more.
This paper aims at finding a fixed-time solution to the Sylvester equation by using a gradient neural network (GNN). To reach this goal, a modified sign-bi-power (msbp) function is presented and applied on a linear GNN as an activation function. Accordingly, a fixed-time convergent GNN (FTC-GNN) model is developed for solving the Sylvester equation. The upper bound of the convergence time of such an FTC-GNN model can be predetermined if parameters are given regardless of the initial conditions. This point is corroborated by a detailed theoretical analysis. In addition, the convergence time is also estimated utilizing the Lyapunov stability theory. Two examples are then simulated to demonstrate the validation of the theoretical analysis, as well as the superior convergence performance of the presented FTC-GNN model as compared to the existing GNN models. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
OptiNET—Automatic Network Topology Optimization
Information 2022, 13(9), 405; https://doi.org/10.3390/info13090405 - 27 Aug 2022
Viewed by 1532
Abstract
The recent boom of artificial Neural Networks (NN) has shown that NN can provide viable solutions to a variety of problems. However, their complexity and the lack of efficient interpretation of NN architectures (commonly considered black box techniques) has adverse effects on the [...] Read more.
The recent boom of artificial Neural Networks (NN) has shown that NN can provide viable solutions to a variety of problems. However, their complexity and the lack of efficient interpretation of NN architectures (commonly considered black box techniques) has adverse effects on the optimization of each NN architecture. One cannot simply use a generic topology and have the best performance in every application field, since the network topology is commonly fine-tuned to the problem/dataset in question. In this paper, we introduce a novel method of computationally assessing the complexity of the dataset. The NN is treated as an information channel, and thus information theory is used to estimate the optimal number of neurons for each layer, reducing the memory and computational load, while achieving the same, if not greater, accuracy. Experiments using common datasets confirm the theoretical findings, and the derived algorithm seems to improve the performance of the original architecture. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Almost Anti-Periodic Oscillation Excited by External Inputs and Synchronization of Clifford-Valued Recurrent Neural Networks
Mathematics 2022, 10(15), 2764; https://doi.org/10.3390/math10152764 - 04 Aug 2022
Cited by 2 | Viewed by 1092
Abstract
The main purpose of this paper was to study the almost anti-periodic oscillation caused by external inputs and the global exponential synchronization of Clifford-valued recurrent neural networks with mixed delays. Since the space consists of almost anti-periodic functions has no vector space structure, [...] Read more.
The main purpose of this paper was to study the almost anti-periodic oscillation caused by external inputs and the global exponential synchronization of Clifford-valued recurrent neural networks with mixed delays. Since the space consists of almost anti-periodic functions has no vector space structure, firstly, we prove that the network under consideration possesses a unique bounded continuous solution by using the contraction fixed point theorem. Then, by using the inequality technique, it was proved that the unique bounded continuous solution is also an almost anti-periodic solution. Secondly, taking the neural network that was considered as the driving system, introducing the corresponding response system and designing the appropriate controller, some sufficient conditions for the global exponential synchronization of the driving-response system were obtained by employing the inequality technique. When the system we consider degenerated into a real-valued system, our results were considered new. Finally, the validity of the results was verified using a numerical example. Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Article
Sequential Normalization: Embracing Smaller Sample Sizes for Normalization
Information 2022, 13(7), 337; https://doi.org/10.3390/info13070337 - 12 Jul 2022
Viewed by 1427
Abstract
Normalization as a layer within neural networks has over the years demonstrated its effectiveness in neural network optimization across a wide range of different tasks, with one of the most successful approaches being that of batch normalization. The consensus is that better estimates [...] Read more.
Normalization as a layer within neural networks has over the years demonstrated its effectiveness in neural network optimization across a wide range of different tasks, with one of the most successful approaches being that of batch normalization. The consensus is that better estimates of the BatchNorm normalization statistics (μ and σ2) in each mini-batch result in better optimization. In this work, we challenge this belief and experiment with a new variant of BatchNorm known as GhostNorm that, despite independently normalizing batches within the mini-batches, i.e., μ and σ2 are independently computed and applied to groups of samples in each mini-batch, outperforms BatchNorm consistently. Next, we introduce sequential normalization (SeqNorm), the sequential application of the above type of normalization across two dimensions of the input, and find that models trained with SeqNorm consistently outperform models trained with BatchNorm or GhostNorm on multiple image classification data sets. Our contributions are as follows: (i) we uncover a source of regularization that is unique to GhostNorm, and not simply an extension from BatchNorm, and illustrate its effects on the loss landscape, (ii) we introduce sequential normalization (SeqNorm) a new normalization layer that improves the regularization effects of GhostNorm, (iii) we compare both GhostNorm and SeqNorm against BatchNorm alone as well as with other regularization techniques, (iv) for both GhostNorm and SeqNorm models, we train models whose performance is consistently better than our baselines, including ones with BatchNorm, on the standard image classification data sets of CIFAR–10, CIFAR-100, and ImageNet ((+0.2%, +0.7%, +0.4%), and (+0.3%, +1.7%, +1.1%) for GhostNorm and SeqNorm, respectively). Full article
(This article belongs to the Topic Advances in Artificial Neural Networks)
Show Figures

Figure 1

Back to TopTop