Topical Collection "Deep Vision Algorithms and Applications"

A topical collection in Applied Sciences (ISSN 2076-3417). This collection belongs to the section "Computing and Artificial Intelligence".

Viewed by 4864

Editors

Department of IT Engineering, Sookmyung Women’s University, Seoul 04310, Korea
Interests: image/video signal processing; pattern recognition; computer vision; deep learning; artificial intelligence
Special Issues, Collections and Topics in MDPI journals
Department of Computer Science and Engineering, Indian Institute of Technology Roorkee, Uttarakhand 247667, India
Interests: computer vision; deep learning; pattern recognition

Topical Collection Information

Dear Colleagues,

We are pleased to invite you to contribute on the topic of intelligent vision systems and their various applications. To achieve an intelligent vision system, dealing with sophisticated artificial intelligence (AI) technology is an inevitability. In AI technology, the core technology of deep learning has been based on deep neural networks and their training mechanisms. Through deep leaning-based approaches, many problems have been solved regarding various applications, especially in the field of classification and recognition. Despite the rapid development of deep learning approaches, further progress in visual feature extraction and pattern mining remain important for accelerating this kind of development.

This Topical Collection aims to publish scientific papers on various big vision data analysis and vision data-based artificial intelligence technologies. The scope includes various topics from visual feature extraction, visual pattern recognition, deep learning structure, and learning methods to the results of their application in various fields.

For this Topical Collection, both original research articles and reviews are welcome. Specific areas of research interest include (but are not limited to) the following:

  • Big vision data analysis
  • Vision data mining and pattern recognition
  • Visual feature extraction
  • Intelligent vision systems
  • Deep learning structures and optimization mechanisms
  • Lightweight deep vision structures
  • Various applications of deep vision schemes
  • Novel data fusion schemes with vision data
  • Vision sensor network and distributed processing
  • Deep neural networks for multimedia data processing
  • Human-centric vision systems and technologies

We look forward to receiving your valuable contributions.

Prof. Dr. Byung-Gyu Kim
Prof. Dr. Partha Pratim Roy
Collection Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the collection website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Applied Sciences is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2300 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • big vision data analysis
  • deep vision algorithm
  • visual data mining
  • pattern recognition
  • intelligent media
  • visual feature extraction

Published Papers (2 papers)

2022

Article
Predictive Distillation Method of Anchor-Free Object Detection Model for Continual Learning
Appl. Sci. 2022, 12(13), 6419; https://doi.org/10.3390/app12136419 - 24 Jun 2022
Viewed by 806
Abstract
Continual learning (CL) is becoming increasingly important, not only for storage space because of the ever-increasing amount of data being generated, but also for associated copyright problems. In this study, we propose ground truth’ (GT’), which is a combination of ground truth (GT) [...] Read more.
Continual learning (CL) is becoming increasingly important, not only for storage space because of the ever-increasing amount of data being generated, but also for associated copyright problems. In this study, we propose ground truth’ (GT’), which is a combination of ground truth (GT) and a prediction of the teacher model that distills the prediction results of the previously trained model, called the teacher model, by applying the knowledge distillation (KD) technique to an anchor-free object detection model. Among all the objects predicted by the teacher model, an object for which the prediction score is higher than the threshold value is distilled into the current trained model, called the student model. To avoid interference with new class learning, the IoU is obtained between every object of the GT and the predicted objects. Through the continual learning scenario, even if the reuse of past data is limited, if new data are sufficient, the proposed model minimizes catastrophic forgetting problems and enables learning for newly added classes. The proposed model was learned in PascalVOC 2007 + 2012 and tested in PascalVOC2007, with better results of 9.6% p mAP and 13.7% p F1i shown in the scenario 19 + 1. The result in scenario 15 + 5 showed better results than the compared algorithm, with 1.6% p mAP and 0.9% p F1i. The scenario 10 + 10 also outperformed the other alternatives, with 0.9% p mAP and 0.6% p F1i. Full article
Show Figures

Figure 1

Article
A Framework for Pedestrian Attribute Recognition Using Deep Learning
Appl. Sci. 2022, 12(2), 622; https://doi.org/10.3390/app12020622 - 10 Jan 2022
Cited by 4 | Viewed by 2226
Abstract
The pedestrian attribute recognition task is becoming more popular daily because of its significant role in surveillance scenarios. As the technological advances are significantly more than before, deep learning came to the surface of computer vision. Previous works applied deep learning in different [...] Read more.
The pedestrian attribute recognition task is becoming more popular daily because of its significant role in surveillance scenarios. As the technological advances are significantly more than before, deep learning came to the surface of computer vision. Previous works applied deep learning in different ways to recognize pedestrian attributes. The results are satisfactory, but still, there is some scope for improvement. The transfer learning technique is becoming more popular for its extraordinary performance in reducing computation cost and scarcity of data in any task. This paper proposes a framework that can work in surveillance scenarios to recognize pedestrian attributes. The mask R-CNN object detector extracts the pedestrians. Additionally, we applied transfer learning techniques on different CNN architectures, i.e., Inception ResNet v2, Xception, ResNet 101 v2, ResNet 152 v2. The main contribution of this paper is fine-tuning the ResNet 152 v2 architecture, which is performed by freezing layers, last 4, 8, 12, 14, 20, none, and all. Moreover, data balancing techniques are applied, i.e., oversampling, to resolve the class imbalance problem of the dataset and analysis of the usefulness of this technique is discussed in this paper. Our proposed framework outperforms state-of-the-art methods, and it provides 93.41% mA and 89.24% mA on the RAP v2 and PARSE100K datasets, respectively. Full article
Show Figures

Figure 1

Back to TopTop