Novel Methods for Object Detection and Segmentation

A special issue of Electronics (ISSN 2079-9292). This special issue belongs to the section "Computer Science & Engineering".

Deadline for manuscript submissions: 16 August 2024 | Viewed by 5333

Special Issue Editors


E-Mail Website
Guest Editor
Department of Information and Communication Engineering, and Convergence Engineering for Intelligent Drone, Sejong University, Seoul 05006, Republic of Korea
Interests: deep learning; object detection; NLP; pattern recognition; computer vision
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Department of Computer Science and Engineering, Sejong University, Seoul, Republic of Korea
Interests: big data; computer vision; pattern recognition; biometrics; deep learning
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

Object detection and segmentation are critical tasks in computer vision with a wide range of applications, including autonomous driving, robotics, and medical imaging. In recent years, deep learning-based methods have achieved remarkable success in these areas. However, challenges remain, such as handling occlusions, low-resolution images, and diverse object shapes and sizes.

This Special Issue presents novel methods for object detection and segmentation that address these challenges. The articles cover a variety of topics, including the use of attention mechanisms, multi-scale feature fusion, and generative adversarial networks (GANs) for object detection and segmentation. Other articles explore the integration of 3D information, such as point clouds and depth maps, into object detection and segmentation frameworks.

Overall, the articles in this Special Issue offer new insights and approaches for object detection and segmentation, with potential implications for a wide range of industries and fields.

Dr. Lien Minh Dang
Prof. Dr. Hyeonjoon Moon
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Electronics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • object detection
  • object segmentation
  • deep learning
  • computer vision
  • attention mechanisms
  • multi-scale feature fusion
  • generative adversarial networks (GANs)
  • 3D information
  • point clouds
  • depth maps
  • autonomous driving
  • robotics
  • medical imaging
  • image processing

Published Papers (5 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

16 pages, 9203 KiB  
Article
Integration of ShuffleNet V2 and YOLOv5s Networks for a Lightweight Object Detection Model of Electric Bikes within Elevators
by Jingfang Su, Minrui Yang and Xinliang Tang
Electronics 2024, 13(2), 394; https://doi.org/10.3390/electronics13020394 - 18 Jan 2024
Viewed by 961
Abstract
The entry of electric bikes into elevators poses safety risks. This article proposes a lightweight object detection model for edge deployment in elevator environments specifically designed for electric bikes. Based on the YOLOv5s network, the backbone network replaces the original CSPDarknet53 with a [...] Read more.
The entry of electric bikes into elevators poses safety risks. This article proposes a lightweight object detection model for edge deployment in elevator environments specifically designed for electric bikes. Based on the YOLOv5s network, the backbone network replaces the original CSPDarknet53 with a lightweight multilayer ShuffleNet V2 convolutional neural network, achieving a lightweight backbone network. Swin Transformer modules are introduced between layers to enhance the feature expression capability of images, and a SimAM attention mechanism is applied at the end layer to further improve the feature extraction capability of the backbone network. In the neck network, lightweight and depth-balanced GSConv and VoV-GSCSP modules replace several Conv and C3 basic convolutional modules, reducing the parameter count while enhancing the cross-scale connection and fusion capabilities of feature maps. The prediction network uses the faster-converging and more accurate EIOU error function as the position loss function for iterative training. This article conducts various lightweighting comparison experiments and ablation experiments on the improved object detection model. The experimental results demonstrate that the proposed object detection model, with a model size of only 2.6 megabytes and 1.1 million parameters, achieves a frame rate of 106 frames per second and a detection accuracy of 95.5%. This represents an 84.8% reduction in computational load compared to the original YOLOv5s model. The model’s volume and parameter count are reduced by 81.0% and 84.3%, respectively, with only a 0.9% decrease in mAP. The improved object detection model proposed in this paper can meet the real-time detection requirements for electric bikes in elevator scenarios, providing a feasible technical solution for its deployment on edge devices within elevators. Full article
(This article belongs to the Special Issue Novel Methods for Object Detection and Segmentation)
Show Figures

Figure 1

14 pages, 8632 KiB  
Article
Applying Advanced Lightweight Architecture DSGSE-Yolov5 to Rapid Chip Contour Detection
by Bao Rong Chang, Hsiu-Fen Tsai and Fu-Yang Chang
Electronics 2024, 13(1), 10; https://doi.org/10.3390/electronics13010010 - 19 Dec 2023
Viewed by 1081
Abstract
Chip contour detection aims to detect damaged chips in chip slots during IC packaging and testing using vision facilities. However, the operation speed of the new chip transportation machine is too fast, and the current chip contour detection models, such as Yolov5, M3-Yolov5, [...] Read more.
Chip contour detection aims to detect damaged chips in chip slots during IC packaging and testing using vision facilities. However, the operation speed of the new chip transportation machine is too fast, and the current chip contour detection models, such as Yolov5, M3-Yolov5, FGHSE-Yolov5, and GSEH-Yolov5, running on the embedded platform, Jetson Nano, cannot detect chip contours in a timely manner. Therefore, there must be a rapid response for chip contour detection. This paper introduces the DSGSE-Yolov5s algorithm, which can accelerate object detection and image recognition to resolve this problem. Additionally, this study makes a performance comparison between the different models. Compared with the traditional model Yolov5, the proposed DSGSE-Yolov5s algorithm can significantly promote the speed of object detection by 132.17% and slightly increase the precision by 0.85%. As a result, the proposed approach can outperform the other methods. Full article
(This article belongs to the Special Issue Novel Methods for Object Detection and Segmentation)
Show Figures

Figure 1

15 pages, 3654 KiB  
Article
Innovative Cucumber Phenotyping: A Smartphone-Based and Data-Labeling-Free Model
by Le Quan Nguyen, Jihye Shin, Sanghuyn Ryu, L. Minh Dang, Han Yong Park, O New Lee and Hyeonjoon Moon
Electronics 2023, 12(23), 4775; https://doi.org/10.3390/electronics12234775 - 25 Nov 2023
Viewed by 916
Abstract
Sustaining global food security amid a growing world population demands advanced breeding methods. Phenotyping, which observes and measures physical traits, is a vital component of agricultural research. However, its labor-intensive nature has long hindered progress. In response, we present an efficient phenotyping platform [...] Read more.
Sustaining global food security amid a growing world population demands advanced breeding methods. Phenotyping, which observes and measures physical traits, is a vital component of agricultural research. However, its labor-intensive nature has long hindered progress. In response, we present an efficient phenotyping platform tailored specifically for cucumbers, harnessing smartphone cameras for both cost-effectiveness and accessibility. We employ state-of-the-art computer vision models for zero-shot cucumber phenotyping and introduce a B-spline curve as a medial axis to enhance measurement accuracy. Our proposed method excels in predicting sample lengths, achieving an impressive mean absolute percentage error (MAPE) of 2.20%, without the need for extensive data labeling or model training. Full article
(This article belongs to the Special Issue Novel Methods for Object Detection and Segmentation)
Show Figures

Figure 1

18 pages, 7701 KiB  
Article
Boosting the Response of Object Detection and Steering Angle Prediction for Self-Driving Control
by Bao Rong Chang, Hsiu-Fen Tsai and Fu-Yang Chang
Electronics 2023, 12(20), 4281; https://doi.org/10.3390/electronics12204281 - 16 Oct 2023
Cited by 1 | Viewed by 740
Abstract
Our previous work introduced the LW-YOLOv4-tiny and the LW-ResNet18 models by replacing traditional convolution with the Ghost Conv to achieve rapid object detection and steering angle prediction, respectively. However, the entire object detection and steering angle prediction process has encountered a speed limit [...] Read more.
Our previous work introduced the LW-YOLOv4-tiny and the LW-ResNet18 models by replacing traditional convolution with the Ghost Conv to achieve rapid object detection and steering angle prediction, respectively. However, the entire object detection and steering angle prediction process has encountered a speed limit problem. Therefore, this study aims to significantly speed up the object detection and the steering angle prediction simultaneously. This paper proposes the GhostBottleneck approach to speed the frame rate of feature extraction and add the SElayer method to maintain the existing precision of object detection, which constructs an enhanced object detection model abbreviated as LWGSE-YOLOv4-tiny. In addition, this paper also conducted depthwise separable convolution to simplify the Ghost Conv as depthwise separable and ghost convolution, which constructs an improved steering angle prediction model abbreviated as LWDSG-ResNet18 that can considerably speed up the prediction and slightly increase image recognition accuracy. Compared with our previous work, the proposed approach shows that the GhostBottleneck module can significantly boost the frame rate of feature extraction by 9.98%, and SElayer can upgrade the precision of object detection slightly by 0.41%. Moreover, depthwise separable and ghost convolution can considerably boost prediction speed by 20.55% and increase image recognition accuracy by 2.05%. Full article
(This article belongs to the Special Issue Novel Methods for Object Detection and Segmentation)
Show Figures

Figure 1

14 pages, 7299 KiB  
Article
Real Pseudo-Lidar Point Cloud Fusion for 3D Object Detection
by Xiangsuo Fan, Dachuan Xiao, Dengsheng Cai and Wentao Ding
Electronics 2023, 12(18), 3920; https://doi.org/10.3390/electronics12183920 - 18 Sep 2023
Viewed by 1061
Abstract
Three-dimensional object detection technology is an essential component of autonomous driving systems. Existing 3D object detection techniques heavily rely on expensive lidar sensors, leading to increased costs. Recently, the emergence of Pseudo-Lidar point cloud data has addressed this cost issue. However, the current [...] Read more.
Three-dimensional object detection technology is an essential component of autonomous driving systems. Existing 3D object detection techniques heavily rely on expensive lidar sensors, leading to increased costs. Recently, the emergence of Pseudo-Lidar point cloud data has addressed this cost issue. However, the current methods for generating Pseudo-Lidar point clouds are relatively crude, resulting in suboptimal detection performance. This paper proposes an improved method to generate more accurate Pseudo-Lidar point clouds. The method first enhances the stereo-matching network to improve the accuracy of Pseudo-Lidar point cloud representation. Secondly, it fuses 16-Line real lidar point cloud data to obtain more precise Real Pseudo-Lidar point cloud data. Our method achieves impressive results in the popular KITTI benchmark. Our algorithm achieves an object detection accuracy of 85.5% within a range of 30 m. Additionally, the detection accuracies for pedestrians and cyclists reach 68.6% and 61.6%, respectively. Full article
(This article belongs to the Special Issue Novel Methods for Object Detection and Segmentation)
Show Figures

Figure 1

Back to TopTop