Next Article in Journal
Pulsed Current Constructs 3DM Cu/ZnO Current Collector Composite Anode for Free-Dendritic Lithium Metal Batteries
Next Article in Special Issue
Material and Waste Flow Analysis for Environmental and Economic Impact Assessment of Inorganic Acid Leaching Routes for Spent Lithium Batteries’ Cathode Scraps
Previous Article in Journal
All-Solid-State Thin Film Li-Ion Batteries: New Challenges, New Materials, and New Designs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Accurate Activate Screw Detection Method for Automatic Electric Vehicle Battery Disassembly

1
School of Mechanical Engineering, Shanghai Jiaotong University, No. 800, Dongchuan Road, Shanghai 200240, China
2
School of Computer and Information Engineering, Central South University of Forestry Technology, No. 498, Shaoshan South Road, Changsha 410004, China
3
Vision and AI Lab, Intel Labs China, the 8F, Raycom InfoTech Park Tower A, No. 2, Kexueyuan Nanlu, Beijing 100190, China
4
GEM (Wuhan) Industrial Park, Yangluo Economic Development Zone, Wuhan 430413, China
*
Author to whom correspondence should be addressed.
Batteries 2023, 9(3), 187; https://doi.org/10.3390/batteries9030187
Submission received: 19 January 2023 / Revised: 4 March 2023 / Accepted: 14 March 2023 / Published: 21 March 2023
(This article belongs to the Special Issue Recycling of Lithium-Ion Batteries: Current Status and Future Outlook)

Abstract

:
With the increasing popularity of electric vehicles, the number of end-of-life (EOF) electric vehicle batteries (EVBs) is also increasing day by day. Efficient dismantling and recycling of EVBs are essential to ensure environmental protection. There are many types of EVBs with complex structures, and the current automatic dismantling line is immature and lacks corresponding dismantling equipment. This makes it difficult for some small parts to be disassembled precisely. Screws are used extensively in batteries to fix or connect modules in EVBs. However, due to the small size of screws and differences in installation angles, screw detection is a very challenging task and a significant obstacle to automatic EVBs disassembly. This research proposes a systematic method to complete screw detection called “Active Screw Detection”. The experimental results show that with the YOLOX-s model, the improved YOLOX model achieves  95.92 % and  92.14 % accuracy for both  m A P 50 and  m A P 75 positioning after autonomous adjustment of the robotic arm attitude. Compared to the method without autonomous adjustment of the robotic arm,  m A P 50 and  m A P 75 improved by  62.81 % and  57.67 % , respectively. In addition, the improved YOLOX model improves  m A P 50 and  m A P 75 by  0.19 % and  3.59 % , respectively, compared to the original YOLOX model.

1. Introduction

With the rapid growth of EVBs [1], the EVB industry has developed rapidly in recent years [2]. EOF batteries contain a large amount of renewable resources and have gathered considerable attention [3]. The recycling [4] and dismantling [5] of EOF EVBs have become a hot topic. EVBs, a class of electronic waste (E-Waste), contain heavy metals and other harmful substances, which can contaminate the soil and water, thus damaging the environment and affecting human health [6,7,8,9]. The safety of the battery itself also needs to be considered [10]. At the same time, in recycling EOF EVBs, the control circuit, thermal management system, and other standardized parts connected to the battery can be recycled intact [11]. The electrode materials of battery modules contain valuable elements, such as Li, Co, Ni, and Mn [12]. The current mainstream recovery method uses wet metallurgy to separate these elements and make various pure oxides or carbonates [13]. Therefore, in terms of environmental protection and circular economy, recycling EOF electric vehicle batteries is necessary. Due to the wide variety of EOF EVBs and complex structures, the task of recycling and disassembling EVBs is currently mainly manual [14,15]. This is inefficient and exposes workers to dangerous working conditions [16], making it difficult to guarantee the safety of dismantled power batteries [17]. Compared to the manual disassembly of the screws using human hands, an integrated robotic arm disassembly system may save a lot of costs and avoid some safety accidents. Most of the current research on intelligent dismantling of EVBs has focused on the recycling of individual batteries, with few studies mentioning the process of dismantling and recycling parts in the form of bags/stacks. This is because of the problems of non-uniform standardization, different battery pack sizes, and different embedded battery specifications in the intelligent disassembly task [18]. So there is an urgent need to automate intelligent disassembly. Screws, the most numerous fasteners in electric vehicle battery cells, have been a challenge to locate and classify in intelligent disassembly tasks accurately. Unlike automated production, the location of screws on electric vehicle battery cells is uncertain, and the screws must be identified and positioned on site. The lack of identification accuracy restricts the use and popularity of automated disassembly assembly lines. So it is necessary to design an algorithm for the accurate location and classification of screws.
The screw detection task aims to detect the screws in the sensory field while obtaining information about the category of the screw and the accurate position for the subsequent disassembly task. In recent years, artificial intelligence technology has been developing rapidly in the field of object detection [19,20,21,22,23], and it only needs some annotated images for training the model to achieve high target localization accuracy [6,24,25,26]. However, the result of the direct application of these methods of screw detection is not as good as expected. For example: using the global camera to obtain RGB images and complete the screw detection task will result in many false and missed detections. First of all, the screws are installed at very different angles due to the complex structure of the battery pack. Moreover, the features of screws significantly change when observing them from different angles. Even worse, if the image capture position is inappropriate, some screws will be occluded and cannot be detected. So it is difficult to find a suitable global location, capture one image, and accurately identify all the screws in the battery pack. Few researchers have focused on this issue. On the other hand, the screw itself is a small object compared to the other parts in an EVB. The pixel area of small objects is small, and feature information is easily lost in common feature extraction operations [27,28,29]. This increases the difficulty of the screw detection task.
Most of the mechanically related tasks rely on visual information [30]. The role of vision sensors is similar to that of human eyes and is the main source of external information for robots. In tasks related to screw detection and localization based on visual information, most researchers have envisioned experiments in an ideal situation, paying little attention to the problem of increased difficulty in screw detection due to changes in the field of perception of the visual sensor during continuous disassembly. Throughout the process of screw detection, Yildiz, E. et al. [24] assumed that the camera had a vertical surface to the surface of the electronic device to be inspected, with the height set to 60 cm. This scenario is too idealistic and difficult to implement in real production because some screws are located in a plane that is not parallel to the horizontal plane or on the side of the battery pack. It is obvious that real production lines would prefer to develop a coherent screw positioning method that can be actively adjusted.
There are currently two main approaches to screw detection tasks based on visual information. One is based on a pre-programmed robot approach, where a predefined scan of an electric vehicle battery cell is performed to obtain a point cloud model of the pack. Then, operations such as feature matching are performed to obtain information on the position and size of the screws. L. Jurjević et al. [31] used the Two Xiaomi Yi camera to obtain a 3D point cloud of the object to construct a 3D model, but the model was not sufficiently detailed. Adjigble, M. et al. [32] obtained a point cloud of the object to be disassembled by scanning, calculated the zero moment displacement features on the point cloud, and extracting the same features on the point cloud of the fixture model, which were then used together to calculate the local shape between the object and the gripper similarity metric. The main idea is to find the point that maximizes the contact surface and to use only the area of the object that matches the surface curvature of the gripper fingers for gripping. Finally, a feasibility analysis was performed to select the subset of point pairs that returned from previous actions and were kinematically feasible for the gripper. Li, R. et al. [33] used a vision system to extract the approximate position and centre line of the screw in the part from an applicable CAD model to locate the screw. This type of method is currently mostly used on industrial production lines. However, this approach does not adapt to the disassembly strategy to the actual condition of the battery pack. The researchers mentioned that the method can be used only under ideal conditions. In practice, the screws themselves may fail (screw stuck, screw head damaged, etc.), which can affect the effectiveness of the method. In addition, the appearance of damaged cells or new structures of cells can affect the accuracy of the disassembly task. Another approach is to use machine learning methods to detect and locate screws based on 2D images captured by visual sensing. Gil, P. et al. [34] used the Canny operator to extract contours from images with screws and then used feature matching to classify and locate components. Cha, Y. J. et al. [35] performed Circle Hough Transform (CHT) after cropping the images to locate the circles on screw heads and washers in binarized images. Park, J. H. et al. [36] used the Canny operator to detect edges in images and then used Hough Transform (HT) and CHT to find lines and circle lines in images, respectively, to locate screws. The Canny operator was used to detect the edges in the image, and then HT and CHT were used to find the lines as well as the circle lines in the image to localize the screw, respectively. Pan, X. et al. [37] proposed that the HT algorithm using Canny, Prewitt, and Log methods could not accurately detect the lines and circles in complex images. The reason is that the HT algorithm has difficulty identifying the edges of the external hexagonal screws, and factors, such as washers, image shadows, and noisy backgrounds can cause the edges of the screws to be unrecognized. This shows that traditional machine learning methods cannot achieve high robustness in complex real-time environments.
Deep learning in object detection [19,20,38,39], which combines the tasks of both classification and localization of targets with models based on convolutional neural networks (CNN) neural networks, has developed rapidly in recent years. Traditional methods often require human intervention in feature extraction from images, thus limiting the scalability and effectiveness of the methods, whereas deep learning methods can autonomously extract feature parameters from image datasets [26]. Ren, W. et al. [40] classified the current sleeve state in a simulation environment based on RGB images using a VGG16 deep learning model to distinguish whether the sleeve was aligned with the screw to be dismantled. Yildiz, E. et al. [24] used a Hough change to divide the image into circles as screw alternates and then used a two-way convolutional neural network containing Xception and InceptionV3 to classify the alternate circles. Huynh, T. C. et al. [41] used a Fast R-CNN neural network model for screw detection based on RGB images. Poschmann, H. et al. [42] used a Fast R-CNN deep learning model structured by Inception ResNet V2 based on a stereo vision system to identify and locate different types of screws in images. Li, X. et al. [6] used Faster R-CNN to initially locate and classify the screws in 2D images and generate a series of prediction boxes, which were then adjusted using rotated edge similarity (RES) to accurately locate the centre of the screws. Although these Fast R-CNN model-based methods achieve high accuracy in screw detection tasks, they still have some drawbacks. The YOLO family of networks takes this into account by dividing the image into equally sized segments. The main idea is to divide the image into grids of the same size, with each grid detecting the target individually so that it is easier to focus on small objects. In the YOLOv2 [43] network, the concept of using a prior bounding box (anchor box) was first introduced. It uses k-means clustering analysis on the labeled ground-truth boxes in the training set to count the prior bounding boxes that better match the size of the objects in the sample and then uses the prior bounding boxes for each grid to obtain the prediction boxes based on them. In the YOLOv3 [21] network, a Feature Pyramid Network (FPN) structure [44] is introduced to change the size of grids from one to three. Pan, X. et al. [37] propose a YOLOv3-tiny-based screw detection method that enhances the robustness of the traditional Kanade–Lucas–Tomasi algorithm against illumination changes and background noise. Zhao, K. et al. [45] proposed an improved YOLOv3 model screw detection method to improve the detection accuracy and speed of palletizing robots to locate screws in complex scenes. The method designs an improved prior bounding box mechanism based on the k-means++ algorithm to provide more accurate prior bounding boxes for the YOLOv3 model. This model has been continued in the YOLO family of networks. Although the prior bounding box mechanism allows the network model to generate prediction boxes that better match the dimensions of the target object, it is too dependent on the diversity of the dataset, i.e., the model is poorly generalized. Ge, Z. et al. [23] proposed the YOLOX network, which discards the prior bounding box and generates prediction boxes directly based on feature points. This reduces the dependence of the model prediction results’ dependence on the dataset’s goodness and increases the model’s generalization capability. This makes the YOLOX network model more suitable for screw detection and location tasks in power battery disassembly tasks. However, since the shape of the screw itself is generally symmetrical, in the process of model training convergence, the IoU loss function based on the Cartesian coordinate system cannot provide an accurate prediction box with ground-truth box loss values in some cases. The loss function needs to be improved to ensure more accurate convergence of the model during the training process.
According to the above problems and analysis, we chose an eye-in-hand camera as our basic platform and proposed the “Active Screw Detection” method. This method makes full use of the advantages of the eye-in-hand camera. It actively moves the camera to the most suitable position for screw detection to obtain the best image. At the same time, this method improves the classic object detection algorithm YOLOX according to the characteristic of screw shape symmetry and further enhances the recognition accuracy. Experiments show that the proposed method can effectively improve the accuracy of screw detection significantly. There are two main improvements in this method.
  • It takes advantage of the vision sensor on the robot arm to gradually control the arm to reach the best detection position.
  • It proposes a loss function based on the Log Polar coordinate system and improves the IoU-based loss function to improve the accuracy of the YOLOX model for screw detection.

2. Method

Accurate screw detection is vital for automatic EVB disassembly. However, the detection of screws is significantly different from the detection of other objects. Firstly, the features of screws significantly change when observing them from different angles, and the screws are installed at very different angles due to the complex structure of the battery pack. Due to the limitation of image collection (the screw has been installed on the battery pack), most images in the dataset are the screw’s top view, which works well with a trained neural network model whose detection accuracy is the best if the images are the screw’s top view. So an optimal detection position exists for screw detection, that is, the position aligned with the normal vector of the screw surface. In this position, individual features can be observed and are not prone to classification errors. Secondly, during the screw detection, the coordinates of the detection frame have symmetry in the Cartesian coordinate system. This will cause the detection model to have difficulty in overlapping with the ground-truth box during the training process, which in turn will prevent the detection model from converging more completely. This is because the IoU-based loss function does not accurately reflect the difference in position between the prediction box and the ground-truth box. Based on these problems, we chose an eye-in-hand camera as the hardware platform and proposed a systematic method to finish the screw detection, which we name “Active screw detection”.
Our hardware platform is as shown in Figure 1.
We use a passive and compliant pneumatic torque actuator with vision sensing proposed by Zhang H. W. et al. [16] for the experiment (Figure 1). Information on some parameters of the robotic arm is shown in Table 1.
It mainly consists of a pneumatic power subject, a square shaft transmission limit module, a rod limit module, and an eye-in-hand camera module. Particularly, the eye-in-hand camera module includes a metal clamp and an Intel RealSense D435 camera, which can capture both RGB and depth images simultaneously. The “Active Screw Detection” process is shown in Figure 2. Firstly, the eye-in-hand camera module collects RGB and depth images of the robot arm in the initial position. Pre-screw detection is performed on the RGB image using a modified YOLOX model. The approximate pixel position of the screw is obtained. The normal vector of the screw plane is then calculated based on the pixel position of the closest screw to the robot arm, combined with the corresponding depth of the depth image. Here, a RANSAC-based plane fitting algorithm is proposed to calculate the normal vector of the target screw in order to minimise the effect of noise in the depth image. We can then control the robot arm to move to the optimal screw detection position, which is aligned with the normal vector of the target screw. In fact, due to the inherent unevenness of the screw surface and the noise in the depth image, moving to the optimal screw detection position is a process of gradual convergence. The system will continually check that the current position is the best observed position until the error is less than expected. At this point, the eye-in-hand camera module captures the image that is most suitable for screw detection. Post-screw detection is then performed based on the captured RGB images using the improved YOLOX model. The screw prediction box obtained at this point has a higher accuracy compared to the results of pre-screw detection. Finally, the exact spatial coordinates of the screw can be calculated based on the depth information of the corresponding area of the screw prediction box. In addition, compared to the native YOLOX model, this study proposes a LogPolar IoU loss function based on the Log Polar coordinate system to improve the training of the YOLOX model. The prediction box based on the Cartesian coordinate system is converted into a prediction box based on the Log Polar coordinate system. In order to avoid the problem of low convergence accuracy of the model due to the symmetry of the Cartesian coordinate system, the YOLOX model is trained using the self-labelled dataset with LogPolar IoU to make the convergence of the model more accurate. Higher detection and localisation accuracy is achieved.

2.1. Adaptive Adjustment of Robot Arm Angle

The target detection task for the screws is based directly on the RGB images captured by the vision sensor on the robotic arm. The information representation capability of the RGB images is limited by the angle at which the vision sensor is captured. For the screw with too large an angle between the normal vector of the upper surface and the vision sensor, Figure 3a, its main features are obscured, making it difficult to identify and localize the model. For the screw whose upper surface is approximately perpendicular to the capture angle of the vision sensor, on the other hand, the entire screw is clearly characterized, and the model is able to identify and localize it well during the detection process with a high confidence level for the foreground, Figure 3b.
In the screw disassembly task, the screw top layer of nuts is of a certain thickness, and the screw spacing between some automotive battery pack screws is not large. The excessive tilt angle of the visual sensor will, on the one hand, cause the screws to block each other or the washers, a feature-rich component, which is not conducive to the detection and positioning of the screws and can easily cause missed detection. On the other hand, the existence of the tilt angle will cause visual deformation of the screw, which in turn will have a greater impact on the accuracy of the screw classification and positioning. In the intelligent disassembly process of battery pack screws, the mechanical arm is constantly moving, which will cause the screw angle in the image collected by the vision sensor to vary greatly, thus affecting the positioning accuracy of the screw. Therefore, before the accurate positioning of the screws, the angle of the vision sensor needs to be corrected, which requires the mechanical arm to adjust the angle adaptively.
This study proposes an adaptive adjustment algorithm for the robotic arm based on the RANSAC algorithm to correct the attitude of the robotic arm so that it is in an attitude favorable to the screw detection and positioning task, see Figure 4. For the robotic arm that is performing the screw disassembly task, pre-screw detection is performed on its captured images. The best normal vector perpendicular to the screw to be disassembled is then calculated using the RANSAC algorithm. Finally, the robotic arm is paired with the best normal vector to adjust the pose of the robotic arm to the best pose for subsequent post-screw detection.

2.1.1. Acquisition of Plane Fitting Point Set

For the RGB map in the initial pose, the improved YOLOX model is used for pre-screw detection to obtain the set of prediction boxes of the screws,  S b o x = { b 1 , b 2 , , b n } , to determine the approximate position of the screws, where each prediction box b contains the position information of the screw pixel points in the image. Then, the nearest prediction box  b m i n to the robotic arm is obtained from  S b o x . The general scenario of  b m i n is shown in Figure 5. Finally, the set of depth information corresponding to a certain number of pixel points,  S p o i n t s = { d 1 , d 2 , , d m } , is filtered from  b m i n S p o i n t s is the set of points used for plane fitting.
At the beginning of the experiment, we only selected a small number of random points in  b m i n for plane fitting and found that the results were not satisfactory. The reason is that the points on the upper surface of the head are not in the same plane as the points in the background part, resulting in a large randomness in the fitted plane. Moreover, there are some noise points in the image captured by the vision sensor, and the plane fitted by selecting only a small number of points will be affected by the noise points. Therefore, it is necessary to use as many points as possible and in the same plane for plane fitting. Considering that the washer and nut of the screw itself are not in the same plane, the points of the screw part in prediction box  b m i n are not suitable for plane fitting. However, the points of the background part are different, and they are generally in the same plane, which is suitable for plane fitting. In this study, the OTSU-based [46,47] binary thresholding method was used to split the screw part and the background part in  b m i n . The depth information corresponding to all the pixel points in the background part is selected to form  S p o i n t s .

2.1.2. Ransac Algorithm for Plane Fitting

The RANSAC (Random Sample Consensus) algorithm is an iterative algorithm that correctly estimates the parameters of a mathematical model from a set of noisy data, and it is an uncertain algorithm that produces results with only one probability, which increases with the number of iterations. It divides the data into two sets of “outer points” and “inner points”. The “outer points” generally refer to noisy data in the data, such as false matches in matching and outliers in estimation curves and are used in plane fitting tasks to represent points that are far from the fitted plane. The “inner points” refer to the data that make up the model parameters, as opposed to the “outer points”. The RANSAC algorithm for plane fitting is shown in Algorithm 1.
Algorithm 1 RANSAC algorithm for plane fitting
Require:
  • S p o i n t s = p 1 , p 2 , , p m is the set of all points used to fit the plane
  • P = 0.99 is the base probability of expecting RANSAC to get the model right
  • n l i m i t is the threshold at which the ‘inner point’ is expected
  • i t e r a t o r s = 0 indicates the number of iterations
  • v b e s t T and  v T are normal vector
Ensure:
  • v T is the plane normal vector obtained from the plane fit
1:
while   ( n n l i n e r s > n l i m i t ) o r ( i t e r a t o r s > k ) do
2:
   random select n points from  S p o i n t to build  S
3:
   Calculate  v T based on  S
4:
   Calculate  n i n l i n e r s in  S p o i n t s
5:
   Update the best plane normal vector  v b e s t T and  n i n l i n e r s
6:
    t = n i n l i n e r s / ( n i n l i n e r s + n o u t l i n e r s )
7:
    k = l o g ( 1 P ) / l o g ( 1 t n )
8:
    i t e r a t o r s = i t e r a t o r s + 1
9:
end while
10:
return   v b e s t T
For each iteration of the RANSAC algorithm, firstly, n points are randomly selected from  S p o i n t s to form a subset  S , and a plane is fitted based on  S using the least squares method. Then, all the points of  S p o i n t s are divided into “inner points” and “outer points” by calculating the distance from the points to the plane and calculating the proportion t of “inner points”, Equation (1).
t = n i n l i n e r s n i n l i n e r s + n o u t l i n e r s
where  n i n l i n e r s denotes the number of interior points, and  n o u t l i n e r s denotes the number of exterior points. Then, the probability that there is at least one “outliner” among the n randomly selected points in each iteration is  1 t n . Further,  ( 1 t n ) k is the probability that at least one “outliner” is sampled in each of the k iterations of the fitting plane, then the probability that at least one “outliner” can be sampled is  1 t n . The probability of sampling only n “inner points” is  1 ( 1 t n ) k , Equation (2).
P = 1 ( 1 t n ) k
The desired number of iterations k for each iteration can be derived from this, as shown in Equation (3). However, it is not completely guaranteed that all points selected each time are “interior points”, and we can only expect P to be as close to 1 as possible, which is 0.99 in this study.
P = l o g ( 1 P ) l o g ( 1 t n )
There are two conditions for the finalization of the RANSAC algorithm. One condition is that the number of “interior points” in the plane based on the current computation reaches  n l i m i t , and the other condition is that the number of iterators exceeds the current desired number of iterations k.
After executing the RANSAC algorithm, the best plane normal vector  v b e s t T corresponding to  S p o i n t s is obtained. Finally, the robotic arm adjusts its pose to match  v b e s t T to achieve the best pose for the subsequent post-screw detection task.

2.1.3. Total Least Square for Plane Fitting

During each iteration of the RANSAC algorithm, the overall least squares method is used to compute the plane for a subset  S of  S p o i n t s . Assume that the unit normal vector of the plane is  v T = a b c , Equation (4).
a 2 + b 2 + c 2 = 1
Suppose  ( x 0 , y 0 , z 0 ) is any point in the desired plane.Then, the plane can be constructed by the point normal formula, Equation (5).
a ( x x 0 ) + b ( y y 0 ) + c ( z z 0 ) = 0
Based on the calculated plane, the distance of all points of the set  S to the plane is calculated, which is the error of this plane fitting. Based on the fact that the distance from any point in space to the plane is equal to the projection on the normal vector of the vector composed of any point on the plane, the error is Equations (6) and (7).
J 2 = i = 1 n d i 2 = i = 1 n [ ( x x 0 , y y 0 , z z 0 ) ( a , b , c ) ] 2 a 2 + b 2 + c 2
J 2 = i = 1 n [ a ( x x 0 ) + b ( y y 0 ) + c ( z z 0 ) ] 2
where  ( x i , y i , z i ) denotes the i-th point in S’, and  d i denotes the error distance of the i-th point to the plane. Then, the  J 2 at this time is certainly the smaller the better. Then, you need to find the extreme value of  J 2 J 2 calculates the partial derivatives of  x 0 y 0 z 0 , respectively, Equations (8)–(13).
J 2 x 0 = 2 a i = 1 n [ a ( x x 0 ) + b ( y y 0 ) + c ( z z 0 ) ] = 0
J 2 x 0 = 2 a [ a ( x ¯ x 0 ) + b ( y ¯ y 0 ) + c ( z ¯ z 0 ) ] = 0
J 2 y 0 = 2 b i = 1 n [ a ( x x 0 ) + b ( y y 0 ) + c ( z z 0 ) ] = 0
J 2 y 0 = 2 b [ a ( x ¯ x 0 ) + b ( y ¯ y 0 ) + c ( z ¯ z 0 ) ] = 0
J 2 z 0 = 2 c i = 1 n [ a ( x x 0 ) + b ( y y 0 ) + c ( z z 0 ) ] = 0
J 2 z 0 = 2 c [ a ( x ¯ x 0 ) + b ( y ¯ y 0 ) + c ( z ¯ z 0 ) ] = 0
where  x ¯ y ¯ z ¯ denote the average of the three dimensional coordinates of all points in the set  S , respectively. From Equations (9), (11), and (13), we know that the value of all partial derivatives is 0 when  x 0 = x ¯ y 0 = y ¯ , and  z 0 = z ¯ J 2 obtains the extreme value, and  J 2 at this time is Equation (14).
J 2 = i = 1 n [ a ( x i x ¯ ) + b ( y i y ¯ ) + c ( z i z ¯ ) ] 2
Decompose  J 2 into matrix multiplication according to Equation (14) to obtain Equations (15) and (16).
J 2 = v T S v
S = ( x x ¯ ) 2 ( x i x ¯ ) ( y i y ¯ ) ( x i x ¯ ) ( z i z ¯ ) ( x i x ¯ ) ( y i y ¯ ) ( y i y ¯ ) 2 ( y i y ¯ ) ( z i z ¯ ) ( x x ¯ ) ( z i z ¯ ) ( y i y ¯ ) ( z i z ¯ ) ( z z ¯ ) 2
where  v T = a b c is the normal vector of the plane obtained from this subplane fit. It can be found that the matrix S is a positive definite matrix. Then, it can be decomposed into Equation (17).
S = q 1 q 2 q 3 λ 1 λ 2 λ 3 q 1 q 2 q 3 = Q Λ Q T
where the matrix Q is composed of the eigenvectors of the matrix S, and  λ 1 λ 2 λ 3 are the eigenvalues of the matrix S. Equation (18) can be derived from Equations (16) and (17).
J 2 = v T Q Λ Q T v = ( v T Q ) Λ ( v T Q ) T
Define Equation (19):
u 1 = v T q 1 , u 2 = v T q 2 , u 3 = v T q 3 , U T = u 1 u 2 u 3
Equation (19) is brought into Equation (18) to obtain Equation (20).
J 2 = U T Λ U = u 1 u 2 u 3 λ 1 λ 2 λ 3 u 1 u 2 u 3
Equation (20) was further calculated to obtain Equation (21).
J 2 = λ 1 u 1 2 + λ 2 u 2 2 + λ 3 u 3 2
In addition, because:
U T U = v T Q ( v T Q ) T = v T Q Q T v = v T v = E
From Equation (22), matrix U is the unit vector, Equation (23).
u 1 2 + u 2 2 + u 3 2 = 1
It may be set  λ 1 λ 2 λ 3 , then Equations (21)–(23) can be further calculated according to the inequality to obtain Equation (24).
λ 1 = λ 1 ( u 1 2 + u 2 2 + u 3 2 ) J 2 λ 3 ( u 1 2 + u 2 2 + u 3 2 ) = λ 3
From Equation (24), the minimum value of  J 2 is  λ 1 , when  u 1 = 1 u 2 = u 3 = 0 . From Equation (19),  v = q 1 . In summary, when  v = q 1 , the error value  J 2 takes the minimum value  λ 1 , where  λ 1 is the minimum eigenvalue of the matrix S, and  q 1 is the eigenvector corresponding to  λ 1 . Here,  q 1 can be quickly calculated by matrix singular value decomposition. Constructing the matrix A from the set  S , see Equation (25), it is easy to find that  A T A = S . Since  q 1 is the eigenvector corresponding to the smallest eigenvalue of the matrix S, then  q 1 is the eigenvector corresponding to the smallest eigenvalue of the matrix  A T A .
A = x 1 x ¯ y 1 y ¯ z 1 z ¯ x 2 x ¯ y 2 y ¯ z 2 z ¯ . . . . . . . . . x n x ¯ y n y ¯ z n z ¯
The singular value decomposition of the matrix A is shown in Equation (26).
A = U Σ V T
where the matrix V consists of eigenvectors of the matrix  A T A , and each column of eigenvectors from left to right corresponds to the eigenvalues in decreasing order. Then, the eigenvector corresponding to the smallest eigenvalue is the last column of the matrix V, which corresponds to the last row of the matrix  V T . Then,  q 1 is the last row of matrix  V T , and the normal vector  v T of the plane constructed by the set  S can be found quickly.

2.2. Screw Detection Based on YOLOX Model

Active Screw Detection uses the improved YOLOX model proposed in this study for screw detection. The implementation principle of the YOLOX model and the network structure are described in detail in the subsequent sections of this subsection.

2.2.1. Architecture of YOLOX Model

The YOLOX model [23] is used to detect and locate the screws based on the RGB image captured by the vision sensor. Its role is to obtain information about the coordinates of the pixel region where each screw is located in the RGB image. YOLOX is a neural network model for deep learning. The main principle is to use a neural network to extract feature information from the RGB image and then use operations, such as decoupling and convolution, to obtain prediction boxes. These prediction boxes frame the area where the screws are located. The YOLOX model consists of four parts: Input, BackBone, Neck, and Prediction, and the specific network structure is shown in Figure 6. RGB images captured by vision sensors are used as input to YOLOX, and pre-processing is performed at the input. Then, BackBone performs feature extraction on different images fine-grained for the preprocessed images. Then, Neck obtains the features under different perceptual fields and converts the extracted feature information into coordinates, categories, and other information. Finally, the Prediction part is decoupled to obtain the position information, category, and confidence of the detected boxes, respectively.

2.2.2. Input

In the Input section, a preprocessing operation is performed on the RGB images to unify the size of the input images. In addition, high-resolution images contain too much pixel information to be used directly for neural network calculations that would take up a large amount of GPU video memory. To reduce the performance requirements of the model on the hardware, the images are scaled and filled. In the training phase of the model, the RGB images in the dataset are enhanced with Mosaic and MixUp data to expand the quantity and quality of the dataset. The Mosaic data enhancement method, which was applied in the YOLOv3 model, has proved to be beneficial to model learning and training. The Mosaic data augmentation method stitches the calibrated RGB images by random scaling, random cropping, and random arrangement to expand the dataset and the diversity of images in the dataset. The Mosaic method is an additional enhancement strategy, which randomly selects two images for scaling, filling, and merging and then sets a fusion factor for the respective labels of the images to generate a new label. This module turns off the data enhancement method after the model training is finished.

2.2.3. BackBone

The BackBone part is consistent with YOLOv3 and uses the DarkNet53 network model [21], which is a classical deep network mainly composed of residual convolution and SPP modules. The residual convolution consists of convolutional layer jump connections, whose role is to extract the feature information in the input image. The structure of the residual avoids the problems of gradient explosion and gradient disappearance caused by too deep layers of the network. In the middle of the residual module, a convolutional layer with a kernel of 3 × 3, a step size of 2 and a padding of 1 is added to downsample and upgrade the output feature map of each residual block. The SPP structure mainly consists of maximum pooling layers of different sizes, and the feature maps after downsampling of different sizes are stitched together to obtain richer texture features, which is beneficial to the subsequent detection operations.

2.2.4. Neck

The Neck part consists of an FPN structure, a pyramid-shaped feature extraction module. Its function is to extract shallow features, middle features, and high features obtained from the convolution of different depth residuals in the BackBone part and construct three feature maps based on different perceptual fields of the original image using convolution and upsampling operations from top to bottom. This part mainly focuses on the different sizes of screws with different percentages in RGB pictures, which is equivalent to the average chunking of the original picture in three sizes and carries out the subsequent screw detection localization classification task based on the regions of these three sizes. In the screw disassembly process, the height of the robotic arm will constantly change, which will lead to a large difference in the proportion of the same screw in the pictures collected by the vision sensor, and the use of FPN structure can effectively ensure the accuracy of subsequent screw identification and localization.

2.2.5. Prediction Head

Unlike the network structure of the previous YOLO series, YOLOX performs a decoupling operation in the Prediction Head part to calculate the classification task and the regression task separately. The conflict between classification and regression tasks is a very common problem because during the final convolution operation, the convolution kernel parameters are shared, which leads to the feature maps containing both tasks affecting each other. By separating the feature maps of classification and regression tasks, this problem can be avoided and the accuracy of classification and localization can be improved. SimOTA in the Prediction Head section obtains and filters the positive sample prediction boxes. The anchor-based approach is used in the network before YOLOv5 to obtain the prediction boxes. The k-means algorithm was used to cluster the sizes of the ground-truth boxes in the dataset, and the sizes of the centroids of the nine categories were taken as the sizes of the prior bounding boxes. The prior bounding boxes of 9 sizes are assigned to the feature maps of 3 sizes of perceptual fields obtained from the BackBone and Neck parts. Then, each cell under each size of perceptual field corresponds to 3 sizes of prior bounding boxes. The role of the prior bounding box is to give an initial value to the prediction box. The output of the model represents the box centroid and the offset of the width and height. The mapping relationship is used to combine the prior bounding boxes and the model output to obtain the final prediction boxes. In the anchor-based approach, each cell is responsible for three sizes of a prior bounding boxes, which depend on the ground-truth box size of the labels in the dataset, making it difficult to guarantee the generalization of the model.The YOLOX anchor-free approach generates prediction boxes directly based on each feature point using the feature maps of the three size senses. The advantage of this is that the generated prediction boxes do not depend on the ground-truth boxes information of the labels in the dataset, avoiding the limitation in terms of data. The number of ground-truth boxes is much smaller than these directly generated prediction boxes, so it is also necessary to select the positive sample prediction boxes among them, and SimOTA is used here. First, the cost matrix between each ground-truth box and each prediction box is calculated, see Equation (27).
c o s t i j = L i j c l s + λ 1 L i j r e g + λ 2 L i j c e n
where  λ 1 and  λ 2 are balance coefficients, i and j are subscripts,  L i j c l s is the classification loss of each ground-truth box and the current feature point prediction box,  L i j r e g is the position loss of each ground-truth box and the current feature point prediction box, and  L i j c e n is whether the center of each ground-truth box falls within a certain radius of the feature point. The IoU of each ground-truth box is calculated and rounded down to obtain  d y n a m i c k . The first  d y n a m i c k boxes with the lowest cost are taken as the positive samples of the ground-truth box, and the rest are all negative samples. YOLOX uses this method together with anchor-free to filter the prediction boxes, which not only reduces the training time but also avoids additional parameters compared to the OTA used in YOLO.

2.2.6. Output of YOLOX

In the screw detection and localization task, the prediction results of  S × S × 7 dimensions are generated under each perceptual field, where S × S denotes the number of cells into which the predicted image is divided, and each cell contains 7 channels of output, which consists of 3 branches, i.e., cls, obj, and reg branches. The cls branch occupies 2 channels to indicate which type of screw the object in the predicted detection box belongs to. The obj branch occupies 1 channel to distinguish whether the prediction box is foreground or background. The reg branch occupies 4 channels, including the location of the center point of the prediction box and the dimensions of the width and height.

2.2.7. Loss Function

The loss function of YOLOX is the sum of all losses averaged over the number of positive labels according to the output cls, obj, and reg branches, and its loss function is shown in Equation (28).
L = 1 N p o s ( L c l s + L o b j + λ r e g L r e g )
where  N p o s is the number of positive samples obtained by SimOTA screening,  λ r e g is the balance coefficient, and  L c l s is the loss value of the classification task, whose cross-entropy loss is calculated for the classification results of the positive samples with the classification labels of the ground-truth boxes, as shown in Equation (29).
L c l s = i N p o s [ t i l o g ( p i ) + ( 1 t i ) l o g ( 1 p i ) ]
where  N p o s denotes the total number of samples,  t i denotes the true label of each target, and  p i denotes the predicted classification result.  L o b j is the confidence loss value, which is calculated as the cross-entropy for the confidence of all samples, see Equation (30).
L o b j = i N [ c i l o g ( c ^ i ) + ( 1 c i ) l o g ( 1 c ^ i ) ]
where N is the number of samples,  c i is the true confidence of each target, which is 1 for foreground and 0 for background, and  c ^ i is the confidence of each prediction box, which is not simply 1 or 0, but multiplied by the IoU of the prediction box and its corresponding ground-truth box.  L r e g is the regression loss value of the ground-truth boxes, which is obtained by calculating the IoU of the prediction box and its corresponding ground truth. Since in the paper presented by YOLOX, the authors only state that  L r e g is obtained by IoU and do not give a specific formula, we derived the formula for the  L r e g part from the open source source code, see Equation (31).
L r e g = 1 I o U 2
The IoU represents the intersection ratio of the area of the prediction box and its corresponding ground-truth box, and its principle and calculation process will be elaborated in Section 2.3.

2.3. Logpolar IoU

In the training process of the regression task for target detection, how to obtain the loss value of the prediction box and the ground-truth box is the key step, and the common method is to use IoU and the improved version of IoU to perform the calculation.

2.3.1. IoU

Intersection over Union (IoU) is the original version of IoU, widely used in the field of target detection, which calculates the area intersection ratio of the prediction box to the ground-truth box, see Equation (32).
I O U = B B g t B B g t
where  B g t is the ground-truth box, B is the prediction box,  B B g t denotes the area of the intersection part of the prediction box and the ground-truth box, and  B B g t denotes the area of the merged part of the prediction box and the ground-truth box. Although IoU can visually represent the overlap of the two boxes, it also has some problems.
On the one hand, if there is no intersection between the prediction box and the ground-truth box, then the value of IoU will be zero, and it will be useless. When screening the positive samples, YOLOX will only keep the prediction box samples whose center point of the prediction box falls within a certain range of the center point of the ground-truth box so that the intersection of the prediction box and the ground-truth box will definitely exist, avoiding this defect of IoU. However, on the other hand, if one box is contained in another box, see Figure 7, and the area of both boxes remains the same, then the values of both  B B g t and  B B g t do not change, and the value of IoU does not change either, which eventually leads to the failure of this part to contribute to the training process. Subsequently, there are some versions of IoU that improve the original defects of IoU.

2.3.2. Improved Version of IoU

To solve the problem that a constant value of 0 for IoU does not contribute to the training process when there is no intersecting part between the prediction box and the ground-truth box, GIoU is proposed [48]. GIoU takes into account two factors, the area of the smallest box containing the prediction box and the detection box and the area of the non-prediction box and detection box area within the smallest box, see Equation (33).
G I o U = I o U C ( B B g t ) C
where  B g t is the ground-truth box, B is the prediction box, C is the area of the smallest box that contains both the prediction box and the ground-truth box,  B B g t denotes the area of the part of the prediction box that merges with the ground-truth box, and then  C ( B B g t ) denotes the area of the non-prediction box and ground-truth box area within the smallest box, see Figure 8.
When the two boxes are very far apart, IoU loses its role by being constant at zero, and GIoU still provides a loss calculation through the relationship between the position of the boxes and their area. However, GIoU still does not solve the case where one box is contained within another, see Figure 9. In this case,  B B g t and  B B g t as well as C will remain constant, and the value of GIoU will remain constant, making no contribution to the training process.
To address the fact that IoU and GIoU cannot handle the case where one box is contained in another, DIoU considers the ratio of the distance between the centre point of the prediction box and the detected box to the distance between the diagonals of the smallest box containing both boxes, Equation (34).
D I o U = I o U ρ 2 ( b , b g t ) c 2
where  b g t is the Cartesian coordinate of the centre point of the ground-truth box, b is the Cartesian coordinate of the centre point of the prediction box,  ρ ( b , b g t ) denotes the Euclidean distance between the prediction box and the midpoint of the ground-truth box, and c denotes the Euclidean distance of the diagonal of the smallest box that contains both the prediction and ground-truth boxes, see Figure 10.
DIoU somewhat avoids the problem of degradation of the regression loss function as the model converges, while still providing loss for training when the prediction box is far away from the ground-truth box without an intersecting part. However, DIoU can fail in some cases. The DIoU degenerates into an IoU when the prediction box coincides with the centroid coordinates of the ground-truth box and one box is contained within the other, see Figure 11, where the value of  ρ ( b , b g t ) is 0. At this point, the DIoU is completely useless if the large box is left unchanged and the small box changes in width and height while keeping the area the same.
The authors of DIoU were also aware of this problem and subsequently proposed CIoU, which adds the aspect ratio of the prediction box to the ground-truth box as a penalty factor, see Equations (35)–(37).
C I o U = I o U ρ 2 ( b , b g t ) c 2 + α v
v = 4 π 2 ( arctan w g t h g t arctan w h )
α = v ( 1 I o U ) + v
where  w g t and  h g t denote the width and height of the ground-truth box, w and h denote the width and height of the prediction box, respectively, v is used to measure the consistency of the aspect ratio, and  α is the weight function. By taking the aspect ratio into account, CIoU compensates for the deficiency of DIoU scoring the same area at different scales. CIoU is also the loss function for the YOLOv5 model to calculate the difference between the prediction box and the ground-truth box.
In the regression task, the loss value as a penalty term provides the direction of convergence for the training of the neural network model. During the training of the neural network, the position of the ground-truth box is constant, and the prediction box will gradually approach the ground-truth box as the training progresses and the loss value becomes progressively smaller. We would like the prediction box to return different loss values at different locations, which would facilitate the convergence of the neural network, as well as improve the detection accuracy. The current versions of IoU do not provide significant differences when calculating the loss of the detection box in the symmetric case, see Figure 12.
When the prediction box is very close to the ground-truth box, the neural network gradually converges and the step of the prediction box adjustment will gradually become smaller, which can easily move to a symmetrical position, thus making the effectiveness of the position loss function weaker and eventually affecting the localization accuracy. Therefore, this study proposes to convert the coordinates of the detection box in the Cartesian coordinate system into the Log Polar coordinate system and use the mode of the vector in the Log Polar coordinate system to calculate the loss instead of the Euclidean distance.

2.3.3. Logpolar IoU

We assume that each pixel point on the image is a point in the Cartesian coordinate system and that information about the position and dimensions of the prediction and ground-truth boxes can be represented by Cartesian coordinates. A point  ( x , y ) in the Cartesian coordinate system is converted to the corresponding coordinate  ( ρ , θ ) in the Log Polar coordinate system, see Equations (38) and (39).
ρ = x 2 + y 2
θ = π 2 i f   x = 0 a n d y > 0 arctan ( y x ) i f   x > 0 a n d y 0
where  ρ denotes the polar diameter, and  θ denotes the polar angle. As the image occupied only the first quadrant in the original Cartesian coordinate system, the value domain of  θ was  [ 0 , π 2 ] . The prediction box is transformed from the ground-truth box through the coordinate system, see Figure 13.
In the Log Polar coordinate system, the distance between any two points is the vector modulus of the difference between the polar diameter and the polar angle, Equations (40)–(43).
Δ ρ = ρ 1 ρ 2
Δ θ = θ 1 θ 2
λ = ( Δ ρ , Δ θ )
λ = ( Δ ρ ) 2 + ( Δ θ ) 2
where  ( ρ 1 , θ 1 ) and  ( ρ 2 , θ 2 ) denote the coordinates of any two points in the Log Polar coordinate system, and  λ denotes the vector mode of the two points. The vector mode in the Log Polar coordinate system is used as the distance between the two points, replacing the distance in the Cartesian coordinate system in CIoU, thus enabling the prediction box in the symmetric case to produce different loss values at different locations and inducing more accurate convergence of the model; logPolar IoU is shown in Equation (44).
L o g P o l a r I o U = I o U ρ l p 2 ( b , b g t ) c l p 2 + α v
where  ρ l p ( b , b g t ) denotes the vector mode of the centroid of the prediction box and the ground-truth box in the corresponding coordinates of the Log Polar coordinate system,  c l p denotes the vector mode of the diagonal coordinates of the smallest box containing both the prediction box and the ground-truth box in the corresponding coordinates of the Log Polar coordinate system, and  α and v continue to follow the penalty terms of the aspect ratio in CIoU.
The Log Polar coordinate system is equivalent to changing the viewing angle of the observation with respect to the Cartesian coordinate system. In the Cartesian coordinate system, the viewpoint is on the ground-truth box, and the prediction box passes through many symmetrical places as it keeps moving, producing the same losses and thus diminishing the effect of the loss function. In contrast, under the Log Polar coordinate system, the viewpoint is at the origin, and the polar diameter and angle of the prediction box change based on the origin after each move, reducing the occurrence of symmetrical situations and allowing the model to converge in a more accurate direction, which in turn increases the accuracy of the localisation.

3. Experiment

3.1. Experiment Based on Adaptive Adjustment of Robot Arm Angle

In this section, the experimental procedure and comparison of experimental results of the robot arm’s adaptive angle adjustment method based on a combination of RGB images and depth images are presented. For a given angle of the robot arm, both an RGB image and a depth image are taken, see Figure 14. The depth image contains the distance from each pixel point in the sensory field to the camera, with darker colours representing closer proximity and, conversely, warmer colours representing further away from the camera. We want to use the adaptive adjustment algorithm of the robot arm based on the RANSAC algorithm to adjust the angle of the robot arm to a plane perpendicular to the surface of the nut, i.e., perpendicular to the surface of the car’s power cell where the screw is located.
The screws are initially detected based on the RGB image using the target detection model to obtain approximate prediction boxes for the screw, which are not accurate at this stage, Figure 15. These prediction boxes are selected to approximate the pixel area of the screw in the image.
The prediction box closest to the robot arm is selected for the plane fitting task, and in Figure 15 the rightmost screw is selected. The selected prediction box includes both the stud part and the background part. Limited by the accuracy of the depth camera and the existence of thickness of the nut itself causing the top surface of the nut and the washer to be not in the same plane, the depth information of the screw part in the prediction box is not favorable for the plane fitting task, and although the background part is regionally cut, it is all on the same plane, which is favorable for plane fitting. So, the depth value of the background part is used for plane fitting. For the selected screw prediction box, the RGB image inside the prediction box is loosely cropped, and the loosening value is set to 10 pixels, see Figure 16a. The cropped image is then subjected to OTSU-based binary thresholding, and the image is divided into a binary map based on the screw part and the background part, Figure 16b. Then, a convex wrapping process is performed on the binary map, which thoroughly divides the image into screw part and background part based on the binary map, Figure 16c.
The set of points selected from the depth image of the background part is used for plane fitting to calculate the normal vector of the plane where the screw is located. The arm then adjusts its own position and attitude according to the calculated plane normal vector, and finally hovers over the target screw and is perpendicular to the plane where the target screw is located, see Figure 17. After the arm is adjusted autonomously, the visual features of the screw surface under its field of perception become clearer, which facilitates the subsequent accurate detection and localization tasks of the screw. Figure 18 shows the results of screw detection under the current sensory field. The model has a high confidence level for each prediction box while accurately locating all the screws in the picture. It is worth noting that the same localization model is used in Figure 15 and Figure 18. This shows that the perceptual field of the vision sensor plays a decisive role in the positioning accuracy in the screw detection and positioning task, as well as the effectiveness of our proposed adaptive adjustment algorithm for the robotic arm based on the RANSAC algorithm.
Moreover, we conducted a large number of experiments on the screw positioning task without attitude adjustment and the autonomous attitude adjustment based on the RANSAC algorithm proposed in this paper, respectively, and compare the results in Table 2.
The average positioning accuracy of the robotic arm for screw positioning was higher after autonomous attitude adjustment, both in the case of  m A P 50 and in the case of  m A P 75 .

3.2. Improved YOLOX Model for Screw Detection Experiments

In this section, experiments are presented for the screw detection and localization task based on the improved YOLOX model.

3.2.1. Dataset for Improved YOLOX

We collected a total of 1719 RGB images on different real EVBs. The information of EVBs is shown in Table 3.
We labeled the screws in all images using LabelImg software to construct a dataset for the screw detection and localization task. The boxes generated by manual labeling are called ground-truth box labels. The dataset included a total of 6992 screws in both categories of external hexagonal screws and 687 hexagonal nuts. These images were used for the training of the improved YOLOX model.

3.2.2. Train of Improved YOLOX

We divided the entire dataset into two parts, the training set and the validation set, with a ratio of 9:1. The data in the training set are involved in the training process of the model, which mainly includes the updating of the model parameters. The data in the validation set are not involved in the training process and are only used for the evaluation of the model. We ensured the randomness in the process of dividing the dataset. The information of all the components used in the experiments is shown in Table 4.
For the training process of the improved YOLOX model, a total of 300 iterations were performed, the initial learning rate was set to 0.01, and the whole process adopted the cosine annealing learning rate decreasing strategy. The batch size was set to 32. The stochastic gradient descent method was used as the optimizer; the momentum was set to 0.937, and the weight decay was set to  5 × 10 4 . Figure 19 shows the loss values of the training and validation sets converge gradually as the number of iterations increases and decreases.
After the training process, the model was evaluated with the validation set data.

3.2.3. Results of Improved YOLOX

The mAP value was chosen as the judging criterion of the model, and the mAP value was mainly calculated under two thresholds of IoU of 0.5 ( m A P 50 ) and 0.75 ( m A P 75 );  m A P 50 and  m A P 75 are the common judging criteria for localization accuracy presentation in the field of target detection. Firstly, for the experimental results of YOLOv2, YOLOv3, and YOLOv5 models, the YOLOX series models have higher localization accuracy on complex datasets due to the abandonment of the dependence of the prior frame on the dataset, and then, we used the original YOLOX model as the baseline model and also replaced the IoU function in the loss function with GIoU,DIoU,CIoU functions for the experiments, and the parameters of all models in the training process were aligned with the parameters of this experiment. The experimental results of these models are compared and shown in Table 5, where tiny, s, and m denote the YOLOX-tiny, YOLOX-s, and YOLOX-m models, respectively.
The IoU threshold is the confidence threshold, i.e., the IoU of the prediction box and the ground-truth box exceeds the specified threshold to be judged as a positive sample, otherwise it is a negative sample.
In the screw disassembly task based on the real machine, we pay more attention to the model at the threshold value of 0.75 compared to the threshold value of 0.5, and we can find that the mAP of our method is higher than other methods and has a greater improvement compared to the original YOLOX model. To show the improvement of the improved YOLOX model for screw detection accuracy, we re-captured some RGB images for the screw detection task. We set the confidence threshold to 0.75 and found that the original YOLOX model has a large number of missed detections on slightly complex images, e.g., when the screw itself is similar to the background color. Figure 20a,b show that the improved YOLOX model has a higher confidence level for the screw compared to the original YOLOX model when all the screws are detected, i.e., the model “trusts” each prediction box more. According to the above experiments, the improved YOLOX model based on LogPolar IoU has higher performance compared to the original YOLOX model, while maintaining high localization accuracy under the constraints of complex environments.

4. Conclusions

In order to solve the problem of poor accuracy of screw detection due to the uncertainty of robot arm posture in continuous screw disassembly tasks. This research proposes an adaptive adjustment method for the robotic arm with “Activate Screw Detection”. It is evaluated on an experimental platform based on an eye-in-hand camera. The experimental results show that the adaptive tuned robotic arm improves  62.81 % and  57.67 % in the screw detection task for  m A P 50 and  m A P 75 , respectively.
Meanwhile, for different IoU loss functions, the YOLOX model cannot converge accurately due to the symmetry of the Cartesian coordinate system, which makes the localization accuracy insufficient. In this study, the loss function of LogPolar IoU based on the Log Polar coordinate system is proposed to improve the training of the YOLOX model. Experimental evaluation is performed on a self-built dataset. The experimental results show that the LogPolar IoU-based loss function, compared with the native YOLOX loss function, improves the three models, YOLOX-tiny, YOLOX-s, and YOLOX-m, by  0.17 % 0.19 % , and  0.51 % , respectively, under the  m A P 50 rubric. Meanwhile, the three models improve by  3.36 % 3.59 % , and  2.38 % under the  m A P 75 criterion, respectively, and the localization accuracy of the models based on LogPolar IoU loss function is higher than that of the models based on other versions of IoU loss function.
According to the above experimental results and conclusions, the screw localization method based on “Activate Screw Detection” proposed in this work has certain effectiveness. In addition, there are still some problems in the whole experimental process. For example, the accuracy of plane fitting depends on the accuracy of the depth image, and the accuracy of the current depth camera is not enough. The self-built dataset contains fewer kinds of screws. In future work, we will improve the robotic arm with a higher precision camera, and we will continue to expand the types and numbers of screws in the screw detection dataset to increase the generalization and robustness of the localization model. At the same time, the robotic arm used in this study has its own limitations of not being able to move. The size of the EVB is too large for the arm span of the robotic arm to cover. Therefore, in order to achieve screw detection of the whole EVB, another way out needs to be found. In future work, we will consider placing the robotic arm on a mobile cart and using the program to achieve real-time control of the cart to enable movement of the robotic arm as a way to compensate for the arm span of the robotic arm.

Author Contributions

Data curation: H.L., H.Z. and Y.P.; formal analysis: Z.W., H.S. and M.C.; funding acquisition: Z.W. and M.C.; investigation: H.L. and Z.W.; methodology: H.L. and Z.W.; project administratio: M.C. and Z.W.; resources: H.L. and H.Z.; software: H.L. and Y.Z.; supervision: H.L. and H.Z.; validation: H.L. and S.Z.; visualization: H.L. and H.Z.; writing—original draft H.L.; writing—review and editing: H.L., H.Z., Z.W., H.S. and M.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by“2021 High Quality Development Project“ (TC210H02C).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors express their sincerest thanks to the Ministry of Industry and Information Technology of China for financing this research within the program “2021 High Quality Development Project (TC210H02C)”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, Z.; Liu, X.; Hao, H.; Zhao, F.; Amer, A.A.; Babiker, H. Research on the critical issues for power battery reusing of new energy vehicles in China. Energies 2020, 13, 1932. [Google Scholar] [CrossRef] [Green Version]
  2. Hao, H.; Cheng, X.; Liu, Z.; Zhao, F. China’s traction battery technology roadmap: Targets, impacts and concerns. Energy Policy 2017, 108, 355–358. [Google Scholar] [CrossRef]
  3. Talele, V.; Patil, M.S.; Panchal, S.; Fraser, R.; Fowler, M.; Gunti, S.R. Novel metallic separator coupled composite phase change material passive thermal design for large format prismatic battery pack. J. Energy Storage 2023, 58, 106336. [Google Scholar] [CrossRef]
  4. Gu, F.; Guo, J.; Yao, X.; Summers, P.A.; Widijatmoko, S.D.; Hall, P. An investigation of the current status of recycling spent lithium-ion batteries from consumer electronics in China. J. Clean. Prod. 2017, 161, 765–780. [Google Scholar] [CrossRef]
  5. Wegener, K.; Andrew, S.; Raatz, A.; Dröder, K.; Herrmann, C. Disassembly of electric vehicle batteries using the example of the Audi Q5 hybrid system. Procedia CIRP 2014, 23, 155–160. [Google Scholar] [CrossRef] [Green Version]
  6. Li, X.; Li, M.; Wu, Y.; Zhou, D.; Liu, T.; Hao, F.; Yue, J.; Ma, Q. Accurate screw detection method based on faster R-CNN and rotation edge similarity for automatic screw disassembly. Int. J. Comput. Integr. Manuf. 2021, 34, 1177–1195. [Google Scholar] [CrossRef]
  7. Weyrich, M.; Wang, Y. Architecture design of a vision-based intelligent system for automated disassembly of E-waste with a case study of traction batteries. In Proceedings of the 2013 IEEE 18th Conference on Emerging Technologies & Factory Automation (ETFA), Cagliari, Italy, 10–13 September 2013; pp. 1–8. [Google Scholar]
  8. Chen, A.; Dietrich, K.N.; Huo, X.; Ho, S.M. Developmental neurotoxicants in e-waste: An emerging health concern. Environ. Health Perspect. 2011, 119, 431–438. [Google Scholar] [CrossRef]
  9. Luo, C.; Liu, C.; Wang, Y.; Liu, X.; Li, F.; Zhang, G.; Li, X. Heavy metal contamination in soils and vegetables near an e-waste processing site, south China. J. Hazard. Mater. 2011, 186, 481–490. [Google Scholar] [CrossRef]
  10. Xie, Y.; Liu, Y.; Fowler, M.; Tran, M.K.; Panchal, S.; Li, W.; Zhang, Y. Enhanced optimization algorithm for the structural design of an air-cooled battery pack considering battery lifespan and consistency. Int. J. Energy Res. 2022, 46, 24021–24044. [Google Scholar] [CrossRef]
  11. Gaines, L. The future of automotive lithium-ion battery recycling: Charting a sustainable course. Sustain. Mater. Technol. 2014, 1, 2–7. [Google Scholar] [CrossRef] [Green Version]
  12. Castelvecchi, D. Electric cars: The battery challenge. Nature 2021, 596, 336–339. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, Y.; Liu, M. Reproduction of Li battery LiNixMnyCo1- x-yO2 positive electrode material from the recycling of waste battery. Int. J. Hydrogen Energy 2017, 42, 18189–18195. [Google Scholar]
  14. Yu, J.; Zhang, H.; Jiang, Z.; Yan, W.; Wang, Y.; Zhou, Q. Disassembly task planning for end-of-life automotive traction batteries based on ontology and partial destructive rules. J. Manuf. Syst. 2022, 62, 347–366. [Google Scholar] [CrossRef]
  15. Vongbunyong, S.; Kara, S.; Pagnucco, M. Application of cognitive robotics in disassembly of products. CIRP Ann. 2013, 62, 31–34. [Google Scholar] [CrossRef]
  16. Zhang, H.; Yang, H.; Wang, H.; Wang, Z.; Zhang, S.; Chen, M. Autonomous Electric Vehicle Battery Disassembly Based on NeuroSymbolic Computing. In Proceedings of the SAI Intelligent Systems Conference, Amsterdam, The Netherlands, 7–8 September 2023; pp. 443–457. [Google Scholar]
  17. Harter, J.J.; Mcintyre, T.J.; White, J.D. Electrical Safety Practices Developed for Automotive Lithium Ion Battery Dismantlement; Technical Report; Oak Ridge National Lab.(ORNL): Oak Ridge, TN, USA, 2020. [Google Scholar]
  18. Zhou, L.; Garg, A.; Zheng, J.; Gao, L.; Oh, K.Y. Battery pack recycling challenges for the year 2030: Recommended solutions based on intelligent robotics for safe and efficient disassembly, residual energy detection, and secondary utilization. Energy Storage 2021, 3, e190. [Google Scholar] [CrossRef]
  19. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Washington, DC, USA, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  20. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  21. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  22. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  23. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  24. Yildiz, E.; Wörgötter, F. Dcnn-based screw detection for automated disassembly processes. In Proceedings of the 2019 15th International Conference on Signal-Image Technology & Internet-Based Systems (SITIS), Sorrento, Italy, 26–29 November 2019; pp. 187–192. [Google Scholar]
  25. Ameperosa, E.T. Bolt Detection and Position Estimation Using Domain Randomization. Ph.D. Thesis, The University of Texas at San Antonio, San Antonio, TX, USA, 2018. [Google Scholar]
  26. Zhang, Y.; Sun, X.; Loh, K.J.; Su, W.; Xue, Z.; Zhao, X. Autonomous bolt loosening detection using deep learning. Struct. Health Monit. 2020, 19, 105–122. [Google Scholar] [CrossRef]
  27. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar]
  28. Wang, S. An Augmentation Small Object Detection Method Based on NAS-FPN. In Proceedings of the 2020 7th International Conference on Information Science and Control Engineering (ICISCE), Changsha, China, 18–20 December 2020; pp. 213–218. [Google Scholar]
  29. Lim, J.S.; Astrid, M.; Yoon, H.J.; Lee, S.I. Small object detection using context and attention. In Proceedings of the 2021 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Changsha, China, 18–20 December 2021; pp. 181–186. [Google Scholar]
  30. Wu, J.; Shao, J.; Zhou, G.; Yang, D.; Cheng, Y. Detection of Bolt Looseness Based on Average Autocorrelation Function. Shock Vib. 2021, 2021, 6662686. [Google Scholar] [CrossRef]
  31. Jurjević, L.; Gašparović, M. 3D data acquisition based on OpenCV for close-range photogrammetry applications. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 377. [Google Scholar] [CrossRef] [Green Version]
  32. Adjigble, M.; Marturi, N.; Ortenzi, V.; Rajasekaran, V.; Corke, P.; Stolkin, R. Model-free and learning-free grasping by local contact moment matching. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 2933–2940. [Google Scholar]
  33. Li, R.; Pham, D.T.; Huang, J.; Tan, Y.; Qu, M.; Wang, Y.; Kerin, M.; Jiang, K.; Su, S.; Ji, C.; et al. Unfastening of hexagonal headed screws by a collaborative robot. IEEE Trans. Autom. Sci. Eng. 2020, 17, 1455–1468. [Google Scholar] [CrossRef]
  34. Gil, P.; Pomares, J.; Diaz, S.v.P.C.; Candelas, F.; Torres, F. Flexible multi-sensorial system for automatic disassembly using cooperative robots. Int. J. Comput. Integr. Manuf. 2007, 20, 757–772. [Google Scholar] [CrossRef] [Green Version]
  35. Cha, Y.; You, K.; Choi, W. Computer-image-based loosened bolt detection using support vector machines. In Proceedings of the International Structural Specialty Conference, Guangzhou, China, 8–11 May 2016. [Google Scholar]
  36. Park, J.H.; Huynh, T.C.; Choi, S.H.; Kim, J.T. Vision-based technique for bolt-loosening detection in wind turbine tower. Wind Struct. 2015, 21, 709–726. [Google Scholar] [CrossRef]
  37. Pan, X.; Yang, T. Image-based monitoring of bolt loosening through deep-learning-based integrated detection and tracking. Comput. Aided Civ. Infrastruct. Eng. 2022, 37, 1207–1222. [Google Scholar] [CrossRef]
  38. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  39. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  40. Ren, W.; Wang, Z.; Yang, H.; Zhang, Y.; Chen, M. NeuroSymbolic Task and Motion Planner for Disassembly Electric Vehicle Batteries. J. Comput. Res. Dev. 2021, 058, 2604. [Google Scholar] [CrossRef]
  41. Huynh, T.C.; Hoang, N.D.; Ho, D.D.; Tran, X.L. An Image-based Algorithm for Automatic Detection of Loosened Bolts. Comput. Sci. Math. Forum 2021, 2, 1. [Google Scholar]
  42. Poschmann, H.; Brüggemann, H.; Goldmann, D. Fostering end-of-life utilization by information-driven robotic disassembly. Procedia CIRP 2021, 98, 282–287. [Google Scholar] [CrossRef]
  43. Redmon, J.; Farhadi, A. YOLO9000: Better, faster, stronger. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7263–7271. [Google Scholar]
  44. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  45. Zhao, K.; Wang, Y.; Zuo, Y.; Zhang, C. Palletizing Robot Positioning Bolt Detection Based on Improved YOLO-V3. J. Intell. Robot. Syst. 2022, 104, 1–12. [Google Scholar] [CrossRef]
  46. Tellman, B.; Sullivan, J.; Kuhn, C.; Kettner, A.; Doyle, C.; Brakenridge, G.; Erickson, T.; Slayback, D. Satellite imaging reveals increased proportion of population exposed to floods. Nature 2021, 596, 80–86. [Google Scholar] [CrossRef] [PubMed]
  47. Sezgin, M.; Sankur, B. Survey over image thresholding techniques and quantitative performance evaluation. J. Electron. Imaging 2004, 13, 146–165. [Google Scholar]
  48. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized intersection over union: A metric and a loss for bounding box regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
Figure 1. A passive and compliant pneumatic torque actuator with vision sensing.
Figure 1. A passive and compliant pneumatic torque actuator with vision sensing.
Batteries 09 00187 g001
Figure 2. Framework of the “Activate Screw Detection”.
Figure 2. Framework of the “Activate Screw Detection”.
Batteries 09 00187 g002
Figure 3. Pictures of screw taken by the robot arm from different angles. The red arrow represents the normal vector on the upper surface of the screw.
Figure 3. Pictures of screw taken by the robot arm from different angles. The red arrow represents the normal vector on the upper surface of the screw.
Batteries 09 00187 g003
Figure 4. Flowchart of the robot arm adaptive adjustment algorithm.
Figure 4. Flowchart of the robot arm adaptive adjustment algorithm.
Batteries 09 00187 g004
Figure 5. Schematic diagram of the screw and background parts in the prediction box, where the gray part is the screw body area, and the orange part is the background area.
Figure 5. Schematic diagram of the screw and background parts in the prediction box, where the gray part is the screw body area, and the orange part is the background area.
Batteries 09 00187 g005
Figure 6. Architecture of YOLOX model.
Figure 6. Architecture of YOLOX model.
Batteries 09 00187 g006
Figure 7. IoU faces the case of a box contained in another box. Blue box is  B g t , and the red box is B. (ac) denote the different B contained in the same  B g t inner where the area of B is constant. The IoU is the same in all three cases.
Figure 7. IoU faces the case of a box contained in another box. Blue box is  B g t , and the red box is B. (ac) denote the different B contained in the same  B g t inner where the area of B is constant. The IoU is the same in all three cases.
Batteries 09 00187 g007
Figure 8. Demo diagram of GIoU calculations.  B B g t is the green part, C is the outermost rectangular part, and  C ( B B g t ) is the orange part.
Figure 8. Demo diagram of GIoU calculations.  B B g t is the green part, C is the outermost rectangular part, and  C ( B B g t ) is the orange part.
Batteries 09 00187 g008
Figure 9. GIoU faces the case of a box contained in another box. (ac) denote the different B contained in the same  B g t inner where the area of B remains unchanged. Both  B B g t (the green part) and C (the orange part) are the same, resulting in the same value for GIoU.
Figure 9. GIoU faces the case of a box contained in another box. (ac) denote the different B contained in the same  B g t inner where the area of B remains unchanged. Both  B B g t (the green part) and C (the orange part) are the same, resulting in the same value for GIoU.
Batteries 09 00187 g009
Figure 10. Calculation of DIoU.  b g t is the centroid of the ground-truth box, b is the centroid of the prediction box, the green line is  ρ ( b , b g t ) , and the black line is c.
Figure 10. Calculation of DIoU.  b g t is the centroid of the ground-truth box, b is the centroid of the prediction box, the green line is  ρ ( b , b g t ) , and the black line is c.
Batteries 09 00187 g010
Figure 11. DIoU faces the case where one box is contained in another and the centroids of the two boxes coincide. (ac) denote the cases where different B coincide with the centroid of the same  B g t and B is contained inside  B g t . Leads to the same result for DIoU.
Figure 11. DIoU faces the case where one box is contained in another and the centroids of the two boxes coincide. (ac) denote the cases where different B coincide with the centroid of the same  B g t and B is contained inside  B g t . Leads to the same result for DIoU.
Batteries 09 00187 g011
Figure 12. The case of box-to-box symmetry. B at each of the three positions is symmetrical about the midline of  B g t , then the value of each version of IoU calculated from B and  B g t is the same.
Figure 12. The case of box-to-box symmetry. B at each of the three positions is symmetrical about the midline of  B g t , then the value of each version of IoU calculated from B and  B g t is the same.
Batteries 09 00187 g012
Figure 13. The ground-truth and prediction boxes are transferred from the Cartesian coordinate system (left) to the polar coordinate system (right).
Figure 13. The ground-truth and prediction boxes are transferred from the Cartesian coordinate system (left) to the polar coordinate system (right).
Batteries 09 00187 g013
Figure 14. RGB images (a) and depth images (b) captured by the vision sensor in the initial state.
Figure 14. RGB images (a) and depth images (b) captured by the vision sensor in the initial state.
Batteries 09 00187 g014
Figure 15. Results of the initial positioning of the screws.
Figure 15. Results of the initial positioning of the screws.
Batteries 09 00187 g015
Figure 16. OTSU-based binary thresholding process: (a) RGB original image after relaxed crop out; (b) binary image after OTSU-based binary thresholding; (c) binary image after convex wrapping of what is considered to be the screw part (white area), and the black part is considered to be the background.
Figure 16. OTSU-based binary thresholding process: (a) RGB original image after relaxed crop out; (b) binary image after OTSU-based binary thresholding; (c) binary image after convex wrapping of what is considered to be the screw part (white area), and the black part is considered to be the background.
Batteries 09 00187 g016
Figure 17. RGB image (a) and depth image (b) captured by the vision sensor after adaptive pose and position adjustment.
Figure 17. RGB image (a) and depth image (b) captured by the vision sensor after adaptive pose and position adjustment.
Batteries 09 00187 g017
Figure 18. Screw positioning results based on images collected after attitude correction by robotic arm.
Figure 18. Screw positioning results based on images collected after attitude correction by robotic arm.
Batteries 09 00187 g018
Figure 19. Change of loss value during training.
Figure 19. Change of loss value during training.
Batteries 09 00187 g019
Figure 20. Effect of the original YOLOX (a) and the improved YOLOX (b) on screw detection tasks.
Figure 20. Effect of the original YOLOX (a) and the improved YOLOX (b) on screw detection tasks.
Batteries 09 00187 g020aBatteries 09 00187 g020b
Table 1. Information on some parameters of the robotic arm.
Table 1. Information on some parameters of the robotic arm.
ItemParameter
Robot typeUR10e
Eye-in-hand cameraRealsense D435
Weight33.3 kg/73.5 lb
Maximum payload10 kg/22 lb
Extension1300 mm/57.2 in
Range of joints ± 360 for all joints
Torque sensor accuracy5.5 N
Degree of freedom6 rotating joints
System update frequency500 Hz
Table 2. Comparison of screw positioning results with and without autonomous adjustment of screw attitude.
Table 2. Comparison of screw positioning results with and without autonomous adjustment of screw attitude.
MethodmA P50(%)mA P75(%)
Normal   32.92   30.88
With self-calibration   95.73   88.55
Table 3. The information of EVBs.
Table 3. The information of EVBs.
Product ModelStandard VoltageRated CapacityQuality
9201AAA357.1 V298 Ah567.4 kg
9201AFA384.0 V204 Ah540.0 kg
9201ABA376.4 V218.1 Ah514.0 kg
GACBEVA1813771.68 V93 Ah528.0 kg
GACBEVA2905384.0 V132 Ah402.0 kg
Table 4. Information about the components used in the experiment.
Table 4. Information about the components used in the experiment.
Software or HardwareParameters
Robot typeUR10e
eye-in-hand CameraRealsense D435
CPUIntel i9
Memory32G
GPUGeForce RTX 3090
Programming LanguagePython 3.7
Deep learning framePytorch 1.10
Operating SystemUbuntu 20.04
Table 5. Experimental results of different YOLO models with different IoU on YOLOX model.
Table 5. Experimental results of different YOLO models with different IoU on YOLOX model.
MethodmA P50(%)mA P75(%)
TinysmTinysm
YOLOv3-91.24--85.76-
YOLOv4-92.47--87.95-
YOLOv5-94.01--88.64-
YOLOX with IoU(baseline)95.4895.7395.1387.8988.5588.92
YOLOX with GIoU95.895.6995.7390.791.0790.73
YOLOX with DIoU95.3295.4995.289.8488.7190.62
YOLOX with CIoU95.5895.9695.8391.3389.9290.67
YOLOX with LogPolarIoU(Proposed)95.6595.9296.2491.2592.1491.3
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, H.; Zhang, H.; Zhang, Y.; Zhang, S.; Peng, Y.; Wang, Z.; Song, H.; Chen, M. An Accurate Activate Screw Detection Method for Automatic Electric Vehicle Battery Disassembly. Batteries 2023, 9, 187. https://doi.org/10.3390/batteries9030187

AMA Style

Li H, Zhang H, Zhang Y, Zhang S, Peng Y, Wang Z, Song H, Chen M. An Accurate Activate Screw Detection Method for Automatic Electric Vehicle Battery Disassembly. Batteries. 2023; 9(3):187. https://doi.org/10.3390/batteries9030187

Chicago/Turabian Style

Li, Huaicheng, Hengwei Zhang, Yisheng Zhang, Shengmin Zhang, Yanlong Peng, Zhigang Wang, Huawei Song, and Ming Chen. 2023. "An Accurate Activate Screw Detection Method for Automatic Electric Vehicle Battery Disassembly" Batteries 9, no. 3: 187. https://doi.org/10.3390/batteries9030187

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop