Next Article in Journal
Induction Motor Stator Winding Inter-Tern Short Circuit Fault Detection Based on Start-Up Current Envelope Energy
Previous Article in Journal
Rapid Method of Wastewater Classification by Electronic Nose for Performance Evaluation of Bioreactors with Activated Sludge
Previous Article in Special Issue
Micro-Vision Based High-Precision Space Assembly Approach for Trans-Scale Micro-Device: The CFTA Example
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Skill-Learning Method of Dual Peg-in-Hole Compliance Assembly for Micro-Device

Research Center of Laser Fusion, China Academy of Engineering Physics, Mianyang 621900, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(20), 8579; https://doi.org/10.3390/s23208579
Submission received: 30 August 2023 / Revised: 30 September 2023 / Accepted: 12 October 2023 / Published: 19 October 2023
(This article belongs to the Special Issue Intelligent Vision Technology/Sensors for Industrial Applications)

Abstract

:
For the dual peg-in-hole compliance assembly task of upper and lower double-hole structural micro-devices, a skill-learning method is proposed. This method combines offline training in a simulation space and online training in a realistic space. In this paper, a dual peg-in-hole model is built according to the results of a force analysis, and contact-point searching methods are provided for calculating the contact force. Then, a skill-learning framework is built based on deep reinforcement learning. Both expert action and incremental action are used in training, and a reward system considers both efficiency and safety; additionally, a dynamic exploration method is provided to improve the training efficiency. In addition, based on experimental data, an online training method is used to optimize the skill-learning model continuously so that the error caused by the deviation in the offline training data from reality can be reduced. The final experiments demonstrate that the method can effectively reduce the contact force while assembling, improve the efficiency and reduce the impact of the change in position and orientation.

1. Introduction

With the development of a Micro-Electro-Mechanical System (MEMS), the demands for micro-devices are gradually increasing in both type and quality. In order to ensure accuracy and quality, micro-assembly technology is often used to control the assembly tasks.
To reduce the contact force while assembling, two strategies are commonly used: active compliance and passive compliance. A passive compliance strategy mainly produces natural compliance through some auxiliary constructions for devices to adapt to the environment, such as Remote Center Compliance (RCC), Spatial Remote Center Compliance (SRCC), Variable Remote Center Compliance (VRCC), etc. Huang H. [1] focused on the passive compliance tools and designed an adaptive path-planning approach for the robotic grinding and polishing system. Koco E. [2] presented a new hybrid compliance control system for an electrically powered quadruped robot leg, including both active and variable passive compliance parts, by using passive compliance for the filtering of sudden impacts during locomotion.
However, the passive compliance strategy cannot solve the contradiction between rigid parts and flexible insertion demands, the application range is restricted and these kinds of devices often face the problems of poor adaptability, high pertinence and a low success rate of assembly. Therefore, researchers have conducted a lot of research on active compliance strategies. Since the 1970s, with the development of sensors and control technology, force feedback control has developed into one of the main directions of robot-assembly research. The active compliance strategy contains various control methods such as impedance control, force/position hybrid control and learning-based compliance control. Impedance control achieves compliance through the position–velocity–force relationship model of the actuator at the end of the manipulator. It adjusts the robot’s action according to the force feedback. Jin Y. [3] proposed an active compliance control method for the aircraft-cleaning arm, aiming at controlling the contact force in aircraft-skin cleaning by the aircraft-cleaning arm. This method is based on impedance control combined with the adaptive control method to improve the accuracy of the tracking position. And through the fuzzy controller, the impedance parameters are adjusted in real time according to the force error, which improves the accuracy and stability of the task. However, classical impedance and admittance control are environment-dependent, and the force-tracking accuracy cannot meet the requirements [4]. Force/position hybrid control can improve the control accuracy by combining position and force information. Binglong W. [5] proposed a force/position-hybrid-control-based precision peg-in-hole assembly method for industrial robots and designed a new force/position hybrid control strategy based on a servo velocity loop, which can better track high-frequency signals and has a better force-tracking performance than traditional force-control methods. Nevertheless, this kind of method has low efficiency, low accuracy and a slow response for unstructured environments or non-linear assembly tasks.
To solve the problem of compliant assembly and improve the adaptive control performance of the system, considering the uncertainty of the system, including modeling errors, structural uncertainties and other unfavorable conditions [6], researchers have also conducted some research on adaptive systems. Qin X. [7] proposed a load-adaptive control system of Marine hydraulic steering gear. The system adopted a three-closed-loop control structure, using steering gear for data acquisition, and ran a fuzzy PID adaptive control algorithm to output the multi-level controlled quantity. This method enables the system to control the rudder of the Angle of Marine hydraulic steering gear according to the load change and accurately track the load sine signal. With the development of artificial intelligence, researchers have also combined deep learning methods with adaptive systems. Hua O.Y. [8] designed an adaptive PID controller based on the neural network for adjusting internal parameters, which solves the problem of pre-setting PID parameters through manual experience. This system can reduce the dependence on the operator’s experience and avoid problems such as error increasing and the extension of response time caused by the changes in the environment. Xin L. [9] proposed an adaptive PID ship-motion controller based on a neural network, which is composed of a PID controller, neural network unit and ship-rudder-Angle control unit. By using the adaptive ability of the RBF neural network, accurate and fast control of the ship-rudder Angle is realized. Patino H.D. and Liu D. [10] implemented a neural network to address the classical problems in the Model Reference Adaptive System (MRAS), especially for the first-order continuous-time non-linear dynamical system. They designed a controller structure that can employ a radial basis function network or a feed-forward neural network to deal with the non-linearities of the system, ensuring the system can successfully assure the convergence of the control error around zero.
Reinforcement learning, as one of the critical technologies of artificial intelligence, is widely used in control systems because of its excellent ability to explore the environment and achieve adaptive and compliant control. And, a control system always contains the vision module, the force-perception module, the motion module and the grasping module so that the real-time status information can be obtained. Industrial robots can use reinforcement learning methods to autonomously learn optimal moving actions for the peg-in-hole assembly tasks to adapt to complex and variable environments and requirements. Senda K. [11] proposed a reinforcement-learning-based skill-generating method and added perturbations to the object poses for training to improve self-adaptation for uncertain environments. Zou P. [12] designed a variable impedance controller by combining fuzzy theory and Q-learning to ensure the safety, stability and efficiency of the assembly task by a force controller and the learning-based optimization algorithm. However, the method requires many fuzzy rules and is very complex to implement. The neural network can transform high-dimensional data into low-dimensional features by adjusting the internal nodes and has the characteristics of multiprogramming and nonlinear operations, showing a strong numeration power, adaptability, fault-tolerant ability and self-organizing ability. Therefore, researchers have combined neural networks with reinforcement learning and proposed deep reinforcement learning methods for assembly tasks. Beltran-Hernandez C. C. [13] proposed a reinforcement learning and adaptive control method for peg-in-hole assembly to solve the problem of uncertain poses of holes, using the robot state as the input and outputting action through neural networks, such as the trajectory and adaptive parameters, which improves the efficiency of insertion. Feng S. [14] proposed a variable-parameter impedance-control algorithm by reinforcement learning; the controller parameters are optimized by the deep deterministic policy gradient (DDPG) algorithm, and the fuzzy reward is used to avoid getting into the local optimum dilemma. Xu J. [15] proposed a feedback-exploration strategy based on the DDPG algorithm, and a simple P-controller is used to actuate learning. The approach ensures that the assembly task can be completed without analyzing the contact state, but the stability and effectiveness still need to be improved.
Weak micro-devices are small in size and request high accuracy and are more prone to deformation, especially in interference-fit assembly tasks where tiny deviations in pre-registration can lead to deformation and damage during part insertion. Therefore, for such assembly tasks, it is often necessary to continuously adjust the inserting action according to the status information to reduce the contact force and improve the assembly accuracy.
In this paper, a skill-learning compliant inserting method for weak micro-devices based on simulation pre-training together with online learning is proposed, mainly taking the micro-target assembly problem in the Inertial Confinement Fusion (ICF) experiment as an example. Moreover, the proposed method builds a dual peg-in-hole simulation model of the upper and lower double-hole structure according to the force analysis results and pre-trains it. Then, the trained model is used for online learning of the assembly task. We utilized a DDPG algorithm to learn the assembly skills by setting the positive and negative rewards and added noise to both the state and action to improve the environmental adaptability and exploration efficiency of the model. The simulation space pre-training combined with the online training method effectively improves the training efficiency. The feasibility of the method is verified through experiments. Also, the stability, efficiency and safety of the method in the micro-device assembly task are demonstrated through comparative experiments with the classical assembly method.

2. Problem Description and Skill-Learning Framework Design

2.1. Dual Peg-in-Hole Task Description of Upper and Lower Double-Hole Structure

A double-hole part is a typical micro-part in the ICF micro-target composition which has two holes on the top and bottom and is usually a thin-wall metal cavity part. The double-hole part needs to be assembled with the metal shaft, and the insertion depth needs to be kept within limits. The assembly process is shown in Figure 1.
The part is internally hollow, and two holes are punched along the radial direction in the middle position. Since the internal structure of the part can not be observed by microscopic vision, it is impossible to guarantee the alignment attitude before the lower hole is loaded, which corresponds to a definite uncertainty in the internal structure. The part is clamped by vacuum adsorption during the loading, which leads to the clamping deflection if the force during the loading process is extensive, corresponding to a definite uncertainty in the part’s degree of freedom. And because of the thin-wall cavity structure of the double-hole part where the thickness of the wall is usually in the order of 30–50 μm, it is very easy to deform. So, the vacuum-adsorption gripper is used to position it, but it is limited in its binding effect, which can lead to errors in position and orientation when the contact force is strong. Additionally, the clearance required is usually in micrometers. The parts’ batch error can lead to the fit mode from clearance to micro-clearance or even micro-interference. In addition, the mechanical arm is widely used to operate micro-parts, which meets the needs of operating parts in a large space, but the uncertainty of the assembly environment is increased by the instability of the degree of freedom of the multi-joint tandem mechanical arm. Therefore, for the compliance assembly of weak micro-parts in this kind of uncertain environment, the reinforcement learning method is an effective approach. The ability to self-adapt and continuously optimize the loading strategy through reinforcement learning can improve the loading success rate and reduce the impact of excessive contact force.

2.2. Skill-Learning Framework for Dual Peg-in-Hole Compliance Assembly

The following challenges still remain while applying the reinforcement learning methods for online learning directly:
  • The binding force is limited, so the position and orientation are unstable.
  • For weak micro-parts, it is costly to train online directly because the deformation can occur when the contact force is too strong.
  • The control precision is limited due to the instability of the multi-joint robot arm.
  • The robot arms may collide during assembly.
To satisfy the actual needs, we established a network for skill learning combining mechanical-force-analysis-based offline training and experiment-based online training. To improve the training efficiency, expert action based on the P-controller and incremental action output by the network are added for training; this achieves the purpose of compliance assembly, and we adjust the insert action according to the force information.
The framework of skill learning is illustrated in Figure 2. Based on the force analysis of the dual peg-in-hole assembly task of the upper and lower double-hole structure to establish the simulation model for offline training, we add noise to obtain the data of the insertion task to avoid falling into the local minima and use the deep reinforcement learning network to train the intelligent agent. The online learning strategy is based on the feedback of multidimensional force/moment information in real space, and it constantly updates the data in the experience pool for agent optimization.

3. Simulation Model Modeling

Because the network training needs many samples and has easy-to-occur characteristics of deformation, it is easy to cause damage to parts and robots when the operation is not appropriate. Therefore, in order to improve the sample acquisition efficiency and reduce the collision risk and the wastage of components, this section establishes a simulation model by analyzing the contact force of the dual peg-in-hole assembly process, which is used to obtain the pre-training data. The simulation training takes the force information and the number of action steps as the basis for judging success. Whenever the force exceeds the threshold range or the number of execution steps exceeds the threshold range, the assembly will be judged as a failure and the simulation model will be initialized to return to the initial state and start a new round of action.

3.1. Force Analysis of the Assembly Process

In order to analyze the contact force of different states in the assembly process, a mechanical model is established, and a rectangular coordinate system is set up with the center of the very bottom circle O as the coordinate origin, as shown in Figure 3a. According to the symmetry, the spatial assembly model can be simplified to a two-dimensional planar dual peg-in-hole assembly model of upper and lower double-hole structures. We assume that the pegs and holes are point contacts, the pegs are strictly rigid and the holes are elastic. The point with the most considerable deformation is considered as the contact point for analysis.
The state before assembly is shown in Figure 3b; the shaft mainly contains the top characteristic circle B 1 and bottom characteristic circle B 2 , and the hole contains four characteristic circles A 1 , A 2 , A 3 and A 4 , where D is the hole diameter, d is the shaft diameter, l z denotes the Z-directional deviation after alignment and  θ denotes the angular deviation. l 1 is the shaft length, l 2 is the wall thickness of the hole and  l b is the inner-space length.
During the assembly process, there are four phases according to the mandrel inserting depth: the upper-hole approaching phase, upper-hole inserting phase, lower-hole approaching phase and lower-hole inserting phase. Since the upper-hole approaching phase is not subject to contact forces, only the last three phases are taken into consideration.
As shown in Figure 4, the assembly process contains a total of nine contact states. According to the number of contact points, the contact states can be classified into single-point contact, double-point contact and three-point contact. And all contact points can be divided into hole-based contact points and shaft-based contact points, where the main feature of the hole-based contact points is that the contact points are located in the characteristic circle of the shaft, and the main feature of the shaft-based contact points is that the contact points are located in the characteristic circle of the hole. As the inserting depth increases, the contact point position changes, and a different characteristic circle needs to be selected to calculate the contact force.
The contact point shown in Figure 5 is the shaft-based contact point. The characteristic circle at the bottom of the shaft is the reference circle. According to F N = E δ , F f = μ F N , the contact force at the point ( x 1 , y 1 , z 1 ) can be calculated by:
F x 1 = y 1 y 1 2 + x 1 2 · F N x y 1 F y 1 = x 1 y 1 2 + x 1 2 · F N x y 1 F z 1 = F f 1 + F N z 1 = μ F N x y 1 + F N z 1
where E is the Elastic Modulus, δ is the deformation perpendicular to the contact surface and  μ is the friction coefficient which can be obtained by pre-calibration. F N x y 1 and F N z 1 are calculated from the radial deformation Δ x y and axial deformation Δ z 1 , respectively.
Similarly, Figure 6 shows that the contact point is based on the hole, and the characteristic circle at the bottom of the hole is the reference circle. The analysis of point 2 can be obtained:
F x 2 = y 2 y 2 2 + x 2 2 · F N x y 2 F y 2 = x 2 y 2 2 + x 2 2 · F N x y 2 F z 2 = F f 2 = μ F N x y 2
In summary, the assembly process contains a total of nine contact states and two types of contact points, and the total contact force F = [ F x , F y , F z ] can be calculated by choosing different calculation methods according to the contact-point types.

3.2. Training-Data-Acquisition Strategy in the Simulation Environment

According to the force analysis results, theoffline training simulation model was built based on Python. Theconstruction of the simulation environment takes the edge circles of the end faces of the parts as the main calculation object and discretizes the circle as points for searching the contact point. Duringthe training process, assoon as the action information is generated according to the current force situation, themodel will calculate and update the force information according to the contact-point situation; return to the network for the next step; andstore the action, force information and the reward of the current step for training.
The contact-force calculation method is shown in Figure 7, which can be divided into the initial pose-generating phase, contact-point searching phase and contact-force calculating phase. To facilitate the calculation, the characteristic circles of the end faces are used to replace the complex side surface for calculation.
During the initial pose-generating phase, N groups of position and orientation errors are randomly generated to simulate the errors after alignment in the real assembly process, and the initial pose of the mandrel can be calculated:
x i y i z i 1 = x i y i z i 1 · R x · R y · R z · T r
where ( x i , y i , z i ) is the coordinate on the characteristic circle without deviation; ( x i , y i , z i ) is the coordinate on the characteristic circle after initial pose generating; and R x , R y and R z are the rotation matrices around the X, Y and Z axes, respectively. T r is the translation matrix.
Subsequently, based on the analysis of the contact-point type, two searching methods can be obtained: the shaft-based searching method and the hole-based searching method. Take the upper-hole inserting phase as an example: the shaft-based searching method is mainly used for searching shaft-based contact point 1 shown in Figure 5. The center of the characteristic circle of shaft B 2 can be considered as the reference point. The elliptic curve projected from the characteristic circle at the bottom of the hole to the end plane of the shaft can be considered as the searching curve, searching for the point with minimum Δ x y k , whereby k is the index of the point:
Δ x y k = x p x b 2 + y p y b 2 R b
where ( x b , y b , z b ) is the coordinate of the reference point and ( x p , y p , z p ) is the coordinate of the point on the searching curve. R b is the radius of the reference circle. Similarly, the hole-based search method is mainly used to search the hole-based contact point 2 shown in Figure 6. In the upper-hole inserting phase, the center of the characteristic circle of the hole is taken as the reference center, and the elliptic curve projected from the characteristic circle of the shaft on the end face of the hole can be considered as the searching curve, searching for the point with maximal Δ x y k . If there is no such point that satisfies the equation Δ x y k 0 while using the shaft-based search method or satisfies Δ x y k 0 while using the hole-based search method, it is deemed that there is no contact between the two parts. In the contact-force calculating phase, the axial force and radial forces can be obtained by Equation (1) or (2) mentioned in Section 2.1. Finally, the current state s t can be obtained through the calculation of the contact force, and the incremental action a t i can be obtained by the action network. The inserting action a t can be calculated by the Jacobian matrix of force, and the new state s t + 1 can be updated. The reward value can be calculated by the current state s t , incremental action a t i and new state s t + 1 , and these data can be saved for network training.

4. Skill Learning

The DDPG extends DQN into a continuous action space, which can output specific actions corresponding to the maximum long-term benefits and improve stability and convergence. Therefore, the DDPG framework is used in this paper.

4.1. Network Based on DDPG

The DDPG network in this paper is built based on the AC framework, as shown in Figure 8. The action network is used to obtain the incremental action, and the critic network can output the action value to update the actor network and optimize the inserting action. Both the action network and the critic network are whole conjunction neural networks. The inserting action a t contains expert actions a t e and incremental action a t i output by the action network:
a t = a t e + a t i
The expert action a t e = [ d x t e , d y t e , d z t e ] T is used to ensure the safety and success of the insertion and is obtained by the P-controller:
d x t e d y t e d z t e = λ J F f x t f y t 0 + 0 0 δ z
where J F is the Jacobi matrix of force and f x t and f y t are the contact-force components along the X F and Y F , respectively. d x t e , d y t e and d z t e are the components of the expert action along the X M , Y M and Z M , respectively. A small δ z is set to ensure the insertion safety. Constant λ [0,1].
The incremental action is obtained from the actor network: a t i = μ ( s t | θ μ ) , where s t is the state at t moment. According to the chain rule, the parameters θ μ of the actor network can be updated by computing the policy gradient:
θ μ J θ μ = E a Q s t , a t i θ Q · θ μ μ s t θ μ
The critic network takes the state s t and incremental action a t i as inputs, outputs the action values Q ( s t , a t i | θ Q ) and optimizes by minimizing the loss:
L θ Q = E Q s t , a t i θ Q y t 2
where y t = R t + γ Q s t + 1 , μ s t + 1 θ Q and γ is the discount factor. The target action network and the target critic network have the same structure as the action network and the critic network. It periodically updates the parameters of the target network:
θ Q = τ θ Q + ( 1 τ ) θ Q θ μ = τ θ μ + ( 1 τ ) θ μ
where the learning rate τ (0, 1).

4.2. State and Action Definition

During the inserting process, since the inserting action is related to the contact force and the insertion depth at that moment, the state and action at the t moment can be defined as:
s t = f x t , f y t , f z t , p z t T a t = d x t , d y t , d z t T
Since the parts’ orientation is guaranteed by the alignment step, the inserting action mainly aims at controlling the position of parts, where, f x t , f y t and f z t denote the contact force along X F , Y F and Z F , respectively. p z t denotes the total insertion depth along the Z M . d x t , d y t and d z t denote the action along X M , Y M and Z M respectively.

4.3. Reward

In order to ensure the safety of the assembly, while taking efficiency into account, this paper uses the contact force to calculate the positive reward R p t , and the negative reward R n t can be calculated according to the inserting depth:
R t = R p t R n t
Since the part is limited by the fixture step along Y M , the displacement generated by the force is small while the position and orientation along X M rely on the adsorption force, which is more likely to lead to displacement, so the contact forces along X and Y are weighted separately; the larger the contact force, the smaller the reward obtained:
R p t = 1 α 1 f x t f X max α 2 f y t f Y max
where f X m a x and f Y m a x are the maximum allowable radial forces along X F and Y F , respectively. α 1 and α 2 are the contact-force weight factors along X F and Y F , respectively, satisfying α 1 + α 2 = 1 . The negative reward R n t is calculated by the difference between the actual inserting depth and the ideal inserting depth. The larger the difference, the larger the reward value.
Additionally, too long an insertion depth has more impact than too short an insertion depth in order to reduce wastage. Since the ideal inserting depth is related to the current radial force and axial force, we can obtain:
R n t = d z t R p ( t 1 ) d max d max , d z R p ( t 1 ) d max R p ( t 1 ) d max d z t ω d max , d z < R p ( t 1 ) d max
where d z t is the inserting depth, d m a x is the maximum inserting depth allowed in the Z-direction, f z t is the axial force in the Z-direction after acting and the scale factor ω > 1 .

4.4. Environment Imitation and Action Optimization

During reinforcement learning training, in order to ensure the efficiency of exploration and reduce the negative effect due to a large number of random explorations, in this paper, environment and action noises are added to train the intelligent body to complete the exploration task efficiently. First, add Gaussian noise to the action a t :
a t n = a t + N 0 , σ 1 2 I
where a t n is the action with noise; σ 1 is the standard deviation of the exploration action, which determines the intensity of the exploration; and  N ( 0 , σ 1 2 I ) is the Gaussian noise. Reducing the intensity of exploration when the insertion is successful results in:
σ 1 = σ min , succeed σ max , failed
Meanwhile, in order to simulate the effect of environmental changes in real space, noise is added to the state after the execution of the action to improve the model’s adaptive capability:
s t n = s t + N 0 , σ 2 2 I
where s t n is the environment with noise and  σ 2 is the standard deviation of the environmental noise. The pseudocode of the algorithm is shown as Algorithm 1.
Algorithm 1 Algorithm based on DDPG for Dual Peg-in-hole Assembly Task.
  • Initialize parameters σ 1 0.1 , σ 2 0.2 , M E , target network update period T, learning rate τ
  • for epoch = 1,2, …, do
  •     Initialize s 0
  •     for t = 1,2, …, do
  •        Calculate expert action a t e and select incremental action a t i
  •        Calculate exploratory action a t n = a t e + a t i + N ( 0 , σ 1 2 I )
  •        Execute action a t n , observe reward R t and observe new state s t + 1
  •        if  s t + 1 is terminal then
  •            Stop epoch and reset environment state
  •        end if
  •        Store ( s t , a t n , R t , s t + 1 ) in the replay buffer M E
  •        Randomly sample N transitions from M E
  •        Set y t = R t + γ Q s t + 1 , μ s t + 1 θ Q
  •        Update the critic network by minimizing the loss:
  •         L θ Q = E Q s t , a t θ Q y t 2
  •        Update the actor network by using the sampled policy gradient:
  •         θ J θ μ = E a Q s t , a t θ Q · θ μ μ s t θ μ
  •        if t mod T = 0 then
  •            Update target networks:      θ Q = τ θ Q + ( 1 τ ) θ Q θ μ = τ θ μ + ( 1 τ ) θ μ
  •        end if
  •         s t s t + 1
  •     end for  σ 1 = σ min , succeed σ max , failed
  •  end for

5. Experiments

The model-driven deep deterministic policy gradient algorithm can improve the efficiency and interpretability of the algorithm. To ensure the training effect and reduce the parts consumption, firstly, the agent is trained in the simulation. And then it optimizes the network by real-space online training.

5.1. Simulation Training Setup

According to the results of the contact-force analysis, build the dual peg-in-hole simulation model of the upper and lower double-hole structure with the following parameter settings as shown in Table 1.
In the simulation environment, the bore diameter D is 1.01 and the shaft diameter d is 1.0. The length l 1 is 4, the upper and lower bore thicknesses l 2 are 0.5 and the distance l b between the upper and the lower hole is 1.5. The δ is 50. The single step length ( d x t , d y t , d z t ) is controlled within the range of [20, 20, 60], respectively; f X m a x is 70 and f Y m a x is 100; and the upper-limit value of the axial contact force is 1000. When the contact force exceeds the allowed range or the maximum number of steps exceeds 1000, the assembly task can be diagnosed as a fail.
We randomly generated 300 groups of the initial state with position deviation within 10 μ m and orientation deviation within 0.3 . The results are shown in Figure 9.

5.2. Online Training Setup

For online training, the assembly object is a decussation target where the shaft material is aluminum with a diameter d of 1 mm and l 1 length of 4 mm, the cavity material is gold with a hole diameter D of 1.01 mm, the distance l b between the upper and lower holes is 2.5 mm, the wall thickness l 2 is 0.5 mm and the gold cavity is fixed by the gripper beyond the force sensor.
The assembly experiment platform mainly includes a six DOF industrial robot and a six DOF operating platform. The detection module includes a variable magnification vision system, vertical servo vision system and micro-force detection system. The variable magnification vision system mainly consists of two horizontal micro-vision channels and is installed on the three DOF motion platform. The vertical servo vision and the gripper are fixed at the end of the robot. The micro-force detection module collects force information by a force sensor.
To improve the training effect, the network parameters are further optimized by online training considering the environment information of the real assembly task, the insertion data are stored and the network is periodically updated.
The online learning method exchanges the state and action information through socket communication technology and continuously stores the data of each actual insertion in the experience replay of size 200 in the form of ( s t , a t i , R t , s ( t + 1 ) ), and it updates the network weights. To ensure the assembly safety, the operation station motion speed is limited to 5 mm/s. To a certain extent, it can reduce the impact of the error between the simulation environment and the actual insertion environment and builds a bridge between the intelligent body and the real assembly environment.

5.3. Assembly Experiment Results Based on the Proposed Skill-Learning Method

Before each insertion, the alignment between two parts is performed by using visual inspection, including the rough alignment based on the robot arm and the fine alignment based on the operation stage, with the final position deviation controlled at 2 μ m and the orientation deviation controlled at 0.1 .
Due to the upper and lower double-hole structure, there are two sections with large contact forces while inserting, as shown in Figure 10. The maximum force along X F is controlled within 100 mN, the maximum force along Y F is controlled within 40 mN, the maximum axial force is controlled within 90 mN and the insert step is 53.
Based on the trained model, 30 inserting experiments were carried out and the model was continuously optimized by online training, and the experimental results are shown in Figure 11. The forces are mainly concentrated in the [0, 44] mN interval, meaning they cannot easily cause the deformation of weak micro-parts and the deflection of posture, and they have a good assembly effect.

5.4. Comparative Experiment Results and Analysis

Since the assembly is a static process, the conventional P-controller was used for comparison experiments, setting the gain to 800 μ m/N, the Z-direction single-step length to 60 μ m, the total loading stroke to 3200 μ m and the finish axial force to 1000 mN. This method is more dependent on the human-set parameters, and the Z-direction step length cannot be adaptively adjusted so that force oscillation may occur during the insertion process, and the three-point contact situation, as shown in Figure 4i, may happen.
Comparing the results of the traditional P-control and the trained model based on a combination of offline and online training for the dual peg-in-hole assembly task, as shown in Figure 12, it can be found that the two methods have similar contact forces during the upper-hole inserting phase, mainly because the fine alignment makes the initial position and orientation deviation smaller, while the traditional method tends to have a much larger contact force during the lower-hole inserting than the reinforcement learning inserting method, on account of the deviation caused by the unavailability of the correct contact force during the lower-hole approaching phase while using the P-controller. Meanwhile, insertion using the reinforcement learning method can adjust the step length along Z adaptively according to the force information and reduce the damage of part deformation. Therefore, for the dual peg-in-hole task of the upper and lower double-hole structure, the combination of the offline and online reinforcement learning method can reduce the contact force and avoid the failure of assembly due to the part posture deviation. In summary, the method proposed in this paper can perform a dual peg-in-hole assembly task of upper and lower double-hole structures well with a higher assembly success rate and less contact force, improves the assembly safety, reduces the hidden danger of the deformation of weak micro-parts by contact force and requires less actual samples for pre-training in the simulation environment, which improves the training efficiency and reduces the cost.

6. Conclusions

For the compliance assembly tasks of micro-devices, to avoid the impact on the assembly results due to the limited positional constraint force, part deformation and instable degrees of freedom of the multi-joint robot arm, a skill-learning model for the dual peg-in-hole task of upper and lower double-hole structures based on the combination of offline training in a simulation space and online learning in a real space is proposed. The method is based on the AC framework, using the DDPG algorithm and designing positive and negative rewards for the agent to train according to the part characteristics, adding action and environmental noise to simulate the assembly environment and improving the exploration efficiency. In order to improve the algorithm interpretability and the efficiency of training data acquisition, a simulation model is built to generate assembly data for pre-training. Then, the model is continuously optimized by online training methods. The effectiveness and advantages of the proposed method were demonstrated through comparison experiments, and the results show that the method proposed in this paper can finish the assembly task with less contact force and has a strong self-adaptive capability for the environment, which is essential for improving the safety and efficiency of the peg-in-hole assembly task of micro-devices. Meanwhile, for other peg-in-hole assembly tasks with complicated structures, the model can be adapted to more assembly tasks by adjusting the simulation model. Subsequently, model matching and selecting can be integrated to complete the skill transfer of multiple assembly tasks.

Author Contributions

Y.W. wrote the paper and conceived and designed the experiments; J.Z. carried out the experimental verification; Y.Y. designed computer programs; W.W. provided some suggestions about the overall work; K.D. supervised the overall work and reviewed the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Key Research and Development Program of China (2018AAA0103005).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, H.; Gong, Z.; Chen, X.; Zhou, L. Robotic grinding and polishing for turbine-vane overhaul. J. Mater. Process. Technol. 2002, 127, 140–145. [Google Scholar] [CrossRef]
  2. Koco, E.; Mirkovic, D.; Kovai, Z. Hybrid Compliance Control for Locomotion of Electrically Actuated Quadruped Robot. J. Intell. Robot. Syst. 2019, 94, 537–563. [Google Scholar] [CrossRef]
  3. Jin, Y.; Tian, Z.; Li, K.; Xiong, H. Research on active compliance control of aircraft’s washing arm based on impedance control. J. Mach. Des. 2022, 39, 84–89. [Google Scholar]
  4. Yin, W.; Lian, D.; Li, K.; Zhao, G. Manipulator force/position hybrid control based on staged adaptation. J. Beijing Univ. Aeronaut. Astronaut. 2022, 1–9. [Google Scholar] [CrossRef]
  5. Wu, B.L.; Qu, D.K.; Xu, F. Industrial robot high precision peg-in-hole assembly based on hybrid force/position control. J. Zhejiang Univ. (Eng. Sci.) 2018, 52, 379–386. [Google Scholar] [CrossRef]
  6. Anavatti, S.G.; Santoso, F.; Garratt, M.A. Progress in adaptive control systems: Past, present, and future. In Proceedings of the 2015 International Conference on Advanced Mechatronics, Intelligent Manufacture, and Industrial Automation (ICAMIMIA), Surabaya, Indonesia, 15–17 October 2015; pp. 1–8. [Google Scholar] [CrossRef]
  7. Qin, X. Design of load adaptive control system for marine gear type hydraulic steering gear. Ship Sci. Technol. 2023, 45, 147–150. [Google Scholar]
  8. Hua, O.Y. The Design of Adaptive PID Controller Based on Neural Network. Mach. Des. Manuf. 2014, 263–265. [Google Scholar] [CrossRef]
  9. Xin, L. Research on adaptive PID ship motion controller based on neural network algorithm. Ship Sci. Technol. 2018, 40, 28–30. [Google Scholar]
  10. Patino, H.; Liu, D. Neural network-based model reference adaptive control system. IEEE Trans. Syst. Man Cybern. Part (Cybern.) 2000, 30, 198–204. [Google Scholar] [CrossRef]
  11. Senda, K.; Tani, Y. Autonomous Robust Skill Generation Using Reinforcement Learning with Plant Variation. Adv. Mech. Eng. 2015, 6, 276264. [Google Scholar] [CrossRef]
  12. Zou, P.; Zhu, Q.; Wu, J.; Xiong, R. Learning-based Optimization Algorithms Combining Force Control Strategies for Peg-in-Hole Assembly. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 25–29 October 2020; pp. 7403–7410. [Google Scholar] [CrossRef]
  13. Beltran-Hernandez, C.C.; Petit, D.; Ramirez-Alpizar, I.G.; Harada, K. Variable Compliance Control for Robotic Peg-in-Hole Assembly: A Deep-Reinforcement-Learning Approach. Appl. Sci. 2020, 10, 6923. [Google Scholar] [CrossRef]
  14. Shuang, F.; Lu, W.; Li, S.; Yuan, X. Robotic Peg-in-Hole Assembly Strategy Research Based on Reinforcement Learning Algorithm. ROBOT 2023, 45, 321–332. [Google Scholar] [CrossRef]
  15. Xu, J.; Hou, Z.; Wang, W.; Xu, B.; Zhang, K.; Chen, K. Feedback Deep Deterministic Policy Gradient with Fuzzy Reward for Robotic Multiple Peg-in-Hole Assembly Tasks. IEEE Trans. Ind. Inform. 2019, 15, 1658–1667. [Google Scholar] [CrossRef]
Figure 1. Diagrams of the assembly process.
Figure 1. Diagrams of the assembly process.
Sensors 23 08579 g001
Figure 2. This is the framework of skill learning.
Figure 2. This is the framework of skill learning.
Sensors 23 08579 g002
Figure 3. Basic information of force analysis model.
Figure 3. Basic information of force analysis model.
Sensors 23 08579 g003
Figure 4. Results of assembly phase analysis: (ac) upper-hole inserting phase; (d,e) lower-hole approaching phase; and (fi) lower-hole inserting phase.
Figure 4. Results of assembly phase analysis: (ac) upper-hole inserting phase; (d,e) lower-hole approaching phase; and (fi) lower-hole inserting phase.
Sensors 23 08579 g004
Figure 5. Shaft-based contact-point force analysis.
Figure 5. Shaft-based contact-point force analysis.
Sensors 23 08579 g005
Figure 6. Hole-based contact-point force analysis.
Figure 6. Hole-based contact-point force analysis.
Sensors 23 08579 g006
Figure 7. Data acquisition flow chart from simulation model.
Figure 7. Data acquisition flow chart from simulation model.
Sensors 23 08579 g007
Figure 8. The framework of DDPG algorithm.
Figure 8. The framework of DDPG algorithm.
Sensors 23 08579 g008
Figure 9. Simulation training result.
Figure 9. Simulation training result.
Sensors 23 08579 g009
Figure 10. Contact force while inserting.
Figure 10. Contact force while inserting.
Sensors 23 08579 g010
Figure 11. Statistical diagram of 30 groups of experimental forces.
Figure 11. Statistical diagram of 30 groups of experimental forces.
Sensors 23 08579 g011
Figure 12. Contact force comparison.
Figure 12. Contact force comparison.
Sensors 23 08579 g012
Table 1. Training parameters.
Table 1. Training parameters.
Parameters 1ValueParameters 3Value
σ m a x , σ m i n 0.1,0.4Experience replay size M E 200
α 1 , α 2 0.7,0.3Learning rate τ 0.001
ω 2Batch size N32
γ 0.99Self-learning rounds200
λ 0.2Network update period T0.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, Y.; Zhang, J.; Yang, Y.; Wu, W.; Du, K. Skill-Learning Method of Dual Peg-in-Hole Compliance Assembly for Micro-Device. Sensors 2023, 23, 8579. https://doi.org/10.3390/s23208579

AMA Style

Wu Y, Zhang J, Yang Y, Wu W, Du K. Skill-Learning Method of Dual Peg-in-Hole Compliance Assembly for Micro-Device. Sensors. 2023; 23(20):8579. https://doi.org/10.3390/s23208579

Chicago/Turabian Style

Wu, Yuting, Juan Zhang, Yi Yang, Wenrong Wu, and Kai Du. 2023. "Skill-Learning Method of Dual Peg-in-Hole Compliance Assembly for Micro-Device" Sensors 23, no. 20: 8579. https://doi.org/10.3390/s23208579

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop