Next Article in Journal
Closed-Loop Current Stimulation Feedback Control of a Neural Mass Model Using Reservoir Computing
Previous Article in Journal
A Comparative Study in Forming Behavior of Different Grades of Steel in Cold Forging Backward Extrusion by Integrating Artificial Neural Network (ANN) with Differential Evolution (DE) Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reinforcement Learning-Based Adaptive Position Control Scheme for Uncertain Robotic Manipulators with Constrained Angular Position and Angular Velocity

School of Mechanical Engineering, Xiangtan University, Xiangtan 411105, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(3), 1275; https://doi.org/10.3390/app13031275
Submission received: 6 December 2022 / Revised: 11 January 2023 / Accepted: 16 January 2023 / Published: 18 January 2023

Abstract

:
Aiming at robotic manipulators subject to system uncertainty and external disturbance, this paper presents a novel adaptive control scheme that uses the time delay estimation (TED) technique and reinforcement learning (RL) technique to achieve a good tracking performance for each joint of a manipulator. Compared to conventional controllers, the proposed control scheme can not only handle the system parametric uncertainty and external disturbance but also guarantee both the angular positions and angular velocities of each joint without exceeding their preset constraints. Moreover, it has been proved by using Lyapunov theory that the tracking errors are uniformly ultimately bounded (UUB) with a small bound related to the parameters of the controller. Additionally, an innovative RL-based auxiliary term in the proposed controller further minimizes the steady state tracking errors, and thereby the tracking accuracy is not compromised by the lack of asymptotic convergence of tracking errors. Finally, the simulation results validate the effectiveness of the proposed control scheme.

1. Introduction

Robotic manipulators have been extensively applied in assisting or even replacing humans to perform the various tasks such as assembling [1], machine operation [2], deburring [3], drilling [4], transportation [5] and manufacturing [6,7]. In order to successfully perform such industrial tasks, it is essential to well control the motion of robotic manipulators. However, the highly nonlinear and uncertain dynamics of robotic manipulators result in the inapplicability of many traditional controllers designed for linear systems such as linear quadratic control (LQC) [8,9] and linear H∞ control [10,11]. Hence, many researchers have been motivated to place efforts on designing advanced controllers for robotic manipulators.
State feedback linearization requiring a known dynamic model can transform the nonlinear system to a linear form [12]. However, the dynamic model of robotic manipulators is always uncertain because of unknown external disturbances and uncertain system parameters. To handle the issue of uncertainty, many efforts have been made such as sliding mode control (SMC) [13,14,15,16,17,18], fuzzy logic system (FLS)-based control [19,20,21], neural network (NN)-based control [22,23,24,25,26], and disturbance observer (DOB)-based control [27,28,29]. More precisely, Zhang et al. [13] propose a fixed time sliding mode control for uncertain robotic manipulators in which a conservative switching gain requiring the upper bound of lumped uncertainty is used. In [21], an adaptive controller-based T-S (Takagi-Sugeno) fuzzy system is designed, and the modified T-S fuzzy system can efficiently approximate the unknown model of robotic manipulators. Hu et al. [26] present a multiple-layer neural network-based controller that can achieve a high accuracy of motion control of robotic manipulators subject to unknown disturbances. In [28], a high-order-sliding-mode differentiator (HOSMD)-based estimator is designed to compensate the mismatched uncertainties without the need of assuming a bounded uncertainty.
Besides the uncertainty of the dynamic model, state constraint is also a common problem of robotic manipulators, which has attracted much attention. For example, Sun et al. [30] proposed an adaptive neural network control scheme considering full-state constraints of robotic manipulators. In [31], an adaptive fuzzy control scheme that can guarantee the constrained output is presented. Yu et al. [32] designed an adaptive fuzzy control scheme working with a disturbance observer for the manipulators with full-state constraints. Nevertheless, those controllers can only achieve either the constrained tracking error of joint angle or the constrained error between the actual angular velocity and the virtual control signal. In other words, controllers in [30,31,32] do not guarantee the angular position and angular velocity for each joint of the robotic manipulator, within the preset constraints. Recently, Yang et al. [33] designed a new adaptive control scheme that can guarantee the angular position for each joint of a robotic manipulator to never exceed the preset constraints. However, the constrained angular velocity of each joint is not guaranteed.
In the light of reviewing the existing literature, the following issues need to be further solved:
  • In order to safely perform robotic manipulators, both the angular position and angular velocity of each joint of robotic manipulators should be controlled to not exceed the preset constraints. More precisely, the angular position (rotation angle) of each joint should be always within a reasonable range to have no risk on physically breaking the joint. Similarly, the angular velocity of each joint should not exceed its maximum related to the maximum rotational speed of the driving motor;
  • For some existing controllers (e.g., [9,23,26,30,31,32]), the tracking accuracy could be compromised due to the bounded result of tracking errors and uncertainty-estimation errors. Therefore, it is needed to avoid the loss of tracking accuracy caused by the lack of asymptotic convergence of tracking errors.
To handle the above issues, this paper proposes a new adaptive control scheme that utilizes time delay estimation (TDE) and reinforcement learning (RL) for n -link robotic manipulators. In the proposed control scheme, the multiple-input-and-multiple-output (MIMO) robotic system is initially decomposed into n single-input-single-output (SISO) subsystems by TDE. Each subsystem is with an unknown bounded TDE error. After that, the novel virtual control law for each subsystem is designed, which can not only achieve the boundness of tracking errors for each joint in the presence of TDE error, but can also guarantee both the angular position and angular velocity for each joint to be not exceeding the preset constraints. To improve the tracking accuracy, an RL-based term in the virtual control law is designed, which automatically learns the optimal parameters of the controller in the different system states. RL is an artificial intelligence technique that gradually explores the optimal policy by interacting with the environment, which has attracted many interests in the control of robotic manipulators such as [34,35,36]. Particularly, the RL-based term in this paper is designed to avoid the violation of the boundness of tracking errors, even if a bad policy is tried by RL, which ensures a safe environment for implementing RL. The tracking errors are proven to be uniformly ultimately bounded (UUB) via Lyapunov theory. Simulation results indicate the proposed control scheme can achieve a high tracking accuracy in the presence of model uncertainty and unknown disturbances.
The major merits of this paper include the following points:
  • Compare with some existing research [30,31,32,33], in addition to the basic achievement of the uniformly ultimately bounded (UUB) tracking error for each joint in the presence of TDE error; the control scheme can guarantee both the angular position and angular velocity for each joint to be not exceeding the preset constraints;
  • The novel adaptive gain in (13) results in the smooth control torques to reduce the chattering effect caused by switching term in (10). Meanwhile, an RL-based term can effectively improve the tracking accuracy, which thereby reduces the possible steady-state tracking errors caused by the lack of asymptotic convergence;
  • The mathematical expression of the controller is simple, meanwhile, any prior knowledge of upper bounds caused by an imprecise model are unnecessary in our control scheme.
The rest of this paper is organized as follows: in Section 2, the dynamics model of n-link robotic manipulators is given, and the control objective is described. In Section 3, the RL-based adaptive control scheme is proposed and the proof of stability is given. In Section 4, the numerical simulation is conducted to verify the effectiveness of the proposed controller. The conclusion is given in Section 5.

2. Dynamical Model and Problem Statement

The dynamic model of n-link robot manipulators is shown as the following:
M ( q ( t ) ) q ¨ ( t ) + C ( q ( t ) , q ˙ ( t ) ) q ˙ ( t ) + G ( q ( t ) ) + F ( q ˙ ( t ) ) = τ ( t ) + τ d ( t )
where M ( q ( t ) ) Rn×n is the inertia matrix, q ( t ) = [ q 1 ( t ) , q 2 ( t ) , .. , q i ( t ) , .. , q n ( t ) ] T Rn is the vector of angular positions of joints of manipulator. C ( q ( t ) , q ˙ ( t ) ) Rn×n is the Coriolis and centrifugal matrix. G ( q ( t ) ) Rn is the gravity vector. F ( q ˙ ( t ) ) Rn is the vector of friction. τ (t) Rn is the vector of torques applied on the joints. τ d ( t )   R n is the external disturbance.
The model (1) can be further written as (2) to indicate the system uncertainty.
q ¨ ( t ) = M ( q ( t ) ) 1 [ C ( q ( t ) , q ˙ ( t ) ) q ˙ ( t ) G ( q ( t ) ) F ( q ˙ ( t ) ) + τ d ( t ) ] + [ M ( q ( t ) ) 1 M ^ ( q ( t ) ) 1 ] τ ( t )         + M ^ ( q ( t ) ) 1 τ ( t ) = Γ ( t ) + M ^ ( q ( t ) ) 1 τ ( t )
where M ^ ( q ( t ) ) is the estimation of M ( q ( t ) ) . Γ ( t ) = M ( q ( t ) ) 1 [ C ( q ( t ) , q ˙ ( t ) ) q ˙ ( t ) G ( q ( t ) ) F ( q ˙ ( t ) ) + τ d ( t ) ] + [ M ( q ( t ) ) 1 M ^ ( q ( t ) ) 1 ] τ ( t ) is the system uncertainty.
The vector of error between the desired angular position and actual angular position is defined in (3):
e ( t ) = q ( t ) q d
where q d = [ q d 1 , q d 2 , , q d n ] T R n is the vector of desired angular position of joints. e = [ e 1 , e 2 , , e n ] T R n is the tracking error vector.
The main control objective is to design the torque τ ( t ) that can drive all the joints of the robotic manipulator system (1) to approach their desired angular positions. Meanwhile, both the angular positions and angular velocities for each joint should be guaranteed to not exceed the given constraints. The control objective can be described by (4)–(6).
| | e ( t ) | | σ ,     t 0
| q i ( t ) | < ε i ,     t 0 ,   i = 1 , 2 , .. , n
| q ˙ i ( t ) | < Λ i ,     t 0 ,   i = 1 , 2 , .. , n
where σ * > σ 0 , and σ * > 0 is a positive constant, ε i > 0 is a positive constant referring to the angular position constraint of the i t h joint, and Λ i > 0 is a positive constant, meaning the angular velocity constraint of the i t h joint.
Remark 1. 
The positive constant σ * is related to the initial state of the system and parameters of the controller. σ reflects the tracking accuracy. Notably, σ ( t ) = 0 means the asymptotic convergence of tracking errors.
Remark 2. 
In this paper, we consider the angular position tracking problem for each joint of the manipulator. Hence, q d i is a constant for i = 1 , 2 , .. , n . , which means q ˙ d i = 0 . The angular trajectory tracking problem will be considered in our future work.

3. Controller Design and Stability Analysis

In this part, the adaptive RL-based controller working with TDE is developed. After that, the stability is proven by using Lyapunov theory.

3.1. Controller Design

The TDE technique is applied to handle the system uncertainty in (2):
Γ ^ ( t ) Γ ( t L ) = q ¨ ( t L ) M ^ 1 ( q ( t L ) ) τ ( t L )
where Γ ^ ( t ) = [ Γ ^ 1 , Γ ^ 2 , .. , Γ ^ n ] T R n is the estimate of Γ ( t ) . L > 0 is the sampling time of TDE.
Lemma 1. 
[14] The TDE error of robotic manipulator (2) is bounded such that | Γ i ( t ) Γ ^ i ( t ) | Γ i * (for i = 1 , 2 , , n ) if the following condition is satisfied:
| | I M 1 ( q ( t ) ) M ^ ( q ( t L ) ) | | 2 < 1
where Γ i * is an unknown positive constant.
The control law working with TDE technique is designed as follows:
τ ( t ) = M ^ ( q ( t ) ) [ Γ ^ ( t ) + u ( t ) ]  
where u ( t ) = [ u 1 ( t ) , u 2 ( t ) , , u n ( t ) ] T R n is the virtual control law.
The virtual control law in (9) is designed as follows:
u i = d ^ i s g n ( q ˙ i ) 1 1 + k y i C i [ k p i e i + k d i q ˙ i + k s i A i q i + k s i B i z i + λ i   q ˙ i + λ i Λ i t a n h ( e i   i ) ] i = 1 ,   2 ,     , n  
where k p i , k s i , k d i ,   i and k y i are the positive constants determined by users. λ i > 0 is a positive variable determined by the fuzzy reinforcement learning mechanism.
A i = ε i 2 z i 2 ( ε i 2 q i 2 ) 2 ,   B i = ε i 2 ε i 2 q i 2 ,   C i = 2 Λ i 2 ( Λ i 2 q ˙ i 2 ) 2
where ε i > 0 is the restricted upper bound of angular position of the i t h joint. Λ i > 0 is the restricted upper bound of angular velocity of the i t h joint.
And the variable z i is defined as follows:
z i = e i + 0 t η i ( Θ ) d Θ ,   η i = β i z i
where β i is a positive constant.
The d ^ i is used to handle the bounded TDE error and the update law of d ^ i is designed as follows.
d ^ ˙ i = { ψ i ( 1 + k y i C i ) | q ˙ i |   ,   i f ( d ^ i 0 )   o r   ( Ω i > Ω ¯ i ) ψ i δ i ( 1 + k y i C i ) | q ˙ i |   ,   i f ( d ^ i > 0 )   a n d   ( Ω i Ω ¯ i )
where ψ i , δ i and Ω ¯ i are positive constants. The variable Ω i is defined as (14).
Ω i = 1 2 q ˙ i 2 + 1 2 k p i e i 2 + 1 2 k s i ε i 2 z i 2 ε i 2 q i 2 + k y i q ˙ i 2 Λ i 2 q ˙ i 2 + λ i i ln [ cosh ( e i   i ) ]  
Remark 3. 
d ^ i is to guarantee the stability in the presence of the bounded TDE errors. A conservative update law of d ^ i is to monotonously increase the value of d ^ i such that d ^ ˙ i = ψ i ( 1 + k y i C i ) | q ˙ i | . Although such adaptive law can achieve the bounded tracking errors and angular velocities, the great value of d ^ i could result in the chattering effect on the calculated control torques. Therefore, a novel adaptive law for d ^ i shown in (13) is proposed to mitigate the chattering effect by decreasing the value of d ^ i without the breach of stability of system. The proof will be given later. Moreover, d ^ i > 0 holds because d ^ i 0 leads to d ^ ˙ i = ψ i ( 1 + k y i C i ) | q ˙ i | 0 .
Remark 4. 
Similar to [33], the terms A i q i and B i z i in (10) are to guarantee the angular position of each joint of manipulator to not exceed the preset constraint ± ε i . While the term 1 1 + k y i C i in (10) is to guarantee the angular velocity of each joint to not exceed the preset constraint ± Λ i , which was not achieved in [33]. The proof will be detailed later.

3.2. Stability Analysis

Theorem 1. 
If the initial angular position and velocity of all joints are within their preset constraints such that | q i ( 0 ) | < ε i and | q ˙ i ( 0 ) | < Λ i , and (8) in lemma 1 holds, the control law consisting of (7) and (9)–(14) can achieve the uniformly ultimately bounded tracking errors of robotic manipulator system (1). Meanwhile, the angular velocity and angular position of each joint of manipulator are within the preset constraints such that | q i ( t ) | < ε i and | q ˙ i ( t ) | < Λ i ,   t > 0 . Namely, the control target (4)–(6) is achieved.
Proof. 
By inserting (9) into (2) and using the fact of M ^ 1 M ^ = I with I = d i a g ( [ 1 , 1 , .. , 1 ] ) R n × n , the MIMO robotic manipulator system can be decoupled into n uncertain SISO subsystems (15).
q ¨ i ( t ) = u i ( t ) + d i ( t ) i = 1 , 2 , , n
where q i q is the angular position of the i t h joint. u i is designed in (10). d i ( t ) = Γ i ( t ) Γ ^ i ( t ) is the TDE error.
The following Lyapunov candidate is designed:
V = i = 1 n V i = i = 1 n [ 1 2 q ˙ i 2 + 1 2 k p i e i 2 + 1 2 ψ i d ˜ i 2 + k s i ε i 2 z i 2 2 ( ε i 2 q i 2 ) + k y i q ˙ i 2 Λ i 2 q ˙ i 2 + λ i   Λ i i ln ( cos h ( e i i ) )   ]
where d ˜ i = Γ i * d ^ i . Γ i * is the upper bound TDE error defined in Lemma 1.
Remark 5. 
Clearly, k s i ε i 2 z i 2 2 ( ε i 2 q i 2 ) > 0 holds as long as | q i | < ε i . k y i q ˙ i 2 Λ i 2 q ˙ i 2 > 0 as long as | q ˙ i | < Λ i . Furthermore, Λ i i l n ( c o s h ( e i i ) )   0 holds because of the fact of c o s h ( · ) 1 and the fact of l n ( x ) 0 with x 1 . Therefore, the Lyapunov candidate (16) is positive defined as long as | q i | < ε i and | q ˙ i | < Λ i .
Taking the derivative of the Lyapunov function (16) with respect to the time t and using (11), (12) and (15), we have:
V ˙ = i = 1 n [ q ˙ i q ¨ i + k p i e i q ˙ i 1 ψ i d ˜ i d ^ ˙ i +   k s i A i q i q ˙ i + k s i B i z i z ˙ i + k y i C i q ˙ i q ¨ i + λ i Λ i t a n h ( e i   i ) q ˙ i ]                                   = i = 1 n q ˙ i [ ( 1 + k y i C i ) ( u i + d i ) + k p i e i +   k s i A i q i +   k s i B i z i   + λ i Λ i t a n h ( e i   i ) ]                                   i = 1 n ( 1 ψ i d ˜ i d ^ ˙ i +   k s i B i z i   η i )
Substituting (10) and into (17) with the fact of C i > 0 as long as | q ˙ i | < Λ i :
V ˙ = i = 1 n q ˙ i ( 1 + k y i C i ) [ d ^ i s g n ( q ˙ i ) + d i ] i = 1 n 1 ψ i d ˜ i d ^ ˙ i i = 1 n   k s i B i β i z i 2 i = 1 n ( λ i + k d i ) q ˙ i 2   i = 1 n | q ˙ i | ( 1 + k y i C i ) ( Γ i * d ^ i ) i = 1 n 1 ψ i d ˜ i d ^ ˙ i = i = 1 n [ | q ˙ i | ( 1 + k y i C i ) 1 ψ i d ^ ˙ i ] ( Γ i * d ^ i )
Therefore, for each V i , we can derive (19) by combining (13) and (18):
V ˙ i {   0   ,   i f ( d ^ i 0 )   o r   ( Ω i Ω ¯ i )     x i ( Γ i * d ^ i ) ,   i f ( d ^ i > 0 )   a n d   ( Ω i < Ω ¯ i )
where x i = | q ˙ i | ( 1 + k y i C i ) + δ i | q ˙ i | ( 1 + k y i C i ) . Clearly, x i 2 δ i > 0 holds.
In addition, the V i defined in (16) can be written as:
V i = Ω i + 1 2 ψ i d ˜ i 2
where Ω i is defined in (14).
To prove the Lyapunov function, V i is bounded by a positive constant and we assume a sufficiently large constant V i * . Clearly, the sufficiently large V i * requires at least one of the terms ( Ω i or d ˜ i 2 ) to be sufficiently large. If Ω i is sufficiently large such that Ω i > Ω ¯ i , then V ˙ i 0 holds according to (19). If d ˜ i 2 = ( Γ i * d ^ i ) 2 is sufficiently large such that d ˜ i 2 > ( 2 Γ i * ) 2 , then d ^ i > 3 Γ i * will hold because of the facts of Γ i * > 0 (Lemma 1) and d ^ i 0 (remark 3), which further means Γ i * d ^ i < 0 holds. As a result, according to (19) and the fact of x i 2 δ i > 0 , V ˙ i 0 can hold by Ω i > Ω ¯ i or d ˜ i 2 > ( 2 Γ i * ) 2 , which means V ˙ i 0 holds if V i Ω ¯ i + 1 2 ψ i ( 2 Γ i * ) 2 holds. It is thereby easy to conclude that V i is bounded by V i * such that V i V i * = max { V i ( 0 ) ,   Ω ¯ i + 1 2 ψ i ( 2 Γ i * ) 2 } .
Therefore, the tracking error of the i t h subsystem is bounded because of the fact of 1 2 k p i e i 2 V i .
| e i | 2 k p i max { V i ( 0 ) ,   ( Ω ¯ i + 1 2 ψ i ( 2 Γ i * ) 2 ) i = 1 , 2 , .. , n
Then, the norm of the vector of the tracking error is thereby bounded.
| | e ( t ) | | i = 1 n 2 k p i max { V i ( 0 ) , [ Ω ¯ i + 1 2 ψ i ( 2 Γ i * ) 2 ]
Moreover, the bounded V i implies all terms in V i are bounded. The (23) therefore holds, which means | q i | < ε i and | q ˙ i | < Λ i hold.
ε i 2 z i 2 2 ( ε i 2 q i 2 ) ,   q ˙ i 2 Λ i 2 q ˙ i 2 L i = 1 , 2 , , n

3.3. Fuzzy Q Reinforcement Learning Mechanism Determining Parameters of Controller

In this section, a fuzzy Q reinforcement learning mechanism is designed to tune the parameter λ i to improve the tracking accuracy. The motivations of RL are detailed in remark 7.
Lemma 2. 
If (24) holds, the tracking error e i defined in (3) will asymptotically converge to zero with the converging rate satisfying | q ˙ i | < Λ i .
0 = q ˙ i + Λ i t a n h ( e i   i ) ,   i = 1 , 2 , .. n
Proof. 
A simple Lyapunov function is given in (25).
V i = 1 2 e i 2
Combining (24) and derivative of (25), we can obtain (26).
V ˙ i = Λ i e i t a n h ( e i   i ) Λ i i t a n h 2 ( e i   i ) 0
V ˙ i 0 implies V i L . And then, (27) can be obtained by integration on both sides of (26).
V i ( ) V i ( 0 ) Λ i i 0 t a n h 2 ( e i   i )
In the light of Barbalet’s Lemma, (27) implies t a n h ( e i ( )   i ) = 0 , which means e i ( ) = 0 . Moreover, it is clear that | q ˙ i | = | Λ i t a n h ( e i   i ) | < Λ i due to the fact of tanh ( · ) < 1 .
Remark 6. 
The asymptotic convergence of tracking errors in Lemma 2 is stronger than the boundness of tracking errors that we achieved in Theorem 1, which means a better tracking accuracy. It is because asymptotic convergence means e i is eventually going to zero, while boundedness only implies | e i | bounded by a positive constant. Therefore, the proposed controller can be improved with a better tracking accuracy by finding the optimal parameters of the proposed controller that are able to minimize | q ˙ i + Λ i t a n h ( e i   i ) | .
The converging behaviour of tracking errors of (24) is visualized in Figure 1.
Lemma 3. 
For a moment t * , if the system states ( q i ( t * ) , q ˙ i ( t * ) ) satisfy (5) and (6) as well as q ˙ i ( t * ) + Λ i t a n h ( e i ( t * )   i ) 0 holds, there always exists the optimal parameter λ i * > 0 to decrease ( q ˙ i + Λ i t a n h ( e i   i ) ) 2 . Namely, (28) holds.
d [ ( q ˙ i + Λ i t a n h ( e i   i ) ) 2 ] d t | t = t * < 0
Proof. 
Combining (10) and (15), we can obtain the 1st derivative of q ˙ i + Λ i t a n h ( e i   i ) with respect to time, shown as (29).
d ( q ˙ i + Λ i t a n h ( e i   i ) ) d t = d ^ i s g n ( q ˙ i ) + d i ( t ) + Λ i   i t a n h ( e i   i ) e i   i q ˙ i 1 1 + k y i C i [ k p i e i + k d i q ˙ i + k s i A i q i + k s i B i z i ] 1 1 + k y i C i [ λ i   q ˙ i + λ i Λ i t a n h ( e i   i ) ]
Then, we define a negative changing rate of ( q ˙ i + Λ i t a n h ( e i   i ) ) 2 at moment t * in (30).
d d t [ q ˙ i ( t ) + Λ i t a n h ( e i ( t ) i ) ] 2 | t = t * < 0
Using (5), (6), (29) and (30), a solution of λ i * satisfying (30) is obtained in (31).
λ i * > 1 | q ˙ i + Λ i t a n h ( e i   i ) | [ ( 1 + k y i C i ) ( d ^ i + Γ i * + Λ i 2   i ) + k p i | e i | + k d i Λ i + k s i | A i | ε i + k s i | B i | | z i | ]  
Notably, d ^ i , | e i | , | z i | , | A i | and | B i | are all bounded because of the bounded Lyapunov function (16). Therefore, a finite real solution of λ i * satisfying (31) exists as long as q ˙ i + Λ i t a n h ( e i   i ) 0 .
Remark 7. 
According to Lemma 3 and Remark 6, the optimal parameters λ i * leading to a decrease of | q ˙ i + Λ i t a n h ( e i   i ) | at the moment t * can improve the tracking accuracy. Although a large enough λ i can be selected to satisfy (31), an inappropriately large λ i could lead to a significant chattering of control torques and thereby compromise the tracking accuracy. Moreover, an optimal λ i is hard to be deterministically found due to the complexity of the system and the unknown TDE errors. Therefore, a fuzzy Q RL mechanism is designed to automatically determine the optimal λ i * .
Fuzzy Q learning is a common version of RL applicable on continuous systems, which can explore the optimal policy by interacting with the environment [37]. In this paper, the λ i is to be tuned by a fuzzy Q learning mechanism according to the tracking error e i and the angular velocity q ˙ i . The linguistic rules to determine λ i can be given as the following form:
IF   e i ( k )   is   e , i   AND   q ˙ i ( k )   is   q , i ˙ ,   THEN   λ i ( k )   is   λ , i
e , i , q , i ˙ and λ , i are the linguistic descriptions of tracking error e i , angular velocity q ˙ i and parameter λ i respectively. k is the current moment. The linguistic description could be “small”, “medium” and “big”. As a result, an example of linguistic rule could be: IF e i ( k ) is s m a l l AND q ˙ i ( k ) is s m a l l , THEN λ i ( k ) is s m a l l .
Some intuitive linguistic rules can be given as:
  • IF e i ( k ) is small, and q ˙ i ( k ) is small THEN λ i ( k ) is small
  • IF e i ( k ) is large, and q ˙ i ( k ) is large THEN λ i ( k ) is large
  • IF e i ( k ) is large, and q ˙ i ( k ) is small THEN λ i ( k ) is large
  • IF e i ( k ) is small, and q ˙ i ( k ) is large THEN λ i ( k ) is small
Remark 8. 
To carry out the linguistic inference shown by (32) in a numerical form, fuzzy logic inference is required. In detail, initially, numerical variables e i ( k ) and q ˙ i ( k ) at the moment k are fuzzified to the firing rates of the linguistic descriptions by the triangular membership function shown in Figure 2. After that, a group of firing rates of fuzzy rules are obtained by fuzzy reasoning. Then, the numerical values of λ i ( k ) are calculated by the defuzzification according to the firing rates of all fuzzy rules and the numerical value of the action corresponding to each fuzzy rule.
The parameters of membership function shown in Figure 2 are defined as follows:
L i n ( e i ) = { ζ 1 , 1 ( i ) , ,   ζ 1 , 𝒶 ( i ) , , ζ 1 , A ( i ) } ,   𝒶 = 1 , 2 , , A
L i n ( q ˙ i ) = { ζ 2 , 1 ( i ) , ,   ζ 2 , 𝒷 ( i ) , , ζ 2 , ( i ) } ,   𝒷 = 1 , 2 , ,
where A is the number of fuzzy sets ( ζ 1 , 𝒶 ) for the fuzzy input e i , and is the number of fuzzy sets ( ζ 1 , 𝒷 ) for the fuzzy input q ˙ i .
The 𝓃 -th fuzzy rule in fuzzy Q learning is defined as follows:
R 𝓃 , i :   IF   𝓈 1 , i ( k )   is   L 1 , i ( k )   and   𝓈 2 , i ( k )   is   L 2 , i ( 𝓃 )   and     and   𝓈 𝓂 , i ( k )   is   L 𝓂 , i ( 𝓃 ) ,   THEN   u i ( 𝓃 ) U i , 𝓃   that   u i ( 𝓃 ) = u i , 1 ( 𝓃 )   with   𝓆 i ( k ) ( 𝓃 , 1 )   or   u i ( 𝓃 ) = u i , 2 ( 𝓃 )   with   𝓆 i ( k ) ( 𝓃 , 2 )   or   u i ( 𝓃 ) = u i , 3 ( 𝓃 )   with   𝓆 i ( k ) ( 𝓃 , 3 ) , . ,   u i ( 𝓃 ) = u i , p ( 𝓃 )   with   𝓆 i ( k ) ( 𝓃 , p ) , . , u i ( 𝓃 ) = u i , P ( 𝓃 ) .   with   𝓆 i ( k ) ( 𝓃 , P )
where U i , 𝓃 = { u i , 1 ( 𝓃 ) , .. u i , p ( 𝓃 ) , .. , u i , P ( 𝓃 ) } is the set of action candidates in the rule R 𝓃 , i . L i ( 𝓃 ) = { L 1 , i ( 𝓃 ) , , L 𝓂 , i ( 𝓃 ) } is the set of linguistic variables of fuzzy inputs. 𝓈 i ( k ) = { 𝓈 1 , i ( k ) , ,   𝓈 𝓂 , i ( k ) } is the set of fuzzy inputs at the k moment. In this paper, the fuzzy inputs are e i and q ˙ i such that 𝓈 i ( k ) = { e i ( k ) ,   q ˙ i ( k ) } .
The set of fuzzy inputs 𝓈 i ( k ) = { 𝓈 1 , i ( k ) , ,   𝓈 𝓂 , i ( k ) } is fuzzified by the membership function shown in Figure 2 and then matched with the rule antecedents (19), providing the firing rate vector φ ( 𝓈 i ( k ) ) = [ φ 1 ( 𝓈 i ( k ) ) , φ 2 ( 𝓈 i ( k ) ) , .. , φ N ( 𝓈 i ( k ) ) ] . N is the amount of fuzzy rules (There are N fuzzy rules).
For the 𝓃 t h rule in the i t h subsystem ( R 𝓃 , i ), the optimal action at the k moment is defined as the action with the maximum 𝓆 i ( k ) ( 𝓃 , p ) , p { 1 , 2 , .. , P } among P action candidates.
u * i ( 𝓃 ) = a r g max u i ( 𝓃 ) U i , 𝓃 𝓆 i ( k ) ( 𝓃 , u i , p ( 𝓃 ) )
To prevent the selection of u from the local optimum in the learning process, we introduce a greed mechanism:
u ^ i ( 𝓃 ) = { u + i ( 𝓃 ) ,   w i t h   p r o b a b l i t y   ρ   u * i ( 𝓃 )   w i t h   p r o b a b l i t y   1 ρ  
where u + i ( 𝓃 ) is an action randomly selected from U i , 𝓃 . 0 < ρ < 1 is the probability to explore random actions.
The numerical value of λ i is calculated by firing rates and the selected actions:
λ i ( 𝓈 i ( k ) ) = 𝓃 = 1 N φ 𝓃 ( 𝓈 i ( k ) ) u ^ i ( 𝓃 ) 𝓃 = 1 N φ 𝓃 ( 𝓈 i ( k ) )
The update principle of 𝓆 is the most important part in the whole learning process. The 𝓆 -values are updated according to the rewards of the selected actions; the optimal action can achieve higher rewards, therefore, we can finally learn the optimal action.
The Q value at the k moment can be designed as follows:
Q ( 𝓈 i ( k ) ) = 𝓃 = 1 N φ 𝓃 ( 𝓈 i ( k ) ) 𝓆 i ( k ) ( 𝓃 ,   u ^ i ( 𝓃 ) ) 𝓃 = 1 N φ 𝓃 ( 𝓈 i ( k ) )
The target value at the state 𝓈 i ( k ) is calculated as:
V ( 𝓈 i ( k ) ) = 𝓃 = 1 N φ 𝓃 ( 𝓈 i ( k ) ) 𝓆 i ( k ) ( 𝓃 ,   u * i ( 𝓃 ) ) 𝓃 = 1 N φ 𝓃 ( 𝓈 i ( k ) )
When the system state 𝓈 i ( k ) is driven to the next state 𝓈 i ( k + 1 ) , temporal difference (TD) is calculated according to the reward obtained at the k moment 𝓇 i ( k ) :
Q i ( k ) = 𝓇 i ( k ) + α i V ( 𝓈 i ( k + 1 ) ) Q ( 𝓈 i ( k ) )
where α i [ 0 , 1 ] is the discount factor reflecting the contribution of the future reward. The reward 𝓇 i k at the k moment of the i t h subsystem is designed in (42) and the meaning of (42) is explained in Remark 9.
𝓇 i ( k ) = { 0 ,   i f   | 𝓍 i ( k ) | > σ i cos ( π 2 | 𝓍 i ( k ) | σ i ) ,   i f   | 𝓍 i ( k ) | σ i  
where 𝓍 i ( k ) = q ˙ i ( k ) + Λ i t a n h ( e i ( k ) i ) and σ i > 0 is a positive constant.
Finally, the adaptive law of q-values is:
𝓆 i ( k + 1 ) ( 𝓃 , u ^ i ( 𝓃 ) ) = 𝓆 i ( k ) ( 𝓃 , u ^ i ( 𝓃 ) ) + γ i · Q i ( k ) · φ 𝓃 ( 𝓈 i ( k ) ) 𝓃 = 1 N φ 𝓃 ( 𝓈 i ( k ) )
where γ i [ 0 , 1 ] is the learning rate.
Remark 9. 
The reward function (42) indicates that higher values of reward 𝓇 i ( k ) can be obtained by the smaller 𝓍 i ( k ) , which means the the action that satisfies (24) will be given with the highest reward ( 𝓇 i ( k ) = 1 ).
The aforementioned process of fuzzy Q learning to tune λ i can be also concluded in Figure 3. The proposed control scheme can be concluded in Figure 4.
Remark 10. 
All the system states including the tracking errors are proven to be bounded with λ i > 0 . Therefore, the instability will not occur even if an inappropriate positive value of λ i is calculated by the fuzzy Q learning mechanism trying some bad action candidates. Hence, the designed control law (7), (9)–(14) offers the fuzzy Q learning mechanism (32)–(43) a safe environment to learn a optimal λ i . Notably, in order to make sure λ i > 0 always holds, all the action candidates U i , 𝓃 = { u i , 1 ( 𝓃 ) , .. u i , p ( 𝓃 ) , .. , u i , P ( 𝓃 ) } should be positive.
Remark 11. 
The proof of Theorem 1 requires the parameter λ i > 0 to be a constant. However, the parameter λ i is oneline tuned by a fuzzy Q learning mechanism, and therefore λ i is varying in the implementation of the proposed controller. To handle this issue, we design that the time interval between two consecutive fuzzy inferences of λ i in fuzzy Q learning is 20-times greater than the time interval between two consecutive control actions τ (eg. 0.1s between two consecutive u ^ i ( 𝓃 )   (equally, two consecutive λ i ) while 0.001s between two consecutive τ ). Namely, λ i ( t ) = λ i ( t * ) ,     t [ t * , t * + 20 L ] ; τ ( t ) =   τ ( t * ) ,   t [ t * , t * + L ] . t * is any time moment when the algorithm works. Thereby, not only can the time derivative of λ i be negligible in the proof of Theorem 1, but the computational load is also decreased.

4. Simulation Results and Analysis

In this section, similar to works [38,39,40,41,42], we use the simulation to verify the effectiveness of the proposed controller. A 2-rigid-link robotic manipulator shown in Figure 5 is carried in Matlab 2018a. The sampling time of simulating the real dynamics of robotic manipulators is set as 1 × 10 4 s. The sampling time of TDE and controller is set as 1 × 10 3 s (10 times to 1 × 10 4 ) to show the discrete nature of using controllers in practice. The sampling time of fuzzy Q learning is set as 0.01 s according to Remark 10. The dynamic model of a 2-rigid-link robotic manipulator is given as follows, which can be also found in [26].
M ( q ) = [ m 2 l 2 2 + 2 l 1 l 2 m 2 cos ( q 2 ) + ( m 1 + m 2 ) l 1 2 m 2 l 2 2 + l 1 l 2 m 2 cos ( q 2 ) m 2 l 2 2 + l 1 l 2 m 2 cos ( q 2 ) m 2 l 2 2 ]
C ( q , q ˙ ) q ˙ = [ m 2 l 1 l 2 sin ( q 2 ) q ˙ 2 2 2 m 2 l 1 l 2 sin ( q 2 ) q ˙ 1 q ˙ 2 m 2 l 1 l 2 sin ( q 2 ) q ˙ 1 2 ]
G ( q ) = [ ( m 1 + m 2 ) l 1 c o s ( q 2 ) g + m 2 l 2 c o s ( q 1 + q 2 ) g m 2 l 2 cos ( q 1 + q 2 ) g ]
F ( q ˙ ) = [ F 1 F 2 ]
τ d = [ τ d 1 τ d 2 ]
The system parameters are given in Table 1, which are the same as [26].
The initial angular positions and velocities of two joints are set as: q 1 = 0 o , q 2 = 0 o , q ˙ 1 = 0 o / s and q ˙ 2 = 0 o / s . The desired angular positions of two joints are set as: q r 1 = 30 o and q r 2 = 45 o .
The upper bounds of the angular position and angular velocity for the i t h joint of manipulator ( ε i and Λ i ) are selected by the user dependent on the specific mission requirement. To successfully implement the proposed controller, users can select any values satisfying ε i > | q r i | and Λ i > 0 . Therefore, in the simulation of this paper, we select ε 1 = 50 o , ε 2 = 60 o ,   Λ 1 = 10 o / s and Λ 2 = 12 o / s .
Remark 12. 
k p i and k d i are the proportional coefficient and differential coefficient, respectively. The great k p i could decrease steady state error, but an excessively great could result in a significant overshoot. While the great k d i could improve the robustness to disturbance/uncertainty, but an over-great k d i could compromise the tracking accuracy. The selection of k p i and k d i can be based on the tuning rules of PID controllers.
Remark 13. 
Great k s i can amplify the effect of the terms A i and B i in (10) to keep the angular positions from hitting their constraints. An inappropriately small k s i could lead to a | q i | too close to ε i and then result in an over-great u i in (10). Therefore, the selection of k s i could be started at a small value, and then users could gradually increase k s i until the magnitude of u i is acceptable. Great k y i can amplify the effect of the term C i in (10) to keep the angular velocities from hitting their constraints. An excessively great k y i could lead to an over great ( 1 + k y i C i ) in (10) to decline the converging rate of the tracking error even if the angular velocity is not close to its constraint, which could increase the settling time. Therefore, the trial of selecting k y i could start at a small value, and then users could decrease it until the converging rate of the tracking error is acceptable. β i determines the converging rate of the auxiliary variable z i that has a significant effect on steady state error in the controller [33]. However, in the proposed controller, the RL-based term is introduced to improve the tracking performance and thereby the effect of β i is decreased. We suggest to offer β i with a medium value ranging between 0.1–1.
Remark 14. 
large values of ψ i and δ i can lead to a fast adaption of d ^ i to handle system uncertainty and external disturbance. However, the inappropriately large ψ i and δ i could result in a sharp variation of d ^ i and then a chattering effect on control torques. Therefore, the selection of ψ i and δ i could start at a big value, and then users could decrease them until no chattering effect occurs on the control torques. Over-small values of Ω ¯ i could lead to an insufficient decrease of d ^ i that still brings up a chattering effect on control torques. Meanwhile, over-great values of Ω ¯ i could negatively influence the robustness to uncertainty and disturbance. Hence, the selection of Ω ¯ i could start at some small values, and then users could increase Ω ¯ i until d ^ i is significantly decreased at the final stage of the control to have a satisfactory chattering attenuation.
Remark 15. 
Small/great values of i can amplify/reduce the effect of the RL term λ i Λ i t a n h ( e i   i ) in (10). The over-small values of i could result in the over-great magnitude of u i , while the over-great values of i could bring up an insufficient improvement on tracking performance. Therefore, the trial of selecting i is suggested to start at a great value, and then users could decrease i until a satisfying improvement on the tracking performance.
According to Remark 11~Remark 15, the parameters of the proposed controller are selected in Table 2.
Notably, it does not require an extensive trial to select parameters in Table 2 as the satisfying tracking performance is mainly obtained by the optimal λ i in the RL term λ i Λ i t a n h ( e i   i ) . In other words, users can select some medium values for parameters (not optimal) in Table 2 by an acceptable amount of parameters-selection trails. Then, the fuzzy Q learning can automatically explore the λ i matching the selected parameters to have a satisfying tracking performance.
Remark 16. 
Similar to [36], the parameters of membership function in (33) and (34), L i n ( e i ) = { ζ 1 , 1 ( i ) ,   ζ 1 , 2 ( i ) , , ζ 1 , A ( i ) } and L i n ( q ˙ i ) = { ζ 2 , 1 ( i ) ,   ζ 2 , 2 ( i ) , , ζ 2 , ( i ) } , are used to fuzzify e i and q ˙ i . To well present the e i and q ˙ i in the form of firing rate, we suggest to offer ζ 1 , 1 ( i ) and ζ 2 , 1 ( i ) with some small values and provide some great values to ζ 1 , A ( i ) and ζ 2 , ( i ) . The ζ 1 , 𝒶 ( i ) and ζ 2 , 𝒷 ( i ) are suggested to be evenly distributed among ( ζ 1 , 1 ( i ) ,   ζ 1 , A ( i ) ) and ( ζ 2 , 1 ( i ) ,   ζ 2 , ( i ) ), respectively. The great A and increase the potential to well present e i and q ˙ i at the cost of increasing the computational load. Therefore, the trial of selecting A and could start at some great values (e.g., 20), and then users could decrease them until a reasonable computational load.
Remark 17. 
σ i is the threshold of obtaining rewards. The inappropriately small values of σ i could result in a difficulty in obtaining high values of reward, while the over-great values of σ i could result in the different action candidates to be offered with the similarly high rewards even if they lead to the different control performances. Therefore, both of the over-great and over-small values of σ i will negatively influence the convergence of q-values in (43) and thereby compromise the performance of reinforcement learning. The selection of σ i could start at a small value, and the users could increase σ i until a satisfying convergence of q values (the convergence of q-values can be also reflected by the convergence of obtained rewards). The selection of mutation probability ρ , learning rate γ i and discount factor α 2 can be based on the strategy mentioned in [36].
Remark 18 
Action candidates U i , 𝓃 = { u i , 1 ( 𝓃 ) , .. u i , p ( 𝓃 ) , .. , u i , P ( 𝓃 ) } are the most important parameters in the proposed control scheme because the optimal λ i that brings up a satisfying control performance is calculated based on them. To make the optimal action included in the group of action candidates, u i , 1 ( 𝓃 ) should be given a small value (e.g., 0) while u i , P ( 𝓃 ) should be given a great value. The rest candidates u i , p ( 𝓃 ) ( p = 1 , 2 , .. , p ) are suggested to be evenly distributed between u i , 1 ( 𝓃 ) and u i , P ( 𝓃 ) . Users could initially give u i , P ( 𝓃 ) with a small value (but greater than u i , 1 ( 𝓃 ) ), and an increase of u i , P ( 𝓃 ) until a sufficient improvement on the tracking performance is achieved. p is the amount of action candidates of each fuzzy rule for each subsystem. The selection of p could start at a great number (e.g., 20), and then users could decrease p until the computational load is acceptable.
According to Remark 16~Remark 18, the parameters of the fuzzy Q learning mechanism are given as follows. The amount of fuzzy sets in (17) are A = = 8 . Therefore, the amount of fuzzy rules are N = 64 . The parameters of membership function to do the fuzzification are: L i n ( e 1 ) = { ζ 1 , 1 ( 1 ) ,   ζ 1 , 2 ( 1 ) , , ζ 1 , 8 ( 1 ) } = {−0.0017, −0.0014, −0.001, −0.0006, −0.0002, 0.0002, 0.0006, 0.001, 0.0014, 0.0017}. L i n ( q ˙ 1 ) = { ζ 2 , 1 ( 1 ) ,   ζ 2 , 2 ( 1 ) , , ζ 2 , 8 ( 1 ) } = {−0.175, −0.138, −0.097, −0.058, −0.019, 0.019, 0.058, 0.097, 0.138, 0.175} × 10 3 . L i n ( e 2 ) = { ζ 1 , 1 ( 2 ) ,   ζ 1 , 2 ( 2 ) , , ζ 1 , 8 ( 2 ) } = {−0.0017, −0.0014, −0.001, −0.0006, −0.0002, 0.0002, 0.0006, 0.001, 0.0014, 0.0017}. L i n ( q ˙ 2 ) = { ζ 2 , 1 ( 2 ) ,   ζ 2 , 2 ( 2 ) , , ζ 2 , 8 ( 2 ) } = {−0.175, −0.138, −0.097, −0.058, −0.019, 0.019, 0.058, 0.097, 0.138, 0.175} × 10 3 . The action candidates for each fuzzy rule are: U i , 𝓃 = { 0 , 22 , 44 , 66 , 88 , 111 , 133 , 155 , 177 , 200 } for all i = 1 , 2 and 𝓃 = 1 , 2 , . ,   64 . Therefore, the amount of action candidates for each fuzzy rule is P = 10 . The discount factor is α 1 = α 2 = 0.01 and the learning rate is γ 1 = γ 2 = 0.2 . The threshold of obtaining rewards is σ 1 = σ 2 = 0.01 .
To show the superiority, four existing controllers, Refs. [26,33,43,44] are compared with the proposed controller.
The controller from [26] is given as (44)–(52)
τ = τ 1 + τ 2
τ 1 = M ^ ( q ) ( q ¨ r K V e ˙ K P e ) + C ^ ( q , q ˙ ) q ˙ + G ^ ( q )
τ 2 = M ^ ( q ) f ^ + M ^ ( q ) u r
u r = [ ξ ^ tan h ( a 1 p / ρ 1 ) ξ ^ tan h ( a 2 p / ρ 1 ) ]
f ^ = W ^ T σ ( V ^ T X )
ξ ^ ˙ = a 1 p t a n h ( a 1 p ρ 1 ) + a 1 p t a n h ( a 2 p ρ 1 ) K ξ ^
W ^ ˙ = ( σ σ V ^ T X ) x T P B Υ W W ^
V ^ ˙ = X x T P B W ^ T σ Υ V V ^
p = 1 + | | X | | + | | V ^ | | · | | X | | + | | W ^ | | · | | X | |
where σ = [ σ 1 , σ 2 , , σ m ] T is the vector of hidden neurons with the activation function σ i ( s ) = 1 / ( 1 + e s ) . σ is the vector of partial derivative of σ such that σ = σ ( s ) / s . x = [ e 1 , e 2 , q ˙ 1 , q ˙ 2 ] T and X = [ q r 1 , q r 2 , q 1 , q 2 , q ˙ 1 , q ˙ 2 , q ¨ 1 , q ¨ 2 ] T . P satisfies P A + A T P = Q with A = [ 0 I K P K V ] . [ a 1 , a 2 ] = x T P B .
In this simulation, we let the controller from [26] to fully know the system parameters such that M ^ ( q ) = M ( q ) , C ^ ( q , q ˙ ) = C ( q , q ˙ ) and G ^ ( q ) = G ( q ) .
The parameters in (44)–(52) are given as:
K V = [ 300 0 0 300 ] ,   K P = [ 200 0 0 200 ]
B = [ 0 0 0 0 1 0 0 1 ] ,   Q = [ 1 0 0 0 0 1 0 0 0 0 1 0 0 0 0 1 ]
ρ 1 = 0.01 , K = 0.005 , Υ W = 0.15 , Υ V = 0.15 , ξ ^ ( 0 ) = 0.01
W ^ ( 0 ) = [ 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 ]
V ^ ( 0 ) = [ 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 ]
The controller from [33] is given as (53)–(58).
τ = G ( q ) p ^ K d q ˙ K p e K s d i a g { X 1 , X 2 } z
X i = ζ i 2 ζ i 2 q i 2 + ζ i 2 q i z i ( ζ i 2 q i 2 ) 2 ,   i = 1 , 2
z i = e i + 0 t η i ( Θ ) d Θ ,   i = 1 , 2
η i = β i z i ,   i = 1 , 2
p ^ ˙ = Ψ G T ( q ) q ˙
G ( q ) = [ cos ( q 1 + q 2 ) cos ( q 2 ) cos ( q 1 + q 2 ) 0 ]
where p ^ = [ p ^ 1 , p ^ 2 ] T , z 1 ( 0 ) = e 1 ( 0 ) and z 2 ( 0 ) = e 2 ( 0 ) . ζ 1 and ζ 2 are the specific constraint (designed upper bound) of the 1st and 2nd joint respectively, therefore, ζ 1 = ε 1 = 50 o and ζ 2 = ε 2 = 60 o .
The parameters in (53)–(58) are given as follows:
K d = [ 60 0 0 20 ] ,   K p = [ 24 0 0 24 ] ,   K s = [ 20 0 0 20 ]
β 1 = 30 ,   β 2 = 30 ,   Ψ = [ 0.1 0 0 0.1 ]
The controller from [43] is given as (59)–(61).
τ = K D 1 M   ^ [ K P e ˙ K I e + K D q ¨ r + β s ˙ K D Γ ^ K D λ s K D k s s g n ( s ˙ ) ]
s ˙ + β s = K P e + K I 0 t e + K D e ˙
k ˙ s = γ | | s ˙ | |
where s ˙ ( 0 ) = [ 0 , 0 ] T , k s ( 0 ) = 0 , γ = 0.1 , Γ ^ is the known part of lumped uncertainty Γ . In this simulation, we let Γ ^ = 0.7 Γ . The parameters K P , K D , K I and β are given as follows.
K D = [ 20 0 0 20 ] ,   K P = [ 10 0 0 10 ] ,   K I = [ 0.01 0 0 0.01 ] ,   β = [ 0.1 0 0 0.1 ]
The controller from [44] is given as (62)–(65).
{ D ^ = z + Y q ˙ z ˙ = Y M ^ 1 z + Y M ^ 1   ( τ Y q ˙ D ¯ )
τ = M ^ ( q ¨ r Λ q ˙ + Λ q ˙ r ) D ¯ K D s a t ( s ) D ^
s a t ( s i ) = { s g n ( s i ) ,   | s i | σ s i / σ , | s i | < σ
where σ = 0.1 , s a t ( s ) = [ s 1 , s 2 ] T . z ( 0 ) = [ 0 , 0 ] T and s = e ˙ + Λ e . D ¯ is known part of M ^ 1 Γ , D ¯ = 0.7 M ^ 1 Γ in the simulation. The parameters Y , Λ and K D are given as follows.
K D = [ 40 0 0 16 ] ,   Y = [ 1 0.06 0 0 1 0.06 ] ,   Λ = [ 40 0 0 20 ]
Furthermore, two cases are considered to run the simulation. In the 1st case, the parameters of the dynamic model are fully known, and there is no external disturbance and no friction. In the 2nd case, the parametric uncertainty is considered and the unknown external disturbance and frictions are applied on the dynamics model.
Remark 19. 
The selections of parameters for the proposed controller, the controller from [26] and the controller from [33] are all carried out in case 1. In other words, all the parameters for the three different controllers are fine-tuned to have a good performance in case 1. In case 2, all the selected parameters of the three controllers remain unchanged to test the robustness to the lumped uncertainty.
Case 1. 
In the absence of system uncertainty, external disturbance and friction.
In this case, the system parameters used in the three controllers (proposed controller, controller from [26] and controller from [33]) are the same as the parameters in the dynamics model of manipulator, which means no parametric uncertainty. Meanwhile, the friction and external disturbance are null, shown as follows:
F ( q ˙ ) = [ F 1 F 2 ] = [ 0 0 ] ,   τ d = [ τ d 1 τ d 2 ] = [ 0 0 ]
The comparisons of angular position tracking are shown in Figure 6 and Figure 7, while the comparisons of angular velocity are shown in Figure 8 and Figure 9. Cleanly, in the absence of unknown disturbance and system parametric uncertainty, the proposed controller shows an inferior performance in terms of a greater steady state error and a slower error converging rate, compared to the controllers from [26,33]. In comparison to [43,44], it provides a response faster than the sliding mode controller [44] but slower than the disturbance observer-based controller [44]. However, the steady state error of the proposed controller is smaller than that of [43,44].
In Figure 8 and Figure 9, the angular velocities of the proposed control scheme are within the preset constraint (red lines). While the controllers from [26,33,43,44] result in the angular velocities of two joints exceeding the constraints a t = 0 ~ 2   s .
case 2: 
In the presence of system uncertainty, external disturbance and friction.
In this case, the system parameters used in all the five controllers are different to that in the dynamics model of the manipulator to indicate the parametric uncertainty. Namely, M , C and G is taken as 20% of M , C and G . Moreover, the friction and external disturbance, which are not known by the controllers, are applied on the dynamics model.
The friction model is from [33].
F ( q ˙ ) = [ f s 1 [ tanh ( f s 2 q ˙ 1 ) tan h ( f s 3 q ˙ 1 ) ] + f c 1 tanh ( f c 2 q ˙ 1 ) + f v q ˙ 1 f s 1 [ tanh ( f s 2 q ˙ 2 ) tan h ( f s 3 q ˙ 2 ) ] + f c 1 tanh ( f c 2 q ˙ 2 ) + f v q ˙ 2 ]
The parameters of the friction model are given as: f s 1 = 20 , f s 2 = 5 , f s 3 = 3 , f c 1 = 10 , f c 2 = 2 and f v = 10 .
It is widely seen to use triangular functions as the unknown disturbance in the literature of the robotic system control, such as [13,36,45,46]. Therefore, in this paper, the external disturbance is in the form of the triangular functions.
τ d = [ 0.78 sin ( π 3 t + π 4 ) + 0.065 sin ( π 10 t + π 4 ) 0.58 cos ( π 3 t + π 4 ) + 0.091 sin ( π 10 t + π 4 ) ]
The comparisons of tracking performance and tracking error in the presence of system parametric uncertainty, friction and external disturbance are shown in Figure 10, Figure 11, Figure 12 and Figure 13, respectively. Clearly, the proposed controller can achieve the smallest steady state errors, which means the robustness to the lumped uncertainty and unknown disturbance. The converging rate of tracking errors of the proposed controller is faster than the controller [33,43] but slower than the controller [26,44]. It is because the preset constraint of angular velocity (black lines in Figure 14 and Figure 15) limits the converging rate of tracking errors. Therefore, the converging rate of tracking errors could be increased by increasing the value of velocity constraint ( Λ i ) (e.g., applying some better driving motors that have a greater maximum rotational speed to drive the joints of manipulator).
The computed control torques are shown in Figure 16 and Figure 17. The chattering effect of the proposed controller occurs at the initial stage because of the two following factors: 1. the increasing value of switching gain d ^ i to handle disturbance at the initial stage, 2. the fuzzy Q learning mechanism tried some bad action candidates that lead to an undesirable consequence. After the initial stage ( t > 8   s ), it is observed that the proposed controller shows the smoothest control torque compared to [26,33,43,44]. Figure 18 shows the values of switching gain d ^ 1 and d ^ 2 , it is clear the d ^ 1 and d ^ 2 will decrease to a small value to avoid a chattering in steady state regardless of the disturbance and uncertainty.
Notably, in both case 1 and case 2, the proposed controller can make the angular positions and angular velocities of two joints to be within their constraints during the whole period of position tracking. More precisely, the angular position of each joint is always between the two black dash lines in Figure 6, Figure 7, Figure 10 and Figure 11. Meanwhile, the angular velocity of each joint is always between the two black dash lines in Figure 8, Figure 9, Figure 14 and Figure 15.

5. Conclusions

This paper proposed a novel adaptive control scheme utilizing TDE and RL for the angular position tracking control of robotic manipulators. The proposed control scheme can achieve a good tracking accuracy and a fast tracking performance even when subject to the system uncertainty and unknown disturbance. Moreover, the angular position and angular velocity of each joint of the manipulator are guaranteed to be within their preset constraints. The boundness of tracking errors and the stability of the robotic system controlled by the proposed controller are proven by Lyapunov theory. Notably, the stability will not be breached by the RL trying some bad action candidates, which ensures a safe environment for RL to explore the optimal policy. Simulation results validate the effectiveness of the proposed control scheme.

Author Contributions

Z.X.: Conceptualization, data curation, writing original draft. Q.L.: Formal analysis. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

The author agrees to publication.

Data Availability Statement

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare that there are no conflict of interest regarding the publication of this paper.

References

  1. Chan, S.P.; Liaw, H.C. Generalized impedance control of robot for assembly tasks requiring compliant manipulation. IEEE Trans. Ind. Electron. 2002, 43, 453–461. [Google Scholar] [CrossRef]
  2. Naito, J.; Obinata, G.; Nakayama, A.; Hase, K. Development of a Wearable Robot for Assisting Carpentry Workers. Int. J. Adv. Robot. Syst. 2007, 4, 48. [Google Scholar] [CrossRef]
  3. Kazerooni, H.; Bausch, J.J.; Kramer, B.M. An Approach to Automated Deburring by Robot Manipulators. J. Dyn. Syst. Meas. Control 1986, 108, 354–359. [Google Scholar] [CrossRef]
  4. Lee, W.Y.; Shih, C.L. Control and breakthrough detection of a three-axis robotic bone drilling system. Mechatronics 2006, 16, 73–84. [Google Scholar] [CrossRef]
  5. Takei, T.; Imamura, R.; Yuta, S.I. Baggage Transportation and Navigation by a Wheeled Inverted Pendulum Mobile Robot. IEEE Trans. Ind. Electron. 2009, 56, 3985–3994. [Google Scholar] [CrossRef]
  6. Datta, S.; Ray, R.; Banerji, D. Development of autonomous mobile robot with manipulator for manufacturing environment. Int. J. Adv. Manuf. Technol. 2008, 38, 536–542. [Google Scholar] [CrossRef]
  7. Kim, Y.G.; Jeong, K.S.; Lee, J.W. Development of the composite third robot arm of the six-axis articulated robot manipulator. Compos. Struct. 1996, 35, 331–342. [Google Scholar] [CrossRef]
  8. Gao, M.C.; Hou, J.C. Finite time linear quadratic control for weakly regular linear systems. IMA J. Math. Control Inf. 2001, 18, 405–425. [Google Scholar] [CrossRef]
  9. Dabiri, A.; Chahrogh, L.K.; Machado, J.A.T. Closed-form Solution for The Finite-horizon Linear-quadratic Control Problem of Linear Fractional-order Systems. In Proceedings of the 2021 American Control Conference (ACC), New Orleans, LA, USA, 25–28 May 2021. [Google Scholar]
  10. Shi, S. H output feedback stabilization for continuous-time switched linear systems. In Proceedings of the 2014 International Conference on Mechatronics and Control (ICMC), Jinzhou, China, 3–5 July 2014. [Google Scholar]
  11. Chang, X.; Yang, G. New Results on Output Feedback H Control for Linear Discrete-Time Systems. IEEE Trans. Autom. Control 2014, 59, 1355–1359. [Google Scholar] [CrossRef]
  12. Kim, E. Output feedback tracking control of robot manipulators with model uncertainty via adaptive fuzzy logic. IEEE Trans. Fuzzy Syst. 2004, 12, 368–378. [Google Scholar] [CrossRef]
  13. Zhang, L.; Wang, Y.; Hou, Y.; Li, H. Fixed-time sliding mode control for uncertain robot manipulators. IEEE Access 2019, 7, 149750–149763. [Google Scholar] [CrossRef]
  14. Baek, J.; Jin, M.; Han, S. A New Adaptive Sliding-Mode Control Scheme for Application to Robot Manipulators. IEEE Trans. Ind. Electron. 2016, 63, 3628–3637. [Google Scholar] [CrossRef]
  15. Islam, S.; Liu, P.X. Robust sliding mode control for robot manipulators. IEEE Trans. Ind. Electron. 2011, 58, 2444–2453. [Google Scholar] [CrossRef]
  16. Ahmed, S.; Wang, H.; Tian, Y. Adaptive fractional high-order terminal sliding mode control for nonlinear robotic manipulator under alternating loads. Asian J. Control 2021, 23, 1900–1910. [Google Scholar] [CrossRef]
  17. Feng, Y.; Zhou, M.; Yu, X.; Han, F. Full-order sliding-mode control of rigid robotic manipulators. Asian J. Control 2019, 21, 1228–1236. [Google Scholar] [CrossRef]
  18. Qi, W.; Zong, G.; Karimi, H.R. Sliding mode control for nonlinear stochastic semi-Markov switching systems with application to space robot manipulator model. IEEE Trans. Ind. Electron. 2020, 67, 3955–3966. [Google Scholar] [CrossRef]
  19. Sun, F.; Li, L.; Li, H.; Liu, H. Neuro-fuzzy dynamic inversionbased adaptive control for robotic manipulators—Discrete time case. IEEE Trans. Ind. Electron. 2007, 54, 1342–1351. [Google Scholar] [CrossRef]
  20. Fateh, S.; Fateh, M.M. Adaptive fuzzy control of robot manipulators with asymptotic tracking performance. J. Control Autom. Electr. Syst. 2020, 31, 52–61. [Google Scholar] [CrossRef]
  21. Fan, Y.; An, Y.; Wang, W.; Yang, C. TS Fuzzy Adaptive Control Based on Small Gain Approach for an Uncertain Robot Manipulators. Int. J. Fuzzy Syst. 2020, 22, 930–942. [Google Scholar] [CrossRef]
  22. He, W.; Dong, Y.; Sun, C. Adaptive neural impedance control of a robotic manipulator with input saturation. IEEE Trans. Syst. Man Cybern. Syst. 2016, 46, 334–344. [Google Scholar] [CrossRef]
  23. Zhou, Q.; Zhao, S.; Li, H.; Lu, R.; Wu, C. Adaptive neural network tracking control for robotic manipulators with dead zone. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3611–3620. [Google Scholar] [CrossRef] [PubMed]
  24. Nikdel, N.; Nikdel, P.; Badamchizadeh, M.A.; Hassanzadeh, I. Using Neural Network Model Predictive Control for Controlling Shape Memory Alloy-Based Manipulator. IEEE Trans. Ind. Electron. 2013, 61, 1394–1401. [Google Scholar] [CrossRef]
  25. Yen, V.T.; Nan, W.Y.; Van Cuong, P.; Quynh, N.X.; Thich, V.H. Robust adaptive sliding mode control for industrial robot manipulator using fuzzy wavelet neural networks. Int. J. Control Autom. Syst. 2017, 15, 2930–2941. [Google Scholar] [CrossRef]
  26. Hu, J.; Wang, P.; Xu, C.; Zhou, H.; Yao, J. High accuracy adaptive motion control for a robotic manipulator with model uncertainties based on multilayer neural network. Asian J. Control 2021, 24, 1503–1514. [Google Scholar] [CrossRef]
  27. Liu, H.; Sun, J.; Nie, J.; Zou, L. Observer-based adaptive second-order non-singular fast terminal sliding mode controller for robotic manipulators. Asian J. Control 2021, 23, 1845–1854. [Google Scholar] [CrossRef]
  28. Xiao, B.; Yang, X.; Karimi, H.R.; Qiu, J. Asymptotic tracking control for a more representative class of uncertain nonlinear systems with mismatched uncertainties. IEEE Trans. Ind. Electron. 2019, 66, 9417–9427. [Google Scholar] [CrossRef]
  29. Song, T.; Fang, L.; Wang, H. Model-free finite-time terminal sliding mode control with a novel adaptive sliding mode observer of uncertain robot systems. Asian J. Control 2021, 24, 1437–1451. [Google Scholar] [CrossRef]
  30. Sun, W.; Wu, Y.; Lv, X. Adaptive Neural Network Control for Full-State Constrained Robotic Manipulator With Actuator Saturation and Time-Varying Delays. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 3331–3342. [Google Scholar] [CrossRef]
  31. Zhou, Q.; Wang, L.; Wu, C.; Li, H.; Du, H. Adaptive fuzzy control for nonstrict-feedback systems with input saturation and output constraint. IEEE Trans. Syst. Man Cybern. Syst. 2016, 47, 2209–2217. [Google Scholar] [CrossRef]
  32. Yu, X.; Zhang, S.; Fu, Q.; Xue, C.; Sun, W. Fuzzy Logic Control of an Uncertain Manipulator with Full-State Constraints and Disturbance Observer. IEEE Access 2020, 8, 24284–24295. [Google Scholar] [CrossRef]
  33. Yang, T.; Sun, N.; Fang, Y.; Xin, X.; Chen, H. New adaptive control methods for n-link robot manipulators with online gravity compensation: Design and experiments. IEEE Trans. Ind. Electron. 2021, 69, 539–548. [Google Scholar] [CrossRef]
  34. Tang, L.; Liu, Y.-J.; Tong, S. Adaptive neural contro lusing reinforcement learning for a class of robot manipulator. Neural Comput. Appl. 2013, 25, 135–141. [Google Scholar] [CrossRef]
  35. Li, Y.; Chen, L.; Tee, K.P.; Li, Q. Reinforcement learning control for coordinated manipulation of multi-robots. Neurocomputing 2015, 170, 168–175. [Google Scholar] [CrossRef] [Green Version]
  36. Xie, Z.; Sun, T.; Kwan, T.H.; Mu, Z.; Wu, X. A New Reinforcement Learning Based Adaptive Sliding Mode Control Scheme for Free-Floating Space Robotic Manipulator. IEEE Access 2020, 8, 127048–127064. [Google Scholar] [CrossRef]
  37. Kumar, A.; Sharma, R. Linguistic Lyapunov reinforcement learning control for robotic manipulators. Neurocomputing 2018, 272, 84–95. [Google Scholar] [CrossRef]
  38. Yih, C.C.; Wu, S.J. Adaptive task-space manipulator control with parametric uncertainties in kinematics and dynamics. Appl. Sci. 2020, 10, 8806. [Google Scholar] [CrossRef]
  39. Han, S.H.; Tran, M.S.; Tran, D.T. Adaptive sliding mode control for a robotic manipulator with unknown friction and unknown control direction. Appl. Sci. 2021, 11, 3919. [Google Scholar] [CrossRef]
  40. Gao, M.; Ding, L.; Jin, X. ELM-Based Adaptive Faster Fixed-Time Control of Robotic Manipulator Systems. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 1–13. [Google Scholar] [CrossRef]
  41. Liu, C.; Wen, G.; Zhao, Z.; Sedaghati, R. Neural-network-based sliding-mode control of an uncertain robot using dynamic model approximated switching gain. IEEE Trans. Cybern. 2020, 51, 2339–2346. [Google Scholar] [CrossRef]
  42. Doan, Q.V.; Vo, A.T.; Le, T.D.; Kang, H.J.; Nguyen, N.H.A. A novel fast terminal sliding mode tracking control methodology for robot manipulators. Appl. Sci. 2020, 10, 3010. [Google Scholar] [CrossRef]
  43. Mobayen, S.; Mofid, O.; Din, S.U.; Bartoszewicz, A. Finite-time tracking controller design of perturbed robotic manipulator based on adaptive second-order sliding mode control method. IEEE Access 2021, 9, 71159–71169. [Google Scholar] [CrossRef]
  44. Yin, F.; Wen, C.; Ji, Q.; Zhang, H.; Shao, H. A compensation sliding mode control for machining robotic manipulators based on nonlinear disturbance observer. Trans. Inst. Meas. Control 2022, 44, 01423312221083771. [Google Scholar] [CrossRef]
  45. Jia, S.; Shan, J. Continuous integral sliding mode control for space manipulator with actuator uncertainties. Aerosp. Sci. Technol. 2020, 106, 106192. [Google Scholar] [CrossRef]
  46. Jia, S.; Shan, J. Finite-time trajectory tracking control of space manipulator under actuator saturation. IEEE Trans. Ind. Electron. 2019, 67, 2086–2096. [Google Scholar] [CrossRef]
Figure 1. Converging behaviour of tracking error satisfying (24).The red line is the velocity constraint, and the black line with arrow is the converging trajectory.
Figure 1. Converging behaviour of tracking error satisfying (24).The red line is the velocity constraint, and the black line with arrow is the converging trajectory.
Applsci 13 01275 g001
Figure 2. Membership function of fuzzy input of the i t h subsystem.
Figure 2. Membership function of fuzzy input of the i t h subsystem.
Applsci 13 01275 g002
Figure 3. Flow chart of fuzzy Q learning to determine λ i .
Figure 3. Flow chart of fuzzy Q learning to determine λ i .
Applsci 13 01275 g003
Figure 4. Proposed RL-based adaptive control scheme.
Figure 4. Proposed RL-based adaptive control scheme.
Applsci 13 01275 g004
Figure 5. 2-rigid-link robotic manipulator.
Figure 5. 2-rigid-link robotic manipulator.
Applsci 13 01275 g005
Figure 6. Comparison of angular position to [26,33] in case 1: [26] scheme (yellow solid line), [33] scheme (red dash line), the proposed scheme (blue solid line), the reference angular position (green dashed line), the angular position constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Figure 6. Comparison of angular position to [26,33] in case 1: [26] scheme (yellow solid line), [33] scheme (red dash line), the proposed scheme (blue solid line), the reference angular position (green dashed line), the angular position constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g006
Figure 7. Comparison of the angular position to [43,44] in case 1: [43] scheme (yellow solid line), [44] scheme (red dash line), the proposed scheme (blue solid line), the reference angular position (green dashed line), the angular position constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Figure 7. Comparison of the angular position to [43,44] in case 1: [43] scheme (yellow solid line), [44] scheme (red dash line), the proposed scheme (blue solid line), the reference angular position (green dashed line), the angular position constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g007aApplsci 13 01275 g007b
Figure 8. Comparison of angular velocity to [26,32] in case 1: [26] scheme (yellow solid line), [33] scheme (red dash line), the proposed scheme (blue solid line), the angular velocity constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Figure 8. Comparison of angular velocity to [26,32] in case 1: [26] scheme (yellow solid line), [33] scheme (red dash line), the proposed scheme (blue solid line), the angular velocity constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g008
Figure 9. Comparison of angular velocity to [43,44] in case 1: [43] scheme (yellow solid line), [44] scheme (red dash line), the proposed scheme (blue solid line), the angular velocity constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Figure 9. Comparison of angular velocity to [43,44] in case 1: [43] scheme (yellow solid line), [44] scheme (red dash line), the proposed scheme (blue solid line), the angular velocity constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g009
Figure 10. Comparison of angular position to [26,33] in case 2: the [26] scheme (yellow solid line), the [33] scheme (red dash line), the proposed scheme (blue solid line), the reference angular position (green dashed line), the angular position constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Figure 10. Comparison of angular position to [26,33] in case 2: the [26] scheme (yellow solid line), the [33] scheme (red dash line), the proposed scheme (blue solid line), the reference angular position (green dashed line), the angular position constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g010aApplsci 13 01275 g010b
Figure 11. Comparison of angular position error to [26,33] in case 2: the [26] scheme (yellow solid line), the [33] scheme (red dash line), the proposed scheme (blue solid line); (a) Joint 1. (b) Joint 2.
Figure 11. Comparison of angular position error to [26,33] in case 2: the [26] scheme (yellow solid line), the [33] scheme (red dash line), the proposed scheme (blue solid line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g011
Figure 12. Comparison of angular position error to [43,44] in case 2: the [43] scheme (yellow solid line), the [44] scheme (red dash line), the proposed scheme (blue solid line); (a) Joint 1. (b) Joint 2.
Figure 12. Comparison of angular position error to [43,44] in case 2: the [43] scheme (yellow solid line), the [44] scheme (red dash line), the proposed scheme (blue solid line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g012
Figure 13. Comparison of angular position to [43,44] in case 2: The [43] scheme (yellow solid line), the [44] scheme (red dash line), the proposed scheme (blue solid line), the reference angular position (green dashed line), the angular position constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Figure 13. Comparison of angular position to [43,44] in case 2: The [43] scheme (yellow solid line), the [44] scheme (red dash line), the proposed scheme (blue solid line), the reference angular position (green dashed line), the angular position constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g013
Figure 14. Comparison of angular velocity to [26,33] in case 2: the [26] scheme (yellow solid line), the [33] scheme (red dash line), the proposed scheme (blue solid line), the angular velocity constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Figure 14. Comparison of angular velocity to [26,33] in case 2: the [26] scheme (yellow solid line), the [33] scheme (red dash line), the proposed scheme (blue solid line), the angular velocity constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g014
Figure 15. Comparison of angular velocity to [43,44] in case 2: the [43] scheme (yellow solid line), the [44] scheme (red dash line), the proposed scheme (blue solid line), the angular velocity constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Figure 15. Comparison of angular velocity to [43,44] in case 2: the [43] scheme (yellow solid line), the [44] scheme (red dash line), the proposed scheme (blue solid line), the angular velocity constrains (black dashed line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g015
Figure 16. Comparison of torque to [26,33] in case 2: the [26] scheme (yellow solid line), the [33] scheme (red dash line), the proposed scheme (blue solid line); (a) Joint 1. (b) Joint 2.
Figure 16. Comparison of torque to [26,33] in case 2: the [26] scheme (yellow solid line), the [33] scheme (red dash line), the proposed scheme (blue solid line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g016
Figure 17. Comparison of torque to [43,44] in case 2: the [43] scheme (yellow solid line), the [44] scheme (red dash line), the proposed scheme (blue solid line); (a) Joint 1. (b) Joint 2.
Figure 17. Comparison of torque to [43,44] in case 2: the [43] scheme (yellow solid line), the [44] scheme (red dash line), the proposed scheme (blue solid line); (a) Joint 1. (b) Joint 2.
Applsci 13 01275 g017
Figure 18. Parameter d ^ i handling TDE errors in case 2: Joint 1 (blue solid line). Joint 2 (orange solid line).
Figure 18. Parameter d ^ i handling TDE errors in case 2: Joint 1 (blue solid line). Joint 2 (orange solid line).
Applsci 13 01275 g018
Table 1. Parameters of robotic manipulator.
Table 1. Parameters of robotic manipulator.
Robotic manipulator ParameterValue
l 1 ( m )0.5
l 2 ( m ) 0.5
m 1 ( k g ) 5
m 2 ( k g ) 2
g ( m / s 2 )9.8
Table 2. Parameters of proposed controller.
Table 2. Parameters of proposed controller.
ParameterValue ( i = 1 ) Value ( i = 2 )
k p i 1010
k d i 11
k s i 0.10.1
i 0.10.1
k y i 0.10.1
β i 0.10.1
ψ i 0.010.01
Ω ¯ i 0.050.05
δ i 0.10.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, Z.; Lin, Q. Reinforcement Learning-Based Adaptive Position Control Scheme for Uncertain Robotic Manipulators with Constrained Angular Position and Angular Velocity. Appl. Sci. 2023, 13, 1275. https://doi.org/10.3390/app13031275

AMA Style

Xie Z, Lin Q. Reinforcement Learning-Based Adaptive Position Control Scheme for Uncertain Robotic Manipulators with Constrained Angular Position and Angular Velocity. Applied Sciences. 2023; 13(3):1275. https://doi.org/10.3390/app13031275

Chicago/Turabian Style

Xie, Zhihang, and Qiquan Lin. 2023. "Reinforcement Learning-Based Adaptive Position Control Scheme for Uncertain Robotic Manipulators with Constrained Angular Position and Angular Velocity" Applied Sciences 13, no. 3: 1275. https://doi.org/10.3390/app13031275

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop