Next Article in Journal
Asymptotic Chow Semistability Implies Ding Polystability for Gorenstein Toric Fano Varieties
Next Article in Special Issue
Modeling and Control of a DC-DC Buck–Boost Converter with Non-Linear Power Inductor Operating in Saturation Region Considering Electrical Losses
Previous Article in Journal
High-Quality Reversible Data Hiding Based on Multi-Embedding for Binary Images
Previous Article in Special Issue
Robust Sliding-Mode Control Design of DC-DC Zeta Converter Operating in Buck and Boost Modes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Tracking as Constrained Optimization by Uncertain Dynamic Plant: Mirror Descent Method and ASG—Version of Integral Sliding Mode Control

by
Alexander Nazin
1,*,†,
Hussain Alazki
2,† and
Alexander Poznyak
3,†
1
V. A. Trapeznikov Institute of Control Sciences, Russian Academy of Sciences, Moscow 117997, Russia
2
Facultad de Ingeniería, Universidad Autónoma del Carmen (UNACAR), Playa del Carmen 24180, Mexico
3
Automatic Control Department, Centro de Investigacion y Estudios Avanzados del Instituto Politecnico Nacional, Ciudad de Mexico 07360, Mexico
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2023, 11(19), 4112; https://doi.org/10.3390/math11194112
Submission received: 18 July 2023 / Revised: 21 September 2023 / Accepted: 26 September 2023 / Published: 28 September 2023
(This article belongs to the Special Issue Dynamics and Control Theory with Applications)

Abstract

:
A class of controlled objects is considered, the dynamics of which are determined by a vector system of ordinary differential equations with a partially known right-hand side. It is presumed that the state variables and their velocities can be measured. Designing a robust tracking controller under some constraints to admissible state variables is the research goal. This construction, which extends the results for the average subgradient technique (ASG), and is an update of the subgradient descent technique (SDM) and integral sliding mode (ISM) approach, is realized by using the Legendre–Fenchel transform. A two-link robot manipulator with three revolute joints, powered by individual PMDC motors, is presented as an illustrative example of the suggested approach implementation.

1. Introduction

1.1. Brief Survey

Constrained optimization is the process of optimizing an objective function with respect to some variables, subject to constraints on those variables. The objective function is either a cost function or energy function to be minimized, or a reward function or utility function to be maximized. Constraints can be either hard constraints, which set conditions on the variables that must be satisfied, or soft constraints, which have some variable values that are penalized in the objective function if and depending on the extent to which the conditions on the variables are not satisfied (see, for example, [1,2,3,4,5,6]).
All control strategies in most publications, treated as Static Optimization Methods (SOM), in continuous time may be represented in the following form
F ( x t ) t F : = min x X a d m R n F ( x ) ,
where F : R n R is a convex (not obligatory strongly convex) mapping, X a d m is the admissible convex set of arguments, and the process x t is generated by the simple ordinary differential equation (ODE)
x ˙ t = u t , x 0 is fixed , t 0 ,
with any initial conditions x 0 R n . The relation (1) is referred to hereafter as a static plant. All known procedures of SOM differ only in the designing of the control action u t (or an optimization algorithm) as a function of the current state x t (Markov’s strategy) or more profound available history, namely, u t = u ( t , x τ τ 0 , t ) .
Here we consider a more general, and hence, more complex situation when the process x t is generated by the dynamic plant
x ¨ t = f t , x t , x ˙ t + u t , x 0 , x ˙ 0 are fixed , t 0 , x t , u t R n ,
where the vector function f in the right-hand side is supposed to be unknown but belongs to some class C of nonlinearities.
This problem is more closed to the, so-called, Extremum Seeking Problem [7,8,9,10] and includes the first-order derivatives only. Thus, in [11], several optimization schemes are considered and it is shown that under appropriate conditions, these schemes achieve the extremum point from an arbitrarily large domain of initial conditions if the parameters in the controller are appropriately adjusted. This approach was applied in [12] for two levels of plant’s economic optimization. Many advanced process control systems use some form of model predictive control approach [13,14]. Article [15] describes a new algorithm for finding an extremum using stochastic online gradient estimation. Article [16] considers the optimization problem with constraints in dynamic linear time-invariant (LTI) systems characterized by the dimension of the control vector being smaller than the dimension of the system state vector. Convergence in finite time to a neighborhood of order ε of the optimal equilibrium point is proven. Ref. [17] presents variable structure convex programming control for a class of linear uncertain accessible state systems.
It is also crucial to discuss the relationship between the problem under consideration and the Model Predictive Control (MPC) approach, commonly referred to as moving or receding horizon control; see the two known surveys [18,19] as well as some new recent papers on Robust MPC (RMPC) [20,21,22,23]. MPC is a group of model-based control theories that anticipate system behavior using linear or nonlinear process models. The quality of the open-loop predictions, which in turn depends on the correctness of the process models, determines how well the MPC control performance operates. The estimated trajectory could not match the behavior of the plant in real life. The performance of the control system may be sluggish or unstable due to the mismatch between the plant and the model, sometimes referred to as model uncertainty. Robust predictive controllers are those that explicitly take the process and model uncertainty into account while establishing the best control strategies. Similar to H-infinity controllers, the core principle behind these controllers is to minimize the worst possible disruption to the behavior of the process. This idea is very different from the one suggested in this paper:
The prediction process cannot be accomplished precisely since the right-hand side of the ODE, representing the object model, is considered to be unknown (only dimensions of states and control are available);
Because the control action should be implemented in real-time online utilizing feedback (but not open-loop control), it is difficult to test-repeat the appropriate produced trajectories for various potential uncertainties.
In this paper, we consider a class of controlled plants with dynamics governed by a vector system of the second-order ordinary differential equations (ODE) with an unknown right-hand side. All mechanical Lagrange models belong to this class. The state variables and their velocities are assumed to be measurable. We design a controller minimizing a loss function subjected to a set of constraints to the state of the controlled plant. The designed control action is admitted to be a function of the current subgradients of loss function and constraints only, which are also supposed to be measurable online. The control is designed based on the SDM (subgradient descent method) version [24,25] of the integral sliding mode (ISM) concept [26,27,28] aimed at minimizing “on average” a given convex (not obligatory strongly convex) cost function of the current state under a set of given constraints. An optimization type algorithm is developed and analyzed using ideas from the SDM technique [1]. We prove the reachability of the “desired regime” (nonstationary analogue of sliding surface) [27] from the beginning of the process and obtain an explicit upper bound for the cost function decrement; that is, the convergence is proven and the rate of convergence is estimated as O ( t 1 ) . This paper generalizes the approach, suggested in [29] for unconstrained dynamic optimization, to the constraint optimization problem realized by an uncertain second-order dynamic plant.

1.2. Main Contributions

  • The robust tracking problem is reformulated as a constrained optimization realized by a dynamic plant with an unknown (but bounded) right-hand side. When we refer to “robust tracking”, we imply two distinct characteristics that are connected to imperfect a priori knowledge. While the exact control plant models and tracking trajectories are unavailable, a robust controller should nevertheless be able to successfully operate. It is just necessary to measure states and corresponding velocities online.
  • The cost as well as the constraints are admitted to be convex but not obligatory strictly or strongly convex.
  • The mirror descent method (MDM) and ASG version of sliding mode control are suggested and realized.
  • The convergence of the obtained trajectories of the controlled uncertain plant to the corresponding admissible zone close to the minimal point is realized.

2. Uncertain Plant Description and Admitted Dynamic Zone

2.1. Dynamic Model

The second-order dynamic model (2) can be represented in the following extended format
x ˙ 1 , t x ˙ 2 , t = x 2 , t f t , x 1 , t , x 2 , t + 0 n × n I n × n u t , x 1 , t 0 = x ˚ 1 R n , x 2 , t 0 = x ˚ 2 R n , u t R n .
Here the extended state variables x 1 , t = x t , x 2 , t = x ˙ t are the current coordinates and their velocities at time t 0 . Function f t , x 1 , t , x 2 , t is piecewise continuous in all arguments and admits to being unknown but is bounded as
f t , x 1 , x 2 k x x 1 , x 2 : = c 0 + c 1 x 1 + c 2 x 2
with final positive constants c 0 , c 1 , and c 2 . Hereafter the symbol · means the Euclidean norm.

2.2. Reference Trajectory, Tracking Error Dynamics, and Admissible Zone

The aim of the controller (which will be exactly formulated below) is to realize the tracking of the state x t for the given reference trajectory x t t 0 . Define the tracking error  δ 1 , t as
δ 1 , t : = x 1 , t x 1 , t , δ 2 , t = δ ˙ 1 , t = x 2 , t x 2 , t ,
where x 1 , t is the continuously differentiable trajectory to be tracked satisfying
x ˙ 1 , t = x 2 , t = φ t , x 1 , t , t 0 , x 1 , 0 is known .
In view of that, the error tracking dynamics can be represented as follows
δ ˙ 1 , t δ ˙ 2 , t = δ 2 , t f δ t , δ 1 , t , δ 2 , t + 0 n × n I n × n u t , f δ t , δ 1 , t , δ 2 , t : = f t , δ 1 , t + x 1 , t , δ 2 , t + x 2 , t x ˙ 2 , t .
Let us require that the dynamics of δ 1 , t should be realized after time t 0 0 within a bounded admissible zone D a d m .
This paper’s primary objective is to build a control that minimizes the tracking error δ 1 . The minimization of the assumed convex loss function F ( δ 1 ) can be used to represent this. For example, the class of convex loss functions under consideration includes the following functions:
1 ) F δ 1 = i = 1 n δ 1 , i , 2 ) F δ 1 = i = 1 n δ 1 , i ε + , z ε + : = z ε if z ε z ε if z ε 0 if z < ε .

2.3. Basic Assumptions

A1
The current states x t , x ˙ t of the plant (3) are supposed to be measurable (available) online for all t 0 .
A2
The function f t , x t , x ˙ t , satisfying (4), is piecewise continuous in all arguments and admits to being unknown.
A3
The current states x t , x ˙ t of the reference trajectory are also supposed to be available online for any t 0 .
A4
Here we assume that the subgradient (Recall that a vector a ( x ) R n , satisfying the inequality F ( x + y ) F ( x ) + a ( x ) y for all y R n , is called the subgradient of the function F ( x ) at the point x R n and is denoted by a ( x ) F ( x ) , which is the set of all subgradients of F at the point x. If F ( x ) is differentiable at a point x, then a ( x ) = F ( x ) . In the minimal point x , we have 0 F ( x ) .) of the loss function F ( δ 1 , t ) is available online for a current time t 0 , and the set of minimizers δ 1 of F ( · ) on the set D a d m includes the origin δ 1 = 0 ; that is,
0 A r g min δ 1 D a d m F ( δ 1 ) .
A5
The admissible set D a d m is nonempty convex compact, i.e., D a d m .

3. Desired Dynamics

3.1. Mirror Descent Method in Continuous Time

Let us apply the mirror descent approach using the Legendre–Fenchel transformation [30] as follows. For any ζ R n define
U ζ = max z D a d m ζ z U z , U z = 1 2 z 2 ,
so that (see, for instance, [31,32])
U ζ = arg max δ 1 D a d m ζ δ 1 U δ 1 .
Define the dynamics for the vector function ζ t R n as
ζ ˙ t = a ( δ 1 , t ) , a ( δ 1 , t ) F ( δ 1 , t ) , ζ t 0 = 0 , t + θ δ ˙ 1 , t + δ 1 , t = U ζ t η , t t 0 0 , η R n .
Remark 1. 
The second differential equation in (11) can be integrated as follows
t + θ δ 1 , t t 0 + θ δ 1 , t 0 = τ = t 0 t U ζ τ η d τ , δ 1 , t = λ t δ 1 , t 0 + 1 λ t 1 t t 0 τ = t 0 t U ζ τ η d τ D a d m , λ t : = t 0 + θ t + θ .
Therefore, δ 1 , t D a d m for all t t 0 because of convexity and due to (9) and (10).

3.2. Why the Dynamics δ 1 , t Are Desired

The following theorem explains why the dynamics δ 1 , t may be considered as desired.
Theorem 1. 
Under Assumptions A1–A5 on the trajectories δ 1 , t , generated by (11), for all t t 0 0 the following property holds
F ( δ 1 , t ) F ( δ 1 η ) + t 0 + θ t + θ F ( δ 1 , t 0 ) F ( δ 1 η ) ,
where
δ 1 η = arg min δ 1 D a d m η δ 1 + U δ 1 .
Proof. 
Defining μ t : = t + θ , δ 1 : = δ 1 η , we have from (11)
d d t U ζ t η ζ t η δ 1 = ζ ˙ t U ζ t η δ 1 = a ( δ 1 , t ) μ t δ ˙ 1 , t + δ 1 , t δ 1 = a ( δ 1 , t ) δ 1 , t δ 1 μ t a ( δ 1 , t ) δ ˙ 1 , t .
Due to the convexity property for F ( δ 1 ) , we have
a ( δ 1 , t ) ( δ 1 , t ) δ 1 , t δ 1 F ( δ 1 , t ) F ( δ 1 ) ,
and in view of the relation
a ( δ 1 , t ) δ ˙ 1 , t = d d t F ( δ 1 , t ) ,
it follows
d d t U ζ t η ζ t η δ 1 = ζ ˙ t U ζ t η δ 1 = a ( δ 1 , t ) μ t δ ˙ 1 , t + δ 1 , t δ 1 F ( δ 1 , t ) F ( δ 1 ) μ t a ( δ 1 , t ) δ ˙ 1 , t ,
or equivalently,
d d t U ζ δ , t η δ ζ δ , t η δ δ 1 F ( δ 1 , t ) F ( δ 1 ) μ t d d t F ( δ 1 , t ) .
After integration, we obtain
τ = t 0 t F ( δ 1 , τ ) F ( δ 1 ) d τ U ζ τ η ζ τ η δ 1 τ = t 0 τ = t τ = t 0 t μ τ d d τ F ( δ 1 , τ ) F ( δ 1 ) d τ = U ζ t η ζ t η δ 1 + U η + η δ 1 μ τ F ( δ 1 , τ ) F ( δ 1 ) τ = t 0 τ = t + τ = t 0 t F ( δ 1 , τ ) F ( δ 1 ) d τ ,
which implies
μ t F ( δ 1 , t ) F ( δ 1 ) U ζ t η ζ t η δ 1 + U η + η δ 1 + μ t 0 F ( δ 1 , t 0 ) F ( δ 1 ) .
Using (9), we obtain
U ζ t η ζ t η δ 1 U δ 1 ,
U ζ t η ζ t η δ 1 U δ 1 = 1 2 δ 1 2 ,
and
μ t F ( δ 1 , t ) F ( δ 1 ) 1 2 δ 1 2 + U η + η δ 1 + μ t 0 F ( δ 1 , t 0 ) F ( δ 1 ) .
Since by (9) and (10)
U η = arg max δ 1 D a d m η δ 1 U δ 1 , U δ 1 = 1 2 δ 1 2 ,
and defining
δ 1 η : = arg max δ 1 D a d m η δ 1 U δ 1 = U η ,
we obtain
U η + η δ 1 = U δ 1 = 1 2 δ 1 2 .
Therefore, we obtain
μ t F ( δ 1 , t ) F ( δ 1 ) 1 2 δ 1 2 1 2 δ 1 2 + μ t 0 F ( δ 1 , t 0 ) F ( δ 1 ) .
F ( δ 1 , t ) F ( δ 1 η ) + μ t 0 μ t F ( δ 1 , t 0 ) F ( δ 1 η ) .
Example 1. 
Assume that
D a d m : = δ 1 R n : δ 1 r .
To calculate δ 1 , according to (13), it is sufficient to note that the solution of the problem
2 η δ 1 + δ 1 2 = δ 1 + η 2 η 2 min δ 1 r ,
is
δ 1 η = η i f η r η η r i f η > r .

4. Robust Controller Design

4.1. Auxiliary Sliding Variable and Its Dynamics

Introduce a new auxiliary variable (sliding variable)
s t = t + θ δ 2 , t + δ 1 , t U ζ t η , t t 0 0 .
Notice that the function s t is measurable online, and that the situation when
s t = 0 for all t t 0
corresponds exactly to the desired regime (11), starting from the moment t 0 . Then for V s t = 1 2 s t 2 in view of (7) and the first equation in (11) we have
d d t V s t = s t s ˙ t = s t 2 δ ˙ 1 , t + t + θ δ ˙ 2 , t d d t U ζ t η = s t 2 δ 2 , t + t + θ f t , δ 1 , t + x 1 , t , δ 2 , t + x 2 , t x ˙ 2 , t + u t 2 U ζ t η ζ ˙ t = t + θ s t f t , δ 1 , t + x 1 , t , δ 2 , t + x 2 , t + t + θ s t 2 t + θ δ 2 , t x ˙ 2 , t + u t + 1 t + θ 2 U ζ t η a ( δ 1 , t ) k t Sign s t t + θ s t f t , δ 1 , t + x 1 , t , δ 2 , t + x 2 , t t + θ k t s t Sign s t t + θ s t c 0 + c 1 δ 1 , t + x 1 , t + c 2 δ 2 , t + x 2 , t k x , t : = k x ( δ 1 , t + x 1 , t , δ 2 , t + x 2 , t ) k t s t Sign s t .
Here
Sign s t = sign s 1 , t , , sign s n , t , sign s i , t = + 1 if s i , t > 0 = 1 if s i , t < 0 1 , + 1 if s i , t = 0 .

4.2. Robust Control Structure

Since
s t Sign s t = i = 1 n s i , t s t
and taking
k t = k x , t + ρ , ρ > 0 ,
we obtain
d d t V s t t + θ s t k x , t k t = t + θ ρ 2 V s t ,
which implies
d V s t V s t t + θ 2 ρ d t , 2 V s t V s t 0 2 2 ρ t + θ 2 t 0 + θ 2 , 0 V s t V s t 0 2 4 ρ t + θ 2 t 0 + θ 2 .
This means that for all t t r e a c h , where
t r e a c h : = t : V s t 0 2 4 ρ t + θ 2 t 0 + θ 2 = 0 = 2 ρ s t 0 + t 0 + θ 2 θ .
Finally, the robust control is
u t = 2 t + θ δ 2 , t + x ˙ 2 , t 1 t + θ 2 U ζ t η a ( δ 1 , t ) k t Sign s t = u c o m p , t + u d i s c , t ,
where
u c o m p , t : = 2 t + θ δ 2 , t + x ˙ 2 , t 1 t + θ 2 U ζ t η a ( δ 1 , t ) , u d i s c , t : = k t Sign s t .
Remark 2. 
If we wish to obtain t r e a c h = t 0 = 0 , we need to complete the identity
s 0 = θ δ 2 , 0 + δ 1 , 0 U η = ( 14 ) θ δ 2 , 0 + δ 1 , 0 δ 1 η = 0 .
Since δ 1 η D a d m , we may conclude that parameters θ > 0 , η and initial conditions δ 1 , 0 , δ 2 , 0 should be consistent in the sense that
θ δ 2 , 0 + δ 1 , 0 D a d m .
Remark 3. 
For example, with Euclidean r-ball in R n being the admissible set D a d m , from (9) and (10) one has
U ζ = arg max δ 1 D a d m ζ δ 1 U δ 1 = ζ if ζ r r ζ ζ if ζ > r ,
δ 1 η = arg min δ 1 D a d m η δ 1 + U δ 1 = arg min δ 1 D a d m η δ 1 + 1 2 δ 1 2 = η if η r .
From (18) it follows that
θ δ 2 , 0 + δ 1 , 0 = η , η r ,
and
2 U ζ = I n × n if ζ r r ζ I n × n ζ ζ T ζ 2 if ζ > r .
Notice that U -function (10) is nondifferential in the points of r-sphere of the ball, and it is continuously differential in all other points of R n . The formulas in (20) and (23) are presented as their continuous versions on the ball, including the r-sphere.

4.3. Main Result

We are ready to formulate the main result.
Theorem 2. 
Under Assumptions A1–A5, the robust control (17) and (18) with parameter η , satisfying (19), provides the property
F ( δ 1 , t ) F ( δ 1 η ) + θ t + θ F ( δ 1 , 0 ) F ( δ 1 η )
for all t 0 and any regularizing parameter θ > 0 .
Proof. 
In view of the relation (19) of the parameter η and initial conditions δ 1 , 0 , δ ˙ 1 , 0 , the auxiliary variable s t = 0 for all t 0 starting from the beginning of the control process. Using Formula (12) for t 0 = 0 we obtain (24). □

5. Discussion

Equations (14) and (19) hold under θ > 0 , η R n for the following cases:
  • Zero initial conditions δ 1 , 0 = 0 , δ 2 , 0 = 0 . Thus, η = 0 for arbitrary θ > 0 (see, as an example, the first item in the loss function (8)).
  • Nonzero initial conditions δ 1 , 0 , δ 2 , 0 are collinear oppositely directed vectors. Therefore, θ > 0 and η = 0 exist (see, as an example, the first item in the loss function (8)).
  • Equation (22) holds under nonzero vector η with a sufficiently small η ϵ and for θ > 0 (see, as an example, the second item in the loss function (8)).

6. Numerical Example

A two-link robot manipulator with three revolute joints powered by individual PMDC motors is presented below as an illustrative example of the suggested approach.

6.1. Model Description

A dynamic model of a Lagrangian mechanical system with n degrees of freedom in standard form driven by n independent Permanent Magnet Direct Current (PMDC) motors [33] is defined by the following system of differential equations:
D q t q ¨ t + C q t , q ˙ t q ˙ t + G q t = τ t + ϑ t , τ t = W K a I a t , L a I ˙ a t + R a I a t + K e W q ˙ t = v a t ,
where q t , q ˙ t R n are the state and velocity vectors, τ t R n is a vector of external torques, I a t R n is the armature current vector, W R n × n is the matrix of electromotive force constants (possibly taking into account engine gear ratios), K a R n × n is the matrix of constants of direct electromotive forces, D q t = M q t + W J W is a positive definite inertia matrix, that is D q t = D q t d I n × n , d > 0 and, therefore, invertible for all q t , J = diag J 1 , J 2 , , J n is the rotor inertia matrix, M q t is the matrix of the Lagrangian system corresponding to the armature inertia matrix in the original coordinates, C q t , q ˙ t R n × n is the matrix corresponding to the generalized nonpotential forces C q t , q ˙ t q ˙ t , which can describe friction, hysteresis, Coriolis, damping, centripetal effects, etc., G q t R n is a vector corresponding to the generalized potential forces, K e = diag K e 1 , K e 2 , , K e n is the matrix of reverse electromotive force constants, L a = diag L a 1 , L a 2 , , L a n and R a = diag R a 1 , R a 2 , , R a n are armature inductance and resistant positive matrices, respectively, ϑ t R n is the disturbance (or uncertainty) vector, and v a t R n is the armature voltage vector, which is treated below as a control designed to achieve the desired behavior. In fact, the third equation in (25) describes the dynamics of the actuator implementing the applied control action v a t . Equation (25) assumes fully allocated control.
We assume that q t , q ˙ t , and I a t are available online. From (25) it follows that
I a t I a t 0 = L a 1 R a τ = t 0 t I a τ d τ L a 1 K e W q t q t 0 + L a 1 τ = t 0 t v a τ d τ
( t 0 0 —any fixed time), and selecting (neglecting the Joule effect, related to the dependence of the winding motor resistance)
v a t = v a t ( 1 ) + v a t ( 2 ) ,
with
v a t ( 1 ) = R a I a t + K e W q ˙ t ,
the relation (26) becomes
I a t = I a t 0 + L a 1 τ = t 0 t v a τ 2 d τ .
Substituting (29) into (25) gives
D q t q ¨ t + C q t , q ˙ t q ˙ t + G q t = u t + ϑ ˜ t ,
u t : = W K a L a 1 τ = t 0 t v a τ 2 d τ , ϑ ˜ t : = W K a I a t 0 + ϑ t .
Note that in the standard matrix format (3) with new state vectors  x 1 = q R n and velocity vectors  x 2 = q ˙ R n , the Lagrange dynamics under consideration (30) have the following form:
x ˙ t = x ˙ 1 t x ˙ 2 t = H x 1 t , x 2 t x 1 t x 2 t + B u t + ξ t , H x 1 , x 2 = 0 I n × n 0 D 1 x 1 C x 1 , x 2 , B = 0 n × n D 1 x 1 , ξ t = 0 n × n D 1 x 1 t ϑ ˜ t G x 1 .
Thus, in this representation, the dimension of control vector u is n and the extended state dimension x = x 1 , x 2 is 2 n .

6.2. Intended Moving Point

The considered mechanical construction is depicted in Figure 1.
The challenge is to move the robot’s cueing point such that it follows the intended moving point x 1 , t = ( x t , y t , z t ) R 3 , maintaining a tiny space between them. Here, we suppose that the point x 1 , t accomplishes travel along a certain ellipse  E x ˚ 1 , r 0 at a constant module speed v:
E l x ˚ 1 , r 0 : = x 1 = x , y , z : x t = x ˚ + r 0 sin ω t , y t = y ˚ + r 0 cos ω t , z t = z ˚ , x ˙ t = v = r 0 ω .
The corresponding cost function F (8) to be minimized is the function of the tracking error δ 1 , t = x t x t , y t y t , z t z t , and is selected as
F δ 1 , t = F 1 x t x t + F 2 y t y t + F 3 z t z t , F 1 x t x t = x t x t , F 2 y t y t = y t y t , F 3 z t z t = z t z t , a δ 1 , t = sign x t x t , sign y t y t , sign z t z t F δ 1 , t .

6.3. Relation between Cartesian and Angular Coordinates

The relation between vectors x 1 and q
x 1 = F q , x 1 : = x y z , q : = φ 1 φ 2 φ 3
is as follows:
x t = l 1 cos φ 2 cos φ 1 + l 2 cos φ 2 + φ 3 cos φ 1 , y t = l 1 cos φ 2 sin φ 1 + l 2 cos φ 2 + φ 3 sin φ 1 , z t = l 1 sin φ 2 + l 2 sin φ 2 + φ 3 .
To implement the simulation, the following form of immeasurable nonpotential forces (friction, hysteresis, Coriolis, damping, centripetal effects, and others) was modeled as
C q t , q ˙ t q ˙ t = k r e s q ˙ t Sign q ˙ t q ˙ t , k r e s > 0 .

6.4. Applied Robust Controller Structure

In this particular case, the suggested robust controller (17) and (18) is as follows:
u t = u c o m p , t + u d i s c , t ,
with the compensation control part u c o m p , t equal to
u c o m p , t : = 2 t + θ δ 2 , t + x ˙ 2 , t 1 t + θ 2 U ζ t η a ( δ 1 , t ) ,
where a ( δ 1 , t ) is defined in (34), and
δ 2 , t = δ ˙ 1 , t = x ˙ 1 , t x ˙ 1 , t , x ˙ 2 , t = x ¨ 1 , t = Q x t x ˚ , ζ t = τ = 0 t a δ 1 , τ d τ , η = θ δ 2 , 0 + δ 1 , 0 , η r , 2 U ζ t η = I n × n if ζ t η r , r ζ t η I n × n ζ t η ζ t η T ζ t η 2 if ζ t η > r .
The discontinuous control u d i s c , t is designed as
u d i s c , t : = k t Sign s t , s t = t + θ δ 2 , t + δ 1 , t U ζ t η , U ζ t η = ζ t η if ζ t η r r ζ t η ζ t η if ζ t η > r .

6.5. Parameters of Simulation

The following are the computer simulation parameters:
ParameterNumerical ValueDescription
k r e s 1.1  × 10 6 environmental (air) resistance
g9.81 m/s 2 Gravitational acceleration
m 1 , m 2 1 kgMass
l 1 , l 2 0.35 m, 0.67 mLength
W = 0.05 0.04 0.06 0.04 0.044 0.022 0.06 0.022 0.057 , L a = 2 0 0 0 2 0 0 0 2 , J = 0.021 0.0006 0 0.0006 0.021 0 0 0 0.0407 , K a = 0.02 0.04 0.06 0.04 0.08 0.1 0.06 0.1 0.04 , r 0 = 0.4 , ω = 1 ,
ϑ t is generated by uniformly distributed random numbers at an interval [−0.8, 0.8], and
K e = diag 70 , 140 , 151 , θ = 10 3 , Q = diag 0.5 , 0.9 , 2 , x ˚ = 0.35 , 2.35 , 0.8 , r = 0.6 , δ 1 , 0 = 0.4001 0.7576 0.5236 , δ 2 , 0 = 0.3990 0.1802 0.0012 η = 0.3996 0.7570 0.6283 calculated as η = θ δ 2 , 0 + δ 1 , 0 ,

6.6. Results of Numerical Simulations

The simulation results are shown in Figure 2, Figure 3, Figure 4, Figure 5 and Figure 6.
As one can see, the suggested method demonstrates a successful workability in the presence of essential model uncertainties and external perturbations.

7. Conclusions

-
The constrained optimization problem is addressed in this study using a second-order differential controlled plant with an unknown (but bounded) right side of the model.
-
The desired dynamics in the tracking error variables is designed based on the mirror descent method.
-
The continuous time convergence to the set of minimizing points is established, and the associated rate of convergence is analytically evaluated.
-
The robust controller, containing both the continuous (compensating) u c o m p and the discontinuous u d i s c , is proposed using the ASG version of the integral sliding mode approach.
-
The suggested controller, under the special relations of it parameters with the initial conditions, is proved to provide the desired regime from the beginning of the control process.
-
This method may have several applications in the development of robust control in mechanical systems, including soft robotics and moving dynamic plants.

Author Contributions

Conceptualization, A.N. and A.P.; methodology, H.A.; formal analysis, A.N. and A.P.; data curation: H.A.; resources: H.A.; software: H.A.; validation: A.N. and A.P.; visualization: A.N., H.A. and A.P.; writing—original draft preparation, A.N. and A.P.; writing—review and editing, A.N. and A.P.; supervision, A.P.; project administration, A.P.; funding acquisition, A.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ASGAverage subgradient
SDMSubgradient descent Method
ISMIntegral sliding mode
SOMStatic Optimization Methods
ODEOrdinary differential equation

References

  1. Bertsekas, D.P. Constrained Optimization and Lagrange Multiplier Methods; Academic Press: New York, NY, USA, 1982; ISBN 0-12-093480-9. [Google Scholar]
  2. Dechter, R. Constraint Processing; Morgan Kaufmann: Burlington, MA, USA, 2003; ISBN 1-55860-890-7. [Google Scholar]
  3. Leader, J.J. Numerical Analysis and Scientific Computation; Addison Wesley: Boston, MA, USA, 2003; ISBN 0-201-73499-0. [Google Scholar]
  4. Posser, M.; Posser, M.J. Basic Mathematics for Economists; Routledge: New York, NY, USA, 1993; ISBN 0-415-08424-5. [Google Scholar]
  5. Rossi, F.; van Beek, P.; Walsh, T. Chapter 1—Introduction. In Foundations of Artificial Intelligence: Handbook of Constraint Programming; Rossi, F., van Beek, P., Walsh, T., Eds.; Elsevier: Amsterdam, The Netherlands, 2006; Volume 2, pp. 3–12. [Google Scholar] [CrossRef]
  6. Sun, W.; Yua, Y.-X. Optimization Theory and Methods: Nonlinear Programming; Springer: New York, NY, USA, 2010; ISBN 978-1441937650. [Google Scholar]
  7. Rastrigin, L.A. Systems of Extremal Control; Nauka: Moscow, Russia, 1974. (In Russian) [Google Scholar]
  8. Krstic, M.; Wang, H.H. Stability of extremum seeking feedback for general nonlinear dynamic systems. Automatica 2000, 36, 595–601. [Google Scholar] [CrossRef]
  9. Ariyur, K.B.; Krstic, M. Real-Time Optimization by Extremum-Seeking Control; John Wiley & Sons: Hoboken, NJ, USA, 2003. [Google Scholar]
  10. Tan, Y.; Moase, W.H.; Manzie, C.; Nešić, D.; Mareels, I.M.Y. Extremum seeking from 1922 to 2010. In Proceedings of the 29th Chinese Control Conference, Beijing, China, 29–31 July 2010; pp. 14–26. [Google Scholar]
  11. Tan, Y.; Nešić, D.; Mareels, I. On non-local stability properties of extremum seeking control. Automatica 2006, 42, 889–903. [Google Scholar] [CrossRef]
  12. Rawlings, J.B.; Amrit, R. Optimizing process economic performance using model predictive control. In Nonlinear Model Predictive Control; Magni, L., Raimondo, D.M., Allgöwer, F., Eds.; Lecture Notes in Control and Information Sciences; Springer: Berlin/Heidelberg, Germany, 2009; Volume 384, pp. 119–138. [Google Scholar]
  13. Dehaan, D.; Guay, M. Extremum-seeking control of state-constrained nonlinear systems. Automatica 2005, 41, 1567–1574. [Google Scholar] [CrossRef]
  14. Chunlei, Z.; Ordóñez, R. Robust and adaptive design of numerical optimization-based extremum seeking control. Automatica 2009, 45, 634–646. [Google Scholar]
  15. Solis, C.U.; Clempner, J.B.; Poznyak, A.S. Extremum seeking by a dynamic plant using mixed integral sliding mode controller with synchronous detection gradient estimation. Int. J. Robust Nonlinear Control 2018, 29, 702–714. [Google Scholar] [CrossRef]
  16. Ferrara, A.; Utkin, V.I. Sliding Mode Optimization in Dynamic LTI Systems. J. Optim. Theory Appl. 2002, 115, 727–740. [Google Scholar] [CrossRef]
  17. Ferrara, A. A variable structure convex programming based control approach for a class of uncertain linear systems. Syst. Control Lett. 2005, 54, 529–538. [Google Scholar] [CrossRef]
  18. Bemporad, A.; Morari, M. Robust model predictive control: A survey. In Robustness in Identification and Control; Garulli, A., Tesi, A., Eds.; Lecture Notes in Control and Information Sciences; Springer: London, UK, 1999; Volume 245, pp. 207–226. [Google Scholar] [CrossRef]
  19. Jalali, A.A.; Nadimi, V. A Survey on Robust Model Predictive Control from 1999–2006. In Proceedings of the 2006 International Conference on Computational Inteligence for Modelling Control and Automation and International Conference on Intelligent Agents Web Technologies and International Commerce (CIMCA’06), Sydney, NSW, Australia, 28 November–1 December 2006; p. 207. [Google Scholar] [CrossRef]
  20. Li, H.; Wang, S.; Shi, H.; Su, C.; Li, P. Two-Dimensional Iterative Learning Robust Asynchronous Switching Predictive Control for Multiphase Batch Processes with Time-Varying Delays. IEEE Trans. Syst. Man Cybern. Syst. 2023, 53, 6488–6502. [Google Scholar] [CrossRef]
  21. Wang, L.; Zhang, W.; Zhang, Q.; Shi, H.; Zhang, R.; Gao, F. Terminal constrained robust hybrid iterative learning model predictive control for complex time-delayed batch processes. Nonlinear Anal. Hybrid Syst. 2023, 47, 101276. [Google Scholar] [CrossRef]
  22. Liu, X.; Ma, L.; Kong, X.; Lee, K.Y. An efficient iterative learning predictive functional control for nonlinear batch processes. IEEE Trans. Cybern. 2020, 52, 4147–4160. [Google Scholar] [CrossRef] [PubMed]
  23. Shi, H.; Li, P.; Cao, J.; Su, C.; Yu, J. Robust fuzzy predictive control for discrete-time systems with interval time-varying delays and unknown disturbances. IEEE Trans. Fuzzy Syst. 2019, 28, 1504–1516. [Google Scholar] [CrossRef]
  24. Simpson-Porco, J.W. Input/output analysis of primal-dual gradient algorithms. In Proceedings of the 54th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 27–30 September 2016; pp. 219–224. [Google Scholar]
  25. Nazin, A.V. Algorithms of Inertial Mirror Descent in Convex Problems of Stochastic Optimization. Autom. Remote Control 2018, 79, 78–88. [Google Scholar] [CrossRef]
  26. Utkin, V. Sliding Modes in Control and Optimization; Springer: Berlin/Heidelberg, Germany, 1992. [Google Scholar]
  27. Fridman, L.; Poznyak, A.; Bejarano, F.J. Robust Output LQ Optimal Control via Integral Sliding Modes; Birkhäuser: Basel, Switzerland; Springer Science and Business Media: New York, NY, USA, 2014. [Google Scholar]
  28. Utkin, V.; Poznyak, A.; Orlov, Y.V.; Polyakov, A. Conclusions. In Road Map for Sliding Mode Control Design; SpringerBriefs in Mathematics; Springer International Publishing: Cham, Switzerland, 2020; pp. 125–127. [Google Scholar]
  29. Poznyak, A.S.; Nazin, A.V.; Alazki, H. Integral Sliding Mode Convex Optimization in Uncertain Lagrangian Systems Driven by PMDC Motors: Averaged Subgradient Approach. IEEE Trans. Autom. Control 2021, 66, 4267–4273. [Google Scholar] [CrossRef]
  30. Rockafellar, R.T. Convex Analysis; Princeton University Press: Princeton, NJ, USA, 1970. [Google Scholar]
  31. Ben-Tal, A.; Nemirovski, A. The Conjugate Barrier Mirror Descent Method for Non-Smooth Convex Optimization; Minerva Optimization Center, Technion Institute of Technology: Haifa, Israel, 1999. [Google Scholar]
  32. Juditsky, A.B.; Nazin, A.V.; Tsybakov, A.B.; Vayatis, N. Recursive aggregation of estimators by the mirror descent algorithm with averaging. Probl. Inf. Transm. 2005, 41, 368–384. [Google Scholar] [CrossRef]
  33. Patel, D.K. Mathematical modeling of open loop PMDC motor using MATLAB/Simulink. Int. J. Eng. Develop. Res 2015, 3, 495–500. [Google Scholar]
Figure 1. Robot with two links.
Figure 1. Robot with two links.
Mathematics 11 04112 g001
Figure 2. Tracking ellipse.
Figure 2. Tracking ellipse.
Mathematics 11 04112 g002
Figure 3. x t tracking of x t .
Figure 3. x t tracking of x t .
Mathematics 11 04112 g003
Figure 4. Tracking of y t with y t .
Figure 4. Tracking of y t with y t .
Mathematics 11 04112 g004
Figure 5. Tracking of z t with z t .
Figure 5. Tracking of z t with z t .
Mathematics 11 04112 g005
Figure 6. Cost function components.
Figure 6. Cost function components.
Mathematics 11 04112 g006
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nazin, A.; Alazki, H.; Poznyak, A. Robust Tracking as Constrained Optimization by Uncertain Dynamic Plant: Mirror Descent Method and ASG—Version of Integral Sliding Mode Control. Mathematics 2023, 11, 4112. https://doi.org/10.3390/math11194112

AMA Style

Nazin A, Alazki H, Poznyak A. Robust Tracking as Constrained Optimization by Uncertain Dynamic Plant: Mirror Descent Method and ASG—Version of Integral Sliding Mode Control. Mathematics. 2023; 11(19):4112. https://doi.org/10.3390/math11194112

Chicago/Turabian Style

Nazin, Alexander, Hussain Alazki, and Alexander Poznyak. 2023. "Robust Tracking as Constrained Optimization by Uncertain Dynamic Plant: Mirror Descent Method and ASG—Version of Integral Sliding Mode Control" Mathematics 11, no. 19: 4112. https://doi.org/10.3390/math11194112

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop