Chapter 5

PID Control with Neural Compensation

Abstract

The simplest method to decrease the tracking error of PD control is to add an integral action, i.e., change the neural PD control into neural PID control. A natural question is: Why do we not add an integrator instead of increasing derivative gain in the neural PD control?

In this chapter, the well-known neural PD control of robot manipulators is extended to the neural PID control. It is an industrial linear PID controller adding a neural compensator. The semiglobal asymptotic stability of this novel neural control is proven. Explicit conditions for choosing PID gains are given. When the measurement of velocities it is not available, local asymptotic stability is also proven with a velocity observer. Unlike the other neural controllers of robot manipulators, our neural PID does not need a big derivative and integral gains to assure asymptotic stability. We apply this neural control to our 3-DoF exoskeleton robot in CINVESTA-IPN. Experimental results show that this neural PID control has many advantages over classic PID control, neural PD control, and the other neural PID control.

Keywords

Neural networks; Linear PID control

5.1 Stable neural PID control

Without flexible links and high-frequency joint dynamics, rigid robots can be expressed in the Lagrangian form:

M ( q ) q + C ( q , q ) q + G ( q ) + F ( q ˙ ) = u

Image (5.1)

where qRnImage represents the link positions. M(q)Image is the inertia matrix, C(q,q˙)={ckj}Image represents centrifugal force, G(q)Image is a vector of gravity torques, and F(q˙)Image is friction. All terms M(q)Image, C(q,q)Image, G(q)Image, and F(q˙)Image are unknown. uRnImage is the control input. The friction F(q˙)Image is represented by the Coulomb friction model:

F ( q ˙ ) = K f 1 q ˙ + K f 2 tanh ( k f 3 q ˙ )

Image (5.2)

where kf3Image is a large positive constant, such that tanh(kf3q˙)Image can approximate sign(q˙)Image, and Kf1Image and Kf2Image are positive coefficients. In this paper, we use a simple model for the friction as in [85] and [71],

F ( q ˙ ) = K f 1 q ˙

Image (5.3)

When G(q)Image and F(q˙)Image are unknown, we may use a neural network to approximate them as

f ( q , q ) = G ( q ) + F ( q ˙ ) f ˆ ( q , q ) = W ˆ σ ( q , q ) , f ( q , q ) = W σ ( q , q ) + ϕ ( q )

Image (5.4)

where WImage is an unknown constant weight, WˆImage is the estimated weight, ϕ(q,q)Image is the neural approximation error, σ is a neural activation function. Here, we use the Gaussian function such that σ(q,q)0Image.

Since the joint velocity q˙Image is not always available, we may use a velocity observer to approximate it. This linear-in-the-parameter net is the simplest neural network. According to the universal function approximation theory, the smooth function f(q,q)Image can be approximated by a multilayer neural network with one hidden layer in any desired accuracy provided proper weights and hidden neurons:

f ˆ ( q , q ) = W ˆ σ ( V ˆ [ q , q ] ) , f ( q ) = W σ ( V [ q , q ] ) + ϕ ( q , q )

Image (5.5)

where WˆRn×mImage, VˆRm×nImage, m is hidden node number, VˆImage is the weight in the hidden layer. In order to simplify the theory analysis, we first use linear-in-the-parameter net (5.4). Then we will show that the multilayer neural network (5.5) can also be used for the neural control of robot manipulators. The robot dynamics (5.1) have the following standard properties [131], which will be used to prove stability.

P5.1. The inertia matrix M(q)Image is symmetric positive definite, and

0 < λ m { M ( q ) } M λ M { M ( q ) } β , β > 0

Image (5.6)

where λM{M}Image and λm{M}Image are the maximum and minimum eigenvalues of the matrix M.

P5.2. For the centrifugal and Coriolis matrix C(q,q˙)Image, there exists a number kc>0Image such that

C ( q , q ˙ ) q ˙ k c q ˙ 2 , k c > 0

Image (5.7)

and M˙(q)2C(q,q˙)Image is skew symmetric, i.e.,

x T [ M ˙ ( q ) 2 C ( q , q ˙ ) ] x = 0

Image (5.8)

also

M ˙ ( q ) = C ( q , q ˙ ) + C ( q , q ˙ ) T

Image (5.9)

P5.3. The neural approximation error ϕ(q,q)Image is Lipschitz over q and qImage

ϕ ( x ) ϕ ( y ) k ϕ x y

Image (5.10)

From (5.4), we know

G ( q ) + F ( q ˙ ) = W σ ( q , q ) + ϕ ( q , q )

Image (5.11)

Because G(q)Image and F(q˙)Image satisfy the Lipschitz condition, P5.3 is established.

In order to simplify calculation, we use the simple model for the friction as in (5.3). The lower bound of ϕ(q)dqImage can be estimated as

0 t ϕ ( q , q ) d q = 0 t G ( q ) d q + 0 t F ( q ˙ ) d q 0 t W σ ( q ) d q

Image (5.12)

where U(qt)Image is the potential energy of the robot, Uq=G(q)Image. Since σ()Image is a Gaussian function, Wσ(q)>0Image. By U(qt)>0Image,

0 t ϕ ( q , q ) d q > K f 1 q t K f 1 q 0 1 2 π W

Image

where 0tσ(q)=12πerf(q)Image. Since the work space of a manipulator (the entire set of points reachable by the manipulator) is known, min{qt}Image can be estimated. We define the lower bound of 0tϕ(q)dqImage as

k ϕ = K f 1 min { q t } K f 1 q 0 1 2 π W

Image (5.13)

Given a desired constant position qdRnImage, the objective of robot control is to design the input torque u in (5.1) such that the regulation error

q ˜ = q d q

Image (5.14)

q˜0Image and q˜0Image when initial conditions are in arbitrary large domain of attraction.

The classic industrial PID law is

u = K p q ˜ + K i 0 t q ˜ ( τ ) d τ + K d q ˜

Image (5.15)

where KpImage, KiImage, and KdImage are proportional, integral, and derivative gains of the PID controller, respectively.

When the unknown dynamic f(q,q)Image in (5.4) is big, in order to assure asymptotic stability, the integral gain KiImage has to be increased. This may cause big overshoot, bad stability, and integrator windup. Model-free compensation is an alternative solution, where f(q)Image is estimated by a neural network as in (5.4). Normal neural PD control is [85]

u = K p q ˜ + K d q ˜ + f ˆ

Image (5.16)

where fˆ(q,q)=Wˆσ(q,q)Image. With the filtered error r=q˜+Λq˜Image, (5.16) becomes

u = K v r + f ˆ

Image (5.17)

The control (5.17) avoids integrator problems in (5.15). Unlike industrial PID control, they cannot reach asymptotic stability. The stability condition of the neural PD control (5.16) is r>BKvImage, B is a constant [87]. In order to decrease rImage, KdImage has to be increased. This causes a long settling time problem. The asymptotic stability (r0Image) requires KvImage.

An integrator is added into the normal neural PD control (5.16). It has a similar form as the industrial PID in (5.15),

u = K p q ˜ + K d q ˜ + K i 0 t q ˜ ( τ ) d τ + f ˆ

Image (5.18)

Because in the regulation case q˙d=0Image, q˜=q˙Image, the PID control law can be expressed via the following equations:

u = K p q ˜ K d q ˙ + ξ + W ˆ σ ( q , q ) ξ ˙ = K i q ˜ , ξ ( 0 ) = ξ 0

Image (5.19)

We require the PID control part of (5.19) is decoupled, i.e., Kp,KiImage, and KdImage are positive definite diagonal matrices. The closed-loop system of the robot (5.1) is

M ( q ) q ¨ + C ( q , q ˙ ) q ˙ + f ˜ ( q , q ) = K p q ˜ K d q ˙ + ξ ξ ˙ = K i q ˜

Image (5.20)

where f˜=ffˆImage

f ˜ = W σ ( q ) + ϕ ( q ) W ˆ σ ( q ) = W ˜ σ ( q ) + ϕ ( q )

Image (5.21)

Here, W˜=WWˆImage. In matrix form the closed-loop system is

d d t [ ξ q ˜ q ˜ ] = [ K i q ˜ q ˙ q ¨ d + M 1 ( C q ˙ + W ˜ σ ( q ) + ϕ ( q , q ) K p q ˜ + K d q ˙ ξ ) ]

Image (5.22)

The equilibrium of (5.22) is [ξ,q˜,q˜]=[ξ,0,0]Image. Since at equilibrium point q=qdImage and qd=0Image the equilibrium is [ϕ(qd),0,0]Image. We simplify ϕ(qd,0)Image as ϕ(qd)Image.

In order to move the equilibrium to the origin, we define

ξ ˜ = ξ ϕ ( q d )

Image (5.23)

The final closed-loop equation becomes

M ( q ) q ¨ + C ( q , q ˙ ) q ˙ + W ˜ σ ( q , q ) + ϕ ( q , q ) = K p q ˜ K d q ˙ + ξ ˜ + ϕ ( q d ) ξ ˜ = K i q ˜

Image (5.24)

The following theorem gives the stability analysis of the neural PID control. From this theorem, we can see how to choose the PID gains and how to train the weight of the neural compensator in (5.19). Another important conclusion is that the neural PID control (5.19) can force the error q˜Image to zero.

Theorem 5.1

Consider the robot dynamic (5.1) controlled by the neural PID control (5.19). The closed-loop system (5.24) is semiglobally asymptotically stable at the equilibrium x=[ξϕ(qd),q˜,q˜]T=0Image, provided that control gains satisfy

λ m ( K p ) 3 2 k ϕ λ M ( K i ) β λ m ( K p ) λ M ( M ) λ m ( K d ) β + λ M ( M )

Image (5.25)

where β=λm(M)λm(Kp)3Image, kϕImage satisfies (5.10), and the weight of the neural networks (5.4) is tuned by

W ˆ = K w σ ( q , q ) ( q ˙ + α q ˜ ) T

Image (5.26)

where α is positive design constant, it satisfies

1 3 λ m ( M ) λ m ( K p ) λ M ( M ) α 3 λ m ( K i 1 ) λ m ( K p )

Image (5.27)

Proof

We construct a Lyapunov function as

V = 1 2 q ˙ T M q ˙ + 1 2 q ˜ T K p q ˜ + 0 t ϕ ( q , q ) d q k ϕ + q ˜ T ϕ ( q d ) + 3 2 ϕ ( q d ) T K p 1 ϕ ( q d ) + α 2 ξ ˜ T K i 1 ξ ˜ + q ˜ T ξ ˜ α q ˜ T M q ˙ + α 2 q ˜ T K d q ˜ + 1 2 t r ( W ˜ T K w 1 W ˜ )

Image (5.28)

where kϕImage is defined in (5.13) such that V(0)=0Image. α is a design positive constant. We first prove V is a Lyapunov function, V0Image. The term 12q˜TKpq˜Image is separated into three parts, and V=i=14ViImage:

V 1 = 1 6 q ˜ T K p q ˜ + q ˜ T ϕ ( q d ) + 3 2 ϕ ( q d ) T K p 1 ϕ ( q d ) V 2 = 1 6 q ˜ T K p q ˜ + q ˜ T ξ ˜ + α 2 ξ ˜ T K i 1 ξ ˜ V 3 = 1 6 q ˜ T K p q ˜ α q ˜ T M q ˙ + 1 2 q ˙ T M q ˙ V 4 = 0 t ϕ ( q ) d τ k ϕ + α 2 q ˜ T K d q ˜ + 1 2 t r ( W ˜ T K w 1 W ˜ ) 0

Image (5.29)

It is easy to find

V 1 = 1 2 [ q ˜ ϕ ( q d ) ] T [ 1 3 K p I I 3 K p 1 ] [ q ˜ ϕ ( q d ) ]

Image (5.30)

Since Kp0Image, V1Image is a semipositive definite matrix, V10Image. When α3λm(Ki1)λm(Kp)Image,

V 2 1 2 ( 1 3 λ m ( K p ) q ˜ 3 λ m ( K p ) ξ ˜ ) 2 0

Image (5.31)

Because

y T A x y A x y A x | λ M ( A ) | y x

Image (5.32)

when α13λm(M)λm(Kp)λM(M)Image,

V 3 1 2 ( λ m ( M ) q ˙ 1 3 λ m ( K p ) q ˜ ) 2 0

Image (5.33)

Obviously, if

1 3 λ m ( K i 1 ) λ m 3 2 ( K p ) λ m 1 2 ( M ) λ M ( M )

Image (5.34)

there exists

1 3 λ m ( M ) λ m ( K p ) λ M ( M ) α 3 λ m ( K i 1 ) λ m ( K p )

Image (5.35)

This means if KpImage is sufficiently large or KiImage is sufficiently small, (5.34) is established, and V(q˙,q˜,ξ˜)Image is globally positive definite. Using ddt0tϕ(q,q)dq=0tϕ(q,q)dqqqt=q˙Tϕ(q,q)Image, ddtϕ(qd)=0Image and ddt[q˜Tϕ(qd)]=q˜Tϕ(qd)Image, the derivative of V is

V ˙ = q ˙ T M q ¨ + 1 2 q ˙ T M q ˙ + q ˜ T K p q ˜ + ϕ ( q , q ) T q ˙ + q ˜ T ϕ ( q d ) + t r ( W ˜ T K w 1 W ˜ ) + α ξ ˜ T K i 1 ξ ˜ + q ˜ T ξ ˜ + q ˜ T ξ ˜ α ( q ˜ T M q ˙ + q ˜ T M q ˙ + q ˜ T M q ¨ ) + α q ˜ T K d q ˜

Image (5.36)

Using (5.8) the first three terms of (5.36) become

q ˙ T ϕ ( q ) q ˙ T K d q ˙ + q ˙ T ξ ˜ + q ˙ T ϕ ( q d ) + q ˙ T W ˜ σ ( q , q )

Image (5.37)

And

V ˙ [ λ m ( K d ) α λ M ( M ) α k c q ˜ ] q ˙ 2 [ α λ m ( K p ) λ M ( K i ) α k g ] q ˜ 2

Image (5.38)

If

q ˜ λ M ( M ) α k c

Image (5.39)

and

λ m ( K d ) ( 1 + α ) λ M ( M ) λ m ( K p ) 1 α λ M ( K i ) + k g

Image (5.40)

then V˙0Image, q˜Image decreases. Then (5.40) is established. Using (5.34) and λm(Ki1)=1λM(Ki)Image, (5.40) is (5.25).

V˙Image is negative semidefinite. Define a ball Σ of radius σ>0Image centered at the origin of the state space, which satisfies these conditions:

Σ = { q ˜ : q ˜ λ M ( M ) α k c = σ }

Image (5.41)

V˙Image is negative semidefinite on the ball Σ. There exists a ball Σ of radius σ>0Image centered at the origin of the state space on which V˙0Image. The origin of the closed-loop equation (5.24) is a stable equilibrium. Since the closed-loop equation is autonomous, we use LaSalle's theorem. Define Ω as

Ω = { x ( t ) = [ q ˜ , q ˙ , ξ ˜ ] R 3 n : V ˙ = 0 } = { ξ ˜ R n : q ˜ = 0 R n , q ˙ = 0 R n }

Image (5.42)

From (5.36), V˙=0Image if and only if q˜=q˙=0Image. For a solution x(t)Image to belong to Ω for all t0Image, it is necessary and sufficient that q˜=q˙=0Image for all t0Image. Therefore it must also hold that q¨=0Image for all t0Image. We conclude from the closed-loop system (5.24) that if x(t)ΩImage for all t0Image, then

ϕ ( q , q ) = ϕ ( q d , 0 ) = ξ ˜ + ϕ ( q d , 0 ) ξ ˜ = 0

Image (5.43)

implies that ξ˜=0Image for all t0Image. So x(t)=[q˜,q˙,ξ˜]=0R3nImage is the only initial condition in Ω for which x(t)ΩImage for all t0Image.

Finally, we conclude from all of this that the origin of the closed-loop system (5.24) is locally asymptotically stable. Because 1αλm(Ki1)λm(Kp)Image, the upper bound for q˜Image can be

q ˜ λ M ( M ) k c λ M ( K i ) λ m ( K p )

Image (5.44)

It establishes the semiglobal stability of our controller, in the sense that the domain of attraction can be arbitrarily enlarged with a suitable choice of the gains. Namely, increasing KpImage the basin of attraction will grow. □

Remark 5.1

From the above stability analysis, we see that the gain matrices of the neural PID control (5.19) can be chosen directly from the conditions (5.25). The tuning procedure of the PID parameters is more simple than [2,4,71,101,117]. No modeling information is needed. The upper or lower bounds of PID gains need the maximum eigenvalue of M in (5.25); it can be estimated without calculating M. For a robot with only revolute joints [131],

λ M ( M ) β , β n ( max i , j | m i j | )

Image (5.45)

where mijImage stands the ijth element of M, MRn×nImage. A β can be selected such that it is much bigger than all elements.

Remark 5.2

The main difference between our neural PID control with the other neural PD controllers is that the stability condition is changed. We require the regulation error:

q ˜ < k 1 λ M ( K i ) λ m ( K p )

Image (5.46)

The other neural PD controllers need

q ˜ > k 2 K v

Image (5.47)

where k1Image and k2Image are positive constants. Obviously, if the initial condition is not worse and satisfies (5.46), (5.46) is always satisfied, and q˜Image will decrease to zero. But (5.47) cannot be satisfied when q˜Image becomes small, so KvImage has to be increased.

Remark 5.3

If the unknown f(q)Image is estimated by the multilayer neural network (5.5) the modeling error (5.21) becomes

f ˜ = f f ˆ = W σ ( V [ q , q ] ) + ϕ ( q , q ) W ˆ σ ( V ˆ [ q , q ] ) = W ˜ σ ( V ˆ [ q , q ] ) W σ ( V ˆ [ q , q ] ) + W σ ( V [ q , q ] ) + ϕ ( q , q ) = W ˜ σ ( V ˆ [ q , q ] ) + W σ V ˜ [ q , q ] + ϵ 1 + ϕ ( q , q ) = W ˜ [ σ ( V ˆ [ q , q ] ) + σ V ˜ [ q , q ] ] + W ˆ σ V ˜ [ q , q ] + ϕ 1 ( q , q )

Image (5.48)

where ϕ1(q)=ϵ1+ϕ(q,q)Image, ϵ1Image is the Taylor approximation error. The closed-loop equation (5.24) becomes

M ( q ) q ¨ + C ( q , q ˙ ) q ˙ + W ˜ { σ ( V ˆ [ q , q ] ) + σ V ˜ [ q , q ] } + W ˆ σ V ˜ [ q , q ] + ϕ 1 ( q , q ) = K p q ˜ K d q ˙ + ξ ˜ + ϕ ( q d ) ξ ˜ = K i q ˜

Image (5.49)

If the Lyapunov function in (5.28) is changed as

V m = V + 1 2 t r ( V ˜ T K v 1 V ˜ )

Image (5.50)

then the derivative of (5.50) is

V m = V ˙ q ˙ T W ˜ σ ( q [ q , q ] ) + q ˙ T W ˜ [ σ ( V ˆ [ q , q ] ) + σ V ˜ [ q , q ] ] + t r ( V ˜ T K v 1 V ˜ )

Image (5.51)

If the training rule (5.26) is changed as

W ˆ = K w { σ ( V ˆ [ q , q ] ) + σ V ˜ [ q , q ] } ( q ˙ + α q ˜ ) T V ˆ = K v W ˆ σ q ( q ˙ + α q ˜ ) T

Image (5.52)

Theorem 5.1 is also established.

One common problem of the linear PID control (5.18) is integral windup, where the rate of integration is larger than the actual speed of the system. The integrator's output may exceed the saturation limit of the actuator. The actuator will then operate at its limit no matter what the process outputs. This means that the system runs with an open loop instead of a constant feedback loop. The solutions of antiwindup schemes are mainly classified into two types [142]: conditional integration and back-calculation. It has been shown that none of the existed methods is able to provide good performance over a wide range of processes [5]. In this paper, we use the conditional integration algorithm. The integral term is limited to a selected value

u = K p q ˜ + K d q ˜ + s a t [ K i 0 t q ˜ ( τ ) d τ , ν max ] + f ˆ

Image (5.53)

where sat[x,νmax]={xif x<νmaxνmaxif xνmaxImage. νmaxImage is a prescribed value to the integral term when the controller saturates. This approach is also called preloading [123]. Now the linear PID controller becomes nonlinear PID. The semiglobal asymptotic stability has been analyzed by [2]. When νmaxImage is the maximum torque of all joint actuators, νmax=ksmaxi(|uimax|)Image, uimax=max(|ui|)Image, ks1Image. A necessary condition is

ν max 3 G ¯ , G ( q ) G ¯

Image

where G(q)Image is the gravity torque of the robot (5.1), G¯Image is the upper bound of G(q)Image. ksImage is a design factor in the case that not all PID terms are subjected to saturation. For the controller (5.53), ksImage can selected as ks=14Image.

Following the process from (5.4) to (5.13), the neural PID with an antiwindup controller (5.53) requires

ν max 3 ϕ ¯ W σ ( q , q ) + ϕ ( q , q ) W σ ( q , q ) + ϕ ( q , q ) ϕ ¯

Image (5.54)

where ϕ¯Image is the upper bound of the neural estimator, WImage, σ(q,q)Image and ϕ(q,q)Image are defined in (5.4).

We can see that the first additional condition for the neural PID with antiwindup is that the neural estimator must be bounded, while the linear neural PID only requires that the neural estimation error satisfy the Lipschitz condition (5.10).

Since uimaxImage (or νmaxImage) is a physical requirement for the actuator, it is not a design parameter. In order to satisfy the condition (5.54), we should force ϕ¯Image as small as possible. A good structure of the neural estimator may make the term Wσ(q,q)Image smaller, such as the multilayer neural network in Remark 5.3. There are several methods that can be used to find a good neural network, such as the genetic algorithm [7] and pruning [130]. Besides structure optimization, the initial condition for the gradient training algorithm (5.26) also affects ϕ¯Image. Since the initial conditions for WˆImage and VˆImage in (5.52) do not affect the stability property, we design an off-line method to find a better value for Wˆ(0)Image and Vˆ(0)Image. If we let Wˆ(0)=W0Image, V(0)=V0Image, the algorithm (5.52) can make the identification error convergent, i.e., Wˆ(t)Image and Vˆ(t)Image will make the identification error smaller than that of W0Image and V0Image. Wˆ(0)Image and Vˆ(0)Image are selected by the following steps:

  1. 1.  Start from any initial value for Wˆ(0)=W0Image, Vˆ(0)=V0Image.
  2. 2.  Do training with (5.52) until T0Image.
  3. 3.  If the q˜(T0)<q˜(0)Image, let Wˆ(T0)Image and Vˆ(T0)Image as a new Wˆ(0)Image and Vˆ(0)0Image, i.e., Wˆ(0)=Wˆ(T0)Image, Vˆ(0)=Vˆ(T0)Image, go to 2 to repeat the training process.
  4. 4.  If the q˜(T0)q˜(0)Image, stop this off-line identification, now Wˆ(T0)Image and Vˆ(T0)Image are the final value for Wˆ(0)Image and Vˆ(0)Image.

5.2 Neural PID control with unmeasurable velocities

The neural PID control (5.19) uses the joint velocities q˙Image. In contrast to the high precision of the position measurements by the optical encoders, the measurement of velocities by tachometers may be quite mediocre in accuracy, specifically for certain intervals of velocity. The common idea in the design of PID controllers, which requires velocity measurements, has been to propose state observers to estimate the velocity. The simplest observer may be the first-order and zero-relative position filter [131]

υ i ( s ) = b i s s + a i q i ( s ) , i = 1 n

Image (5.55)

where υi(s)Image is an estimation of q˙iImage, aiImage, and biImage are the elements of diagonal matrices A and B, A=diag{ai}Image, B=diag{bi}Image, ai>0Image, bi>0Image. The transfer function (5.55) can be realized by

{ x ˙ = A ( x + B q ) q ˙ ˜ = x + B q

Image (5.56)

The linear PID control (5.19) becomes

u = K p q ˜ K d υ + ξ + W ˆ σ ( q ) ξ ˙ = K i q ˜ , ξ ( 0 ) = ξ 0 x ˙ = A ( x + B q ) υ = x + B q

Image (5.57)

where KpImage, KiImage, and KdImage are positive definite diagonal matrices, aiImage and biImage in (5.55) are positive constants. The closed-loop system of the robot (5.1) is

d d t [ ξ υ q ˙ ] = [ K i q ˜ A υ + B q ˙ M 1 [ C ( q , q ˙ ) q ˙ W ˜ σ ( q ) ϕ ( q ) + K p q ˜ K d υ + ξ ˜ + ϕ ( q d ) ] ]

Image (5.58)

The equilibrium of (5.58) is [ξ˜,υ,q˙]=[0,0,0]Image.

The following theorem gives the asymptotic stability of the neural PID control with the velocity observer (5.55). This theorem also provides a training algorithm for neural weights, and an explicit selection method of PID gains.

Since the velocities are not available, the input of the neural networks becomes

f ˆ ( q , q ) = W ˆ σ ( q , v ) or  f ˆ ( q , q ) = W ˆ σ ( V ˆ [ q , v ] )

Image (5.59)

Theorem 5.2

Consider the robot dynamic (5.1) controlled by the neural PID controller (5.57), if A and B of the velocity observer (5.55) satisfy

λ M ( A ) λ m 2 ( A ) λ m ( K d ) λ m ( B ) λ M ( B ) λ M ( B ) 1 4 λ m ( K d ) λ m ( M ) λ M 2 ( M ) λ m ( B α I ) 1 2 λ m ( A )

Image (5.60)

where α is positive design constant, provided that the PID control gains of (5.57) satisfy

λ m ( K p ) 1 2 λ M ( K p ) 1 α [ λ M ( K i ) + λ M ( A 1 B K i ) + 1 + 2 α 2 k ϕ + α 2 2 λ M ( K d ) + α 2 λ M ( A 1 K i ) ] λ m ( K d ) k g + 1 2 α λ M ( A 1 K i ) + 1 2 α λ M ( K p ) + κ ( M ) λ M ( M ) λ M ( A ) 2 λ m ( A B 1 I ) 1 λ M ( K i ) α 3 λ m ( K p )

Image (5.61)

where kϕImage satisfies (5.10), κ(M)Image is the condition number of M, and the weight of neural networks is tuned by

W ˆ = K w σ ( q , υ ) [ α q ˜ + υ + B 1 ( υ ˙ + A v ) ] T

Image (5.62)

then the closed-loop system (5.58) is locally asymptotically stable at the equilibrium:

x = [ ξ ϕ ( q d ) , q ˜ , q ˜ ] T = 0

Image (5.63)

in the domain of attraction

q ˜ λ m ( M ) α k c [ λ m ( B α I ) 1 2 λ m ( A ) ] + 1 α υ

Image (5.64)

Proof

We construct a Lyapunov function as

V c = 1 2 q ˙ T M q ˙ + 1 2 q ˜ T K p q ˜ + 0 t ϕ ( q ) d τ k ϕ + q ˜ T ϕ ( q d ) + 3 2 ϕ ( q d ) T K p 1 ϕ ( q d ) + α 2 ξ ˜ T K i 1 ξ ˜ α q ˜ T M q ˙ + q ˜ T ( I + A 1 B ) ξ ˜ + 1 2 υ T B 1 K d υ υ T M q ˙ + υ T A 1 ξ ˜ + 1 2 t r ( W ˜ T K w 1 W ˜ )

Image (5.65)

where the definition of kϕImage is the same as Theorem 5.1. α is a design positive constant. We first prove V is a Lyapunov function, V0Image. The term 12q˜TKpq˜Image is separated into three parts, and V=i=16ViImage:

V 1 = 1 6 q ˜ T K p q ˜ + q ˜ T ϕ ( q d ) + 3 2 ϕ ( q d ) T K p 1 ϕ ( q d ) V 2 = 1 6 q ˜ T K p q ˜ + q ˜ T ξ ˜ + α 2 ξ ˜ T K i 1 ξ ˜ V 3 = 1 6 q ˜ T K p q ˜ α q ˜ T M q ˙ + 1 4 q ˙ T M q ˙ V 4 = 1 4 υ T ( B 1 K d ) υ + υ T A 1 ξ ˜ + ξ ˜ T ( A 1 B ) ξ ˜ V 5 = 1 4 υ T ( B 1 K d ) υ υ T M q ˙ + 1 4 q ˙ T M q ˙ V 6 = 0 t ϕ ( q ) d τ k ϕ + 1 2 t r ( W ˜ T K w 1 W ˜ ) 0

Image (5.66)

Here, V1Image and V2Image are the same as (5.29), i.e.,

λ M ( K i ) α 3 λ m ( K p )

Image (5.67)

For V3Image, if α16λm(Kp)λm(M)λM(M)Image

V 3 1 2 ( 1 2 λ m ( M ) q ˙ 1 3 λ m ( K p ) q ˜ ) 2 0

Image (5.68)

Because λm(AB)λm(B1)λM(A)Image and λm(B1)=1λM(B)Image, it is easy to find that, if λM(A1)λm(B1Kd)λm((A1B))Image or λM(A)λm2(A)λm(Kd)λm(B)λM(B)Image,

V 4 1 2 ( 1 2 λ m ( B 1 K d ) υ 2 2 λ M ( A 1 ) υ ξ ˜ + 2 λ m ( ( A 1 B ) ) ξ ˜ 2 ) 0

Image (5.69)

If λM(M)12λm((B1Kd))λm(M)Image or λM(B)14λm(Kd)λm(M)λM2(M)Image

V 5 = 1 2 [ 1 2 υ T K d B 1 υ + 2 υ T M q ˙ + 1 2 q ˙ T M q ˙ ] 0

Image (5.70)

Because V60Image, obviously, there exist α, A, and B such that

α 2 1 6 λ m ( K p ) λ m ( M ) λ M 2 ( M ) λ M ( A ) λ m 2 ( A ) λ m ( K d ) λ m ( B ) λ M ( B ) λ M ( B ) 1 4 λ m ( K d ) λ m ( M ) λ M 2 ( M )

Image (5.71)

This means if KpImage is sufficiently large or KiImage is sufficiently small, (5.34) is established, and VcImage is globally positive definite. Now we compute its derivative. The derivative of VcImage is

V ˙ q ˙ T B M q ˙ υ T A B 1 K d υ α q ˜ T K p q ˜ + α k g q ˜ 2 + α q ˙ T M q ˙ + k c α q ˜ υ q ˙ 2 + q ˜ T K i q ˜ + q ˜ T A 1 B K i q ˜ + υ T K d υ + 1 2 k g υ 2 + 1 2 k g q ˜ 2 + q ˜ T ( α K d K p A 1 K i ) υ + q ˙ T A M υ + t r [ W ˜ T ( K w 1 W ˜ + σ ( q , v ) q ˙ T + α σ ( q , v ) q ˜ T + σ ( q , v ) υ T ) ]

Image (5.72)

Because υ˙=Aυ+Bq˙Image and B=diag{bi}Image, the last term is zero if we apply the updating law (5.62). Using (5.32), (5.72) is

V ˙ q ˙ T { λ m ( B M α M ) k c α q ˜ υ 1 2 κ ( M ) λ M ( A M ) } q ˙ υ T ( λ m ( A B 1 K d K d ) 1 2 k g 1 2 α λ M ( K e ) κ ( M ) 2 λ M ( A M ) ) υ q ˜ T ( λ m ( α K p K i A 1 B K i ) α k g 1 2 k g α 2 λ M ( K e ) ) q ˜

Image (5.73)

Using λi(A)λM(B)λi(AB)λi(A)λm(B)Image, i can be “m” or “M,” the last condition of (5.71) can be replaced by

q ˜ 1 α υ 1 α k c [ λ m ( B α I ) λ m ( M ) 1 2 κ ( M ) λ M ( M ) λ m ( A ) ]

Image

It is the attraction area (5.64).

Using λi(A)+λM(B)λi(A+B)λi(A)+λm(B)Image, the second condition of (5.71) is

λ m [ ( A B 1 I ) K d ] λ m ( A B 1 I ) λ m ( K d ) 1 2 k g + 1 2 α λ M ( K e ) + κ ( M ) 2 λ M ( A M )

Image (5.74)

It is the condition for KdImage in (5.61). Also

λ m ( α K p ) λ M ( K i ) + λ M ( A 1 B K i ) + 1 + 2 α 2 k g + α 2 λ M ( K e )

Image

It is the condition for KpImage in (5.61). The condition for KiImage in (5.61) is obtained from (5.67). The remaining part of the proof is the same as Theorem 5.1. □

Remark 5.4

The conditions (5.60) and (5.61) decide how to choose the PID gains. The first condition of (5.61) is

λ m ( K p ) 1 α λ M ( K i ) + Ω Ω = 1 α [ λ M ( A 1 B K i ) + 1 + 2 α 2 k g + α 2 λ M ( K d ) + 1 2 λ M ( A 1 K i ) ] + 1 2 α λ M ( K p )

Image (5.75)

the third conditions of (5.61) is λm(Kp)3αλM(Ki)Image; they are compatible. When KiImage is not big, these conditions can be established. The second condition of (5.61) and the third condition of (5.60) are not directly compatible. We first let α as small as possible, and KpImage as big as possible. So KiImage cannot be big. These requirements are reasonable for our real control. If we select B=βA+αIImage, form the third condition of (5.60), β12Image. The second condition of (5.61) requires λm(AB1I)>12Image, there exists 1>β12Image and a small α such that λm[A(βA+αI)1I]>12Image. After A and B are decided, we use the second condition of (5.61) to select KdImage.

5.3 Neural PID tracking control

We modify the reference signal and add a filter, such that the PID regulation control is extended to PID tracking control. Most important, the global and semiglobal asymptotic stability of this PID tracking control are proved.

Model-based compensation is an alternative method for PID control to overcome the problems of PID control. The intelligent compensation does not need a mathematical model; it is a model-free compensator, such as fuzzy PID with the fuzzy compensator [153] and neural PID with the neural compensator [154]. From the best of our knowledge, theory analysis for neural PID tracking control is still not published. In this paper, we will analyze the neural PID tracking control. We prove that the robot can follow a filtered reference asymptotically.

Given a desired time-varying position qd(t)RnImage, the objective of robot control is to design the input torque u in (5.1) such that the tracking error is

Δ ( t ) = q d ( t ) q ( t )

Image (5.76)

e(t)0Image when initial conditions are in arbitrary large domain of attraction. The classic industrial PID law is (5.15). In order to finish the tracking job, we define a filtered error as

s = q ˙ d q ˙ = Δ ˙ + K a Δ + K b Δ d t

Image (5.77)

So

q ˙ d = q ˙ + K a Δ + K b Δ d t

Image (5.78)

We define a matrix KaImage as

K a = K a T = K d 1 K p , K b = K d 1 K i

Image (5.79)

In order to decouple the PID tracking control (5.19), we choose Kp,KiImage, and KdImage as positive definite diagonal matrices. The closed-loop system of the robot (5.1) is

M ( q ) q ¨ + C ( q , q ˙ ) q ˙ + G ( q ) + F ( q ˙ ) = K p q ˜ K d q ˙ + z z ˙ = K i q ˜

Image

Using the filtered error s, the final closed-loop equation becomes

M s ˙ = C s + f K D s

Image (5.80)

where f(q,q)=G(q)+F(q˙)Image. The following theorem gives the stability analysis of the PID tracking control.

Theorem 5.3

Consider the robot dynamic (5.1) controlled by the PID tracking control:

u = K p Δ + K i 0 t Δ ( τ ) d τ + K d Δ ˙ + f

Image (5.81)

the closed-loop system (5.80) is globally asymptotically stable at the equilibrium, provided that control gains are positive.

Proof

Propose the following Lyapunov function:

V = 1 2 s T M s

Image (5.82)

Using (5.80) the derivative of the proposed function yields

V ˙ = s T M s ˙ + 1 2 s T M ˙ s = s T C s + s T f s T K D s + 1 2 s T M ˙ s

Image (5.83)

Since xT(M˙2C)x=0Image,

V ˙ = s T K D s + s T f

Image

Using the PID control law (5.81), then

V ˙ = s T K D s

Image

So

V ˙ λ m ( K D ) s 2

Image

V˙Image is negative semidefinite on the ball Σ. There exists a ball Σ of radius σ>0Image centered at the origin of the state space on which V˙0Image. The origin of the closed-loop equation (5.80) is a stable equilibrium. Since the closed-loop equation is autonomous, we use LaSalle's theorem. Define Ω as

Ω = { x ( t ) = [ e , e ˙ , f ˜ ] R 3 n : V ˙ = 0 } = { f ˜ R n : e = 0 R n , e ˙ = 0 R n }

Image

From (5.83), V˙=0Image if and only if e=e˙=0Image. For a solution x(t)Image to belong to Ω for all t0Image, it is necessary and sufficient that e=e˙=0Image for all t0Image. □

In many cases, in (5.1) the gravity G(q)Image and the friction F(q˙)Image are not available. We can use a feedforward neural network to model them as (5.4)

N ( q , q ) = G ( q ) + F ( q ˙ ) N ˆ ( q , q ) = W ˆ σ ( q , q ) , N ( q , q ) = W σ ( q , q ) + η

Image (5.84)

where WImage is unknown constant weight, WˆImage is estimated weight, η is the neural approximation error, and σ is a neural activation function. Here, we use the Gaussian function such that σ(q,q)0Image.

This linear-in-the-parameter net is the simplest neural network. According to the universal function approximation theory, the smooth function N(q,q)Image can be approximated by a multilayer neural network with one hidden layer in any desired accuracy provided proper weights and hidden neurons:

N ˆ ( q , q ) = W ˆ σ ( V ˆ [ q , q ] ) , f ( q ) = W σ ( V [ q , q ] ) + η

Image (5.85)

where WˆRn×mImage, VˆRm×nImage, m is hidden node number, VˆImage is the weight in a hidden layer. In order to simplify the theory analysis, we first use linear-in-the-parameter net (5.84), then we will show that the multilayer neural network (5.85) can also be used for the neural control of robot manipulators. According to universal function approximation property, f(x)Image can be shown that as long as x is restricted to a compact set SRnImage. There exist synaptic weights and thresholds such that

f ( x ) = W σ ( V [ x ] ) + η

Image (5.86)

The functional approximation error ϕ(x)Image of the neural network generally decreases as m increases. In fact, for any positive number ϕNImage, we can find a feedforward neural network such that

η < η N

Image (5.87)

Or

η T K a η η ¯ , η ¯ > 0

Image (5.88)

When the unknown dynamic f(q,q)Image in (5.84) is big, in order to assure asymptotic stability, the integral gain KiImage has to be increased. This may cause big overshoot, bad stability, and integrator windup. Model-free compensation is an alternative solution, where f(q)Image is estimated by a neural network as in (5.84). Normal neural PD control is

u = K p e + K d e + N ˆ

Image (5.89)

where Nˆ(q,q)=Wˆσ(q,q)Image. With the filtered error,

r = Δ + Λ Δ ˙

Image (5.90)

(5.16) becomes

u = K v r + N ˆ

Image (5.91)

The control (5.91) avoids integrator problems in (5.15). Unlike industrial PID control, they cannot reach asymptotic stability. The stability condition of the neural PD control (5.89) is r>BKvImage, B is a constant. In order to decrease rImage, KdImage has to be increased. This causes a long settling time problem. The asymptotic stability (r0Image) requires KvImage.

An integrator is added into the normal neural PD control (5.89); it has a similar form as the industrial PID in (5.15),

u = K p e Δ + K d Δ ˙ + K i 0 t Δ ( τ ) d τ + N ˆ

Image (5.92)

So

q ˙ d = q ˙ d + K a Δ + K b Δ d t

Image (5.93)

The closed-loop system of the robot (5.1) is

M ( q ) q ¨ + C ( q , q ˙ ) q ˙ + N ˜ = K p q ˜ K d q ˙ + z z ˙ = K i q ˜

Image (5.94)

where N˜=NNˆImage

N ˜ = W σ ( q ) + ϕ ( q ) W ˆ σ ( q ) = W ˜ σ ( q ) + η

Image (5.95)

here, W˜=WWˆImage. With the filtered error, the final closed-loop equation becomes

M s ˙ = C s + W ˜ σ ( q , q ˙ ) + ϕ K D s

Image (5.96)

The following theorem gives the stability analysis of the neural PID control. From this theorem, we can see how to train the weight of the neural compensator in (5.92).

Theorem 5.4

Consider the robot dynamic (5.1) controlled by the neural PID control (5.92). The closed-loop system (5.96) is semiglobally asymptotically stable at the equilibrium, provided that control gains are positive; weight of the neural networks (5.84) is tuned by

W ˆ = K W σ ( q , q ˙ ) s T

Image (5.97)

Proof

We propose the following Lyapunov function:

V = 1 2 s T M s + 1 2 t r { W ˜ T K W 1 W ˜ }

Image (5.98)

Using (5.96) and M˙(q)2C(q,q˙)Image is skew symmetric, i.e.,

x T [ M ˙ ( q ) 2 C ( q , q ˙ ) ] x = 0

Image (5.99)

the derivative of the proposed function yields

V ˙ = s T M s ˙ + 1 2 s T M ˙ s + t r { W ˜ T K W 1 W ˜ } = s T C s + s T W ˜ σ ( q , q ˙ ) + s T ϕ s T K D s + 1 2 s T M ˙ s + t r { W ˜ T K W 1 W ˜ }

Image (5.100)

Since xT(M˙2C)x=0Image,

V ˙ = s T K D s + s T η + s T W ˜ T σ ( q , q ˙ ) + t r { W ˜ T K W 1 W ˜ }

Image

Using the adaptation law (5.97), then

V ˙ = s T K D s + s T η

Image (5.101)

By the inequality,

y T A x y A x y A x | λ m ( A ) | y x

Image

Using (5.101),

V ˙ λ m ( K D ) s 2 + η N s

Image (5.102)

Since ηNImage is constant, V˙Image is a negative semidefinite provided that

s > η N λ m ( K D )

Image

Define a ball Σ of radius σ>0Image centered at the origin of the state space, which satisfies these conditions:

Σ = { s : s η N λ m ( K D ) }

Image (5.103)

V˙Image is negative semidefinite on the ball Σ. There exists a ball Σ of a radius σ>0Image centered at the origin of the state space on which V˙0Image. The origin of the closed-loop equation (5.96) is a stable equilibrium. Since the closed-loop equation is autonomous, we use LaSalle's theorem. Define Ω as

Ω = { x ( t ) = [ e , e ˙ , f ˜ ] R 3 n : V ˙ = 0 } = { f ˜ R n : e = 0 R n , e ˙ = 0 R n }

Image (5.104)

From (5.102), V˙=0Image if and only if e=e˙=0Image. For a solution x(t)Image to belong to Ω for all t0Image, it is necessary and sufficient that e=e˙=0Image for all t0Image. Therefore it must also hold that e¨=0Image for all t0Image. We conclude that from the closed-loop system (5.96), if x(t)ΩImage for all t0Image, then

η ( q , q ) = η ( q d , 0 ) = N ˜ + η ( q d , 0 ) d d t N ˜ = 0

Image (5.105)

implies that N˜=0Image for all t0Image. So x(t)=[e,e˙,f˜]=0R3nImage is the only initial condition in Ω for which x(t)ΩImage for all t0Image. Finally, we conclude from all of this that the origin of the closed-loop system (5.96) is locally asymptotically stable. The tracking error s and its derivative s˙Image are bounded, which implies boundedness of e and e˙Image. Given that the desired trajectories qdImage and q˙dImage are also bounded, joint position q and joint velocity q˙Image are bounded. □

5.4 Experimental results of the neural PID

Recently, we construct the portable exoskeleton robot, shown in Fig. 5.1. It has two servomotors. By our special design, it can move freely in 3D space; see Appendix A. The control computer is a Pentium Core2Duo with 2GB of RAM and a 120GB hard disk. The control software is under the operating system Linux Mint 15 “Olivia.” We use this operating system, because the development software has a GNU General Public License. The development environment is Qt4 with C++, which is less runtime in a normal PC, and improves control performance compared with MATLAB Simulink®. Another advantage is that our system can generate executable programs, which can run on any PC without the installing development environment.

Image
Figure 5.1 The portable exoskeleton robot.

This upper limb exoskeleton is fixed on the human arm. The behavior of the exoskeleton is the same as the human arm; see Fig. 5.2. The exoskeleton is light, and the height can be adjusted for each user. The users left hand is an enable button, which released the brakes on the device and engaged the motor. The reference signals are generated by admittance control in task space. These references are sent to joint space. The robot in joint space can be regarded as free motion without human constraints. The whole control system is shown in Fig. 5.3. The objective of neural PID control is make the transient performance faster with less overshoot, such that human feel comfortable.

Image
Figure 5.2 The exoskeleton robot vs. human arms.
Image
Figure 5.3 The neural PID control of the exoskeleton robot.

All of the controllers employed a sampling frequency of 1 kHz. The two theorems in this paper give sufficient conditions for the minimal values of proportional and derivative gains and maximal values of integral gains. We use (5.45) to estimate the upper and the lower bounds of the eigenvalues of the inertia matrix M(q)Image, and kgImage in (5.10). We select λM(M)<2.7Image, λm(M)>0.5Image, kg=4.8Image. We choose α=4λM(Ki)λm(Kp)Image such that λM(Ki)α3λm(Kp)Image is satisfied. α=0.01Image, A is chosen as A=diag(3)Image, β=13Image, so B=diag(2)Image. The joint velocities are estimated by the standard filters:

q ˙ ˜ ( s ) = b s s + a q ( s ) = 18 s s + 30 q ( s )

Image (5.106)

The PID gains are chosen as

K p = d i a g [ 12 , 1 ] K i = d i a g [ 2 , 1 ] K d = d i a g [ 31 , 31 ]

Image (5.107)

such that the conditions of Theorem 5.2 are satisfied. The initial elements of the weight matrix WR7×7Image are selected randomly from −1 to 1. The active function in (5.26) is a Gaussian function:

σ = exp { ( q i m i ) 2 / 100 } , i = 1 , 2

Image (5.108)

where miImage is selected randomly from 0 to 2. The weights are updated by (5.62) with Kw=10Image.

The control results of Joint-1 with neural PID control is shown in Fig. 5.4, marked “Neural PID.” We compare our neural PID control with the other popular robot controllers. First, we use the linear PID (5.15). The PID gains are the same as (5.107), and the control result is shown in Fig. 5.4, marked “Linear PID-2.” Because the steady-state error is so big, the integral gains are increased as

K i = d i a g [ 50 , 20 ]

Image (5.109)

The control result is shown in Fig. 5.4, marked “Linear PID-1,” the transient performance is poor. There still exists regulation error. Further increasing KiImage causes the closed-loop system to be unstable. Then we use a neural compensator to replace the integrator. It is normal neural PD control (5.16). In order to decrease steady-state error, the derivative gains are increased as

K d = d i a g [ 200 , 300 ]

Image (5.110)

the control result is shown in Fig. 5.4, marked “Neural PD.” The response becomes very slow.

Image
Figure 5.4 Comparison of several PID controllers for Joint 1.

Now we use Joint-2 to compare our neural PID control with the other two types of neural PID control. The control results of these three neural PID controllers are shown in Fig. 5.5. Here, we use a three-layer neural network with three nodes, which have integral, proportional, and derivative properties. A backpropagation like the training algorithm is used to ensure closed-loop stability [29]; it is marked “Neural Net in PID form.” Then we use a one-hidden layer neural network to tune the linear PID gains as in [117]; it is marked “PID tuning via neural net.” It can be found that the “Neural net in PID form” can assure stability, but the transient performance is not good. The “PID tuning via neural net” is acceptable except its slow response.

Image
Figure 5.5 Comparison of several PID controllers for Joint 2.

Clearly, neural PID control can successfully compensate the uncertainties such as friction, gravity, and the other uncertainties of the robot. Because the linear PID controller has no compensator, it has to increase its integral gain to cancel the uncertainties. The neural PD control does not apply an integrator, its derivative gain is big.

The structure of the neural compensator is very important. The number of hidden nodes m in (5.5) constitutes a structural problem for neural systems. It is well known that increasing the dimension of the hidden layer can cause the “overlap” problem and add to the computational burden. The best dimension to use is still an open problem for the neural control research community. In this application, we did not use a hidden layer, and the control results are satisfied. The learning gain KwImage in (5.62) will influence the learning speed, so a very large gain can cause unstable learning, while a very small gain produces a slow learning process.

Now we will carry out some experimental tests by using the neural PID controller in a tracking case. To accomplish this, we will define conveniently the neural gain KW=0.5Image, and by using [2]. We obtain the following gain matrices:

K P = d i a g [ 700 , 1200 ] K I = d i a g [ 20 , 20 ] K D = d i a g [ 44 , 40 ]

Image

The active function in (5.26) is a Gaussian function:

σ = exp { ( q i m i ) 2 / 100 } , i = 1 , 2

Image

The neural PID tracking control for the first and the second joints are shown in Fig. 5.6 and Fig. 5.7. The position error is reasonably small (between −2 and 2 degrees in both cases). The second joint carries out the extension, the neural network compensates the friction and gravity effects. The actual velocity satisfactorily follows the desired velocity.

Image
Figure 5.6 The first joint – horizontal flexion/extension of shoulder.
Image
Figure 5.7 The second joint – vertical flexion/extension of shoulder.

Overall, we can notice that, though we have used a small gain KWImage, the robot performance is acceptable when using the neural PID. Unfortunately, we cannot increase too much such gain, since the controller tries to compensate every single disturbance that affects the robot. Then it produces a torque that makes the joints to vibrate continuously. In fact, this vibration affects especially the small joints.

The tracking neural PID control proposed in this paper is compared with the other two types of controllers. We first use normal NN with three-layers, and each layer has 3 nodes. The backpropagation is applied to train the NN. Then we use a one-hidden layer NN to tune the linear PID gains. We find that our neural tracking PID is to ensure stability, but the transient performance is not good, while the PID tuning via neural net is acceptable except for its slow response.

Clearly, neural PID control can successfully compensate the uncertainties such as friction, gravity, and the other uncertainties of the robot. Because the linear PID controller has no compensator, it has to increase its integral gain to cancel the uncertainties. The neural PD control does not apply an integrator, and its derivative gain is big.

5.5 Conclusions

The neural PID proposed in this chapter solves the problems of large integral and derivative gains in the linear PID control and the neural PD control. It keeps good properties of the industrial PID control and neural compensator. Semiglobal asymptotic stability of this neural PID control is proven. When the joint velocities of robot manipulators are not available, local asymptotic stability is assured with filtered positions. The stability conditions give explicit methods to select PID gains. We also apply our neural PID to the exoskeleton robot in CINVESTA-IPN. Theory analysis and experimental study show the validity of the neural PID control.

A novel PID tracking controller is proposed to solve the problem of large integral and derivative gains in the tracking case with linear PID control. Global and semiglobal asymptotic stability of this PID tracking control are proven.

Bibliography

[2] J. Alvarez-Ramirez, R. Kelly, I. Cervantes, Semiglobal stability of saturated linear PID control for robot manipulators, Automatica 2003;39:989–995.

[4] S. Arimoto, Fundamental problems of robot control: part I, innovations in the realm of robot servo-loops, Robotica 1995;13(1):19–27.

[5] Karl J. Astrom, T. Hagglund, PID Controllers: Theory, Design and Tuning. Research Triangle Park, North Carolina: ISA Press; 1995.

[7] J. Arifovica, R. Gencay, Using genetic algorithms to select architecture of a feedforward artificial neural network, Physica A 2001;289:574–594.

[29] S. Cong, Y. Liang, PID-like neural network nonlinear adaptive control for uncertain multivariable motion control systems, IEEE Trans. Ind. Electron. 2009;56(10):3872–3879.

[71] R. Kelly, Global positioning of robot manipulators via PD control plus a class of nonlinear integral actions, IEEE Trans. Autom. Control 1998;43(7):934–938.

[85] F.L. Lewis, K. Liu, A. Yesildirek, Neural net robot controller with guaranteed tracking performance, IEEE Trans. Neural Netw. 1995;6(3):703–715.

[87] Frank L. Lewis, Thomas Parisini, Neural network feedback control with guaranteed stability, Int. J. Control 1998;70(3):337–339.

[101] E.V.L. Nunes, L. Hsu, F. Lizarralde, Arbitrarily small damping allows global output feedback tracking of a class of Euler–Lagrange systems, 2008 American Control Conference. Seattle, USA. 2008:378–382.

[117] P. Rocco, Stability of PID control for industrial robot arms, IEEE Trans. Robot. Autom. 1996;12(4):606–614.

[123] F. Greg Shinskey, Process-control Systems: Application, Design, Adjustment. New York, USA: McGraw-Hill; 1996.

[130] K. Suzuki, I. Horina, N. Sugie, A simple neural network pruning algorithm with application to filter synthesis, Neural Process. Lett. 2001;13:43–53.

[131] M.W. Spong, M. Vidyasagar, Robot Dynamics and Control. Canada: John Wiley & Sons Inc.; 1989.

[142] A. Visioli, Modified anti-windup scheme for PID controllers, IEE Proc., Control Theory Appl. 2003;161(1):49–54.

[153] W. Yu, J. Rosen, A novel linear PID controller for an upper limb exoskeleton, 49th IEEE Conference on Decision and Control, CDC'10. Atlanta, USA. 2010:3548–3553.

[154] X. Li, W. Yu, A systematic tunning method of PID controller for robot manipulators, 9th IEEE International Conference on Control & Automation, ICCA11. Santiago, Chile. 2011:274–279.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.137.171.121