Professional Documents
Culture Documents
Adaptive Control
Girish Chowdhary and Eric Johnson
Abstract We present an approach for combining standard
recursive least squares based regression with proven direct
model reference adaptive control using a recursively updated
modification term. This approach is applicable to adaptive
control problems where the uncertainty can be linearly parameterized. The combined training law drives the adaptive weights
smoothly to a recursively updated least squares estimate of the
ideal weights and is shown to have a stability proof. Expected
improvement in performance of the adaptive law is validated
through simulation.
I. INTRODUCTION
This paper presents a combined direct and indirect approach to adaptive control for a class of adaptive control
problems where the uncertainty can be linearly parameterized using a weighted combination of known nonlinear
functions. The presented approach augments any baseline
direct adaptive law by smoothly incorporating a recursively
updated least squares estimate of the ideal weights through
a modification term. The inclusion of the least squares based
estimate provides additional information to the adaptive law
and aids in convergence of the weights to optimal estimates
of the ideal weights. The unique idea behind this approach
is that it combines proven recursive least squares based
methods such as Kalman filter based least squares regression
[5] with standard gradient based adaptive laws while ensuring
that the system states remain uniformly ultimately bounded.
In plants where the plant uncertainty can be linearly
parameterized using a weighted combination of known nonlinear functions, the standard approach to the design of
adaptive controllers is to use a weight training law that
attempts to estimate the ideal weights in order to cancel
the uncertainty. In the direct adaptive control methodology,
adaptive laws are designed using a Lyapunov based approach
and can be thought of as attempting to minimize a quadratic
cost function of the instantaneous tracking error by using
a gradient descent type method [9]. These adaptive laws
however, are susceptible to local adaptation and weight drift
[11],[4],[15]. Furthermore, it can be shown that adaptive
laws formulated using the gradient descent methodology are
always at most rank 1 [2]. In order to ensure that adaptive
weights remain bounded, Ioannou et al. [4] suggested the use
of a modification term which adds damping to the adaptive
law. Narendra et al. [11] suggested the e modification term
which uses the norm of the tracking error to scale the added
damping. The methods are primarily designed to ensure
G. Chowdhary and Professor E. N. Johnson are with the Daniel
Guggenheim school of aerospace engineering at the Georgia Institute
of Technology, Atlanta GA, Girish.Chowdhary@gatech.edu,
Eric.Johnson@ae.gatech.edu
where W is a positive definite matrix of appropriate dimensions results in ultimate boundedness of the weights and
e(t) 0 if the reference input r(t) is persistently exciting
[11],[4],[15]. Equation 8 will be referred to as the baseline
adaptive law.
Case 2: Unstructured Uncertainty: In this more general
case it is only known that the uncertainty (x) can be
linearly parameterized; hence the adaptive part of the control
law is represented using a Radial Basis Function (RBF)
Neural Network (NN):
(1)
(2)
(3)
(4)
(5)
(6)
(7)
= W
On examining equation 4 we see that if W
then the tracking error dynamics are globally exponentially
stable. Hence, we desire to design an adaptive law such that
(t) W asymptotically.
W
T (x).
uad (x) = W
(9)
i (x) = ekxci k
/i
(10)
(11)
Ax(k) Bu(k)].
(12)
Since A, B, x(k), u(k) are known, the problem of estimating system uncertainty can be reduced to that of estimation
min T .
=W
W , tr denote the trace operator, and
Proof: Let W
consider the following Lyapunov candidate function:
(13)
X(k) =
1 (x(k)) 2 (x(k)) ...
m (x(1))
m (x(2))
. (14)
m (x(k))
= (k 1),
(15)
(k)
(16)
(x(k))(k).
(k+1)
= (k)+K(k+1)[(k+1)
(k+1)(k)],
(17)
T W 1 W
).
) = 1 eT P e + 1 tr(W
V (e, W
2
2
(20)
(21)
(22)
W )T (W
)
(W
T
W ) ( W
W)
(W
W )T .
+(W
(25)
Remark I : The above proof shows uniform ultimate boundedness of the tracking error and adaptive weights. It can be clearly seen that if 0
then tracking error e 0. This condition will be
achieved when W , that is when the Kalman
filter estimate of the ideal weights in 17 converges.
The convergence of the Kalman filter estimate is
related to choice of S(0), R and uniform complete
observability of the process model [3]. The uniform
complete observability for the process model in 15
can
if for some T > 0 and > 0
R t+Tbe guaranteed
T
(x(
))
(x(
))d I
t
Remark II : The above proof can be easily extended to the case where the structure of the
uncertainty is unknown (case 2 from section II)
by using Radial Basis Function Neural Networks
for approximating the uncertainty. The following
adaptive law will result in uniform ultimate boundedness of all states:
= ((x)rT (W
))W .
W
(27)
Furthermore, referring to equation 11 and noting
that in this case = , it can be shown that if
the Kalman filter estimates of the ideal weights
converge, then the weights will approach a neighborhood of the best linear approximation of the
uncertainty.
Remark III : Although in this paper we restricted
our attention to nonlinear systems which have
a linear part and a scalar linearly parameterized
uncertainty, it is straight forward to extend the
analysis to more general nonlinear systems with the
uncertainty in vector form as long as it is linearly
parameterized.
Remark IV : The increased computational burden
when using the adaptive law of equation 20 consists mainly of evaluating equations 17,18, and 19.
It should be noted that since (x) <m , the
inversion in equation 18 is reduced to a division
by a scalar.
Ref
Model
u
+
Plant
Estimation of
Adaptive Law
Least Squares
estimation
Fig. 1.
Schematics of adaptive controller with recursive least squares
Modification
= p
(28)
= a + (x).
(29)
parameter convergence was observed despite using a nonpersistently exciting reference input (r(t) = 0t), this is
because (x(t)) was uniformly completely observable.
1.2
adaptive weights
true weights
1
0.8
0.6
0.4
0.5
0.2
p deg/sec
0
0.2
0.4
0.6
0.5
0.8
baseline low gain
baseline high gain
0.2
0.4
0.6
degrees
0.8
1.2
10
15
time (sec)
Fig. 4.
1.5
0.2
1.4
0.8
Fig. 2. Phase portrait of system states with only baseline adaptive control
0.6
adaptive weights
true weights
0.4
0.2
W
0.5
0
0.2
p deg/sec
0.4
0
0.6
0.8
0.5
10
15
time (sec)
LS mod low gain
LS mod high gain
1.5
0.2
0.2
0.4
0.6
degrees
0.8
1.2
1.4
Fig. 3. Phase portrait of system states with combined direct and indirect
adaptive controller
V. CONCLUSION
In this work we presented an approach to combined
direct and indirect adaptive control for plants with a linearly
parameterized uncertainty by augmenting a baseline direct
adaptive law with recursive least squares. The presented
method ensures that the adaptive weights arrive smoothly
at the estimates of the ideal weights updated recursively
using a Kalman filter based parameter estimator. We showed
that the proposed combined adaptive law does not affect
the uniform ultimate boundedness properties of the model
reference adaptive controller. Furthermore, we indicated that
if the Kalman filter estimates of the ideal weights converge,
asymptotic stability can be guaranteed. We showed that the
roll angle
1.5
actual
ref model
pirad
1
0.5
0
0.5
10
15
10
15
time (sec)
roll rate
xDot (pirad/s)
1
actual
ref model
0.5
0.5
5
time (sec)
roll angle
1.4
1.2
actual
ref model
pirad
1
0.8
0.6
0.4
0.2
0
10
15
time (sec)
roll rate
xDot (pirad/s)
1
actual
ref model
0.5
0.5
10
15
time (sec)
Fig. 7.
Numerical differentiation for estimation of state derivatives suffers from high sensitivity to noise. An alternate
method is to use a Kalman filter based approach. Let x,
be the state of the system and x be its first derivative, and
consider the following system:
x
0 1
x
=
(30)
x
0 0
x
Suppose x is available as sensor measurement, then optimal fixed point smoothing can be used for estimating x
from available noisy measurements using the above system
[3]. Optimal Fixed Point Smoothing is a non real time
method for arriving at a state estimate at some time t, where
0 t T , by using all available data up to time T . For
ease of implementation on modern avionics, we present the
relevant equations in the discrete form. Let x
(k|N ) denote
the estimate of the state x = [ x x ]T , let Zk denote the
measurements, () denote predicted values, and (+) denote
corrected values, dt denote the discrete time step, Q and
R denote the process and measurement noise covariance
matrices respectively, while P denotes the error covariance
matrix. Then the forward Kalman filter equations can be
given as follow:
R EFERENCES
[1] Otto Bretscher. Linear Algebra with Applications. Prentice Hall, 2001.
[2] Girish V. Chowdhary and Eric N. Johnson. Theory and flight
test validation of long term learning adaptive flight controller. In
Proceedings of the AIAA GNC Conference, held at Honolulu, HI, 2008.
[3] Arthur Gelb. Applied Optimal Estimation. MIT Press, 1974.
[4] Petros A. Ioannou and Jung Sun. Robust Adaptive Control. Prentice
hall, out of print, 1996.
[5] Ravindra V. Jategaonkar. Flight Vehicle System Identification A
Time Domain Approach. American Institute of Aeronautics and
Astronautics, 2006.
[6] E. N. Johnson. Limited Authority Adaptive Flight Control. PhD thesis,
Georgia Institute of Technology, Atlanta Ga, 2000.
[7] Suresh K. Kannan. Adaptive Control of Systems in Cascade with
Saturation. PhD thesis, Georgia Institute of Technology, Atlanta Ga,
2005.
[8] Nakawan Kim. Improved Methods in Neural Network Based Adaptive
Output Feedback Control, with Applications to Flight Control. PhD
thesis, Georgia Institute of Technology, Atlanta Ga, 2003.
[9] Lewis F. L. Nonlinear network structures for feedback control.
Asian Journal of Control, 1:205228, 1999. Special Issue on Neural
Networks for Feedback Control.
0
0
1
0
dt
Zk
[ 1
x
k ()
k x
k1 ,
Pk ()
Kk
0 ]
,
x
,
x
(31)
(32)
(33)
k Pk1 k + Qk ,
T
(34)
T
x
k (+)
= x
k () + Kk [Zk Hk x
k ()],
Pk (+)
[I Kk Hk ]Pk ().
, (35)
(36)
(37)
Letting x
k|k = x
k the smoothed state estimate can be given
as:
x
k|N = x
k|N 1 + BN [
xN (+) x
N ()].
(38)